WO2015194423A1 - 文字に基づく音声生成のためのコントローラ及びシステム - Google Patents
文字に基づく音声生成のためのコントローラ及びシステム Download PDFInfo
- Publication number
- WO2015194423A1 WO2015194423A1 PCT/JP2015/066659 JP2015066659W WO2015194423A1 WO 2015194423 A1 WO2015194423 A1 WO 2015194423A1 JP 2015066659 W JP2015066659 W JP 2015066659W WO 2015194423 A1 WO2015194423 A1 WO 2015194423A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- voice
- character
- pitch
- selector
- characters
- Prior art date
Links
- 238000000034 method Methods 0.000 claims description 67
- 210000003811 finger Anatomy 0.000 claims description 22
- 230000004044 response Effects 0.000 claims description 15
- 238000000926 separation method Methods 0.000 claims description 15
- 230000002441 reversible effect Effects 0.000 claims description 11
- 210000003813 thumb Anatomy 0.000 claims description 9
- 238000003860 storage Methods 0.000 claims description 4
- 230000002194 synthesizing effect Effects 0.000 claims description 2
- 230000015572 biosynthetic process Effects 0.000 claims 1
- 238000003786 synthesis reaction Methods 0.000 claims 1
- 239000011295 pitch Substances 0.000 description 237
- 230000008569 process Effects 0.000 description 55
- 230000006870 function Effects 0.000 description 42
- 230000008859 change Effects 0.000 description 30
- 238000012545 processing Methods 0.000 description 29
- 238000012937 correction Methods 0.000 description 22
- 238000003825 pressing Methods 0.000 description 14
- 230000010354 integration Effects 0.000 description 12
- 238000006243 chemical reaction Methods 0.000 description 10
- 230000014509 gene expression Effects 0.000 description 9
- 230000005236 sound signal Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 5
- 238000001228 spectrum Methods 0.000 description 5
- 241001342895 Chorus Species 0.000 description 4
- HAORKNGNJCEJBX-UHFFFAOYSA-N cyprodinil Chemical compound N=1C(C)=CC(C2CC2)=NC=1NC1=CC=CC=C1 HAORKNGNJCEJBX-UHFFFAOYSA-N 0.000 description 4
- 230000007704 transition Effects 0.000 description 4
- 238000001514 detection method Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000003252 repetitive effect Effects 0.000 description 3
- 235000016496 Panda oleosa Nutrition 0.000 description 2
- 240000000220 Panda oleosa Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000005520 cutting process Methods 0.000 description 2
- 238000009826 distribution Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 210000005224 forefinger Anatomy 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000005070 sampling Methods 0.000 description 2
- 230000002459 sustained effect Effects 0.000 description 2
- WDJUZGPOPHTGOT-UHFFFAOYSA-N 3-[3-[5-[5-(4,5-dihydroxy-6-methyloxan-2-yl)oxy-4-hydroxy-6-methyloxan-2-yl]oxy-4-hydroxy-6-methyloxan-2-yl]oxy-14-hydroxy-10,13-dimethyl-1,2,3,4,5,6,7,8,9,11,12,15,16,17-tetradecahydrocyclopenta[a]phenanthren-17-yl]-2h-furan-5-one Chemical compound C1C(O)C(O)C(C)OC1OC1C(C)OC(OC2C(OC(OC3CC4C(C5C(C6(CCC(C6(C)CC5)C=5COC(=O)C=5)O)CC4)(C)CC3)CC2O)C)CC1O WDJUZGPOPHTGOT-UHFFFAOYSA-N 0.000 description 1
- 125000002066 L-histidyl group Chemical group [H]N1C([H])=NC(C([H])([H])[C@](C(=O)[*])([H])N([H])[H])=C1[H] 0.000 description 1
- 241001274197 Scatophagus argus Species 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 239000002994 raw material Substances 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 239000002689 soil Substances 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10G—REPRESENTATION OF MUSIC; RECORDING MUSIC IN NOTATION FORM; ACCESSORIES FOR MUSIC OR MUSICAL INSTRUMENTS NOT OTHERWISE PROVIDED FOR, e.g. SUPPORTS
- G10G1/00—Means for the representation of music
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/02—Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos
- G10H1/04—Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos by additional modulation
- G10H1/053—Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos by additional modulation during execution only
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/033—Voice editing, e.g. manipulating the voice of the synthesiser
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/033—Voice editing, e.g. manipulating the voice of the synthesiser
- G10L13/0335—Pitch control
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/08—Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
- G10L13/10—Prosody rules derived from text; Stress or intonation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2220/00—Input/output interfacing specifically adapted for electrophonic musical tools or instruments
- G10H2220/155—User input interfaces for electrophonic musical instruments
- G10H2220/315—User input interfaces for electrophonic musical instruments for joystick-like proportional control of musical input; Videogame input devices used for musical input or control, e.g. gamepad, joysticks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H2250/00—Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
- G10H2250/315—Sound category-dependent sound synthesis processes [Gensound] for musical use; Sound category-specific synthesis-controlling parameters or control means therefor
- G10H2250/455—Gensound singing voices, i.e. generation of human voices for musical applications, vocal singing sounds or intelligible words at a desired pitch or with desired vocal effects, e.g. by phoneme synthesis
Definitions
- the present invention relates to a technology for generating text-based speech at a designated pitch.
- Patent Document 1 discloses a technique of performing update control of the singing position in the lyrics indicated by the lyrics data in response to the reception of the performance data (pitch data). That is, there is disclosed a technique of playing a melody by user operation on an operation unit such as a keyboard and advancing a lyric in synchronization with the progression of the melody performance.
- controllers of various shapes have been developed in electronic musical instruments, and provided with a grip portion by protruding from the main body of the keyboard musical instrument, in order to detect an arbitrary operation portion or an appropriate manual operation in the grip portion. It is known to provide a detection unit of (see, for example, Patent Documents 2 and 3).
- Patent Document 4 a technique of displaying a plurality of lyrics on the display means, selecting an arbitrary section of the lyrics by the operation of the operation means, and outputting the selected section as singing voice of the designated pitch It is disclosed.
- the disclosed syllable is pronounced at the pitch designated by the keyboard. It is done.
- the present invention has been made in view of the above-mentioned point, and in the technology of generating a voice based on a predefined character string such as a lyrics according to the pitch to be played, a change of the voice to be generated, etc. It is an object of the present invention to make it possible to easily perform an ad-liberation performance, and thereby to make it possible to widen the range of expression in character-based speech generation. Another object of the present invention is to make it possible to select repeat objects without relying on vision.
- a controller for a voice generation device said voice generation device being a voice corresponding to specified one or more characters in a predefined character string
- a character selector configured to be operable by a user to designate the one or more characters in the string, and the controller is generated by the voice generation device.
- a voice control operator configured to be operable by a user to control the state of the voice.
- a system comprising the controller and the audio generation device.
- a voice corresponding to the one or more characters in the character string designated according to the operation of the character selector is generated from the voice generation device, and the generated voice is operated by the voice control operator. Since the control can be arbitrarily performed in accordance with the above, it is possible to easily change the sound to be generated in response to the user's operation, etc., even though the sound is generated based on a predetermined character string. Therefore, in the case of generating the voice corresponding to the character such as the lyrics in synchronization with the music performance, the controllability by the user can be enhanced, and the ad-lib performance of the lyric voice generation can be easily performed. This makes it possible to widen the range of expression in character-based speech generation.
- the controller comprises a grip suitable for being grasped by the user's hand, and the character selector and the voice control operator are respectively provided on the grip.
- the character selector and the voice control operator are respectively provided on the grip in an arrangement that can be operated by different fingers of the user holding the grip.
- one of the character selector and the voice control operator is operated by the thumb of the user, and the other is operated by the other finger of the user.
- the character selector and the voice control operator are respectively disposed on different sides of the grip. As described above, in the configuration in which the character selector and the voice control operator are arranged on one grip, the character selector and the voice control operator can be operated by using either of the fingers of the user holding the grip. It is suitable to operate both properly. Therefore, it is possible to easily operate the character selector and the voice control operator on the grip while playing the keyboard instrument or the like with another hand.
- a character information acquisition unit for acquiring information specifying one or more characters in a predetermined character string, and the specified one or more characters based on the acquired information.
- the voice generation unit repeatedly generates a voice generation unit that generates a corresponding voice, a repeat target reception unit that receives information specifying the voice being generated as a repeat target, and the voice specified as the repeat target
- An audio generation device comprising a processor configured to function as a repeat control unit to control. According to this, by listening to the audio sequentially generated by the audio generation unit as an audible sound, the user aurally quickly determines whether the audio generated in real time is suitable for designation as a repeat target. And it can be specified (selected). Therefore, characters to be repeated can be selected without relying on vision.
- the flowchart which shows an example (key on processing) of voice generation processing.
- the flowchart which shows an example (key off processing) of voice generation processing.
- FIG. 6 is a flowchart illustrating an example of a repeat target selection process.
- the figure which shows the example of a change of the grip shape of a controller.
- the figure which shows an example of the character string of the Japanese lyrics.
- the top view which shows another example of the character selector provided in a controller.
- FIG. 8 is a view showing an example of syllable integration processing and syllable separation processing according to the operation of the character selector of FIG. 7;
- FIG. 1A is a view schematically showing an electronic keyboard instrument 10 as a system provided with a controller 10a and an audio generation device 10b according to an embodiment of the present invention.
- the keyboard instrument 10 includes a rectangular parallelepiped main body 10b and a prismatic controller 10a.
- the main unit 10 b of the keyboard instrument 10 functions as an example of a sound generating device that electronically generates arbitrary musical tones and sounds, and includes a pitch selector 50 and an input / output unit 60.
- the pitch selector 50 is an operator operated by the user to designate the pitch of the musical tone or voice to be played, and is constituted by a plurality of keys including, for example, a white key and a black key.
- Shoulder straps are connected to the attachment positions P 1 and P 2 at both ends of the main body 10 b of the keyboard instrument 10 in the present embodiment.
- the user can place the keyboard instrument 10 in front of the body with the shoulder strap held on the shoulder, and perform playing by operating the pitch selector (keyboard) 50 with one hand.
- FIG. 1A upper, lower, left, and right directions viewed from the user when the user plays the keyboard instrument 10 in such a manner are noted.
- the directions referred to in the present specification refer to the directions of up, down, left, right, front, and rear as viewed from the user who plays the keyboard instrument 10.
- the pitch selector 50 is not limited to the keyboard type pitch operation specification operation elements, and any type of performance operation elements may be used. In short, it is necessary to specify some pitch in accordance with the user's operation. What is necessary is just to be configured to be able to
- the input / output unit 60 includes an input unit for inputting an instruction or the like from the user, and an output unit (display and speaker) for outputting various information (image information and audio information) to the user.
- an output unit display and speaker
- FIG. 1A as an example, a rotation switch as an input unit and a display as an output unit included in the keyboard instrument 10 are shown in broken lines.
- the controller 10a is in a direction (left direction seen from the user playing the keyboard instrument 10: see FIG. 1A) substantially perpendicular to the side (the left side in the example of FIG. 1A) of the main body (voice generation device) 10b. It protrudes.
- the outer shape of the controller 10a is substantially columnar. The size of the outer periphery of the substantially columnar portion is such that the user can hold it with one hand, so that the portion of the controller 10a protruding from the main body 10b constitutes a grip G.
- the shape of the cross section in the direction perpendicular to the axis extending in the longitudinal direction (left and right direction in FIG. 1A) of the grip G is constant regardless of the cutting position.
- the controller 10a may be integrally and indivisiblely coupled to the main body (voice generation device) 10b, or may be configured to be detachable from the main body (voice generation device) 10b. Alternatively, it may be separated from the main body (voice generation device) 10b and can communicate with the main body (voice generation device) 10b in a wired or wireless manner.
- FIG. 1B is a schematic view showing a state in which the controller 10a is viewed with the direction from the left to the right shown in FIG. 1A as the viewing direction, and shows an example in which the user grips the grip G.
- the cross section in the direction perpendicular to the axis of the grip G is shaped like a rounded corner of a rectangle. That is, while the field which constitutes the front and back and the upper and lower sides of grip G is a plane, it is in the state where a curved surface or a slope was formed between each plane (the state where it was chamfered).
- the grip G of the controller 10a is provided with a character selector 60a, an audio control operator 60b, and a repeat operator 60c that can function as part of the input / output unit 60 of the keyboard instrument 10. That is, a signal and / or information generated in response to the operation of any of the character selector 60a, the voice control operator 60b, and the repeat operator 60c provided in the controller 10a is the main body (voice generation device) of the keyboard instrument 10. 10b and treated as input signals and / or information by the user.
- the character selector 60a is configured to be operable by the user to designate one or more characters in a predetermined character string (for example, lyrics), and includes a push button type switch as described later.
- the character selector 60a is disposed on a curved surface or a bevel (a chamfered portion) formed between the upper surface and the rear surface of the grip (see FIG. 1B). By arranging the character selector 60 a in this manner, it becomes easy to operate the character selector 60 a with the thumb of the hand holding the grip G.
- the repeat operator 60c is an operator for performing an input related to repeat performance.
- the repeat operator 60c is also a push button type switch, and the repeat operator 60c is disposed on a curved surface or a slope (a chamfered portion) formed between planes forming the upper and the rear of the grip G. (See FIG. 1B).
- the direction in which the grip G extends in the respective buttons Mcf, Mcb, Mpf, and Mpb of the character selector 60a and the button of the repeat operator 60c on the curved surface or inclined surface (the chamfered portion) (FIG. 1A) It is arranged in a line along the left and right direction).
- the voice control operator 60b is configured to be operable by the user to control the state of the voice generated by the voice generation device 10b. As an example, according to the operation of the voice control operator 60b, the pitch of the generated voice can be controlled.
- the voice control operator 60b is disposed on a plane that forms the front of the grip G (see FIG. 1B). As an example, the voice control operator 60b is a long thin film touch sensor, and detects a touch operation position (for example, a one-dimensional position in the longitudinal direction) of a detection target (a finger in this embodiment) on the operation surface It is configured to be able to.
- the voice control operator 60b is disposed at the front of the grip G so that the short side of the rectangular touch sensor is parallel to the vertical direction and the long side of the rectangular is parallel to the horizontal direction (see FIG. 1A). It is attached to the surface.
- the user operates the character selector 60a, the voice control operator 60b and the repeat operator 60c while holding the grip G of the controller 10a with the left hand as shown in FIG. 1B. Specifically, the user holds the grip G of the controller 10a from below with the palm of the left hand, and holds the grip G with the thumb placed backward and the other fingers placed forward.
- the character selector 60a and the repeat operator 60c exist on the curved surface or the slope between the rear surface and the upper surface of the grip G, the character selector 60a and the repeat selector 60c can be easily operated by the thumb as shown in FIG. 1B.
- the repeat operator 60c is disposed.
- the voice control operator 60b when the user grips the grip G, the voice control operator 60b is present on the front face of the grip G, so fingers (forefinger etc.) other than the thumb as shown in FIG. 1B.
- the voice control operator 60b is disposed at a position where it is easy to operate. Therefore, in the present embodiment, when the user operates the character selector 60a or the repeat operator 60c with the thumb while gripping the grip G, the voice control operator 60b is formed at the portion where the other finger is disposed. It will be.
- the user while holding the grip G of the controller 10a with one hand, the user can operate the character selector 60a and the repeat operator 60c with the thumb of the hand, and the other finger of the hand Can operate the voice control operator 60b. Therefore, the voice control operator 60b and the character selector 60 (or the repeat operator 60c) can be easily operated simultaneously with one hand. Furthermore, the operation with respect to the voice control operator 60b with one hand as described above is similar to the operation when pressing the fret of the guitar, and the user performs the voice control operation in the same operation as the operation with the fret of the guitar. By touching the child 60b, the generation mode can be controlled in accordance with the contact position.
- the portion where the hand contacts the controller 10a when the user holds the controller 10a is a flat surface or a curved surface or a slope, and the pointed portion does not touch the hand. Therefore, the user can slide the hand repeatedly along the longitudinal direction (left and right direction shown in FIG. 1A) of the voice control operator 60b without hurting the hand.
- the arrangement for facilitating simultaneous operation of the character selector 60a and the voice control operator 60b is not limited to the example shown in the drawings. The point is that the character selector 60a and the character selector 60a may be operated by a finger holding the grip G It may be arranged such that while one of the voice control operators 60b is being operated, the other can be operated by another finger of the hand.
- FIG. 1C is a block diagram showing a configuration for generating and outputting sound in the keyboard instrument 10.
- the keyboard instrument 10 includes a CPU 20, a non-volatile memory 30, a RAM 40, a pitch selector 50, an input / output unit 60, and a sound output unit 70.
- the sound output unit 70 may include a circuit for outputting sound and a speaker (not shown in FIG. 1A).
- the CPU 20 can execute the program recorded in the non-volatile memory 30 using the RAM 40 as a temporary storage area.
- the voice generation program 30a, the character information 30b, and the voice segment database 30c are recorded in advance.
- the character information 30 b is information of a predefined character string such as a lyric, and includes, for example, information of a plurality of characters constituting the character string and information indicating the order of each character in the character string.
- the character information 30 b is text data in which a code indicating a character is described in accordance with the order.
- the data of the lyrics stored in advance in the non-volatile memory 30 may be for only one song, may be for a plurality of songs, or may be only one phrase of a part of the songs.
- the voice segment database 30c is data for reproducing a person's singing voice, and in the present embodiment, the voice waveform when the voice indicated by the character is pronounced at the reference pitch is collected in advance, and the voice segment database 30c is short. It is generated by dividing into speech segments of a period and converting the waveform data indicating the speech segments into a database. That is, the voice
- the speech segment database 30c includes CV (transition from consonant to vowel), VV (transition from vowel to other vowel), VC (transition from vowel to consonant), etc. It is a collection of waveform data such as a voice transition part (Articulation) and a vowel v unison sound (Stationary). That is, the voice
- These speech segment data are data created on the basis of speech segments extracted from speech waveforms uttered by real people.
- speech segment data to be combined when reproducing an arbitrary character or an arbitrary character string is determined in advance and recorded in the non-volatile memory 30 (not shown) ).
- the CPU 20 refers to the non-volatile memory 30 according to an arbitrary character or character string indicated by the character information 30b, and selects voice segment data to be combined. Then, when the voice segment data selected by the CPU 20 are combined, waveform data for reproducing voice indicated by an arbitrary character or an arbitrary character string is generated.
- the speech segment database 30c may be prepared for various languages, or may be prepared according to gender of the speaker, characteristics of the speech, and the like.
- the waveform data constituting the voice segment database 30c may be data obtained by dividing a sample sequence obtained by sampling the waveform of the voice segment at a predetermined sampling rate into frames of a predetermined time length, or for the data It may be spectrum data (amplitude spectrum and phase spectrum) for each frame obtained by performing FFT (Fast Fourier Transform).
- FFT Fast Fourier Transform
- the CPU 20 can execute the voice generation program 30 a recorded in the non-volatile memory 30.
- the CPU 20 performs processing of the voice generation program 30a to generate a voice signal corresponding to the character defined as the character information 30b at the pitch designated by the user with the pitch selector 50. Do. Then, the CPU 20 outputs, to the sound output unit 70, an instruction to output a sound in accordance with the generated sound signal.
- the sound output unit 70 generates an analog waveform signal for outputting the sound, amplifies the signal, and outputs the sound from the speaker.
- the character string defined in advance is not limited to the lyrics of an existing song previously associated with a predetermined music, and may be any character such as a poem, a rhyme, an ordinary sentence, etc. It may consist of a row. However, in the embodiment described below, it is assumed that the sound corresponding to the character string of the lyrics related to the specific music is generated. As is well known, the progression of notes and the progression of lyrics in the music are associated in advance with a predetermined relationship. In that case, one note may correspond to one syllable, may correspond to multiple syllables, or may be a sustained portion of a syllable generated corresponding to the immediately preceding note.
- the units (numbers) of characters that can be associated with one note also differ.
- one syllable can be expressed by one kana character
- the lyrics can be associated with each note in one kana character unit.
- one syllable is generally represented by one or more characters, and thus it is possible to correspond to individual notes in syllable units instead of one character units.
- the number of characters constituting one syllable may be one or more. The concept derived from this is that the number of characters for specifying the speech to be generated corresponding to one syllable is one or more regardless of the character under any language system. In this sense, in the present invention, one or more characters designated for speech generation is sufficient to identify one or more syllables (including syllables containing only consonants) necessary for speech generation. It is.
- one or more characters in a character string (lyric) are sequentially advanced according to the character progression order in the character string (lyric) in synchronization with the user's pitch designation operation using the pitch selector 50. Is adopted. To that end, each character in the character string (lyric) is divided into groups of one or more characters in association with the individual note to which it is assigned, and each group is ranked in order of progress. 6A and 6B show an example of such character group ranking.
- FIG. 6A shows an example of a Japanese lyric string, and the notes of the corresponding melody are shown in staff notation.
- FIG. 6B shows an example of an English lyric string, and the notes of the corresponding melody are shown in staff notation.
- the character information 30b recorded in the volatile memory 30 includes character data stored readably in such a state that each character in the lyrics character string is divided into a group consisting of one or more characters, and And rank data indicating the rank.
- each character group corresponding to ranks 1, 2, 3, 4, 5, 6, 9, 10 consists of one character
- each character group corresponding to ranks 7 and 8 has multiple characters.
- each character group corresponding to the order 1, 2, 4, 5, 6, 8, 9, 10, 11 is composed of a plurality of characters, and each character group corresponding to the order 3, 7 is 1 It consists of letters.
- FIGS. 3A to 3C show an example of basic voice generation processing executed by the CPU 20.
- FIG. FIG. 3A shows an example of the process of starting voice generation.
- the CPU 20 determines that a song is selected in step S100, proceeds to step S101, and selects the lyrics of the selected song.
- the character information 30 b of the character string is acquired from the non-volatile memory 30 and buffered in the RAM 40.
- the character information 30b of the lyrics character string of the selected song to be buffered and stored in the RAM 40 is character data for each group consisting of one or more characters and rank data indicating the rank of the group And.
- the CPU 20 sets the value of a pointer j (variable) for indicating the order of the character groups to be output to the initial value "1" (step S102).
- the pointer j is maintained in the RAM 40.
- a voice (syllable) indicated by the character data of one character group in the lyrics character string having the rank data corresponding to the value of the pointer j is generated at the next pronunciation opportunity.
- the next sounding opportunity is that the user designates a desired pitch by means of the pitch selector 50.
- the value 1 of the pointer j indicates the character group of the first rank 1 and the value 2 indicates the character group of the second rank 2 from the first.
- FIG. 3B shows an example (key-on process) of a sound generation process of generating a sound in accordance with the pitch designation information.
- the CPU 20 determines in step S103 that the key is on, and proceeds to step S104. Based on the output information of the sensor included in the high selector 50, the operation status (pitch designation information indicating the designated pitch and information indicating the velocity or strength at the time of the operation) is acquired. Next, the CPU 20 generates a voice corresponding to the output target character group designated by the pointer j with the designated pitch and volume strength (step S105).
- the CPU 20 acquires, from the speech segment database 30c, speech segment data for reproducing the speech of the syllable indicated by the output target character group. Further, the CPU 20 executes pitch conversion processing on data corresponding to vowels in the acquired voice segment data, and converts the data into vowel voice segment data having the pitch designated by the pitch selector 50. Do. Furthermore, the CPU 20 substitutes data corresponding to vowels in the voice segment data for reproducing the speech of the syllable indicated by the character group to be output with the vowel voice segment data after the pitch conversion processing. The inverse FFT is performed on data obtained by combining these speech segment data. As a result, a speech signal (digital speech signal in the time domain) that reproduces the speech of the syllable indicated by the character group to be output is synthesized.
- the pitch conversion process may be any process for converting a voice of a specific pitch to a voice of another pitch, and indicates, for example, the pitch designated by the pitch selector 50 and voice segment data. It is possible to obtain a difference from the reference pitch in the voice and move the spectrum distribution indicated by the waveform of the voice segment data in the frequency axis direction by a frequency corresponding to the difference.
- the pitch conversion process can be realized by various other processes, and the process may be performed on the time axis.
- the state (for example, pitch) of the voice to be synthesized is configured to be controlled according to the operation of the voice control operator 60b, which will be described later. explain.
- various aspects (pitch, volume, timbre, etc.) of the voice to be synthesized may be adjustable, and for example, voice control to apply vibrato may be executed. .
- the CPU 20 When an audio signal is generated, the CPU 20 outputs the audio signal to the sound output unit 70. As a result, the sound output unit 70 converts the audio signal into an analog waveform signal, amplifies it, and outputs it. Therefore, the sound output unit 70 outputs the sound of the syllable indicated by the character group to be output, and having the pitch and the volume strength designated by the pitch selector 50.
- step S106 it is determined whether the repeat function is turned on according to the operation of the repeat operator 60c. Details of this will be described later. Normally, the repeat function is off, the process proceeds from step S106 NO to step S120 where the pointer j is incremented by one. Thus, the output target character group designated by the value of the pointer j corresponds to the voice to be generated at the next pronunciation opportunity.
- FIG. 3C shows an example of a process (key-off process) of stopping the generation of the sound generated according to the pitch designation information.
- the CPU 20 determines, based on the output information of the sensor included in the pitch selector 50, whether or not the key-off, that is, the pressing operation on the pitch selector 50 is released (step S107). If it is determined that the key is turned off, the CPU 20 stops (or attenuates) the sound being generated so that the sound signal output from the sound output unit 70 is muted (S108). As a result, the sound output from the sound output unit 70 is stopped.
- 3B and 3C the key-on process and the key-off process
- the CPU 20 causes the sound of the pitch and the intensity designated by the pitch selector 50 to be continuously output for the period designated by the pitch selector 50.
- the CPU 20 increments a variable (pointer j) for specifying the output target character group each time the pitch selector 50 is operated once (step S120).
- a variable (pointer j) for specifying the output target character group each time the pitch selector 50 is operated once (step S120).
- the output target character group is a character group corresponding to the voice to be generated and output according to the next pronunciation instruction, in other words, a character group waiting for generation and output.
- the CPU 20 may display the output target character group and at least the character group in the front or rear order thereof on the display of the input / output unit 60.
- the display of the input / output unit 60 is provided with a lyric display frame for displaying a predetermined number (for example, m) of characters.
- the CPU 20 refers to the RAM 40 and obtains a total of m characters before and / or after the character string including the one character group of the order indicated by the pointer j, and displays these characters as the lyric display of the display. Display in the frame.
- the CPU 20 may display on the display of the input / output unit 60 in order to distinguish the output target character group from other characters.
- the display can be realized by various modes, such as highlighting a character group to be output (flickering, changing color, adding an underline, etc.), or clearly indicating characters before or after the character group to be output (flickering, It is possible to employ changing of color, addition of underline, etc.
- the CPU 20 switches the display content so that the output target character group is always displayed on the display of the input / output unit 60.
- the switching can be realized in various modes, such as scrolling the display content of the display in response to the change of the output character group according to the change of the value of the pointer j, or setting a plurality of characters as a unit. It is possible to adopt such as switching the display content.
- FIG. 2A is a diagram showing a basic example of character-based speech generation.
- the horizontal axis is a time axis
- the vertical axis is an axis indicating a pitch.
- pitches corresponding to several floor names (Do, Re, Mi, Fa, So) in a certain scale are shown on the vertical axis.
- the symbols L 1 , L 2 , L 3 , L 4 , L 4 , L 5 , L are used. 6 and L 7 show. Furthermore, in the graph shown in FIG.
- the generated and output voices are indicated by rectangular regions, and the length in the horizontal direction (time axis direction) in the rectangle corresponds to the output duration of the voice, and is rectangular
- the vertical position corresponds to the pitch.
- the vertical center position of each rectangle corresponds to the pitch of the rectangle.
- the user is in the order of floor names de, re, mi, fa, re, mi.
- the sound generated and output when the pitch selector 50 is operated is shown.
- the output target character group is L 1 , L 2 , L in synchronization with the user operating the pitch selector 50 of de, re, mi, f, de, mi, and mi. 3 , L 4 , L 5 , L 6 , L 7 sequentially change. Therefore, in the example shown in FIG.
- L 1 , L 2 , L 3 , L 4 are synchronized in synchronization with the user operating the pitch selector 50 of D, M, F, D, R, M. , L 5, L 6, voice corresponding to each character group L 7 is indicated de, Re, Mi, fa, de, Re, will be sequentially outputted by Mi pitch.
- the singing voice according to the lyrics in the predetermined order is as intended by the user Can be generated (automatically sing) at a pitch of
- the characters in the character string advance in order in synchronization with the operation on the pitch selector 50 the operation of the pitch selector 50 is incorrect, and the progress of the actual song When different unplanned operations are performed, the singing voice progresses faster or slower than the progress of the song. For example, in the example shown in FIG.
- the character selector 60a is provided in the controller 10a of the keyboard instrument 10 according to the present embodiment, and even if an unexpected operation is performed by the pitch selector 50, the user By operating the character selector 60a, the output target character group to be voice-generated can be returned to the character group according to the original music progression even if there is an erroneous operation. Further, by intentionally operating the combination of the pitch selector 50 and the character selector 60a by the user, it is possible to perform an ad-lib performance in which the original music progression is appropriately deformed.
- the character selector 60a advances a character advance selection button Mcf for advancing the output object character group by one character group (one rank) in accordance with the progression order of the lyrics character string;
- a character backward selection button Mcb for returning one character group (one rank) in the reverse direction to the progression order, and further, for advancing the output object character group in units of one phrase in accordance with the progression order of the lyrics character string
- a phrase forward selection button Mpf and a phrase backward selection button Mpb for returning in phrase units in the reverse direction to the progress order are provided.
- the phrase is a series of a plurality of characters, and the phrase is defined in advance by the delimitation of each phrase being described in the character information 30b of the lyric character string.
- a code for example, a code indicating a space or the like
- the order of the first character group of the immediately preceding phrase and the order of the first character group of the immediately following phrase are easily determined from the phrase definition possessed by the character information 30b of the corresponding lyric character string. Prove.
- the character forward selection button Mcf and the phrase forward selection button Mpf correspond to a forward selector for advancing one or more characters according to the progression order of the character string
- the character backward selection button Mcb and the phrase backward selection button Mpb It corresponds to a reverse selector for returning one or more characters in the reverse direction to the order of progress.
- the character selection process is executed when any selection button of the character selector 60a is operated (when the push operation is canceled after the push operation).
- the CPU 20 determines the operated character selector 60a (step S200). Specifically, when any one of the character forward selection button Mcf, the character reverse selection button Mcb, the phrase forward selection button Mpf, and the phrase reverse selection button Mpb in the character selector 60a is operated, the selection operated from each selection button A signal indicating the type of button and the operation content is output. Therefore, the CPU 20 determines which of the character forward selection button Mcf, the character reverse selection button Mcb, the phrase forward selection button Mpf, and the phrase reverse selection button Mpb the selection button operated based on the signal.
- the CPU 20 advances the rank of the output target character group by one (step S205). That is, the CPU 20 increments the value of the pointer j by one.
- the CPU 20 returns the rank of the output target character group to the first rank (step S210). That is, the CPU 20 decrements the value of the pointer j by one.
- the CPU 20 advances the rank of the output target character group by one phrase (step S215). That is, the CPU 20 refers to the character information 30b of the lyric character string, and searches for the nearest phrase break existing between the character groups in the order (the numerical value indicating the order is large) earlier than the current output object character group. Do. Then, when the division is detected, the CPU 20 sets, in the pointer j, a numerical value indicating the rank of the character group located next to the division (that is, the rank of the first character group of the immediately following phrase).
- the CPU 20 When the operated operator operated is the phrase back selection button Mpb, the CPU 20 returns the order of the output target character group by one phrase (step S220). That is, the CPU 20 refers to the character information 30b of the lyric character string, and searches for the nearest phrase segment existing between the character groups in the rank (the numerical value indicating the rank is small) before the current output target character group. Do. Then, when the break is detected, the CPU 20 sets, in the pointer j, a numerical value indicating the order of the character group located next to the break (that is, the order of the first character group of the immediately preceding phrase).
- the user appropriately operates the pitch selector 50 at an appropriate timing substantially simultaneously with or after the value of the pointer j is appropriately advanced or returned.
- the CPU 20 executes the processing of FIG. 3B, and it is determined YES in the step S103.
- the processing after step S104 described above is executed, and a voice corresponding to the character group (one or more characters) designated according to the operation of the character selector 60a is generated and output.
- Example of correction of erroneous operation As described above, if the order of character groups to be voice-generated can be corrected by the operation of the character selector 60a, even if the pitch designation operation by the pitch selector 50 is erroneous, The order of character groups to be sound-generated can be restored to the proper order along the music progression.
- FIG. 2B shows an example in the case where the operation by the pitch selector 50 is mistaken in the process of playing the same music as FIG. 2A and an example for correcting this erroneous operation. Specifically, in the example illustrated in FIG.
- the user can change the output target character group in units of one character group or in units of phrases in accordance with the order indicated by the character information by operating the character selector 60a. Therefore, the output target character group can be corrected with a simple configuration, and it is also possible to correct the output target character group by blind touch if the user correctly stores the order of the lyrics character strings.
- a voice corresponding to the output target character group is generated in synchronization with the operation on the pitch selector 50, and thereafter the pointer j indicating the order of the output target character group is incremented. Therefore, when a voice is generated in response to an operation on the pitch selector 50, a character group next to a character group relating to the voice becomes an output target.
- the user can grasp the progress of the singing voice by listening to the voice output at the current point, and therefore, when operating any character selector 60a at the current point, the voice of what kind of lyric character is next You can easily figure out if you can For example, if the character reverse selection button Mcb is operated, the character group relating to the currently output voice (or the voice output at the end of the voice whose output is completed) is output by returning the output target character group to one rank. It can be recognized that the output target character group can be set again. Therefore, the user can change the output target character group by operating the character selector 60a based on the information acquired by hearing, and it becomes easier to correct the output target character group by the blind touch.
- the user controls the features of the generated voice by operating the voice control operator 60b (for example, It is configured to be able to adjust the pitch).
- the voice control operator 60b is operated by the user's finger during generation of voice according to the operation of the pitch selector 50
- the CPU 20 acquires the touch position of the finger on the voice control operator 60b.
- the CPU 20 acquires a correction amount associated in advance with the contact position. According to this correction amount, the feature (one of pitch, volume, timbre, etc.) of the voice being generated is controlled.
- FIG. 4A shows an example of adjusting the pitch in accordance with the operation of the voice control operator 60b, as an example of the voice control process executed by the CPU 20 by the voice generation program 30a.
- This voice control process is executed when the voice control operator 60b is operated (when a finger is touched).
- the CPU 20 determines whether a voice is being generated (step S300). For example, between the time when the CPU 20 outputs from the pitch selector 50 a signal indicating that the pressing operation for pitch designation has been performed and immediately before the signal indicating that the pressing operation has been canceled is output, It is determined that voice is being generated. If it is not determined in step S300 that a voice is being generated, there is no voice to be controlled, and the CPU 20 ends the voice control process.
- step S300 If it is determined in step S300 that a voice is being output, the CPU 20 acquires a touch position (step S305). That is, the CPU 20 obtains a signal indicating the touch position output from the voice control operator 60b. Next, the CPU 20 acquires a correction amount (step S310). That is, the CPU 20 uses the pitch designated by the pitch selector 50 as a reference pitch, and acquires the correction amount for the reference pitch based on the touch position of the finger on the voice control operator 60b.
- the voice control operator 60b is a sensor including an elongated rectangular surface as a finger contact detection surface, and is configured to detect at least one-dimensional operation position (linear position).
- the center position in the long side direction of the voice control operator 60b corresponds to the position of the reference pitch, and the contact position is farther from the center position in the long side direction of the voice control operator 60b.
- the correction amount for each touch position is determined in advance so that the correction amount of the pitch becomes large.
- the correction amount in the case of raising the pitch is associated with each contact position on one side across the center position of the voice control operator 60b, and the center position of the voice control operator 60b is interposed.
- a correction amount in the case of lowering the pitch is associated with each contact position on the other side.
- the positions of both ends in the long side direction of the voice control manipulator 60b become the position indicating the highest pitch and the position indicating the lowest pitch.
- the center position in the long side direction of the voice control operator 60b corresponds to the standard pitch, and one end in the long side direction.
- a pitch corresponding to four semitones higher than the reference pitch is associated with the part, and a pitch two semitones higher than the reference pitch is associated with the middle position between the one end and the center position.
- the other end in the long side direction of the voice control operator 60b is associated with a pitch that is four semitones lower than the reference pitch, and the reference sound is located in the middle between the other end and the center position.
- a pitch that is two semitones lower than high is associated.
- the CPU 20 acquires a signal indicating the contact position from the voice control operator 60b, the CPU 20 sets the contact position to the contact position.
- the difference in frequency between the corresponding pitch and the reference pitch is obtained as a correction amount.
- step S315) the CPU 20 sets the pitch designated by the pitch selector 50 during the pressing operation, that is, the pitch during voice generation in step S300 as the reference pitch, and according to the correction amount acquired in step S310, Adjust the pitch of the voice being generated (pitch conversion). Specifically, the CPU 20 outputs the voice at the corrected pitch by, for example, moving the spectrum distribution indicated by the waveform of the voice segment data for outputting the voice at the reference pitch in the direction of the frequency axis. Execute pitch conversion processing to generate speech segment data for Further, the CPU 20 generates an audio signal based on the voice segment data after the pitch conversion process, and outputs the audio signal to the sound output unit 70.
- the sound output unit 70 outputs a sound whose pitch has been corrected.
- the operation of the voice control operator 60b is detected to obtain the correction amount and pitch conversion processing, but the voice control operator 60b is operated before voice output is started.
- the pitch selector 50 is operated, during the generation of the voice according to the operation of the pitch selector 50, the operation of the voice control operator 60b immediately before the generation of the voice is reflected to generate the correction amount. Acquisition or pitch conversion may be performed.
- FIG. 2C combines the singing ad lib performance by the operation of the character selector 60a and the voice control by the operation of the voice control operator 60b in the process of playing the same song as FIG. 2A.
- the operation for the character backward selection button Mcb character selector 60a shows an example performed twice at time t b.
- pitch of pitch selector 50 fa when at time t 4 pitch of pitch selector 50 fa is operated, become sound corresponding to the character group of L 4 is generated at pitch fa, and, output target character group indicated by the pointer j becomes L 5.
- Repeated subsequent operations for the character backward selection button Mcb at time t b is 2 times, accordingly, is returned rank 2 rank output target character group, L 3 is the output target character group.
- the sound corresponding to the character group L 3 is generated in real pitch.
- the generation of the sound corresponding to the character group L 3 is started, the output target character group indicated by the pointer j is changed to L 4 in the following order of L 3.
- the generation period of the speech corresponding to the character group L 3 is, the pushing operation starting pitch selector 50 that specifies the Mi pitch from (time t 5) until the pressing operation is released (time t 6) It is a period.
- audio corresponding to the output target character group L 4 is generated at pitch fa.
- the voices indicated by the character groups L 5 and L 6 should be output at the pitch of the do and the re during the period from time t 5 to time t 7 .
- the voices indicated by the character groups L 3 and L 4 are output at the pitches of mi and fa in the period from time t 5 to time t 7 .
- These character groups and pitch is a character group and pitch at time t 3 ⁇ time t 5 immediately before, that it repeats the same words and pitch even in the period of time t 5 ⁇ time t 7 become.
- Such a performance example is the case where the part that outputs the voices indicated by the character groups L 3 and L 4 at the pitches of mi and fa is the chorus of the song, and a chorus that repeats the same content following the singing of the main vocal is inserted It is used when it swells up in the process of playing. In this way, singing ad-lib can be performed appropriately.
- the state of the singing voice to be repeated in a period of the first time t 5 ⁇ time t 7, singing in the period following time t 3 ⁇ time t 5
- the degree of completion of the performance is improved if the state of the voice is different.
- the user since the keyboard instrument 10 is provided with the voice control operator 60b, the user operates the voice control operator 60b to change the state of the singing voice at the first and second times of the repetitive performance. It can be done easily.
- FIG. 2C vibrato is performed to change the pitch up and down in the period from time t 5 to time t 7 which is a repetitive performance. That is, the user between times t c1 ⁇ time t 6 and, between the time t c2 ⁇ time t 7, the voice control manipulator 60b in a state where the finger touches the voice control operator 60b at the longitudinal center of the The contact position was moved in the lateral direction shown in FIG. 1A centering on the position. In this case, as shown in FIG.
- the user can play the same lyric part sound in different control modes in the first and second times of the repetitive performance.
- the user can flexibly perform the correction of the lyrics and the control of the voice.
- FIG. 2C shows an example in which the user has performed an operation (push operation and release of the push operation) on the character advance selection button Mcf twice at time t f . That is, since the output target character groups by operating the pitch selector 50 at time t 6 has become L 5, time t output target character group if the user operates the character forward selection button Mcf 2 times in f is L 7 It becomes. As a result, the user, by operating the pitch of the pitch selector 50 of the Mi at time t 7, the sound indicated by the letter L 7 is outputted by Mi pitch, returns to the order and pitch of the original lyric character Can make the song progress.
- a character forward selection button Mcf and audio It is easy to simultaneously operate the control operator 60b. That is, in the controller 10a according to the present embodiment, the voice control operator 60b is provided on a plane that constitutes the front surface of the grip seen by the user, and character advancing selection is performed between the planes that constitute the upper and rear of the grip. A button Mcf is provided. Therefore, as shown in FIG. 1B, the user can operate the character advance selection button Mcf with his / her thumb (for example) and the voice control operator 60b with another finger (such as a forefinger) while holding the grip G with one hand. Can be operated at the same time.
- the voice control operator 60b it is possible to play singing voice with more various variations. For example, as in the present embodiment, even if the order of character groups advances each time one pitch selector 50 is operated once, two or more consecutive voices indicated by one character group are displayed. It becomes possible to generate at the pitch of. For example, a song to play a character group L 1 de, Les the character group L 2, character group L 3 a real and files, character group L 4 soil, the character group L 5 Les, a character group L 6 in the order that Mi Assume. In this case, the user operates the pitch selector 50 of De, Re, and Mi at each of times t 1 , t 2 , and t 3 shown in FIG.
- the reference pitch is set by the voice control operator 60b. Perform an operation to raise the pitch of Mi, which is a semitone, that is, a fan.
- the character group L 1 is sound indicates is generated by de pitch
- voice represented by character group L 2 is generated by the record of the pitch
- the sound indicated by the character group L 3 is generated in real pitches After that, it is generated at the pitch of the fa.
- the voice indicated by the character group L 4 is output at the pitch of the de
- voice represented by character group L 6 is output at Mi pitch.
- the user can output the sound indicated by one character group at two or more consecutive pitches.
- the change in pitch from Mi to Fa is continuously performed according to the speed at which the user operates the voice control operator 60b. Therefore, it is possible to generate a voice closer to the voice when singing in human voice.
- the user can use the controller 10a to instruct to generate voice based on characters in various expressions. Furthermore, in the process of the user playing the keyboard instrument 10 and outputting the voice, the user repeats arbitrary lyrics such as chorus and chorus and changes intonation according to the increase of the tune, etc. And the control of the generation mode of. Moreover, when the same lyric is repeated by correction of a lyric, it is also possible to change intonation of the same lyric by controlling a generation
- the user can operate the repeat operator 60c to make it possible to easily perform ad-lib performance of lyrics in more various ways. It is configured to be able to indicate the range (start and end). Specifically, when the push operation on the repeat operator 60c is performed, the CPU 20 starts selection of a character group to be repeated. Further, when the push operation on the repeat operator 60c is released, the CPU 20 ends the selection of the character group to be repeated. The CPU 20 sets the range of the character group selected while the repeat operator 60c is pressed as a repeat target.
- FIG. 4B shows an example in which processing for selecting a repeat target is performed.
- FIG. 2E shows an example in which a character to be repeated is set in the process of playing the same song as that of FIG. 2A and a performance to repeat the character to be repeated is performed. Specifically, in FIG. 2E, the pushing operation on repeat operating element 60c is performed at time t s , the operation for releasing the pushing operation on repeat operating element 60c is performed at time t e , and the repeat operation at time t t An example in which the pressing operation on the child 60c is performed is shown.
- step S400 the CPU 20 determines whether the repeat function is off. That is, the CPU 20 refers to the repeat flag recorded in the RAM 40 and determines whether the repeat function is off.
- step S400 If it is determined in step S400 that the repeat function is off, the CPU 20 turns on the repeat function (step S405). That is, in the present embodiment, when the user performs the pressing operation of the repeat operator 60c in a state where the repeat function is off, the CPU 20 considers that the repeat function is switched on and the repeat flag recorded in the RAM 40 Is rewritten to a value indicating that the repeat function is on. Then, after the repeat function is turned on, the CPU 20 performs processing for setting a range of character groups to be repeated in a period until the push operation of the repeat operator 60c is released.
- the CPU 20 sets the output target character group as the first character group to be repeated (step S410). That is, the CPU 20 acquires the current value of the pointer j and records it in the RAM 40 as a numerical value indicating the order of the first character group to be repeated.
- the output target character group indicated by the current value of the pointer j indicates the voice generated at the next pronunciation opportunity (when the pitch selector 50 is operated next). For example, in the example shown in FIG. 2E, together with the generation of sounds corresponding to the character group L 2 by the operation of the pitch selector 50 at time t 2 is started, the output target character group is updated to L 3. Therefore, the step S410 in response to the pushing operation of the repeat operator 60c at time t s is performed, character groups L 3 indicated by pointer j is set to the first character group in the repeat object.
- the CPU 20 stands by until it is determined that the push operation of the repeat operator 60c is released (step S415). Even during the standby, the CPU 20 executes the above-described sound generation processing (FIGS. 3B and 3C) according to the operation on the pitch selector 50. Therefore, when the pitch selector 50 is operated, the characters to be output advance in the order indicated by the character information 30b in synchronization with the operation. For example, when the pitch selector is operated at time t 3 or t 4 after time t s , the output target character group changes to L 4 or L 5 .
- step S420 the CPU 20 sets the character group one before the output object character group as the last character group to be repeated (step S420). That is, the CPU 20 acquires the current value of the pointer j, and stores a numerical value (j-1) obtained by subtracting 1 from the numerical value in the RAM 40 as a numerical value indicating the order of the last character group to be repeated.
- the immediately preceding character group of the output target character group designated by j-1 corresponds to the currently generated speech or the last generated speech.
- the output target character group is updated to L 5. Therefore, when the step S420 is executed in response to release of the pressing operation of the repeat operator 60c at time t e, the character group L 4 indicating the sound being generated is set as the last character group repeat target. Therefore, in the example shown in FIG. 2E, the first character group to be repeated is L 3 , the last character group to be repeated is L 4 , and the repeat object is set to the range of character groups L 3 and L 4.
- the voice of the character group range to be repeated may be repeated one or more times until the repeat function is turned off. it can. Therefore, it is possible to repeat the voice of the character group range to be repeated as many times as desired by the user. Therefore, as shown in FIG. 2E, the voice indicated by the character to be repeated is repeated once (the same lyric is repeated twice), and not only the performance but also a specific phrase according to the excitement of the audience at the live performance etc. It will be possible to use it over and over again.
- the CPU 20 sets the first character group to be repeated as the output character group (step S425). That is, the CPU 20 refers to the RAM 40 to obtain a numerical value indicating the order of the first character group to be repeated, and sets the numerical value to the pointer j. As a result, when the pitch designation information is acquired next according to the operation of the pitch selector 50, a voice corresponding to the first character group to be repeated is generated.
- step S425 when a pitch designation operation is performed by the pitch selector 50, the CPU 20 goes from YES in step S103 of FIG. 3B to step S104, and indicates the designated pitch. Get specified information. Then, in step S105, the voice corresponding to the character group of the order designated by the pointer j (that is, the first character group to be repeated) is generated at the designated pitch. Next, in step S106, the CPU 20 determines whether the repeat function is on. In this case, since the repeat function is turned on, step S106 is YES, and the process proceeds to step S110.
- step S110 the CPU 20 determines whether the output target character group indicated by the pointer j is the last character group to be repeated. If it is not the last character group to be repeated, the process proceeds from the step S110 NO to the step S120, and the value of the pointer j is incremented by one.
- step S110 is determined as YES, and the process proceeds to step S115.
- step S115 the value of the pointer j is set to the order of the first character group to be repeated.
- step S105 voices from the first to the last character group to be repeated are sequentially generated each time a pitch designation operation is performed, and then the voice generation is repeated by returning to the first character group.
- Such repeat sound generation processing is repeated as long as the repeat function is turned on.
- step S430 the repeat function is turned off. That is, the CPU 20 considers that the repeat function is switched to the off state when the user performs the pressing operation of the repeat operator 60c in the on state of the repeat function, and the repeat function is off in the repeat flag recorded in the RAM 40. Rewrite to a value that indicates that there is.
- the CPU 20 clears the setting of the character group range to be repeated (step S435). That is, the CPU 20 erases from the RAM 40 a numerical value indicating the order of the first character group to be repeated and the last character group.
- the value of the pointer j that is, the output target character group is not changed. Therefore, for example, in the example shown in FIG. 2E, when the repeat function is turned off in response to pressing operation on the repeat operator 60c at time t t , the output target character group remains L 5. .
- the user in the example shown in FIG. 2E, the voice of L 4 sound output when performing the pushing operation for repeat operator 60c listening to, in the example shown in the output target character group (FIG. 2E, L 5 Can be grasped, it is possible to set a desired character group as an output object character group by operating the character selector 60a until the next sound generation timing.
- the user at the timing before time t 7 by operating the character forward selection button Mcf 2 times, it is possible to set the output target character group L 7.
- the user at time t 7 the user by operating the pitch selector 50, audio indicated by the character group L 7 is output.
- the character information 30b if between the character groups L 6 and character group L 7 is set to delimit phrases, the user, a phrase forward selection button Mpf at the timing before time t 7 1 once the operation by, it is possible to set the output target character group L 7.
- the audio corresponding to the character group L 7 is output.
- the CPU 20 may automatically move the value of the pointer j to the original advancing position.
- the CPU 20 may be configured to sequentially advance the reference pointer assuming that the repeat is not performed during the repeat performance according to the pitch designation operation.
- the CPU 20 uses the reference pointer as a pointer output target character group to be indicated by j is identified as the L 7.
- various methods can be adopted for automatically moving the value of the pointer j to the original advancing position when the repeat function is off.
- the CPU 20 counts the number of times of operation of the pitch operator 50 during the period when the repeat function is on, and uses the count value and the value of the pointer j at the start of the repeat to make the pointer j at the end of the repeat You may want to correct the value of.
- FIG. 2F is a diagram showing an example in the case of performing the same performance as FIG. 2C using the repeat operator 60c and the voice control operator 60b. Specifically, in FIG.
- the push operation on repeat operator 60c is performed at time t s
- the operation to release the push operation on repeat operator 60 c is performed at time t e
- vibrato is applied by voice control operating elements 60b
- pushing operation to repeat operator 60c is an example performed at time t t.
- the character groups L 3 and L 4 are repeated twice as in FIG. 2C, and the performance is performed in a second vibrato-applied state.
- the CPU 20 repeatedly generates a voice corresponding to the character group range to be repeated, which is arbitrarily set, in accordance with the operation on the repeat operator 60c. Further, in the present embodiment, it is possible to control the repetition timing of the sound indicated by the character to be repeated according to the user's instruction (the operation of the pitch selector 50).
- the user can designate an arbitrary character range in the lyric character string as a repeat target and repeatedly output the voice, it is possible to repeat the performance at the same place for learning or storage of musical instrument performance. The user can easily specify the repeat range, and can play repeatedly.
- this repeat function not only for musical instrument performance, for example, for acquisition of a foreign language, etc.
- a desired character range is repeated voice generation It can be done. Furthermore, when creating the character information 30b, creation of the second and subsequent character groups to be repeated can be omitted. Therefore, the work of creating the character information 30b can be simplified, and the capacity of the character information 30b can be reduced. Furthermore, in the process of generating speech on the basis of the character information 30b by the speech generation device, it is possible to select and repeat an arbitrary part from a character string of a predetermined order defined as the character information 30b. It is possible to modify the existing order of the columns for speech generation. In addition, various modes are assumed as a mode of correction of the existing order of character strings.
- a mode of performing accompaniment repeating a rising part (sabi) in a song, repeating a scat such as "La La La", or repeating a part having a high degree of difficulty in playing for practice.
- the repeat operator 60c which is one push button type switch. Therefore, it becomes possible to perform designation of a character range to be repeated and control of repeat performance timing by extremely simple operation.
- the user can select the character to be repeated in real time by listening to the voice sequentially output from the sound output unit 70. Therefore, characters to be repeated can be selected without relying on vision.
- FIGS. 5A to 5E are views of various shapes of the grip G of the controller 10a as viewed from one end of the grip G.
- FIG. 5 (A) is a parallelogram
- (B) is a triangle
- (E) is an example of a rectangle
- FIG. 5C may be an example of an ellipse
- FIG. 5D may be an example of a semicircle.
- the shape and size of the cross section do not need to be constant regardless of the cutting position, and the cross sectional area and the curvature may be changed as the main body 10b is approached.
- the character selector 60a or the repeat operator 60c when the character selector 60a or the repeat operator 60c is operated with an arbitrary finger, these operators are formed at positions where the voice control operator 60b can be operated with another finger. good.
- the character selector 60a (or the repeat operator 60c) and the voice control operator 60b are formed in the portion where the finger is disposed when the grip G is held by one hand.
- the character selector 60a (or the repeat operator 60c) and the voice control operator 60b are not on the same plane but on different planes. It is possible to adopt a configuration in which and are formed. With this configuration, erroneous operations on the character selector 60a (or the repeat operator 60c) and the voice control operator 60b are suppressed, and the user can easily perform simultaneous operations on these operators. .
- the character selector 60a (or the repeat operator 60c) and the voice control operator 60b are positioned opposite to each other across the center of gravity of the grip G in order for the user to hold the grip stably with one hand. It is preferable not to exist in two surfaces (for example, the surface which comprises front and back in FIG. 5 (A), (E)). According to this configuration, it is possible to suppress that the user erroneously operates the character selector 60a (or the repeat operator 60c) or the voice control operator 60b in accordance with the operation of gripping the grip G.
- connection aspect of the controller 10a and the main body 10b is not limited to the aspect shown to FIG. 1A.
- the connection position between the controller 10a and the main body 10b is not limited to one, and the controller 10a is configured of a bent columnar member such as a U-shaped member, and both ends of the columnar member are connected to the main body 10b.
- the controller 10 a may be detachable from the keyboard instrument 10. In this case, the operation output of the operation element of the controller 10a is transmitted to the CPU 20 of the main body 10b by wired or wireless communication.
- the application object of the present invention is not limited to the keyboard instrument 10, and may be another type of electronic musical instrument provided with the pitch selector 50.
- it may be a singing voice generation device that automatically sings the lyrics defined by the character information 30b according to the created pitch information (MIDI information etc.), or it is a reproduction device of recording information and recording information.
- the CPU 20 acquires pitch designation information (MIDI event etc.) to be reproduced automatically according to the automatic performance sequence, and the pointer is designated by the pitch designated by the acquired pitch designation information (MIDI event etc.)
- the voice of the character group designated by j may be generated, and the value of the pointer j may be advanced according to the acquired pitch designation information (MIDI event etc.).
- the embodiment for acquiring the pitch specification information of such an automatic performance method when the character selector 60a is operated, the acquisition of the pitch specification information according to the automatic performance sequence is suspended temporarily, and instead, the user operation is performed. Accordingly, the pitch designation information given from the pitch selector 50 is acquired, and the voice of the character group designated by the pointer j changed by the operation of the character selector 60a is generated at the pitch according to the acquired pitch designation information. You may do it.
- the automatic performance is performed according to the change of the value of the pointer j according to the operation of the character selector 60a.
- the pitch selector 50 is unnecessary.
- the instruction means for that purpose is not limited to the pitch selector 50, and may be another appropriate switch or the like.
- the information indicating the pitch of the voice to be generated may be acquired from the automatic sequence data of the music, and the tone generation timing may be designated according to the operation of the appropriate switch by the user.
- the CPU 20 may be configured to obtain the rate of change of the pitch from the reference pitch based on the contact position of the voice control operator 60b, and change the pitch based on the rate of change. Furthermore, in the state where the voice is output at the standard pitch, the CPU 20 regards the position where the user first touched the voice control operator 60b as the standard pitch, and the touch position is determined from the position. When changing, the correction amount of the pitch or the change rate of the pitch may be specified based on the distance between the two positions.
- the correction amount of the pitch per unit distance and the change rate of the pitch are specified in advance.
- the CPU 20 acquires the change distance of the touch position from the position where the user first touched. Further, the CPU 20 specifies the change amount and the change rate by multiplying the value obtained by dividing the change distance by the unit distance by the correction amount of the pitch per unit distance and the change rate of the pitch. Furthermore, in the configuration in which the CPU 20 specifies the correction amount of the pitch and the rate of change of the pitch based on the change (the moving speed etc.) of the contact position of the voice control operator 60b instead of the contact position to the voice control operator 60b It may be.
- the width of the pitch that can be changed by the voice control operator 60b various examples (for example, one octave) can be adopted other than the above example. Further, the width may be variable according to the user's instruction or the like. Furthermore, the control target by the voice control operator 60b may be selectable from among the pitch, the volume, and the properties of the voice (such as the gender of the speaker and the characteristics of the voice) by the user's instruction.
- the voice control operator 60b may be disposed separately from the grip G provided with the character selector 60a without being disposed on the grip G provided with the character selector 60a.
- an existing tone control operator provided in the input / output unit 60 of the main body 10b of the keyboard instrument 10 may be used as the voice control operator 60b.
- the acquisition method of the character information 30b is not limited to that described above.
- the keyboard instrument 10 may be captured from an external recording medium in which the character information 30 b is recorded, through wired or wireless communication.
- the singing voice being sung in real time may be picked up by a microphone and buffered and stored in the RAM 40 in the keyboard instrument 10, and the character information 30b may be acquired based on the buffered audio waveform data.
- the character information 30b defining a previously defined character string such as lyrics may be information that can substantially define the order of a plurality of characters and each character, and the data expression format is text data, image data , Audio data, etc. may consist of any data representation. For example, it may be expressed by code information that indicates a time-series change of syllable corresponding to a character, or may be expressed by time-series audio waveform data. Regardless of which data representation form the character string in the character information 30b is, the point is that each character group (one or more characters corresponding to syllables) in the character string can be individually identified. It may be encoded as such, and may be configured to be able to generate an audio signal according to such a code.
- the voice generation device may be any device having a function of generating voice indicated by the character according to the order of the characters, that is, as long as the pronunciation of the word indicated by the character can be reproduced as voice based on the character information. Good. Furthermore, any of various methods can be arbitrarily adopted as a method for generating speech corresponding to a character group, and the character indicated by the character information is pronounced based on the waveform information indicating the pronunciation of various syllables. It is possible to adopt a configuration for generating a waveform to be used.
- the voice control operator can change the element to be controlled, specify the change from the reference of the element to be controlled, the numerical value of the element to be controlled, the state after the change of the element to be controlled, etc. It may be a sensor capable of The voice control operator is not limited to the touch sensor, but may be a push button switch or the like. Furthermore, in the voice control operator, the generation mode of the voice indicated by the character may be controlled at least for the character selected as the output target by the character selector. However, the present invention is not limited to this. It may be possible to control the speech generation mode independently.
- the character selector 60a may include means for selecting (specifying) other types of characters in addition to the four types of selection buttons Mcf, Mcb, Mpf, and Mpb described above.
- FIG. 7 shows a modification of such a character selector 60a.
- the character selector 60a includes a syllable separation selector Mcs and a syllable integration selector Mcu, in addition to the four types of selection buttons Mcf, Mcb, Mpf, and Mpb described above.
- the syllable separation selector Mcs is for instructing separation and advance of a predetermined one-character group into, for example, two syllables.
- the syllable integration selector Mcu is for instructing integration of, for example, two consecutive character groups to produce one sound. For example, assuming a case of generating voice according to a lyric character string as shown in FIG. 6B, an example of syllable separation and integration control by the syllable separation selector Mcs and the syllable integration selector Mcu is shown in FIG. FIG. 8 shows an example in which the syllable integration selector Mcu is turned on before voice generation of the character group "won" of the order "4" is started.
- the CPU 20 sets an "integration" flag as additional information, and performs syllable integration processing according to the acquisition of the pitch designation information immediately thereafter.
- the process of step S105 (FIG. 3B) is modified to the character group "won" designated by the current value "4" of the pointer j and the characters corresponding to the next rank "5". Integrate the group "der” to generate a polysyllabic voice “wonder” and modify the process of step S120 (FIG. 3B) to add "2" to the current value "4" of the pointer j. And advance the value of the pointer j two places.
- the syllable integration selector Mcu integrates the successive plural character groups included in the predefined character string and instructs to generate the speech of the integrated plural character groups at one sounding timing. Act as an integrated selector.
- FIG. 8 shows an example in which the syllable separation selector Mcs is turned on before voice generation of the character group “why” of the order “6” is started.
- the CPU 20 sets a "separation" flag as additional information according to the turning on of the syllable separation selector Mcs, and performs syllable separation processing according to the acquisition of the pitch designation information immediately thereafter.
- step S105 FIG. 3B
- step S105 is modified to divide the character group "why" designated by the current value "6" of the pointer j into two, "wh-" and "y".
- step S120 The voice of the first syllable (character group) "wh-" separated into syllables and separated is generated, and the process of step S120 (FIG. 3B) is modified to generate the current value "6" of the pointer j.
- the value 0.5 is added, and the value of the pointer j is set to a half value "6.5".
- the voice of the separated second syllable (character group) “y” is generated, and the current value “6.5” of the pointer j is “0”. .5 ", and the value of the pointer j is" 7 ".
- the syllable separation processing is completed, and in accordance with the acquisition of the next pitch designation information, the voice of the character group "I" corresponding to the value "7" of the pointer j is generated.
- the character group to be syllable separated consists of one character (for example, “I”). If it can be separated into two syllables (eg, "a” and "i"), it is separated as such to generate speech. Also, if syllables can not be separated by any means, only the voice of the first syllable should be generated, and at the timing of sounding the second syllable, it may be silent or the voice of the first syllable should be sustained. .
- the syllable separation selector Mcs separates the voice of one character group consisting of one or more characters included in a predefined character string into a plurality of syllables, and generates the voices of the separated syllables at different pronunciation timings. Act as a separate selector to direct
- the CPU 20 advances or retracts the pointer j artificially in response to the operation of the character selector 60a and / or automatically in accordance with the progress of the automatic performance sequence.
- the character string j is configured to specify (acquire) one character group consisting of one or more characters (S102, S105, S200 to S220, etc.), and a function performed by such a CPU 20 is a character string defined in advance. It corresponds to a function as an information acquisition unit that acquires information specifying one or more characters.
- the CPU 20 is configured to generate a voice corresponding to the character group of the order designated by the pointer j at such designated pitch (S105), and the voice thus generated is output as a voice. It is outputted from the unit 70.
- Such a function performed by the CPU 20 corresponds to a function as a voice generation unit that generates a voice corresponding to the designated one or more characters based on the acquired information.
- the CPU 20 performs processing for arbitrarily setting the range of the character string to be repeated according to the user operation by the processing of FIG. 4B.
- Such a function performed by the CPU 20 corresponds to a function as a repeat target reception unit that receives information specifying the sound being generated as a repeat target.
- the CPU 20 sets the order of the first character group to be repeated to the pointer j by the process of step S425 (FIG. 4B), and returns from the end of the repeat object to the beginning to generate voice. It functions to repeat (S105).
- Such a function performed by the CPU 20 corresponds to a function as a repeat control unit that controls the sound generation unit to repeatedly generate the sound designated as the repeat target.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Electrophonic Musical Instruments (AREA)
Abstract
Description
図1Aは、本発明の一実施形態にかかるコントローラ10a及び音声生成装置10bを備えたシステムとしての電子的鍵盤楽器10を模式的に示す図である。鍵盤楽器10は、直方体状の本体10bと角柱状のコントローラ10aとを備えている。鍵盤楽器10の本体10bは、任意の楽音及び音声を電子的に生成する音声生成装置の一例として機能するもので、音高セレクタ50と入出力部60とを備えている。音高セレクタ50は演奏すべき楽音又は音声の音高を指定するためにユーザによって操作される操作子であり、例えば白鍵および黒鍵からなる複数の鍵によって構成される。本実施形態における鍵盤楽器10の本体10bの両端の取付位置P1,P2には、図示しないショルダーストラップが接続されるように構成されている。ユーザは、当該ショルダーストラップを肩にかけた状態で鍵盤楽器10を身体の前方に配置し、片手で音高セレクタ(鍵盤)50を操作することで演奏を行うことができる。図1Aにおいては、このような態様でユーザが鍵盤楽器10を演奏する際にユーザから見た上下左右方向を付記してある。以下、本明細書において言及する方向は、鍵盤楽器10を演奏するユーザから見た上下左右前後の方向を言う。なお、音高セレクタ50は、鍵盤タイプの音高指定用演奏操作子に限らず、任意のタイプの演奏操作子を用いてよく、要は、ユーザの操作に応じて何らかの音高を指定することができるような構成からなるものであればよい。
本発明において、予め規定された文字列とは、予め所定の楽曲と関連づけられている既存の歌の歌詞に限らず、詩、韻文、通常の文章等、任意の文字列からなるものであってよい。しかし、以下説明する実施例においては、特定の楽曲に関連した歌詞の文字列に対応する音声を生成するものとする。公知のように、楽曲における音符進行と歌詞進行とは予め所定の関係に対応づけられている。その場合、1つの音符は、1音節に対応することもあれば、複数音節に対応することもあり、また、直前の音符に対応して発生した或る音節の持続部分であることもある。公知のように、言語のタイプに応じて、1つの音符に対応づけられ得る文字の単位(数)も異なる。例えば、日本語では、一般に、1音節が1つの仮名文字で表現され得るので、歌詞は1つの仮名文字単位で個々の音符に対応づけられ得る。これに対して、その他の多くの言語、例えば英語、においては、一般に、1音節は1又は複数の文字で表現されるので、1文字単位ではなく音節単位で個々の音符に対応づけられることになり、そして、1音節を構成する文字数は1又は複数であり得る。ここから導き出される概念は、どのような言語体系下の文字にあっても、1音節に対応して生成すべき音声を特定するための文字数は1又は複数である、ということである。この意味で、本発明において、音声生成のために指定される1又はまたは複数文字とは、音声生成のために必要な1又は複数の音節(子音のみの音節も含む)を特定するに足るものである。
図3A~3Cは、CPU20によって実行される基本的な音声生成処理の一例を示す。図3Aは、音声生成の開始処理の一例を示す。ユーザが入出力部60を操作して音声生成の対象となる曲を選択すると、CPU20は、ステップS100で曲選択がなされたことを判定して、ステップS101に進み、当該選択された曲の歌詞文字列の文字情報30bを不揮発性メモリ30から取得し、RAM40にバッファ記憶する。なお、RAM40にバッファ記憶される前記選択された曲の歌詞文字列の文字情報30bは、前述したように、1又は複数文字からなる各グループ毎の文字データと、該グループの順位を示す順位データとを含む。次に、CPU20は、出力対象の文字グループの順位を指示するためのポインタj(変数)の値を初期値「1」に設定する(ステップS102)。該ポインタjはRAM40において維持される。該ポインタjの値に対応する順位データを持つ前記歌詞文字列中の1文字グループの前記文字データによって示される音声(音節)が、次の発音機会において生成されることになる。次の発音機会とは、ユーザが音高セレクタ50によって所望の音高を指定することである。例えば、該ポインタjの値1が最初の順位1の文字グループ、値2が最初から2番目の順位2の文字グループを示す。
なお、本実施形態において、CPU20は、出力対象文字グループと、少なくともその前方または後方の順序の文字グループを入出力部60のディスプレイに表示するようにしてよい。例えば、入出力部60のディスプレイには、既定の数(例えばm)の文字を表示するための歌詞表示枠が設けられている。CPU20は、RAM40を参照し、文字列の中からポインタjが示す順位の1文字グループを含む、その前及び/又は後の合計m個の文字を取得し、これらの文字を前記ディスプレイの歌詞表示枠内に表示する。
図2Aは、文字に基づく音声生成の基本的な実例を示す図である。同図2Aにおいて横軸は時間軸であり、縦軸は音高を示す軸である。図2Aにおいては、ある音階におけるいくつかの階名(ド、レ、ミ、ファ、ソ)に相当する音高が縦軸に示されている。また、図2Aにおいては、音声生成されるべき文字列の1番目の順位の文字グループから7番目の順位の文字グループまでを、符号L1,L2,L3,L4,L5,L6,L7で示している。さらに、同図2Aに示すグラフにおいては、生成及び出力される音声を矩形の領域で示しており、矩形における横方向(時間軸方向)の長さが音声の出力継続期間に相当し、矩形の縦方向の位置が音高に相当する。なお、ここでは、各矩形の縦方向の中央の位置が当該矩形の音高に該当する。
そこで、本実施形態にかかる鍵盤楽器10のコントローラ10aには文字セレクタ60aが設けられており、音高セレクタ50で予定外の操作が行われたとしても、ユーザが文字セレクタ60aを操作することによって、誤操作があつとしても、音声生成すべき出力対象文字グループを本来の楽曲進行に従う文字グループに戻すことができるように構成されている。また、ユーザが意図的に音高セレクタ50と文字セレクタ60aを組み合わせて操作することによって、本来の楽曲進行を適宜変形したアドリブ演奏を行うことができるようにもなっている。
図3Dに従い、CPU20が音声生成プログラム30aによって実行する文字選択処理の一例を説明する。文字選択処理は、文字セレクタ60aのいずれかの選択ボタンが操作されると(押し込み操作後に押し込み操作の解除が行われると)実行される。文字選択処理において、CPU20は、操作された文字セレクタ60aを判定する(ステップS200)。具体的には、文字セレクタ60aの中の文字前進選択ボタンMcf、文字後退選択ボタンMcb、フレーズ前進選択ボタンMpf、フレーズ後退選択ボタンMpbのいずれかが操作されると、各選択ボタンから操作した選択ボタンの種類および操作内容を示す信号が出力される。そこで、CPU20は、当該信号に基づいて操作された選択ボタンが文字前進選択ボタンMcf、文字後退選択ボタンMcb、フレーズ前進選択ボタンMpf、フレーズ後退選択ボタンMpbのいずれであるのかを判定する。
このように、音声生成する文字グループの順序を文字セレクタ60aの操作によって修正することができると、音高セレクタ50による音高指定操作を誤った場合であっても、音声生成する文字グループの順序を、楽曲進行に沿う適正な順序に戻すことができる。図2Bは、図2Aと同様の曲を演奏する過程で音高セレクタ50による操作を誤った場合の例及びこの誤操作を修正する例を示している。具体的には、図2Bに示す例においては、時刻t5~t6の期間においてドの音高音高セレクタ50のみを操作すべきところ、ユーザがドの音高の音高セレクタ50に対する押し込み操作を行った直後(時刻t0)において、ドの音高の音高セレクタ50に対する押し込み操作を解除してレの音高の音高セレクタ50の押し込み操作を行ってしまった場合の例を示している。
さらに、本実施形態においては、鍵盤楽器10の楽器としての性能を高めるため、ユーザが音声制御操作子60bを操作することによって、生成される音声の特徴を制御する(例えば音高を調整する)ことができるように構成されている。具体的には、音高セレクタ50の操作に応じた音声の生成中に音声制御操作子60bがユーザの指で操作されると、CPU20は、音声制御操作子60bに対する指の接触位置を取得する。そして、CPU20は、当該接触位置に対して予め対応づけられた補正量を取得する。この補正量に応じて生成中の音声の特徴(音高、音量、音色等のいずれか)を制御する。
図2Cは、図2Aと同様の曲を演奏する過程で、文字セレクタ60aの操作による歌唱アドリブ演奏と音声制御操作子60bの操作による音声制御とを組み合わせて行う例を示している。具体的には、図2Cにおいては、時刻tbにおいて文字セレクタ60aの文字後退選択ボタンMcbに対する操作(押し込みおよび押し込み操作の解除)が2回行われた例を示している。図2Cに示す例においては、時刻t4にてファの音高の音高セレクタ50が操作されると、L4の文字グループに対応する音声がファの音高で生成されるようになり、かつ、ポインタjによって指示される出力対象文字グループはL5となる。その後の時刻tbにおいて文字後退選択ボタンMcbに対する操作が2回繰り返され、これに応じて、出力対象文字グループの順位が2順位戻されて、L3が出力対象文字グループとなる。
本実施形態においては、さらに、より多様な手法で容易に歌詞のアドリブ演奏をできるようにするため、ユーザがリピート操作子60cを操作することによって、リピート対象とする文字グループの範囲(開始および終了)を指示できるように構成されている。具体的には、リピート操作子60cに対する押し込み操作が行われると、CPU20は、リピート対象の文字グループの選択を開始する。また、CPU20は、リピート操作子60cに対する押し込み操作が解除されるとリピート対象の文字グループの選択を終了する。CPU20は、リピート操作子60cが押されている間において選択された文字グループの範囲をリピート対象として設定する。
以上の実施形態は本発明を実施するための一例であり、他にも種々の実施形態を採用可能である。例えば、コントローラ10aの形状は、図1Aに示す態様に限定されない。図5(A)~(E)は、コントローラ10aのグリップGの種々の形状について、該グリップGの一端から見た図である。これらの図に示すように、グリップGの断面は、多角形(図5(A)は平行四辺形、(B)は三角形、(E)は長方形の例)であってもよいし、閉曲線(図5(C)は楕円の例)、直線と曲線で構成される図形(図5(D)は半円の例)であってもよい。むろん、断面の形状や大きさが切断位置によらず一定である必要もなく、本体10bに近づくにつれ断面積や曲率が変化するように構成してもよい。
Claims (21)
- 音声生成装置のためのコントローラであって、前記音声生成装置は、予め規定された文字列中の指定された1または複数文字に対応する音声を生成するように構成されており、前記コントローラは、
前記文字列中の前記1または複数文字を指定するためにユーザによって操作可能なように構成された文字セレクタと、
前記音声生成装置によって生成される前記音声の状態を制御するためにユーザによって操作可能なように構成された音声制御操作子と
を備えるコントローラ。 - ユーザの手によって握られるのに適したグリップを備えており、前記文字セレクタと前記音声制御操作子は、前記グリップ上にそれぞれ設けられて、請求項1のコントローラ。
- 前記文字セレクタと前記音声制御操作子は、前記グリップを握ったユーザの異なる指でそれぞれ操作可能な配置で、前記グリップ上にそれぞれ設けられている、請求項2のコントローラ。
- 前記文字セレクタと前記音声制御操作子の一方が前記ユーザの親指で操作され、他方が前記ユーザの他の指で操作されるように構成されている、請求項3のコントローラ。
- 前記文字セレクタと前記音声制御操作子は、前記グリップの異なる面にそれぞれ配置されている、請求項2乃至4のいずれかのコントローラ。
- 前記音声制御操作子は、操作面に対する接触操作位置を検出するように構成された、タッチセンサからなる、請求項1乃至5のいずれかのコントローラ。
- 前記文字セレクタは、前記文字列の進行順序に従って1または複数文字だけ進めるための前進セレクタと、前記進行順序とは逆向きに1または複数文字だけ戻すための後退セレクタとを含む、請求項1乃至6のいずれかのコントローラ。
- 前記文字セレクタは、前記文字列内に含まれる1または複数文字からなる1文字グループの音声を複数の音節に分離して、分離した各音節の音声を異なる発音タイミングで生成するよう指示するための分離セレクタと、前記文字列内に含まれる連続する複数文字グループを統合して、該統合した複数文字グループの音声を1回の発音タイミングで生成するよう指示するための統合セレクタとを含む、請求項1乃至7のいずれかのコントローラ。
- 前記指定された1または複数文字に対応する音声を繰り返すことを指示するために、ユーザによって操作可能なように構成されたリピート操作子、をさらに備える、請求項1乃至8のいずれかのコントローラ。
- 請求項1乃至9のいずれかのコントローラと、
前記音声生成装置と
を備えるシステム。 - 前記音声生成装置は、
生成すべき音声の音高を指定する音高指定情報を取得し、
前記文字セレクタの操作に従って指定された前記1または複数文字の音声を、前記取得した音高指定情報により指定された音高で合成し、かつ、
合成される前記音声の状態を、前記音声制御操作子の操作に従って制御する、
ように構成されたプロセッサを備える、請求項10のシステム。 - 前記プロセッサは、さらに、
前記音声の合成のために指定されるべき1または複数文字の前記文字列における順位を示すポインタを維持し、かつ、
前記音高指定情報が取得されることに応じて、前記ポインタを順次進める、
ように構成されており、
前記文字セレクタの操作に従って前記1または複数文字を指定することは、前記ポインタによって示される順位を該文字セレクタの操作に応じて前進又は後退させることからなる、請求項11のシステム。 - 前記プロセッサは、前記ポインタが示す順位によって指定される前記1または複数文字の音声を、前記取得した音高指定情報により指定された音高で合成するように構成されている、請求項12のシステム。
- 前記音声生成装置は、さらに、
前記生成すべき音声の音高を指定するためにユーザによって操作可能なように構成された音高セレクタを備える、請求項11乃至13のいずれかのシステム。 - 前記音声生成装置は、電子楽器である、請求項14のシステム。
- コントローラを使用して音声の生成を制御する方法であって、前記コントローラは、予め規定された文字列中の1または複数文字を指定するためにユーザによって操作可能なように構成された文字セレクタと、生成される音声の状態を制御するためにユーザによって操作可能なように構成された音声制御操作子とを備えており、前記方法は、
生成すべき音声の音高を指定する音高指定情報を取得するステップと、
前記文字セレクタから、前記文字列中の1または複数文字を指定するための情報を受け取るステップと、
前記音声制御操作子から、生成されるべき音声の状態を制御するための情報を受け取るステップと、
前記文字セレクタから受け取った前記情報に従って指定される前記1または複数文字の音声を、前記取得した音高指定情報により指定された音高で合成するステップと、
合成される前記音声の状態を、前記音声制御操作子から受け取った前記情報に従って制御するステップと
を備える方法。 - 予め規定された文字列中の1または複数文字を指定する情報を取得する情報取得部と、
前記取得した情報に基づき、前記指定された1または複数文字に対応する音声を生成する音声生成部と、
生成中の音声をリピート対象として指定する情報を受け付けるリピート対象受付部と、
前記リピート対象として指定された前記音声を前記音声生成部が繰り返し生成するように制御するリピート制御部、
として機能するように構成されたプロセッサを備える音声生成装置。 - 前記リピート対象受付部は、1又は複数の音声が時系列的に生成されている間に、ユーザ操作に応じて、前記リピート対象となる最初の音声を指定する情報と、前記リピート対象となる最後の音声を指定する情報とを受け付けるように構成されており、
前記リピート制御部は、時系列的に生成された前記1又は複数の音声のうち、前記指定された最初の音声から最後の音声までをリピート対象として前記音声生成部が繰り返し生成するように制御するように構成されている、請求項17の音声生成装置。 - 前記プロセッサは、さらに、生成すべき音声の音高を指定する音高指定情報を取得する音高指定情報取得部として機能するように構成されており、
前記音声生成部は、前記指定された1または複数文字に対応する音声を、前記取得した音高指定情報により指定された音高で、生成する、請求項17又は18の音声生成装置。 - 予め規定された文字列中の1または複数文字を指定する情報を取得することと、
前記取得した情報に基づき、前記指定された1または複数文字に対応する音声を生成することと、
生成中の音声をリピート対象として指定する情報を受け付けることと、
前記リピート対象として指定された前記音声が繰り返し生成されるように制御すること、
からなる方法。 - 非一過性のコンピュータ読み取り可能な記憶媒体であって、
予め規定された文字列中の1または複数文字を指定する情報を取得することと、
前記取得した情報に基づき、前記指定された1または複数文字に対応する音声を生成することと、
生成中の音声をリピート対象として指定する情報を受け付けることと、
前記リピート対象として指定された前記音声が繰り返し生成されるように制御すること、
からなる音声生成方法を実行するためにプロセッサにより実行可能な命令群を記憶している、記憶媒体。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP15809992.9A EP3159892B1 (en) | 2014-06-17 | 2015-06-10 | Controller and system for voice generation based on characters |
CN201580032999.0A CN106463111B (zh) | 2014-06-17 | 2015-06-10 | 基于字符的话音生成的控制器与系统 |
US15/530,259 US10192533B2 (en) | 2014-06-17 | 2015-06-10 | Controller and system for voice generation based on characters |
JP2016529261A JP6399091B2 (ja) | 2014-06-17 | 2015-06-10 | 文字に基づく音声生成のためのコントローラ及びシステム |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2014124092 | 2014-06-17 | ||
JP2014124091 | 2014-06-17 | ||
JP2014-124091 | 2014-06-17 | ||
JP2014-124092 | 2014-06-17 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015194423A1 true WO2015194423A1 (ja) | 2015-12-23 |
Family
ID=54935410
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/066659 WO2015194423A1 (ja) | 2014-06-17 | 2015-06-10 | 文字に基づく音声生成のためのコントローラ及びシステム |
Country Status (5)
Country | Link |
---|---|
US (1) | US10192533B2 (ja) |
EP (1) | EP3159892B1 (ja) |
JP (2) | JP6399091B2 (ja) |
CN (1) | CN106463111B (ja) |
WO (1) | WO2015194423A1 (ja) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017162176A (ja) * | 2016-03-09 | 2017-09-14 | ヤマハ株式会社 | 歌詞生成装置および歌詞生成方法 |
CN107617214A (zh) * | 2017-09-23 | 2018-01-23 | 深圳市谷粒科技有限公司 | 一种游戏手柄的自动学习控制方法 |
WO2018198379A1 (ja) * | 2017-04-27 | 2018-11-01 | ヤマハ株式会社 | 歌詞表示装置 |
WO2019026233A1 (ja) * | 2017-08-03 | 2019-02-07 | ヤマハ株式会社 | 効果制御装置 |
US20210366448A1 (en) * | 2020-05-21 | 2021-11-25 | Parker J. Wonser | Manual music generator |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6728754B2 (ja) * | 2015-03-20 | 2020-07-22 | ヤマハ株式会社 | 発音装置、発音方法および発音プログラム |
US12020686B2 (en) * | 2017-03-23 | 2024-06-25 | D&M Holdings Inc. | System providing expressive and emotive text-to-speech |
JP6497404B2 (ja) * | 2017-03-23 | 2019-04-10 | カシオ計算機株式会社 | 電子楽器、その電子楽器の制御方法及びその電子楽器用のプログラム |
JP6610714B1 (ja) * | 2018-06-21 | 2019-11-27 | カシオ計算機株式会社 | 電子楽器、電子楽器の制御方法、及びプログラム |
JP6610715B1 (ja) | 2018-06-21 | 2019-11-27 | カシオ計算機株式会社 | 電子楽器、電子楽器の制御方法、及びプログラム |
JP7059972B2 (ja) | 2019-03-14 | 2022-04-26 | カシオ計算機株式会社 | 電子楽器、鍵盤楽器、方法、プログラム |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05341777A (ja) * | 1992-06-08 | 1993-12-24 | Yamaha Corp | 電子楽器のパラメータ制御装置 |
JP2005189454A (ja) * | 2003-12-25 | 2005-07-14 | Casio Comput Co Ltd | テキスト同期音声再生制御装置及びプログラム |
JP2012083569A (ja) * | 2010-10-12 | 2012-04-26 | Yamaha Corp | 歌唱合成制御装置および歌唱合成装置 |
JP2014010190A (ja) * | 2012-06-27 | 2014-01-20 | Yamaha Corp | 歌唱合成を行うための装置およびプログラム |
Family Cites Families (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS6438792A (en) | 1988-02-25 | 1989-02-09 | Yamaha Corp | Electronic keyed instrument |
JP2508937B2 (ja) | 1991-10-25 | 1996-06-19 | ヤマハ株式会社 | 電子鍵盤楽器 |
GB2279172B (en) * | 1993-06-17 | 1996-12-18 | Matsushita Electric Ind Co Ltd | A karaoke sound processor |
KR0135792B1 (ko) * | 1994-12-08 | 1998-05-15 | 김광호 | 노래반주용 곡프로그램 및 이에 적합한 노래반주장치 |
JPH1063287A (ja) * | 1996-08-21 | 1998-03-06 | Brother Ind Ltd | 発音訓練装置 |
US5875427A (en) * | 1996-12-04 | 1999-02-23 | Justsystem Corp. | Voice-generating/document making apparatus voice-generating/document making method and computer-readable medium for storing therein a program having a computer execute voice-generating/document making sequence |
JP3900580B2 (ja) * | 1997-03-24 | 2007-04-04 | ヤマハ株式会社 | カラオケ装置 |
JP3317181B2 (ja) * | 1997-03-25 | 2002-08-26 | ヤマハ株式会社 | カラオケ装置 |
JP3365354B2 (ja) * | 1999-06-30 | 2003-01-08 | ヤマハ株式会社 | 音声信号または楽音信号の処理装置 |
JP2002251185A (ja) * | 2001-02-27 | 2002-09-06 | Casio Comput Co Ltd | 自動演奏装置および自動演奏方法 |
JP4153220B2 (ja) * | 2002-02-28 | 2008-09-24 | ヤマハ株式会社 | 歌唱合成装置、歌唱合成方法及び歌唱合成用プログラム |
JP2004205605A (ja) * | 2002-12-24 | 2004-07-22 | Yamaha Corp | 音声および楽曲再生装置およびシーケンスデータフォーマット |
JP3823930B2 (ja) | 2003-03-03 | 2006-09-20 | ヤマハ株式会社 | 歌唱合成装置、歌唱合成プログラム |
JP4487632B2 (ja) * | 2004-05-21 | 2010-06-23 | ヤマハ株式会社 | 演奏練習装置および演奏練習用コンピュータプログラム |
US20090063152A1 (en) * | 2005-04-12 | 2009-03-05 | Tadahiko Munakata | Audio reproducing method, character code using device, distribution service system, and character code management method |
JP4557919B2 (ja) * | 2006-03-29 | 2010-10-06 | 株式会社東芝 | 音声処理装置、音声処理方法および音声処理プログラム |
JP4735544B2 (ja) | 2007-01-10 | 2011-07-27 | ヤマハ株式会社 | 歌唱合成のための装置およびプログラム |
US7973230B2 (en) * | 2007-12-31 | 2011-07-05 | Apple Inc. | Methods and systems for providing real-time feedback for karaoke |
JP5223433B2 (ja) | 2008-04-15 | 2013-06-26 | ヤマハ株式会社 | 音声データ処理装置およびプログラム |
JP2012150874A (ja) * | 2010-12-28 | 2012-08-09 | Jvc Kenwood Corp | 再生装置、コンテンツ再生方法およびコンピュータプログラム |
US8729374B2 (en) * | 2011-07-22 | 2014-05-20 | Howling Technology | Method and apparatus for converting a spoken voice to a singing voice sung in the manner of a target singer |
JP5821824B2 (ja) | 2012-11-14 | 2015-11-24 | ヤマハ株式会社 | 音声合成装置 |
JP6083764B2 (ja) | 2012-12-04 | 2017-02-22 | 国立研究開発法人産業技術総合研究所 | 歌声合成システム及び歌声合成方法 |
JP6171711B2 (ja) * | 2013-08-09 | 2017-08-02 | ヤマハ株式会社 | 音声解析装置および音声解析方法 |
-
2015
- 2015-06-10 CN CN201580032999.0A patent/CN106463111B/zh active Active
- 2015-06-10 JP JP2016529261A patent/JP6399091B2/ja active Active
- 2015-06-10 WO PCT/JP2015/066659 patent/WO2015194423A1/ja active Application Filing
- 2015-06-10 US US15/530,259 patent/US10192533B2/en active Active
- 2015-06-10 EP EP15809992.9A patent/EP3159892B1/en active Active
-
2018
- 2018-03-09 JP JP2018043095A patent/JP6562104B2/ja active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05341777A (ja) * | 1992-06-08 | 1993-12-24 | Yamaha Corp | 電子楽器のパラメータ制御装置 |
JP2005189454A (ja) * | 2003-12-25 | 2005-07-14 | Casio Comput Co Ltd | テキスト同期音声再生制御装置及びプログラム |
JP2012083569A (ja) * | 2010-10-12 | 2012-04-26 | Yamaha Corp | 歌唱合成制御装置および歌唱合成装置 |
JP2014010190A (ja) * | 2012-06-27 | 2014-01-20 | Yamaha Corp | 歌唱合成を行うための装置およびプログラム |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2017162176A (ja) * | 2016-03-09 | 2017-09-14 | ヤマハ株式会社 | 歌詞生成装置および歌詞生成方法 |
WO2018198379A1 (ja) * | 2017-04-27 | 2018-11-01 | ヤマハ株式会社 | 歌詞表示装置 |
WO2019026233A1 (ja) * | 2017-08-03 | 2019-02-07 | ヤマハ株式会社 | 効果制御装置 |
CN107617214A (zh) * | 2017-09-23 | 2018-01-23 | 深圳市谷粒科技有限公司 | 一种游戏手柄的自动学习控制方法 |
US20210366448A1 (en) * | 2020-05-21 | 2021-11-25 | Parker J. Wonser | Manual music generator |
US12106739B2 (en) * | 2020-05-21 | 2024-10-01 | Parker J Wosner | Manual music generator |
Also Published As
Publication number | Publication date |
---|---|
JP2018112748A (ja) | 2018-07-19 |
CN106463111A (zh) | 2017-02-22 |
JP6399091B2 (ja) | 2018-10-03 |
CN106463111B (zh) | 2020-01-21 |
US20170169806A1 (en) | 2017-06-15 |
US10192533B2 (en) | 2019-01-29 |
EP3159892B1 (en) | 2020-02-12 |
EP3159892A1 (en) | 2017-04-26 |
EP3159892A4 (en) | 2018-03-21 |
JPWO2015194423A1 (ja) | 2017-04-20 |
JP6562104B2 (ja) | 2019-08-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6562104B2 (ja) | 音声生成装置及び方法並びにプログラム | |
US10002604B2 (en) | Voice synthesizing method and voice synthesizing apparatus | |
US6392132B2 (en) | Musical score display for musical performance apparatus | |
US9159307B1 (en) | MIDI controller keyboard, system, and method of using the same | |
JP6465136B2 (ja) | 電子楽器、方法、及びプログラム | |
JP7259817B2 (ja) | 電子楽器、方法及びプログラム | |
JP6728754B2 (ja) | 発音装置、発音方法および発音プログラム | |
US9711133B2 (en) | Estimation of target character train | |
JP7180587B2 (ja) | 電子楽器、方法及びプログラム | |
JP2023118866A (ja) | 電子楽器、方法及びプログラム | |
Ritchie | Before the Chinrest: A Violinist's Guide to the Mysteries of Pre-chinrest Technique and Style | |
JP6977741B2 (ja) | 情報処理装置、情報処理方法、演奏データ表示システム、およびプログラム | |
JP4929604B2 (ja) | 歌データ入力プログラム | |
US20220044662A1 (en) | Audio Information Playback Method, Audio Information Playback Device, Audio Information Generation Method and Audio Information Generation Device | |
JP6809608B2 (ja) | 歌唱音生成装置及び方法、プログラム | |
WO2023153033A1 (ja) | 情報処理方法、プログラム、および情報処理装置 | |
JP7528488B2 (ja) | 電子楽器、方法及びプログラム | |
JPH065455B2 (ja) | 歌唱指導装置 | |
JP6787491B2 (ja) | 音発生装置及び方法 | |
Slatkin | Eight Symphonic Masterworks of the Twentieth Century: A Study Guide for Conductors | |
JP2024089976A (ja) | 電子機器、電子楽器、アドリブ演奏方法及びプログラム | |
JP2022010066A (ja) | 情報処理装置、情報処理方法、およびプログラム | |
WO2018198380A1 (ja) | 歌詞表示装置及び方法 | |
KAMPELA et al. | SARAH BROOKE HORNSBY | |
JP2006251173A (ja) | 楽音制御装置および楽音制御プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15809992 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2016529261 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15530259 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REEP | Request for entry into the european phase |
Ref document number: 2015809992 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2015809992 Country of ref document: EP |