WO2019003350A1 - Singing sound generation device, method and program - Google Patents

Singing sound generation device, method and program Download PDF

Info

Publication number
WO2019003350A1
WO2019003350A1 PCT/JP2017/023786 JP2017023786W WO2019003350A1 WO 2019003350 A1 WO2019003350 A1 WO 2019003350A1 JP 2017023786 W JP2017023786 W JP 2017023786W WO 2019003350 A1 WO2019003350 A1 WO 2019003350A1
Authority
WO
WIPO (PCT)
Prior art keywords
pitch
syllable
instruction
sound
sound generation
Prior art date
Application number
PCT/JP2017/023786
Other languages
French (fr)
Japanese (ja)
Inventor
一輝 柏瀬
桂三 濱野
Original Assignee
ヤマハ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ヤマハ株式会社 filed Critical ヤマハ株式会社
Priority to PCT/JP2017/023786 priority Critical patent/WO2019003350A1/en
Priority to CN201780091652.2A priority patent/CN110709922B/en
Priority to JP2019526039A priority patent/JP6809608B2/en
Publication of WO2019003350A1 publication Critical patent/WO2019003350A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/033Voice editing, e.g. manipulating the voice of the synthesiser

Definitions

  • the present invention relates to a singing sound generation apparatus, method and program for generating a singing sound based on a pronunciation instruction.
  • Patent Document 1 An apparatus that uses voice synthesis technology to synthesize and sing a song according to a performer's performance.
  • This device updates the singing position in the lyrics indicated by the lyrics data according to the performance. That is, according to each performance operation, this device reads the lyrics in advance in the order defined in the lyric data, and generates a single singing voice at the pitch designated by the performance.
  • Patent No. 4735544 gazette
  • a plurality of operation elements may be operated due to a user's mistouch, and a plurality of pitches may be designated.
  • a plurality of sound generations are instructed by mistouch, there is a possibility that the lyrics may be read out unnecessarily. If a singing voice is generated and output corresponding to each of a plurality of instructed pitches, there is a possibility that the audience may clearly recognize a mistouch.
  • An object of the present invention is to provide a singing sound generating apparatus, method, and program capable of determining the sound production pitch of a generated singing sound in a period corresponding to a syllable to be pronounced.
  • a syllable acquisition unit for acquiring syllable information indicating one syllable to be pronounced, and a determination unit for determining a standby time according to syllable information acquired by the syllable acquisition unit
  • the instruction acquiring unit acquires an instruction for sound generation based on the instruction acquiring unit for acquiring the instruction for sound generation or cancellation of sound generation designating the pitch and the instruction for sound generation or cancellation of the pronunciation acquired by the instruction acquiring unit.
  • a determination unit for determining a single tone pitch after a lapse of the waiting time determined by the determination unit, syllable information acquired by the syllable acquisition unit, and the tone pitch determined by the determination unit.
  • a generation unit configured to generate a singing sound based on the above.
  • the tone pitch of the generated singing sound can be determined in a period corresponding to the syllable to be produced.
  • generation apparatus It is a block diagram of an electronic musical instrument. It is a flowchart which shows an example of the flow of a process in case a performance is performed. It is a figure which shows an example of lyric text data. It is a figure which shows an example of the kind of voice
  • FIG. 1 is a schematic view of a singing sound generating apparatus according to an embodiment of the present invention.
  • the singing sound generating apparatus is configured as an electronic musical instrument 100 which is a keyboard instrument as an example, and has a main body 30 and a neck 31.
  • the main body portion 30 has a first surface 30a, a second surface 30b, a third surface 30c, and a fourth surface 30d.
  • the first surface 30a is a keyboard mounting surface on which a keyboard section KB composed of a plurality of keys is disposed.
  • the second surface 30 b is the back surface. Hooks 36 and 37 are provided on the second surface 30 b.
  • a strap (not shown) can be placed between the hooks 36 and 37, and the player usually puts the strap on his shoulder and performs performance such as operating the keyboard KB. Therefore, at the time of use with shoulders, particularly when the scale direction (key arrangement direction) of the keyboard KB is in the left-right direction, the first surface 30a and the keyboard KB face the listener side, the third surface 30c, the fourth The faces 30d face generally downward and upward, respectively.
  • the neck portion 31 is extended from the side of the main body 30.
  • the neck portion 31 is provided with various operators including the advance operator 34 and the return operator 35.
  • a display unit 33 composed of liquid crystal or the like is disposed on the fourth surface 30 d of the main body 30.
  • the electronic musical instrument 100 is a musical instrument that simulates singing in response to an operation on a performance operator.
  • the song simulation is to output a voice simulating a human voice by song synthesis.
  • White keys and black keys are arranged in the order of pitches of the keys of the keyboard section KB, and the keys are associated with different pitches.
  • the user presses a desired key on the keyboard KB.
  • the electronic musical instrument 100 detects a key operated by the user, and produces a singing sound of a pitch corresponding to the operated key.
  • the order of the syllables of the singing voice to be pronounced is predetermined.
  • FIG. 2 is a block diagram of the electronic musical instrument 100.
  • the electronic musical instrument 100 includes a central processing unit (CPU) 10, a timer 11, a read only memory (ROM) 12, a random access memory (RAM) 13, a data storage unit 14, a performance control 15, and other operations.
  • a slave 16 a parameter value setting operator 17, a display unit 33, a sound source 19, an effect circuit 20, a sound system 21, a communication I / F (Interface), and a bus 23.
  • the CPU 10 is a central processing unit that controls the entire electronic musical instrument 100.
  • the timer 11 is a module that measures time.
  • the ROM 12 is a non-volatile memory that stores control programs and various data.
  • the RAM 13 is a volatile memory used as a work area of the CPU 10 and various buffers.
  • the display unit 33 is a display module such as a liquid crystal display panel or an organic EL (Electro-Luminescence) panel.
  • the display unit 33 displays the operation state of the electronic musical instrument 100, various setting screens, a message for the user, and the like.
  • the performance operator 15 is a module that mainly accepts a performance operation that designates a pitch.
  • the keyboard portion KB, the advance operator 34, and the return operator 35 are included in the performance operator 15.
  • the performance operation element 15 when the performance operation element 15 is a keyboard, the performance operation element 15 may be a note on / note off based on sensor on / off corresponding to each key, a key depression strength (speed, velocity), etc.
  • Output performance information may be in the form of a MIDI (musical instrument digital interface) message.
  • the other operator 16 is, for example, an operation module such as an operation button or an operation knob for performing settings other than performance, such as settings relating to the electronic musical instrument 100.
  • the parameter value setting operation unit 17 is an operation module such as an operation button or an operation knob that is mainly used to set parameters for the attribute of the singing voice. Examples of this parameter include harmonics, brightness, resonance, and gender factor.
  • the harmony is a parameter for setting the balance of the harmonic component contained in the voice. Brightness is a parameter for setting the tone of the voice and gives a tone change.
  • the resonance is a parameter for setting timbre and strength of singing voice and musical instrument sound.
  • the gender element is a parameter for setting formants, and changes the thickness and texture of the voice in a feminine or male manner.
  • the external storage device 3 is, for example, an external device connected to the electronic musical instrument 100, and is, for example, a device that stores audio data.
  • the communication I / F 22 is a communication module that communicates with an external device.
  • the bus 23 transfers data between the units in the electronic musical instrument 100.
  • the data storage unit 14 stores singing data 14a.
  • the song data 14a includes lyric text data, a phonological information database, and the like.
  • the lyrics text data is data describing the lyrics.
  • the lyrics of each song are described divided in syllable units. That is, the lyric text data has character information obtained by dividing the lyrics into syllables, and the character information is also information for display corresponding to the syllables.
  • the syllable is a group of sounds output in response to one performance operation.
  • the phonological information database is a database storing speech segment data (syllable information).
  • the voice segment data is data indicating a waveform of voice, and includes, for example, spectrum data of a sample string of the voice segment as waveform data.
  • the speech segment data includes segment pitch data indicating the pitch of the waveform of the speech segment.
  • the lyrics text data and the speech segment data may each be managed by a database.
  • the sound source 19 is a module having a plurality of tone generation channels. Under the control of the CPU 10, one sound generation channel is assigned to the sound source 19 in accordance with the user's performance. In the case of producing a singing voice, the sound source 19 reads voice segment data corresponding to a performance from the data storage unit 14 in the assigned tone generation channel to generate singing voice data.
  • the effect circuit 20 applies the acoustic effect designated by the parameter value setting operator 17 to the singing voice data generated by the sound source 19.
  • the sound system 21 converts the singing sound data processed by the effect circuit 20 into an analog signal by a digital / analog converter. Then, the sound system 21 amplifies the singing sound converted into the analog signal and outputs it from a speaker or the like.
  • FIG. 3 is a flowchart showing an example of the flow of processing when the electronic musical instrument 100 performs a performance.
  • the processing in the case where the user performs the selection of the musical composition and the performance of the selected musical composition will be described. Further, in order to simplify the description, a case where only a single sound is output will be described even if a plurality of keys are simultaneously operated. In this case, only the highest pitch among the pitches of keys operated simultaneously may be processed, or only the lowest pitch may be processed.
  • the processing described below is realized, for example, by the CPU 10 executing a program stored in the ROM 12 or the RAM 13 and functioning as a control unit that controls various components provided in the electronic musical instrument 100.
  • the CPU 10 waits until an operation of selecting a song to be played is received from the user (step S101). Note that if there is no song selection operation even after a certain time has elapsed, the CPU 10 may determine that a song set by default has been selected.
  • the CPU 10 receives the selection of the song, it reads the lyric text data of the song data 14a of the selected song. Then, the CPU 10 sets the cursor position at the top syllable described in the lyric text data (step S102).
  • the cursor is a virtual index indicating the position of the syllable to be pronounced next.
  • the CPU 10 determines whether note-on has been detected based on the operation of the keyboard section KB (step S103).
  • the CPU 10 determines whether the note-off is detected (step S107). On the other hand, when note-on is detected, that is, when a new key depression is detected, the CPU 10 stops the output of the sound if the sound is being output (step S104). Next, the CPU 10 executes an output sound generation process for producing a singing sound according to note-on (step S105).
  • the CPU 10 reads voice segment data of a syllable corresponding to a cursor position, and outputs a sound of a waveform indicated by the read voice segment data at a pitch corresponding to note-on. Specifically, the CPU 10 obtains the difference between the pitch indicated by the segment pitch data included in the voice segment data and the pitch corresponding to the operated key, and the waveform data is obtained by the frequency corresponding to this difference. The spectral distribution shown is moved in the frequency axis direction. Thus, the electronic musical instrument 100 can output a singing sound at the pitch corresponding to the operated key.
  • the CPU 10 updates the cursor position (read position) (step S106), and advances the process to step S107.
  • FIG. 4 is a diagram showing an example of lyrics text data.
  • the lyrics of the five syllables c1 to c5 are described in the lyrics text data.
  • Each character "ha”, “ru”, “yo”, “ko”, "i” indicates one Japanese hiragana character and each character corresponds to one syllable.
  • the CPU 10 updates the cursor position in syllable units.
  • the CPU 10 moves the cursor position to the next syllable c4.
  • the CPU 10 sequentially moves the cursor position to the next syllable in response to the note-on.
  • FIG. 5 is a diagram showing an example of the type of speech segment data.
  • the CPU 10 extracts speech segment data corresponding to syllables from the phonological information database in order to pronounce syllables corresponding to the cursor position.
  • phoneme chain data is data indicating a speech segment when the pronunciation changes, such as "silence (#) to consonant", “consonant to vowel", “vowel to consonant or vowel (of the next syllable)" .
  • the steady part data is data indicating a speech segment when the pronunciation of the vowel continues.
  • the sound source 19 includes voice chain data “# -h” corresponding to “silence ⁇ consonant h”, “consonant h ⁇ vowel a
  • the voice chain data “ha” corresponding to “” and the stationary partial data “a” corresponding to “vowel a” are selected.
  • the CPU 10 operates the singing voice based on the voice chain data “# -h”, the voice chain data “ha,” and the steady part data “a”. Output according to the pitch according to, the velocity according to the operation.
  • the determination of the cursor position and the sounding of the singing sound are performed.
  • step S107 of FIG. 3 if the sound is being output, the CPU 10 stops the output of the sound (step S108), and the process proceeds to step S109.
  • step S109 the CPU 10 determines whether the performance has ended. Then, the CPU 10 returns the process to step S103 when the performance has not ended.
  • step S110 the processing shown in FIG. 3 is ended. Note that the CPU 10 determines whether the performance has ended, for example, whether the last syllable of the selected song has been pronounced, or whether the operation to end the performance has been performed by the other operating element 16 or the like. It can be determined based on
  • FIG. 6 is a schematic view of phoneme type information.
  • the ROM 12 stores phoneme type information Q shown in FIG.
  • the phoneme type information Q designates the type of each phoneme that can be included in the singing voice. Specifically, the phoneme type information Q distinguishes each phoneme constituting the speech segment applied to the speech synthesis process into a first type q1 and a second type q2.
  • the vowel start delay amount differs depending on the syllable.
  • the vowel start delay amount is a delay amount from the start of sound production of the syllable to the start of sound production of the vowel in the syllable, and can also be referred to as the duration (consonant section length) of the consonant located immediately before the vowel.
  • the first type q1 is a type of a phoneme having a relatively large vowel start delay (for example, a phoneme whose vowel start delay exceeds a predetermined threshold)
  • the second type q2 is a vowel start delay having a first type q1
  • the type of a relatively small phoneme for example, a phoneme whose vowel start delay amount falls below a threshold as compared to the phoneme of.
  • half vowels (/ w /, / y /), nasal sounds (/ m /, / n /), tears (/ ts /), frictional noises (/ s /, / f /), roars (/ kja /, Consonants of / kju /, / kjo /, etc. are classified into the first category q1, and vowels (/ a /, / i /, / u /), flowing sounds (/ r /, / l /), plosives ( Phonemes such as / t /, / k /, / p / are classified into the second type q2.
  • a double vowel in which two vowels are made continuous is classified into the first type q1 when there is an accent in the rear vowel, and classified into a second type q2 when there is an accent in the forward vowel. You may adopt the treatment of doing.
  • the CPU 10 refers to the phoneme type information Q, and specifies the phoneme type corresponding to the syllable (the first phoneme when composed of a plurality of phonemes) specified by the read syllable information. For example, the CPU 10 determines which of the first type q1, the second type q2 and the vowel corresponds to the first phoneme of the syllable.
  • the first phoneme can be obtained from phoneme chain data in speech segment data.
  • phoneme type information shown in FIG. 6 may be associated with each of a plurality of syllable information. In that case, the CPU 10 may specify the phoneme type corresponding to the syllable specified by the read syllable information by the phoneme type information associated with the syllable information.
  • the CPU 10 determines the determination time width T based on the phoneme type (for example, determined according to the vowel start delay amount). Further, when the phoneme type can not be specified, the CPU 10 determines the determination time width T based on the waveform data of the volume envelope indicated by the read syllable information.
  • the phoneme type information Q is not stored in the electronic musical instrument 100, and the phoneme type information is not associated with the read syllable information. Case is applicable.
  • the phoneme type information Q is stored, but the phoneme type corresponding to the read syllable information is not registered in the phoneme type information Q, and A case where phoneme type information is not associated with the output syllable information corresponds to the case.
  • FIG. 7 is a diagram showing a volume envelope with respect to an elapsed time when producing syllables.
  • the CPU 10 determines the determination time width T based on, for example, the time from the rising of the waveform of the volume envelope indicated by the read syllable information to the peak.
  • the time from the rise time point t1 to the peak time point t3 of the waveform is tP.
  • the CPU 10 determines the time from the time point t1 to the time point t2 corresponding to a predetermined ratio (for example, 70%) of the time tP as the judgment time width T.
  • a predetermined ratio for example, 70%
  • FIG. 8 is a flowchart of the output sound generation process executed in step S105 of FIG.
  • the CPU 10 reads syllable information (speech segment data) of the syllable corresponding to the cursor position (step S201).
  • the syllable corresponding to the cursor position is the syllable to be pronounced this time. Therefore, the CPU 10 acquires syllable information indicating one syllable to be pronounced from among the plurality of syllable information in a predetermined order.
  • the CPU 10 determines whether the phoneme type can be identified from the read syllable information (step S202).
  • the CPU 10 determines whether the phoneme type corresponding to the syllable specified by the read syllable information is registered in the phoneme type information Q (FIG. 6) or that the phoneme type information corresponds to the syllable information. If it is assigned, it is determined that the phoneme type can be identified. Even when the phoneme type information Q can not be referred to due to some circumstances, it corresponds to the case where the phoneme type corresponding to the syllable specified by the syllable information is not registered in the phoneme type information Q.
  • the CPU 10 specifies syllable information (step S203), and determines the determination time width T based on the specified syllable information (step S204). For example, the CPU 10 determines the determination time width T in accordance with the vowel start delay amount of the syllable to be pronounced (the first phoneme). Specifically, when the phoneme type is vowel, the CPU 10 determines the determination time width T to be 0 since the vowel start delay amount is 0.
  • the CPU 10 determines the determination time width T to be a predetermined value when the phoneme type is the second type q2, and since the vowel start delay amount is relatively large when the phoneme type is the first type q1, the determination time width T The value is determined to be larger than the predetermined value. Thereafter, the process proceeds to step S206.
  • the CPU 10 determines the determination time width T based on the waveform data acquired from the read syllable information (step S205). . That is, as described above, the CPU 10 determines, as the determination time width T, a predetermined ratio (for example, 70%) of the time tP from the rising to the peak in the waveform (FIG. 7) of the volume envelope indicated by the syllable information. . In addition, a predetermined ratio is not limited to the value of an illustration. Further, the determination time width T may be shorter than the time tP, and a value obtained by subtracting a predetermined time from the time tP may be determined as the determination time width T. Thereafter, the process proceeds to step S206.
  • a predetermined ratio for example, 70%
  • step S206 the CPU 10 calculates a determination timing for determining a mistouch based on the note-on detection timing in step S103 and the determined determination time width T.
  • the determination time width T (standby time) is a period provided for determination of an erroneous operation, and the point in time when the determination time width T has elapsed from the note-on detection timing is the determination timing. Note that, when note-on is detected in step S103, clocking is started. When an operation to specify a plurality of pitches is performed between the detection timing of the note-on and the determination timing, the CPU 10 can determine that there is a mistouch.
  • the CPU 10 extracts waveform data from the read syllable information (step S207).
  • the CPU 10 generates and outputs a sound of a waveform indicated by the extracted waveform data at a pitch corresponding to note-on. Specifically, the CPU 10 obtains the difference between the pitch indicated by the segment pitch data included in the speech segment data and the pitch corresponding to the note-on, and the spectrum indicated by the waveform data by the frequency corresponding to this difference. Move the distribution in the frequency axis direction. Thus, the electronic musical instrument 100 can output a singing sound at the pitch corresponding to the operated key.
  • the CPU 10 secures a storage area for storing performance information in the RAM 13 (or the data storage unit 14) (step S209).
  • This storage area is an area for storing information (note-on, note-off) indicating the performance operation of the performance operation element 15 performed until the determination timing comes.
  • the performance operation of the performance operation element 15 corresponds to a sound generation or an instruction to release the sound generation with a specified pitch, and the CPU 10 corresponds to an instruction acquisition unit for acquiring this instruction.
  • the CPU 10 stores information (pitch and timing) on the note-on detected in step S103 in the storage area (step S210). Then, the CPU 10 determines whether or not the determination timing has come (step S211).
  • the CPU 10 determines whether a new performance operation (note on or note off) has been detected (step S212). Then, if a new performance operation is not detected, the CPU 10 returns the process to step S211 (step S212).
  • the CPU 10 stores performance information indicating the new performance operation in the storage area (step S213), and returns the process to step S211. Therefore, each time a new pronunciation instruction or a pronunciation cancellation instruction is detected, the information is accumulated.
  • step S211 when the determination timing has come, the CPU 10 advances the process to step S214.
  • steps S214 to S217 the CPU 10 executes a process of determining a single tone pitch based on the note on detected in step S103 and the note on or note off detected until the determination timing comes. Do.
  • the CPU 10 determines, based on the performance information stored in the storage area, whether to immediately stop the output of sound (step S214). Specifically, the CPU 10 determines that the sound output should be immediately stopped when there is no key (not pressed and released) key in the note on state.
  • step S215 When it is determined that the output of the sound should be immediately stopped (step S215: YES), the CPU 10 stops the sound being output (step S216), and ends the process shown in FIG. On the other hand, when it is determined that the output of the sound should not be stopped immediately, the CPU 10 detects a pitch to be output based on the performance information stored in the storage area (step S217). This determines a single pitch to be output.
  • step S103 the instruction (note-on) for the sound generation detected in the case where the instruction for sound generation for any pitch is not maintained (all keys are in the key release state), and the note-on pitch is Call it the first pitch. Therefore, in step S208, generation of the singing sound is started at the first pitch.
  • the sound generation instruction (note-on) to specify the "second pitch" different from the first pitch is Consider the case.
  • the first pattern is a pattern in which, in order to operate a desired key (for example, C3), another key (for example, D3) is pressed and operated.
  • a desired key for example, C3
  • another key for example, D3
  • the second pattern is a pattern in which an adjacent key (for example, D3) and a desired key are pressed at the same time to operate a desired key (for example, C3).
  • the user releases only the wrong operation key (note off) while maintaining the operation on the desired key among the operations (note on) for the two keys pressed first. Be done.
  • the CPU 10 refers to the performance information stored in the storage area, and when note-on for designating a pitch other than the first note-on is detected by the arrival of the determination timing, it is pressed at the time of arrival of the determination timing. It is determined that the key maintaining the key state is the desired key. Then, the CPU 10 detects a pitch (for example, C3) corresponding to the key determined as the desired key as a pitch to be output. In addition, there may be a mistouch in which the desired key is pressed at first, and then the other key is temporarily pressed and released while the desired key is kept pressed. In this case, the pitch of the first note-on is detected as the pitch to be output.
  • a pitch for example, C3
  • the key pressed first is released and two or more keys different from the first key are pressed and the two or more keys are pressed at the arrival of the determination timing.
  • the pitch of the key pressed last among the two or more keys in the pressed state may be detected as the pitch to be output.
  • the operation to be subjected to the mistouch determination may be limited to the operation of the adjacent key.
  • non-adjacent key operations may be treated as newly performed normal operations. That is, the operation of the remote key may be treated as the operation detected in the process of step S103.
  • the above-described method of determining the desired key is an example, and the CPU 10 may determine the desired key with reference to any information such as note-on, note-off, velocity included in the performance information.
  • the CPU 10 determines whether to correct the currently output pitch (step S218). Specifically, the CPU 10 checks whether the pitch generated in step S208 and the pitch detected in step S217 do not match, and if the two do not match, corrects the pitch being output. Determine that it should be. When the CPU 10 determines that the tone pitch being output should not be corrected, the CPU 10 ends the process shown in FIG. In this case, no pitch correction is made. On the other hand, when it is determined that the pitch being output should be corrected, the CPU 10 adjusts (changes) the pitch of the sound being output to the pitch detected in step S217 (step S219) , And end the process shown in FIG.
  • the CPU 10 changes the pitch by pitch shift, and in this pitch shift, the spectrum distribution indicated by the waveform data is moved in the frequency axis direction by a frequency corresponding to the difference in pitch to be shifted.
  • the CPU 10 may change the pitch gradually, for example, in units of 20 cents.
  • step S214 to S21-7 the process of determining a single tone pitch (steps S214 to S217) can be summarized as follows.
  • a sound generation instruction for instructing a second pitch different from the first pitch is acquired between the sound generation instruction for instructing the first pitch and the arrival of the determination timing, and the second
  • the second pitch is detected as a pitch to be output.
  • the sound production pitch of the generated singing sound is corrected from the first pitch to the second pitch (S219).
  • pitch correction can be performed by re-operation. Since the pitch can be corrected in a short time without muting without changing the syllable (i.e., the lyrics) to be produced in response to the mistouch, it is difficult for the user to feel that there is a mistouch.
  • the sound generation instruction instructing the second pitch different from the first pitch is acquired between the sound generation instruction instructing the first pitch and the arrival of the determination timing. If the sound generation instruction for instructing the first pitch is maintained at the arrival of the determination timing and the sound generation instruction for instructing the second pitch is not maintained, then the sound pitch of the generated singing sound is corrected I will not. As a result, even if a mistake operation is performed within the range of the determination time width T, the original pitch is maintained if it is eliminated before the determination timing comes.
  • CPU 10 determines determination time width T according to the acquired syllable information, and based on the acquired pronunciation or pronunciation cancellation instruction, a single signal is output after determination time width T has elapsed.
  • the tone pitch is determined, and a singing voice is generated based on the acquired syllable information and the determined tone pitch.
  • the tone pitch of the generated singing sound can be determined in a period corresponding to the syllable to be pronounced.
  • the CPU 10 determines the determination time width T based on the phoneme type indicated by the acquired syllable information or determines based on the waveform of the volume envelope indicated by the acquired syllable information.
  • the CPU 10 determines the determination time width T based on the phoneme type indicated by the acquired syllable information or determines based on the waveform of the volume envelope indicated by the acquired syllable information.
  • the CPU 10 determines the determination time width T according to the vowel start delay amount. That is, for a phoneme having a large vowel start delay amount, the CPU 10 sets the determination time width T relatively long to a phoneme having a small vowel start delay amount. As a result, it is possible to secure a long chance of correcting the mistouch for syllables having a large vowel start delay amount, while making the mistouch inconspicuous.
  • the CPU 10 determines the time shorter than the time tP from the rising of the waveform to the peak as the determination time width T decide. As a result, it is possible to correct the generated pitch before the sound production volume rises sufficiently, and make it possible to make the mistouch inconspicuous.
  • the CPU 10 determines a single tone pitch based on the first note-on and the note-on or note-off detected until the determination timing comes. Thus, it is possible to correct the pitch by redoing the operation before the arrival of the determination timing, and it is possible to avoid that the pitch is corrected one by one by a temporary mistake operation.
  • the determination time width T is not limited to defining the absolute time. For example, it may be a relative time according to the tempo.
  • the determination time width T is determined based on the phoneme type, it is not limited to two steps such as the first type q1 and the second type q2, but the determination time width of a value different for each phoneme type or each phoneme type group T may be determined.
  • the performance operation element 15 may have a shape in which strings are arranged side by side like a guitar.
  • the instruction for sound generation or sound generation cancellation designating the pitch is not limited to the configuration in which the input is made with an operator such as a key.
  • the performance control 15 may be a keyboard, a string, or a plurality of buttons for pitch designation displayed on the touch panel. For example, when a pitch is designated by a touch operation on the application, a mistake operation is also assumed in which the pitch changes while the note-on operation continues.
  • the performance control 15 may be one in which operation receiving units for inputting a plurality of different pitches are spatially arranged side by side.
  • the CPU 10 may analyze voice data such as a microphone input and extract a timing of sound generation and a pitch to obtain an instruction of sound generation or sound generation cancellation. Therefore, the apparatus to which the singing sound generating apparatus of the present invention is applied is not limited to the keyboard instrument.
  • Japanese lyrics are exemplified as the lyrics to be sung, but the present invention is not limited to this, and other languages may be used.
  • One letter and one syllable do not necessarily correspond.
  • two letters "ta” (ta) and "" "correspond to one syllable For example, when the English lyrics are “september", it becomes three syllables of "sep" "tem” "ber”.
  • "sep" is one syllable, three characters “s” "e” "p” correspond to one syllable.
  • the CPU 10 sequentially pronounces each syllable at the pitch of the operated key.
  • the storage medium storing the control program represented by the software for achieving the present invention may be read out to the present instrument to achieve the same effect, in which case, the storage medium is read from the storage medium.
  • the program code itself implements the novel functions of the present invention, and the non-transitory computer readable recording medium storing the program code constitutes the present invention.
  • the program code may be supplied via a transmission medium or the like, in which case the program code itself constitutes the present invention.
  • ROMs, floppy disks, hard disks, optical disks, magneto-optical disks, CD-ROMs, CD-Rs, magnetic tapes, non-volatile memory cards, etc. can be used as storage media in these cases.
  • non-transitory computer readable recording medium is a volatile memory (for example, a server or client internal to the computer system when the program is transmitted via a network such as the Internet or a communication line such as a telephone line) It also includes one that holds a program for a fixed time, such as a dynamic random access memory (DRAM).
  • DRAM dynamic random access memory

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Electrophonic Musical Instruments (AREA)

Abstract

Provided is a singing sound generation device capable of defining the sound production pitch of the singing sound to be generated at a period that corresponds to the syllable to be produced. A CPU 10 obtains a sound production or sound production removal instruction specifying pitch, determines the determination duration T according to the obtained syllable information, defines a single sound production pitch after the determination duration T elapses on the basis of the obtained sound production or sound production removal instruction, and generates a singing sound on the basis of the obtained syllable information and the defined sound production pitch.

Description

歌唱音生成装置及び方法、プログラムSing sound generation apparatus and method, program
 本発明は、発音指示に基づき歌唱音を生成する歌唱音生成装置及び方法、プログラムに関する。 The present invention relates to a singing sound generation apparatus, method and program for generating a singing sound based on a pronunciation instruction.
 従来、音声合成技術を用い、演奏者による演奏に応じて歌唱合成して発音する装置が知られている(特許文献1)。この装置は、歌詞データが示す歌詞における歌唱位置を演奏に応じて更新する。すなわちこの装置は、1つ1つの演奏操作に応じて、予め歌詞データにおいて定められている順番で歌詞を読み出し、演奏で指定される音高で単音の歌唱音を発音する。 2. Description of the Related Art Conventionally, there has been known an apparatus that uses voice synthesis technology to synthesize and sing a song according to a performer's performance (Patent Document 1). This device updates the singing position in the lyrics indicated by the lyrics data according to the performance. That is, according to each performance operation, this device reads the lyrics in advance in the order defined in the lyric data, and generates a single singing voice at the pitch designated by the performance.
特許第4735544号公報Patent No. 4735544 gazette
 ところで、鍵盤等の演奏操作子の操作においては、ユーザのミスタッチに起因して複数の操作子が操作され、複数の音高が指定されてしまう場合がある。上記従来の装置で、ミスタッチにより複数の発音が指示されると、意図せずに歌詞が余分に読み出されるおそれがある。仮に、指示された複数の音高のそれぞれに逐一対応して歌唱音が生成・出力されるとすると、聴衆にミスタッチが明確に認識されてしまうおそれがある。 By the way, in operation of performance operation elements, such as a keyboard, a plurality of operation elements may be operated due to a user's mistouch, and a plurality of pitches may be designated. In the above-described conventional apparatus, when a plurality of sound generations are instructed by mistouch, there is a possibility that the lyrics may be read out unnecessarily. If a singing voice is generated and output corresponding to each of a plurality of instructed pitches, there is a possibility that the audience may clearly recognize a mistouch.
 本発明の目的は、生成する歌唱音の発音音高を、発音する音節に応じた期間で確定することができる歌唱音生成装置及び方法、プログラムを提供することである。 An object of the present invention is to provide a singing sound generating apparatus, method, and program capable of determining the sound production pitch of a generated singing sound in a period corresponding to a syllable to be pronounced.
 上記目的を達成するために本発明によれば、発音する1つの音節を示す音節情報を取得する音節取得部と、前記音節取得部により取得された音節情報に応じて待機時間を決定する決定部と、音高を指定した、発音または発音解除の指示を取得する指示取得部と、前記指示取得部により取得された発音または発音解除の指示に基づいて、前記指示取得部により発音の指示が取得されてから前記決定部により決定された待機時間の経過後に、単一の発音音高を確定する確定部と、前記音節取得部により取得された音節情報と前記確定部により確定された発音音高とに基づいて歌唱音を生成する生成部と、を有する歌唱音生成装置が提供される。 In order to achieve the above object, according to the present invention, a syllable acquisition unit for acquiring syllable information indicating one syllable to be pronounced, and a determination unit for determining a standby time according to syllable information acquired by the syllable acquisition unit And the instruction acquiring unit acquires an instruction for sound generation based on the instruction acquiring unit for acquiring the instruction for sound generation or cancellation of sound generation designating the pitch and the instruction for sound generation or cancellation of the pronunciation acquired by the instruction acquiring unit. And a determination unit for determining a single tone pitch after a lapse of the waiting time determined by the determination unit, syllable information acquired by the syllable acquisition unit, and the tone pitch determined by the determination unit. And a generation unit configured to generate a singing sound based on the above.
 なお、上記括弧内の符号は例示である。 In addition, the code in the said parenthesis is an illustration.
 本発明によれば、生成する歌唱音の発音音高を、発音する音節に応じた期間で確定することができる。 According to the present invention, the tone pitch of the generated singing sound can be determined in a period corresponding to the syllable to be produced.
歌唱音生成装置の模式図である。It is a schematic diagram of a song sound production | generation apparatus. 電子楽器のブロック図である。It is a block diagram of an electronic musical instrument. 演奏が行われる場合の処理の流れの一例を示すフローチャートである。It is a flowchart which shows an example of the flow of a process in case a performance is performed. 歌詞テキストデータの一例を示す図である。It is a figure which shows an example of lyric text data. 音声素片データの種類の一例を示す図である。It is a figure which shows an example of the kind of voice | phonetic segment data. 音素種別情報の模式図である。It is a schematic diagram of phoneme type information. 音節を発音する際の経過時間に対する音量エンベロープを示す図である。It is a figure which shows the volume envelope with respect to the elapsed time at the time of producing a syllable. 出力音生成処理のフローチャートである。It is a flowchart of an output sound generation process.
 以下、図面を参照して本発明の実施の形態を説明する。 Hereinafter, embodiments of the present invention will be described with reference to the drawings.
 図1は、本発明の一実施の形態に係る歌唱音生成装置の模式図である。この歌唱音生成装置は、一例として鍵盤楽器である電子楽器100として構成され、本体部30及びネック部31を有する。本体部30は、第1面30a、第2面30b、第3面30c、第4面30dを有する。第1面30aは、複数の鍵から成る鍵盤部KBが配設される鍵盤配設面である。第2面30bは裏面である。第2面30bにはフック36、37が設けられる。フック36、37間には不図示のストラップを架けることができ、演奏者は通常、ストラップを肩に掛けて鍵盤部KBの操作等の演奏を行う。従って、肩掛けした使用時で、特に鍵盤部KBの音階方向(鍵の配列方向)が左右方向となるとき、第1面30a及び鍵盤部KBが聴取者側を向き、第3面30c、第4面30dはそれぞれ概ね下方、上方を向く。ネック部31は本体部30の側部から延設される。ネック部31には、進み操作子34、戻し操作子35をはじめとする各種の操作子が配設される。本体部30の第4面30dには、液晶等で構成される表示ユニット33が配設される。 FIG. 1 is a schematic view of a singing sound generating apparatus according to an embodiment of the present invention. The singing sound generating apparatus is configured as an electronic musical instrument 100 which is a keyboard instrument as an example, and has a main body 30 and a neck 31. The main body portion 30 has a first surface 30a, a second surface 30b, a third surface 30c, and a fourth surface 30d. The first surface 30a is a keyboard mounting surface on which a keyboard section KB composed of a plurality of keys is disposed. The second surface 30 b is the back surface. Hooks 36 and 37 are provided on the second surface 30 b. A strap (not shown) can be placed between the hooks 36 and 37, and the player usually puts the strap on his shoulder and performs performance such as operating the keyboard KB. Therefore, at the time of use with shoulders, particularly when the scale direction (key arrangement direction) of the keyboard KB is in the left-right direction, the first surface 30a and the keyboard KB face the listener side, the third surface 30c, the fourth The faces 30d face generally downward and upward, respectively. The neck portion 31 is extended from the side of the main body 30. The neck portion 31 is provided with various operators including the advance operator 34 and the return operator 35. A display unit 33 composed of liquid crystal or the like is disposed on the fourth surface 30 d of the main body 30.
 電子楽器100は、演奏操作子への操作に応じて歌唱模擬を行う楽器である。ここで、歌唱模擬とは、歌唱合成により人間の声を模擬した音声を出力することである。鍵盤部KBの各鍵は白鍵、黒鍵が音高順に並べられ、各鍵は、それぞれ異なる音高に対応付けられている。電子楽器100を演奏する場合、ユーザは、鍵盤部KBの所望の鍵を押下する。電子楽器100はユーザにより操作された鍵を検出し、操作された鍵に応じた音高の歌唱音を発音する。なお、発音される歌唱音の音節の順番は予め定められている。 The electronic musical instrument 100 is a musical instrument that simulates singing in response to an operation on a performance operator. Here, the song simulation is to output a voice simulating a human voice by song synthesis. White keys and black keys are arranged in the order of pitches of the keys of the keyboard section KB, and the keys are associated with different pitches. When playing the electronic musical instrument 100, the user presses a desired key on the keyboard KB. The electronic musical instrument 100 detects a key operated by the user, and produces a singing sound of a pitch corresponding to the operated key. The order of the syllables of the singing voice to be pronounced is predetermined.
 図2は、電子楽器100のブロック図である。電子楽器100は、CPU(Central Processing Unit)10と、タイマ11と、ROM(Read Only Memory)12と、RAM(Random Access Memory)13と、データ記憶部14と、演奏操作子15と、他操作子16と、パラメータ値設定操作子17と、表示ユニット33と、音源19と、効果回路20と、サウンドシステム21と、通信I/F(Interface)と、バス23と、を備える。CPU10は、電子楽器100全体の制御を行う中央処理装置である。タイマ11は、時間を計測するモジュールである。ROM12は制御プログラムや各種のデータなどを格納する不揮発性のメモリである。RAM13はCPU10のワーク領域及び各種のバッファなどとして使用される揮発性のメモリである。表示ユニット33は、液晶ディスプレイパネル、有機EL(Electro-Luminescence)パネルなどの表示モジュールである。表示ユニット33は、電子楽器100の動作状態、各種設定画面、ユーザに対するメッセージなどを表示する。 FIG. 2 is a block diagram of the electronic musical instrument 100. As shown in FIG. The electronic musical instrument 100 includes a central processing unit (CPU) 10, a timer 11, a read only memory (ROM) 12, a random access memory (RAM) 13, a data storage unit 14, a performance control 15, and other operations. A slave 16, a parameter value setting operator 17, a display unit 33, a sound source 19, an effect circuit 20, a sound system 21, a communication I / F (Interface), and a bus 23. The CPU 10 is a central processing unit that controls the entire electronic musical instrument 100. The timer 11 is a module that measures time. The ROM 12 is a non-volatile memory that stores control programs and various data. The RAM 13 is a volatile memory used as a work area of the CPU 10 and various buffers. The display unit 33 is a display module such as a liquid crystal display panel or an organic EL (Electro-Luminescence) panel. The display unit 33 displays the operation state of the electronic musical instrument 100, various setting screens, a message for the user, and the like.
 演奏操作子15は、主として音高を指定する演奏操作を受け付けるモジュールである。本実施の形態では、鍵盤部KB、進み操作子34、戻し操作子35は演奏操作子15に含まれる。一例として、演奏操作子15が鍵盤である場合、演奏操作子15は、各鍵に対応するセンサのオン/オフに基づくノートオン/ノートオフ、押鍵の強さ(速さ、ベロシティ)などの演奏情報を出力する。この演奏情報は、MIDI(musical instrument digital interface)メッセージ形式であってもよい。他操作子16は、例えば、電子楽器100に関する設定など、演奏以外の設定を行うための操作ボタンや操作つまみなどの操作モジュールである。パラメータ値設定操作子17は、主として歌唱音の属性についてのパラメータを設定するために使用される、操作ボタンや操作つまみなどの操作モジュールである。このパラメータとしては、例えば、和声(Harmonics)、明るさ(Brightness)、共鳴(Resonance)、性別要素(Gender Factor)等がある。和声とは、声に含まれる倍音成分のバランスを設定するパラメータである。明るさとは、声の明暗を設定するパラメータであり、トーン変化を与える。共鳴とは、歌唱音声や楽器音の、音色や強弱を設定するパラメータである。性別要素とは、フォルマントを設定するパラメータであり、声の太さ、質感を女性的、或いは、男性的に変化させる。外部記憶装置3は、例えば、電子楽器100に接続される外部機器であり、例えば、音声データを記憶する装置である。通信I/F22は、外部機器と通信する通信モジュールである。バス23は電子楽器100における各部の間のデータ転送を行う。 The performance operator 15 is a module that mainly accepts a performance operation that designates a pitch. In the present embodiment, the keyboard portion KB, the advance operator 34, and the return operator 35 are included in the performance operator 15. As an example, when the performance operation element 15 is a keyboard, the performance operation element 15 may be a note on / note off based on sensor on / off corresponding to each key, a key depression strength (speed, velocity), etc. Output performance information. This performance information may be in the form of a MIDI (musical instrument digital interface) message. The other operator 16 is, for example, an operation module such as an operation button or an operation knob for performing settings other than performance, such as settings relating to the electronic musical instrument 100. The parameter value setting operation unit 17 is an operation module such as an operation button or an operation knob that is mainly used to set parameters for the attribute of the singing voice. Examples of this parameter include harmonics, brightness, resonance, and gender factor. The harmony is a parameter for setting the balance of the harmonic component contained in the voice. Brightness is a parameter for setting the tone of the voice and gives a tone change. The resonance is a parameter for setting timbre and strength of singing voice and musical instrument sound. The gender element is a parameter for setting formants, and changes the thickness and texture of the voice in a feminine or male manner. The external storage device 3 is, for example, an external device connected to the electronic musical instrument 100, and is, for example, a device that stores audio data. The communication I / F 22 is a communication module that communicates with an external device. The bus 23 transfers data between the units in the electronic musical instrument 100.
 データ記憶部14は、歌唱用データ14aを格納する。歌唱用データ14aには歌詞テキストデータ、音韻情報データベースなどが含まれる。歌詞テキストデータは、歌詞を記述するデータである。歌詞テキストデータには、曲ごとの歌詞が音節単位で区切られて記述されている。すなわち、歌詞テキストデータは歌詞を音節に区切った文字情報を有し、この文字情報は音節に対応する表示用の情報でもある。ここで音節とは、1回の演奏操作に応じて出力する音のまとまりである。音韻情報データベースは、音声素片データ(音節情報)を格納するデータベースである。音声素片データは音声の波形を示すデータであり、例えば、音声素片のサンプル列のスペクトルデータを波形データとして含む。また、音声素片データには、音声素片の波形のピッチを示す素片ピッチデータが含まれる。歌詞テキストデータ、音声素片データは、それぞれ、データベースにより管理されてもよい。 The data storage unit 14 stores singing data 14a. The song data 14a includes lyric text data, a phonological information database, and the like. The lyrics text data is data describing the lyrics. In the lyric text data, the lyrics of each song are described divided in syllable units. That is, the lyric text data has character information obtained by dividing the lyrics into syllables, and the character information is also information for display corresponding to the syllables. Here, the syllable is a group of sounds output in response to one performance operation. The phonological information database is a database storing speech segment data (syllable information). The voice segment data is data indicating a waveform of voice, and includes, for example, spectrum data of a sample string of the voice segment as waveform data. The speech segment data includes segment pitch data indicating the pitch of the waveform of the speech segment. The lyrics text data and the speech segment data may each be managed by a database.
 音源19は、複数の発音チャンネルを有するモジュールである。音源19には、CPU10の制御の基で、ユーザの演奏に応じて1つの発音チャンネルが割り当てられる。歌唱音を発音する場合、音源19は、割り当てられた発音チャンネルにおいて、データ記憶部14から演奏に対応する音声素片データを読み出して歌唱音データを生成する。効果回路20は、音源19が生成した歌唱音データに対して、パラメータ値設定操作子17により指定された音響効果を適用する。サウンドシステム21は、効果回路20による処理後の歌唱音データを、デジタル/アナログ変換器によりアナログ信号に変換する。そして、サウンドシステム21は、アナログ信号に変換された歌唱音を増幅してスピーカなどから出力する。 The sound source 19 is a module having a plurality of tone generation channels. Under the control of the CPU 10, one sound generation channel is assigned to the sound source 19 in accordance with the user's performance. In the case of producing a singing voice, the sound source 19 reads voice segment data corresponding to a performance from the data storage unit 14 in the assigned tone generation channel to generate singing voice data. The effect circuit 20 applies the acoustic effect designated by the parameter value setting operator 17 to the singing voice data generated by the sound source 19. The sound system 21 converts the singing sound data processed by the effect circuit 20 into an analog signal by a digital / analog converter. Then, the sound system 21 amplifies the singing sound converted into the analog signal and outputs it from a speaker or the like.
 図3は、電子楽器100による演奏が行われる場合の処理の流れの一例を示すフローチャートである。ここでは、ユーザにより演奏曲の選択と選択した曲の演奏とが行われる場合の処理について説明する。また、説明を簡単にするため、複数の鍵が同時に操作された場合であっても、単音のみを出力する場合について説明する。この場合、同時に操作された鍵の音高のうち、最も高い音高のみについて処理してもよいし、最も低い音高のみについて処理してもよい。なお、以下に説明する処理は、例えば、CPU10がROM12やRAM13に記憶されたプログラムを実行し、電子楽器100が備える各種構成を制御する制御部として機能することにより実現される。 FIG. 3 is a flowchart showing an example of the flow of processing when the electronic musical instrument 100 performs a performance. Here, the processing in the case where the user performs the selection of the musical composition and the performance of the selected musical composition will be described. Further, in order to simplify the description, a case where only a single sound is output will be described even if a plurality of keys are simultaneously operated. In this case, only the highest pitch among the pitches of keys operated simultaneously may be processed, or only the lowest pitch may be processed. The processing described below is realized, for example, by the CPU 10 executing a program stored in the ROM 12 or the RAM 13 and functioning as a control unit that controls various components provided in the electronic musical instrument 100.
 電源がオンにされると、CPU10は、演奏する曲を選択する操作がユーザから受け付けられるまで待つ(ステップS101)。なお、一定時間経過しても曲選択の操作がない場合は、CPU10は、デフォルトで設定されている曲が選択されたと判断してもよい。CPU10は、曲の選択を受け付けると、選択された曲の歌唱用データ14aの歌詞テキストデータを読み出す。そして、CPU10は、歌詞テキストデータに記述された先頭の音節にカーソル位置を設定する(ステップS102)。ここで、カーソルとは、次に発音する音節の位置を示す仮想的な指標である。次に、CPU10は、鍵盤部KBの操作に基づくノートオンを検出したか否かを判定する(ステップS103)。CPU10は、ノートオンが検出されない場合、ノートオフを検出したか否かを判別する(ステップS107)。一方、ノートオンを検出した場合、すなわち新たな押鍵を検出した場合は、CPU10は、音を出力中であればその音の出力を停止する(ステップS104)。次にCPU10は、ノートオンに応じた歌唱音を発音する出力音生成処理を実行する(ステップS105)。 When the power is turned on, the CPU 10 waits until an operation of selecting a song to be played is received from the user (step S101). Note that if there is no song selection operation even after a certain time has elapsed, the CPU 10 may determine that a song set by default has been selected. When the CPU 10 receives the selection of the song, it reads the lyric text data of the song data 14a of the selected song. Then, the CPU 10 sets the cursor position at the top syllable described in the lyric text data (step S102). Here, the cursor is a virtual index indicating the position of the syllable to be pronounced next. Next, the CPU 10 determines whether note-on has been detected based on the operation of the keyboard section KB (step S103). When the note-on is not detected, the CPU 10 determines whether the note-off is detected (step S107). On the other hand, when note-on is detected, that is, when a new key depression is detected, the CPU 10 stops the output of the sound if the sound is being output (step S104). Next, the CPU 10 executes an output sound generation process for producing a singing sound according to note-on (step S105).
 この出力音生成処理を略説する。CPU10はまず、カーソル位置に対応する音節の音声素片データを読み出し、ノートオンに対応する音高で、読み出した音声素片データが示す波形の音を出力する。具体的には、CPU10は、音声素片データに含まれる素片ピッチデータが示す音高と、操作された鍵に対応する音高との差分を求め、この差分に相当する周波数だけ波形データが示すスペクトル分布を周波数軸方向に移動させる。これにより、電子楽器100は、操作された鍵に対応する音高で歌唱音を出力することができる。次に、CPU10は、カーソル位置(読出位置)を更新し(ステップS106)、処理をステップS107に進める。 This output sound generation process is briefly described. First, the CPU 10 reads voice segment data of a syllable corresponding to a cursor position, and outputs a sound of a waveform indicated by the read voice segment data at a pitch corresponding to note-on. Specifically, the CPU 10 obtains the difference between the pitch indicated by the segment pitch data included in the voice segment data and the pitch corresponding to the operated key, and the waveform data is obtained by the frequency corresponding to this difference. The spectral distribution shown is moved in the frequency axis direction. Thus, the electronic musical instrument 100 can output a singing sound at the pitch corresponding to the operated key. Next, the CPU 10 updates the cursor position (read position) (step S106), and advances the process to step S107.
 ここで、ステップS105、S106の処理に係るカーソル位置の決定と歌唱音の発音について、具体例を用いて説明する。なお、ステップS105の出力音生成処理の詳細については図8でも説明する。まず、カーソル位置の更新について説明する。図4は、歌詞テキストデータの一例を示す図である。図4の例では、歌詞テキストデータには、5つの音節c1~c5の歌詞が記述されている。各字「は」、「る」、「よ」、「こ」、「い」は、日本語のひらがなの1字を示し、各字が1音節に対応する。CPU10は、音節単位でカーソル位置を更新する。例えば、カーソルが音節c3に位置している場合、「よ」に対応する音声素片データをデータ記憶部14から読み出し、「よ」の歌唱音を発音する。CPU10は、「よ」の発音が終了すると、次の音節c4にカーソル位置を移動させる。このように、CPU10は、ノートオンに応じて次の音節にカーソル位置を順次移動させる。 Here, the determination of the cursor position and the sounding of the singing voice according to the processes of steps S105 and S106 will be described using a specific example. The details of the output sound generation process of step S105 will also be described with reference to FIG. First, updating of the cursor position will be described. FIG. 4 is a diagram showing an example of lyrics text data. In the example of FIG. 4, the lyrics of the five syllables c1 to c5 are described in the lyrics text data. Each character "ha", "ru", "yo", "ko", "i" indicates one Japanese hiragana character and each character corresponds to one syllable. The CPU 10 updates the cursor position in syllable units. For example, when the cursor is located at the syllable c3, the voice segment data corresponding to "Y" is read from the data storage unit 14, and the singing voice of "Y" is pronounced. When the sound generation of "Yo" is completed, the CPU 10 moves the cursor position to the next syllable c4. Thus, the CPU 10 sequentially moves the cursor position to the next syllable in response to the note-on.
 次に、歌唱音の発音について説明する。図5は、音声素片データの種類の一例を示す図である。CPU10は、カーソル位置に対応する音節を発音させるために、音韻情報データベースから、音節に対応する音声素片データを抽出する。音声素片データには、音素連鎖データと、定常部分データの2種類が存在する。音素連鎖データとは、「無音(#)から子音」、「子音から母音」、「母音から(次の音節の)子音又は母音」など、発音が変化する際の音声素片を示すデータである。定常部分データは、母音の発音が継続する際の音声素片を示すデータである。例えば、カーソル位置が音節c1の「は(ha)」に設定されている場合、音源19は、「無音→子音h」に対応する音声連鎖データ「#-h」と、「子音h→母音a」に対応する音声連鎖データ「h-a」と、「母音a」に対応する定常部分データ「a」と、を選択する。そして、CPU10は、演奏が開始されて押鍵を検出すると、音声連鎖データ「#-h」、音声連鎖データ「h-a」、定常部分データ「a」に基づく歌唱音を、操作された鍵に応じた音高、操作に応じたベロシティで出力する。このようにして、カーソル位置の決定と歌唱音の発音が実行される。 Next, the pronunciation of the singing sound will be described. FIG. 5 is a diagram showing an example of the type of speech segment data. The CPU 10 extracts speech segment data corresponding to syllables from the phonological information database in order to pronounce syllables corresponding to the cursor position. There are two types of phonetic segment data: phoneme chain data and stationary partial data. The phoneme chain data is data indicating a speech segment when the pronunciation changes, such as "silence (#) to consonant", "consonant to vowel", "vowel to consonant or vowel (of the next syllable)" . The steady part data is data indicating a speech segment when the pronunciation of the vowel continues. For example, when the cursor position is set to “ha” of syllable c 1, the sound source 19 includes voice chain data “# -h” corresponding to “silence → consonant h”, “consonant h → vowel a The voice chain data “ha” corresponding to “” and the stationary partial data “a” corresponding to “vowel a” are selected. Then, when the performance is started and the key depression is detected, the CPU 10 operates the singing voice based on the voice chain data “# -h”, the voice chain data “ha,” and the steady part data “a”. Output according to the pitch according to, the velocity according to the operation. Thus, the determination of the cursor position and the sounding of the singing sound are performed.
 図3のステップS107でノートオフを検出した場合は、CPU10は、音を出力中であればその音の出力を停止して(ステップS108)、処理をステップS109に進める。一方、ノートオフが検出されない場合は、CPU10は処理をステップS109に進める。ステップS109では、CPU10は、演奏が終了したか否かを判別する。そしてCPU10は、演奏を終了していない場合は処理をステップS103に戻す。一方、演奏を終了した場合は、CPU10は、音を出力中であればその音の出力を停止して(ステップS110)、図3に示す処理を終了する。なお、CPU10は、演奏を終了したか否かを、例えば、選択曲の最後尾の音節が発音されたか否か、あるいは他操作子16により演奏を終了する操作が行われた否か、などに基づき判別できる。 When the note-off is detected in step S107 of FIG. 3, if the sound is being output, the CPU 10 stops the output of the sound (step S108), and the process proceeds to step S109. On the other hand, if note-off is not detected, the CPU 10 advances the process to step S109. In step S109, the CPU 10 determines whether the performance has ended. Then, the CPU 10 returns the process to step S103 when the performance has not ended. On the other hand, when the performance is ended, if the sound is being outputted, the CPU 10 stops the output of the sound (step S110), and the processing shown in FIG. 3 is ended. Note that the CPU 10 determines whether the performance has ended, for example, whether the last syllable of the selected song has been pronounced, or whether the operation to end the performance has been performed by the other operating element 16 or the like. It can be determined based on
 図6は、音素種別情報の模式図である。ROM12には図6に示す音素種別情報Qが記憶されている。音素種別情報Qは、歌唱音声に包含され得る各音素の種別を指定する。具体的には、音素種別情報Qは、音声合成処理に適用される音声素片を構成する各音素を第1種別q1と第2種別q2とに区別する。ここで、音節によって母音開始遅延量が異なる。母音開始遅延量は、音節の発音開始から当該音節における母音の発音開始までの遅延量であり、母音の直前に位置する子音の継続長(子音区間長)とも換言され得る。例えば、母音自体(あ[a],い[i],う[M],え[e],お[o])は母音開始遅延量が0である([]内はX-SAMPAに準拠した音素表記)。第1種別q1は、母音開始遅延量が相対的に大きい音素(例えば母音開始遅延量が所定の閾値を上回る音素)の種別であり、第2種別q2は、母音開始遅延量が第1種別q1の音素と比較して相対的に小さい音素(例えば母音開始遅延量が閾値を下回る音素)の種別である。例えば、半母音(/w/,/y/),鼻音(/m/,/n/),破擦音(/ts/),摩擦音(/s/,/f/),拗音(/kja/,/kju/,/kjo/)等の子音は第1種別q1に分類され、母音(/a/,/i/,/u/),流音(/r/,/l/),破裂音(/t/,/k/,/p/)等の音素は第2種別q2に分類される。なお、例えば2個の母音を連続させた2重母音については、後方の母音にアクセントがある場合には第1種別q1に分類し、前方の母音にアクセントがある場合には第2種別q2に分類するという取扱いを採用してもよい。 FIG. 6 is a schematic view of phoneme type information. The ROM 12 stores phoneme type information Q shown in FIG. The phoneme type information Q designates the type of each phoneme that can be included in the singing voice. Specifically, the phoneme type information Q distinguishes each phoneme constituting the speech segment applied to the speech synthesis process into a first type q1 and a second type q2. Here, the vowel start delay amount differs depending on the syllable. The vowel start delay amount is a delay amount from the start of sound production of the syllable to the start of sound production of the vowel in the syllable, and can also be referred to as the duration (consonant section length) of the consonant located immediately before the vowel. For example, vowels themselves (a [a], i [i], u [M], e [o], [o]) have a vowel start delay of 0 (in [] conformed to X-SAMPA Phonetic notation). The first type q1 is a type of a phoneme having a relatively large vowel start delay (for example, a phoneme whose vowel start delay exceeds a predetermined threshold), and the second type q2 is a vowel start delay having a first type q1 The type of a relatively small phoneme (for example, a phoneme whose vowel start delay amount falls below a threshold) as compared to the phoneme of. For example, half vowels (/ w /, / y /), nasal sounds (/ m /, / n /), tears (/ ts /), frictional noises (/ s /, / f /), roars (/ kja /, Consonants of / kju /, / kjo /, etc. are classified into the first category q1, and vowels (/ a /, / i /, / u /), flowing sounds (/ r /, / l /), plosives ( Phonemes such as / t /, / k /, / p / are classified into the second type q2. For example, a double vowel in which two vowels are made continuous is classified into the first type q1 when there is an accent in the rear vowel, and classified into a second type q2 when there is an accent in the forward vowel. You may adopt the treatment of doing.
 CPU10は、音素種別情報Qを参照し、読み出された音節情報で指定される音節(複数の音素で構成される場合は最初の音素)に該当する音素種別を特定する。例えばCPU10は、音節の最初の音素が第1種別q1、第2種別q2、母音の何れに該当するかを判定する。最初の音素は、音声素片データ中の音素連鎖データから取得可能である。なお、歌唱用データ14aには、複数の音節情報のそれぞれに、図6に示す音素種別情報が対応付けられていてもよい。その場合、CPU10は、読み出された音節情報で指定される音節に該当する音素種別を、当該音節情報に対応付けられた音素種別情報により特定してもよい。 The CPU 10 refers to the phoneme type information Q, and specifies the phoneme type corresponding to the syllable (the first phoneme when composed of a plurality of phonemes) specified by the read syllable information. For example, the CPU 10 determines which of the first type q1, the second type q2 and the vowel corresponds to the first phoneme of the syllable. The first phoneme can be obtained from phoneme chain data in speech segment data. In the song data 14a, phoneme type information shown in FIG. 6 may be associated with each of a plurality of syllable information. In that case, the CPU 10 may specify the phoneme type corresponding to the syllable specified by the read syllable information by the phoneme type information associated with the syllable information.
 図8でも述べるが、CPU10は、音素種別を特定(抽出)できた場合、その音素種別に基づいて判定時間幅Tを決定する(例えば、母音開始遅延量に応じて決定)。またCPU10は、音素種別を特定できなかった場合、読み出された音節情報が示す音量エンベロープの波形データに基づき判定時間幅Tを決定する。ここで、音素種別を特定(抽出)できないケースには、例えば、電子楽器100に音素種別情報Qが記憶されておらず、且つ、読み出された音節情報に音素種別情報が対応付けられていないケースが該当する。このほか音素種別を特定(抽出)できないケースには、音素種別情報Qは記憶されているが、読み出された音節情報に対応する音素種別が音素種別情報Qに登録されておらず、且つ読み出された音節情報に音素種別情報が対応付けられていないケース等が該当する。 As described in FIG. 8, when the phoneme type can be specified (extracted), the CPU 10 determines the determination time width T based on the phoneme type (for example, determined according to the vowel start delay amount). Further, when the phoneme type can not be specified, the CPU 10 determines the determination time width T based on the waveform data of the volume envelope indicated by the read syllable information. Here, in the case where the phoneme type can not be specified (extracted), for example, the phoneme type information Q is not stored in the electronic musical instrument 100, and the phoneme type information is not associated with the read syllable information. Case is applicable. Besides, in the case where the phoneme type can not be specified (extracted), the phoneme type information Q is stored, but the phoneme type corresponding to the read syllable information is not registered in the phoneme type information Q, and A case where phoneme type information is not associated with the output syllable information corresponds to the case.
 図7は、音節を発音する際の経過時間に対する音量エンベロープを示す図である。CPU10は、音素種別を特定できなかった場合、例えば、読み出された音節情報が示す音量エンベロープの波形が立ち上がりからピークに達するまでの時間に基づき判定時間幅Tを決定する。波形の立ち上がり時点t1からピーク時点t3までの時間をtPとする。CPU10は、時間tPの所定割合(例えば、70%)に相当する、時点t1から時点t2までの時間を、判定時間幅Tと決定する。 FIG. 7 is a diagram showing a volume envelope with respect to an elapsed time when producing syllables. When the phoneme type can not be specified, the CPU 10 determines the determination time width T based on, for example, the time from the rising of the waveform of the volume envelope indicated by the read syllable information to the peak. The time from the rise time point t1 to the peak time point t3 of the waveform is tP. The CPU 10 determines the time from the time point t1 to the time point t2 corresponding to a predetermined ratio (for example, 70%) of the time tP as the judgment time width T.
 図8は、図3のステップS105で実行される出力音生成処理のフローチャートである。まず、CPU10は、カーソル位置に対応する音節の音節情報(音声素片データ)を読み出す(ステップS201)。カーソル位置に対応する音節が、今回発音する音節である。従ってCPU10は、複数の音節情報の中から、予め定められた順番で、発音する1つの音節を示す音節情報を取得する。次にCPU10は、読み出した音節情報から音素種別を特定可能か否かを判別する(ステップS202)。ここでは上述したように、CPU10は、読み出された音節情報で指定される音節に該当する音素種別が音素種別情報Q(図6)に登録されているか、または音節情報に音素種別情報が対応付けられている場合に、音素種別を特定可能と判別する。なお、何らかの事情により音素種別情報Qを参照できない場合も、音節情報で指定される音節に該当する音素種別が音素種別情報Qに登録されていない場合に該当する。 FIG. 8 is a flowchart of the output sound generation process executed in step S105 of FIG. First, the CPU 10 reads syllable information (speech segment data) of the syllable corresponding to the cursor position (step S201). The syllable corresponding to the cursor position is the syllable to be pronounced this time. Therefore, the CPU 10 acquires syllable information indicating one syllable to be pronounced from among the plurality of syllable information in a predetermined order. Next, the CPU 10 determines whether the phoneme type can be identified from the read syllable information (step S202). Here, as described above, the CPU 10 determines whether the phoneme type corresponding to the syllable specified by the read syllable information is registered in the phoneme type information Q (FIG. 6) or that the phoneme type information corresponds to the syllable information. If it is assigned, it is determined that the phoneme type can be identified. Even when the phoneme type information Q can not be referred to due to some circumstances, it corresponds to the case where the phoneme type corresponding to the syllable specified by the syllable information is not registered in the phoneme type information Q.
 その判別の結果、CPU10は、音素種別を特定可能な場合は音節情報を特定し(ステップS203)、特定した音節情報に基づいて判定時間幅Tを決定する(ステップS204)。例えばCPU10は、発音する音節(最初の音素)の母音開始遅延量に応じて判定時間幅Tを決定する。具体的には、CPU10は、音素種別が母音のときは母音開始遅延量が0であるので、判定時間幅Tを0と決定する。CPU10は、音素種別が第2種別q2のときは判定時間幅Tを所定値に決定し、音素種別が第1種別q1のときは、母音開始遅延量が比較的大きいため、判定時間幅Tを、上記所定値よりも大きい値に決定する。その後、処理はステップS206に進む。 As a result of the determination, when the phoneme type can be specified, the CPU 10 specifies syllable information (step S203), and determines the determination time width T based on the specified syllable information (step S204). For example, the CPU 10 determines the determination time width T in accordance with the vowel start delay amount of the syllable to be pronounced (the first phoneme). Specifically, when the phoneme type is vowel, the CPU 10 determines the determination time width T to be 0 since the vowel start delay amount is 0. The CPU 10 determines the determination time width T to be a predetermined value when the phoneme type is the second type q2, and since the vowel start delay amount is relatively large when the phoneme type is the first type q1, the determination time width T The value is determined to be larger than the predetermined value. Thereafter, the process proceeds to step S206.
 ステップS202の判別の結果、読み出した音節情報から音素種別を特定可能でない場合は、CPU10は、読み出された音節情報から取得される波形データに基づいて判定時間幅Tを決定する(ステップS205)。すなわち、上述したように、CPU10は、音節情報が示す音量エンベロープの波形(図7)における立ち上がりからピークに達するまでの時間tPの所定割合(例えば、70%)を、判定時間幅Tと決定する。なお、所定割合は例示の値に限定されない。また、判定時間幅Tは時間tPより短ければよく、時間tPから所定の時間を差し引いた値を判定時間幅Tとして決定してもよい。その後、処理はステップS206に進む。 As a result of the determination in step S202, when the phoneme type can not be identified from the read syllable information, the CPU 10 determines the determination time width T based on the waveform data acquired from the read syllable information (step S205). . That is, as described above, the CPU 10 determines, as the determination time width T, a predetermined ratio (for example, 70%) of the time tP from the rising to the peak in the waveform (FIG. 7) of the volume envelope indicated by the syllable information. . In addition, a predetermined ratio is not limited to the value of an illustration. Further, the determination time width T may be shorter than the time tP, and a value obtained by subtracting a predetermined time from the time tP may be determined as the determination time width T. Thereafter, the process proceeds to step S206.
 ステップS206では、CPU10は、ステップS103におけるノートオンの検出タイミングと上記決定した判定時間幅Tとに基づいて、ミスタッチを判定するための判定タイミングを算出する。判定時間幅T(待機時間)は、誤操作の判定のために設けられる期間であり、ノートオンの検出タイミングから判定時間幅Tが経過した時点が判定タイミングとなる。なお、ステップS103でノートオンが検出された時に計時が開始されている。ノートオンの検出タイミングから判定タイミングまでの間に音高を複数指定する操作が行われた場合、CPU10は、ミスタッチがあったと判定できる。次に、CPU10は、読み出した音節情報から波形データを抽出する(ステップS207)。次にCPU10は、ノートオンに対応する音高で、抽出した波形データが示す波形の音を生成し、出力する。具体的には、CPU10は、音声素片データに含まれる素片ピッチデータが示す音高と、ノートオンに対応する音高との差分を求め、この差分に相当する周波数だけ波形データが示すスペクトル分布を周波数軸方向に移動させる。これにより、電子楽器100は、操作された鍵に対応する音高で歌唱音を出力することができる。 In step S206, the CPU 10 calculates a determination timing for determining a mistouch based on the note-on detection timing in step S103 and the determined determination time width T. The determination time width T (standby time) is a period provided for determination of an erroneous operation, and the point in time when the determination time width T has elapsed from the note-on detection timing is the determination timing. Note that, when note-on is detected in step S103, clocking is started. When an operation to specify a plurality of pitches is performed between the detection timing of the note-on and the determination timing, the CPU 10 can determine that there is a mistouch. Next, the CPU 10 extracts waveform data from the read syllable information (step S207). Next, the CPU 10 generates and outputs a sound of a waveform indicated by the extracted waveform data at a pitch corresponding to note-on. Specifically, the CPU 10 obtains the difference between the pitch indicated by the segment pitch data included in the speech segment data and the pitch corresponding to the note-on, and the spectrum indicated by the waveform data by the frequency corresponding to this difference. Move the distribution in the frequency axis direction. Thus, the electronic musical instrument 100 can output a singing sound at the pitch corresponding to the operated key.
 次にCPU10は、演奏情報を蓄積するための記憶領域をRAM13(またはデータ記憶部14でもよい)に確保する(ステップS209)。この記憶領域は、判定タイミングが到来するまでになされた演奏操作子15の演奏操作を示す情報(ノートオン、ノートオフ)を記憶するための領域である。演奏操作子15の演奏操作は、音高を指定した、発音または発音解除の指示に該当し、CPU10は、この指示を取得する指示取得部に該当する。次にCPU10は、ステップS103で検出したノートオンに関する情報(音高とタイミング)を記憶領域に記憶する(ステップS210)。そしてCPU10は、判定タイミングが到来したか否かを判別する(ステップS211)。 Next, the CPU 10 secures a storage area for storing performance information in the RAM 13 (or the data storage unit 14) (step S209). This storage area is an area for storing information (note-on, note-off) indicating the performance operation of the performance operation element 15 performed until the determination timing comes. The performance operation of the performance operation element 15 corresponds to a sound generation or an instruction to release the sound generation with a specified pitch, and the CPU 10 corresponds to an instruction acquisition unit for acquiring this instruction. Next, the CPU 10 stores information (pitch and timing) on the note-on detected in step S103 in the storage area (step S210). Then, the CPU 10 determines whether or not the determination timing has come (step S211).
 その判別の結果、判定タイミングが到来していない場合は、CPU10は、新たな演奏操作(ノートオンまたはノートオフ)が検出されたか否かを判別する(ステップS212)。そしてCPU10は、新たな演奏操作が検出されない場合は、処理をステップS211に戻す(ステップS212)。また、新たな演奏操作が検出された場合は、CPU10は、新たな演奏操作を示す演奏情報を記憶領域に記憶して(ステップS213)、処理をステップS211に戻す。従って、新たな発音指示または発音解除指示が検出される毎に、その情報が蓄積される。 As a result of the determination, if the determination timing has not arrived, the CPU 10 determines whether a new performance operation (note on or note off) has been detected (step S212). Then, if a new performance operation is not detected, the CPU 10 returns the process to step S211 (step S212). When a new performance operation is detected, the CPU 10 stores performance information indicating the new performance operation in the storage area (step S213), and returns the process to step S211. Therefore, each time a new pronunciation instruction or a pronunciation cancellation instruction is detected, the information is accumulated.
 ステップS211の判別の結果、判定タイミングが到来した場合は、CPU10は、処理をステップS214に進める。CPU10は、ステップS214~S217で、ステップS103で検出されたノートオンと、判定タイミングの到来までに検出されたノートオンまたはノートオフとに基づいて、単一の発音音高を確定する処理を実行する。まず、CPU10は、記憶領域に蓄積された演奏情報に基づいて、音の出力を直ちに停止すべきか否かを判別する(ステップS214)。具体的には、CPU10は、ノートオン状態の(押下され離操作されていない)鍵が存在していない場合に、音の出力を直ちに停止すべきと判定する。音の出力を直ちに停止すべきと判定した場合(ステップS215:YES)、CPU10は、出力中の音を停止し(ステップS216)、図8に示す処理を終了する。一方、音の出力を直ちに停止すべきでないと判定した場合は、CPU10は、記憶領域に記憶された演奏情報に基づいて、出力すべき音高を検出する(ステップS217)。これにより、出力すべき単一の音高が確定する。 As a result of the determination in step S211, when the determination timing has come, the CPU 10 advances the process to step S214. In steps S214 to S217, the CPU 10 executes a process of determining a single tone pitch based on the note on detected in step S103 and the note on or note off detected until the determination timing comes. Do. First, the CPU 10 determines, based on the performance information stored in the storage area, whether to immediately stop the output of sound (step S214). Specifically, the CPU 10 determines that the sound output should be immediately stopped when there is no key (not pressed and released) key in the note on state. When it is determined that the output of the sound should be immediately stopped (step S215: YES), the CPU 10 stops the sound being output (step S216), and ends the process shown in FIG. On the other hand, when it is determined that the output of the sound should not be stopped immediately, the CPU 10 detects a pitch to be output based on the performance information stored in the storage area (step S217). This determines a single pitch to be output.
 ここで、出力すべき音高の検出について具体的に説明する。まず、ステップS103では、いずれの音高の発音の指示も維持されていない場合(全鍵が離鍵状態)において検出された発音の指示(ノートオン)であり、このノートオンの音高を「第1の音高」と呼称する。従って、ステップS208では第1の音高で歌唱音の生成が開始される。また、第1の音高を指定する発音の指示の後で、判定タイミングの到来前に、第1の音高とは異なる「第2の音高」を指定する発音の指示(ノートオン)があった場合を考える。 Here, the detection of the pitch to be output will be specifically described. First, in step S103, the instruction (note-on) for the sound generation detected in the case where the instruction for sound generation for any pitch is not maintained (all keys are in the key release state), and the note-on pitch is Call it the first pitch. Therefore, in step S208, generation of the singing sound is started at the first pitch. In addition, after the sound generation instruction to specify the first pitch, before the determination timing arrives, the sound generation instruction (note-on) to specify the "second pitch" different from the first pitch is Consider the case.
 一般に、演奏時のミスタッチの訂正が必要となるパターンは複数想定される。ここでは、一例として、第1パターン、第2パターンについて説明する。第1パターンは、所望鍵(例えば、C3)を操作しようとして、別の鍵(例えば、D3)を押下操作してしまうパターンである。ユーザは、ミスタッチを是正するために、最初に行った押下操作(ノートオン)を短い時間で取り止めて(ノートオフ)、改めて所望鍵を押下(ノートオン)することが想定される。そして、ユーザは通常、所望鍵を押下した後はその操作を所望の時間長維持する。第2パターンは、所望鍵(例えば、C3)を操作しようとして、隣接する鍵(例えば、D3)と所望鍵を同時に押下操作してしまうパターンである。ユーザは、ミスタッチを是正するために、最初に押下した2つの鍵に対する操作(ノートオン)のうち、所望鍵に対する操作を維持したまま、誤操作の鍵のみを離操作(ノートオフ)することが想定される。 In general, a plurality of patterns requiring correction of a mistouch at the time of playing are assumed. Here, the first pattern and the second pattern will be described as an example. The first pattern is a pattern in which, in order to operate a desired key (for example, C3), another key (for example, D3) is pressed and operated. In order to correct a mistouch, it is assumed that the user cancels the first pressing operation (note on) in a short time (note off) and presses the desired key again (note on). Then, the user normally maintains the operation for a desired length of time after pressing the desired key. The second pattern is a pattern in which an adjacent key (for example, D3) and a desired key are pressed at the same time to operate a desired key (for example, C3). In order to correct a mistouch, it is assumed that the user releases only the wrong operation key (note off) while maintaining the operation on the desired key among the operations (note on) for the two keys pressed first. Be done.
 CPU10は、記憶領域に記憶された演奏情報を参照し、判定タイミングの到来までに最初のノートオンとは別の音高を指定するノートオンが検出された場合は、判定タイミングの到来時点で押鍵状態を維持した鍵を、所望鍵であると判定する。そしてCPU10は、所望鍵と判定された鍵に対応する音高(例えば、C3)を、出力すべき音高として検出する。このほか、最初は所望鍵を押下したが、その後、所望鍵の押下を継続したまま一時的に他の鍵を押下して離す、というミスタッチも考えられる。この場合は、最初のノートオンの音高が出力すべき音高として検出される。なお、最初に押下した鍵が離鍵されると共に、最初の鍵とは異なる鍵が2以上押下され、それら2以上の鍵が判定タイミングの到来時点で押下状態となっている場合も考えられる。この場合は、押下状態となっている2以上の鍵のうち最後に押下された鍵の音高が、出力すべき音高として検出されるとしてもよい。 The CPU 10 refers to the performance information stored in the storage area, and when note-on for designating a pitch other than the first note-on is detected by the arrival of the determination timing, it is pressed at the time of arrival of the determination timing. It is determined that the key maintaining the key state is the desired key. Then, the CPU 10 detects a pitch (for example, C3) corresponding to the key determined as the desired key as a pitch to be output. In addition, there may be a mistouch in which the desired key is pressed at first, and then the other key is temporarily pressed and released while the desired key is kept pressed. In this case, the pitch of the first note-on is detected as the pitch to be output. It is also conceivable that the key pressed first is released and two or more keys different from the first key are pressed and the two or more keys are pressed at the arrival of the determination timing. In this case, the pitch of the key pressed last among the two or more keys in the pressed state may be detected as the pitch to be output.
 なお、ミスタッチにおいて、所望鍵から空間的に離間した鍵を操作してしまうことは稀であることが想定される。そのため、ミスタッチの判定対象とする操作を、隣接する鍵の操作に限定してもよい。この場合、隣接しない鍵の操作は、新たに行われた正常な操作として扱ってよい。つまり、離間した鍵の操作は、ステップS103の処理で検出された操作として扱ってもよい。また、上述の所望鍵の判定方法は一例であり、CPU10は、演奏情報に含まれるノートオン、ノートオフ、ベロシティなど任意の情報を参照して、所望鍵を判定してもよい。 In the case of mistouch, it is assumed that it is rare to operate a key spatially separated from the desired key. Therefore, the operation to be subjected to the mistouch determination may be limited to the operation of the adjacent key. In this case, non-adjacent key operations may be treated as newly performed normal operations. That is, the operation of the remote key may be treated as the operation detected in the process of step S103. The above-described method of determining the desired key is an example, and the CPU 10 may determine the desired key with reference to any information such as note-on, note-off, velocity included in the performance information.
 次に、CPU10は、出力中の音高を修正すべきか否かを判別する(ステップS218)。具体的には、CPU10は、ステップS208で生成した音高と、ステップS217で検出した音高とが一致しないか否かを確認し、両者が一致しない場合に、出力中の音高を修正すべきと判別する。そしてCPU10は、出力中の音高を修正すべきでないと判別した場合は、図8に示す処理を終了させる。この場合、音高修正はなされない。一方、出力中の音高を修正すべきと判別した場合は、CPU10は、出力中の音の音高を、ステップS217で検出した音高に変更(修正)するよう調整して(ステップS219)、図8に示す処理を終了させる。例えばCPU10は、ピッチシフトにより音高を変更し、このピッチシフトにおいては、シフトする音高の差分に相当する周波数だけ、波形データが示すスペクトル分布を周波数軸方向に移動させる。CPU10は例えば、20セント単位など、段階的に音高を変更してもよい。 Next, the CPU 10 determines whether to correct the currently output pitch (step S218). Specifically, the CPU 10 checks whether the pitch generated in step S208 and the pitch detected in step S217 do not match, and if the two do not match, corrects the pitch being output. Determine that it should be. When the CPU 10 determines that the tone pitch being output should not be corrected, the CPU 10 ends the process shown in FIG. In this case, no pitch correction is made. On the other hand, when it is determined that the pitch being output should be corrected, the CPU 10 adjusts (changes) the pitch of the sound being output to the pitch detected in step S217 (step S219) , And end the process shown in FIG. For example, the CPU 10 changes the pitch by pitch shift, and in this pitch shift, the spectrum distribution indicated by the waveform data is moved in the frequency axis direction by a frequency corresponding to the difference in pitch to be shifted. The CPU 10 may change the pitch gradually, for example, in units of 20 cents.
 このように、図8に示す処理によれば、判定タイミングの到来までは、ノートオンがあっても新たな音節が読み出されることがない。すなわち、誤操作の訂正操作の全てが歌詞の読み出しに反映されて意図しない先の歌詞までが読み出されてしまうということがない。特に、単一の発音音高を確定する処理(ステップS214~S217)を要約すると次のようになる。 As described above, according to the process shown in FIG. 8, new syllables are not read out even if there is note-on until the determination timing comes. That is, there is no possibility that all of the erroneous operation correction operations are reflected in the reading of the lyrics and the unintended lyrics are not read. In particular, the process of determining a single tone pitch (steps S214 to S217) can be summarized as follows.
 まず、第1の音高を指示する発音の指示から判定タイミングの到来までの間に第1の音高とは異なる第2の音高を指示する発音の指示が取得され且つ、その第2の音高を指示する発音の指示が判定タイミングにおいて維持されているときは、第2の音高が、出力すべき音高として検出される。この場合、生成する歌唱音の発音音高が第1の音高から第2の音高に修正される(S219)。これにより、判定時間幅Tの範囲内で、操作し直しによる音高修正が可能となる。ミスタッチに対して、発音する音節(すなわち歌詞)を変えないまま、消音せずに短時間で音高を修正できるので、聴感上、ミスタッチがあったことを感じさせにくい。 First, a sound generation instruction for instructing a second pitch different from the first pitch is acquired between the sound generation instruction for instructing the first pitch and the arrival of the determination timing, and the second When an instruction of sound generation instructing a pitch is maintained at the determination timing, the second pitch is detected as a pitch to be output. In this case, the sound production pitch of the generated singing sound is corrected from the first pitch to the second pitch (S219). As a result, within the range of the determination time width T, pitch correction can be performed by re-operation. Since the pitch can be corrected in a short time without muting without changing the syllable (i.e., the lyrics) to be produced in response to the mistouch, it is difficult for the user to feel that there is a mistouch.
 一方、第1の音高を指示する発音の指示から判定タイミングの到来までの間に第1の音高とは異なる第2の音高を指示する発音の指示が取得された場合であっても、判定タイミングの到来時に第1の音高を指示する発音の指示が維持され且つ第2の音高を指示する発音の指示が維持されていないときは、生成する歌唱音の発音音高は修正されない。これにより、判定時間幅Tの範囲内でミス操作をしてもそれが判定タイミングの到来前に解消されたなら元の音高が維持される。 On the other hand, even if the sound generation instruction instructing the second pitch different from the first pitch is acquired between the sound generation instruction instructing the first pitch and the arrival of the determination timing. If the sound generation instruction for instructing the first pitch is maintained at the arrival of the determination timing and the sound generation instruction for instructing the second pitch is not maintained, then the sound pitch of the generated singing sound is corrected I will not. As a result, even if a mistake operation is performed within the range of the determination time width T, the original pitch is maintained if it is eliminated before the determination timing comes.
 本実施の形態によれば、CPU10は、取得された音節情報に応じて判定時間幅Tを決定し、取得された発音または発音解除の指示に基づいて、判定時間幅Tの経過後に単一の発音音高を確定し、取得された音節情報と確定された発音音高とに基づいて歌唱音を生成する。これにより、生成する歌唱音の発音音高を、発音する音節に応じた期間で確定することができる。 According to the present embodiment, CPU 10 determines determination time width T according to the acquired syllable information, and based on the acquired pronunciation or pronunciation cancellation instruction, a single signal is output after determination time width T has elapsed. The tone pitch is determined, and a singing voice is generated based on the acquired syllable information and the determined tone pitch. Thus, the tone pitch of the generated singing sound can be determined in a period corresponding to the syllable to be pronounced.
 特に、CPU10は、判定時間幅Tを、取得された音節情報が示す音素種別に基づき決定するか、または取得された音節情報が示す音量エンベロープの波形に基づき決定する。これにより、例えば、音程感覚が薄い音節情報に対してミス操作の是正機会を長くすることで、小さな違和感にて正規の音高を発音できる。すなわち、母音や母音開始遅延量が小さい音節は、ミスタッチ是正のための音高修正が目立ちやすい。一方、子音区間は母音区間に比べて音程感が薄い。そこでCPU10は、音素種別に基づき判定時間幅Tを決定する場合、母音開始遅延量に応じて判定時間幅Tを決定する。すなわちCPU10は、母音開始遅延量が大きい音素については、母音開始遅延量が小さい音素に対して相対的に判定時間幅Tを長く設定する。これにより、ミスタッチを目立ちにくくしつつも、母音開始遅延量が大きい音節に関してはミスタッチの是正機会を長く確保できる。 In particular, the CPU 10 determines the determination time width T based on the phoneme type indicated by the acquired syllable information or determines based on the waveform of the volume envelope indicated by the acquired syllable information. As a result, for example, by lengthening the correction opportunity of the mistake operation with respect to syllable information having a thin pitch sense, it is possible to pronounce a regular pitch with small discomfort. That is, a syllable having a small vowel or vowel start delay amount is likely to be prominent in pitch correction for correcting a mistouch. On the other hand, the consonant section has a faint sense of pitch compared to the vowel section. Therefore, when determining the determination time width T based on the phoneme type, the CPU 10 determines the determination time width T according to the vowel start delay amount. That is, for a phoneme having a large vowel start delay amount, the CPU 10 sets the determination time width T relatively long to a phoneme having a small vowel start delay amount. As a result, it is possible to secure a long chance of correcting the mistouch for syllables having a large vowel start delay amount, while making the mistouch inconspicuous.
 また、音節情報が示す音量エンベロープの波形(図7)に基づいて判定時間幅Tを決定する場合、CPU10は、波形の立ち上がりからピークに達するまでの時間tPよりも短い時間を判定時間幅Tとして決定する。これにより、発音音量が十分に立ち上がる前に生成音高を修正可能にして、ミスタッチを目立ちにくくすることができる。 In addition, when determining the determination time width T based on the waveform of the sound volume envelope indicated by the syllable information (FIG. 7), the CPU 10 determines the time shorter than the time tP from the rising of the waveform to the peak as the determination time width T decide. As a result, it is possible to correct the generated pitch before the sound production volume rises sufficiently, and make it possible to make the mistouch inconspicuous.
 またCPU10は、最初のノートオンと、判定タイミングの到来までに検出されたノートオンまたはノートオフとに基づいて、単一の発音音高を確定する。これにより、判定タイミングの到来前における操作のやり直しによる音高修正が可能であると共に、一時的なミス操作で音高が逐一修正されることを回避できる。 Also, the CPU 10 determines a single tone pitch based on the first note-on and the note-on or note-off detected until the determination timing comes. Thus, it is possible to correct the pitch by redoing the operation before the arrival of the determination timing, and it is possible to avoid that the pitch is corrected one by one by a temporary mistake operation.
 なお、判定時間幅Tは絶対時間で規定することに限定されない。例えば、テンポに応じた相対的な時間であってもよい。また、判定時間幅Tを音素種別に基づき決定する場合、第1種別q1と第2種別q2のような2段階に限定されず、音素種類毎、あるいは音素種類グループ毎に異なる値の判定時間幅Tを決定してもよい。 The determination time width T is not limited to defining the absolute time. For example, it may be a relative time according to the tempo. When the determination time width T is determined based on the phoneme type, it is not limited to two steps such as the first type q1 and the second type q2, but the determination time width of a value different for each phoneme type or each phoneme type group T may be determined.
 なお、本実施の形態では、演奏操作子15が鍵盤である場合について説明したが、演奏操作子15は、ギターのように弦が並べて配置された形状であってもよい。また、音高を指定した発音または発音解除の指示は、鍵のような操作子で入力される構成に限定されない。また、演奏操作子15は、鍵盤や弦、或いは、音高指定用の複数のボタンをタッチパネルに表示したものであってもよい。例えば、アプリ上でのタッチ操作により音高が指定される場合、ノートオン操作が継続したまま音高が変化するというミス操作も想定されるからである。また、演奏操作子15は、複数の異なる音高を入力するための操作受付部が空間的に並べて配置されたものであってもよい。また、MIDI形式のデータを扱う場合について説明したがこれに限られない。例えば、CPU10は、マイク入力などの音声データを解析し、発音のタイミングや音高を抽出することで、発音または発音解除の指示を取得してもよい。従って、本発明の歌唱音生成装置が適用される装置は鍵盤楽器に限らない。 In the present embodiment, although the case where the performance operation element 15 is a keyboard has been described, the performance operation element 15 may have a shape in which strings are arranged side by side like a guitar. In addition, the instruction for sound generation or sound generation cancellation designating the pitch is not limited to the configuration in which the input is made with an operator such as a key. Further, the performance control 15 may be a keyboard, a string, or a plurality of buttons for pitch designation displayed on the touch panel. For example, when a pitch is designated by a touch operation on the application, a mistake operation is also assumed in which the pitch changes while the note-on operation continues. In addition, the performance control 15 may be one in which operation receiving units for inputting a plurality of different pitches are spatially arranged side by side. Moreover, although the case where the data of MIDI format were handled was demonstrated, it is not restricted to this. For example, the CPU 10 may analyze voice data such as a microphone input and extract a timing of sound generation and a pitch to obtain an instruction of sound generation or sound generation cancellation. Therefore, the apparatus to which the singing sound generating apparatus of the present invention is applied is not limited to the keyboard instrument.
 また、本実施の形態では、歌唱される歌詞として日本語の歌詞を例示したがこれには限らず、他言語であってもよい。1文字と1音節とは必ずしも対応しない。例えば、濁点を有する「だ」(da)は、「た」(ta)と「"」の2文字が1音節に対応する。また、例えば英語の歌詞が「september」の場合、「sep」「tem」「ber」の3音節となる。「sep」は1音節であるが、「s」「e」「p」の3文字が1音節に対応する。CPU10は、演奏操作子15をユーザが操作する度に、各音節を、操作された鍵の音高で順次発音する。 Further, in the present embodiment, Japanese lyrics are exemplified as the lyrics to be sung, but the present invention is not limited to this, and other languages may be used. One letter and one syllable do not necessarily correspond. For example, in the case of "da" (da) having a cloud point, two letters "ta" (ta) and "" "correspond to one syllable. For example, when the English lyrics are "september", it becomes three syllables of "sep" "tem" "ber". Although "sep" is one syllable, three characters "s" "e" "p" correspond to one syllable. Each time the user manipulates the performance operation element 15, the CPU 10 sequentially pronounces each syllable at the pitch of the operated key.
 以上、本発明をその好適な実施形態に基づいて詳述してきたが、本発明はこれら特定の実施形態に限られるものではなく、この発明の要旨を逸脱しない範囲の様々な形態も本発明に含まれる。 Although the present invention has been described in detail based on its preferred embodiments, the present invention is not limited to these specific embodiments, and various embodiments within the scope of the present invention are also included in the present invention. included.
 なお、本発明を達成するためのソフトウェアによって表される制御プログラムを記憶した記憶媒体を、本楽器に読み出すことによって同様の効果を奏するようにしてもよく、その場合、記憶媒体から読み出されたプログラムコード自体が本発明の新規な機能を実現することになり、そのプログラムコードを記憶した、非一過性のコンピュータ読み取り可能な記録媒体は本発明を構成することになる。また、プログラムコードを伝送媒体等を介して供給してもよく、その場合は、プログラムコード自体が本発明を構成することになる。なお、これらの場合の記憶媒体としては、ROMのほか、フロッピディスク、ハードディスク、光ディスク、光磁気ディスク、CD-ROM、CD-R、磁気テープ、不揮発性のメモリカード等を用いることができる。「非一過性のコンピュータ読み取り可能な記録媒体」は、インターネット等のネットワークや電話回線等の通信回線を介してプログラムが送信された場合のサーバやクライアントとなるコンピュータシステム内部の揮発性メモリ(例えばDRAM(Dynamic Random Access Memory))のように、一定時間プログラムを保持しているものも含む。 Note that the storage medium storing the control program represented by the software for achieving the present invention may be read out to the present instrument to achieve the same effect, in which case, the storage medium is read from the storage medium. The program code itself implements the novel functions of the present invention, and the non-transitory computer readable recording medium storing the program code constitutes the present invention. Also, the program code may be supplied via a transmission medium or the like, in which case the program code itself constitutes the present invention. In addition to ROMs, floppy disks, hard disks, optical disks, magneto-optical disks, CD-ROMs, CD-Rs, magnetic tapes, non-volatile memory cards, etc. can be used as storage media in these cases. The “non-transitory computer readable recording medium” is a volatile memory (for example, a server or client internal to the computer system when the program is transmitted via a network such as the Internet or a communication line such as a telephone line) It also includes one that holds a program for a fixed time, such as a dynamic random access memory (DRAM).
10 CPU(音節取得部、決定部、指示取得部、確定部、生成部)
100 電子楽器
 
 
 
 
 
10 CPU (syllable acquisition unit, determination unit, instruction acquisition unit, determination unit, generation unit)
100 electronic musical instruments



Claims (8)

  1.  発音する1つの音節を示す音節情報を取得する音節取得部と、
     前記音節取得部により取得された音節情報に応じて待機時間を決定する決定部と、
     音高を指定した、発音または発音解除の指示を取得する指示取得部と、
     前記指示取得部により取得された発音または発音解除の指示に基づいて、前記指示取得部により発音の指示が取得されてから前記決定部により決定された待機時間の経過後に、単一の発音音高を確定する確定部と、
     前記音節取得部により取得された音節情報と前記確定部により確定された発音音高とに基づいて歌唱音を生成する生成部と、を有する歌唱音生成装置。
    A syllable acquisition unit for acquiring syllable information indicating one syllable to be pronounced;
    A determination unit that determines a standby time according to syllable information acquired by the syllable acquisition unit;
    An instruction acquisition unit that acquires a sound generation or sound release instruction that specifies a pitch;
    A single tone pitch after the waiting time determined by the determining unit has elapsed after the instruction acquiring unit has acquired the sounding instruction based on the sounding or releasing instruction acquired by the instruction acquiring unit. And a finalization unit to determine
    And a generation unit configured to generate a singing sound based on syllable information acquired by the syllable acquisition unit and the tone pitch determined by the determination unit.
  2.  前記生成部は、いずれの音高の発音の指示も維持されていない場合における第1の音高を指示する発音の指示に基づき前記第1の音高で前記歌唱音の生成を開始し、前記第1の音高を指示する発音の指示から前記待機時間の経過までの間に前記第1の音高とは異なる第2の音高を指示する発音の指示が取得され且つ、前記第2の音高を指示する発音の指示が前記待機時間の経過したタイミングで維持されているときは、生成する前記歌唱音の発音音高を前記第1の音高から前記第2の音高に修正する請求項1に記載の歌唱音生成装置。 The generation unit starts generation of the singing sound at the first pitch based on the sound generation instruction instructing the first pitch when no sound generation instruction of any pitch is maintained. Between the sound generation instruction instructing the first pitch and the elapse of the waiting time, the sound generation instruction instructing the second pitch different from the first pitch is acquired, and the second When the sound generation instruction for instructing the pitch is maintained at the timing when the waiting time has elapsed, the sound production pitch of the singing voice to be generated is corrected from the first pitch to the second pitch. The song sound production | generation apparatus of Claim 1.
  3.  前記生成部は、いずれの音高の発音の指示も維持されていない場合における第1の音高を指示する発音の指示に基づき前記第1の音高で前記歌唱音の生成を開始し、前記第1の音高を指示する発音の指示から前記待機時間の経過までの間に前記第1の音高とは異なる第2の音高を指示する発音の指示が取得された場合であっても、前記待機時間の経過したタイミングで、前記第1の音高を指示する発音の指示が維持され且つ前記第2の音高を指示する発音の指示が維持されていないときは、生成する前記歌唱音の発音音高を修正しない請求項1または2に記載の歌唱音生成装置。 The generation unit starts generation of the singing sound at the first pitch based on the sound generation instruction instructing the first pitch when no sound generation instruction of any pitch is maintained. Even in the case where the sound generation instruction for instructing the second pitch different from the first pitch is acquired between the sound generation instruction for instructing the first pitch and the elapse of the waiting time. When the instruction of sound generation instructing the first pitch is maintained and the instruction of sound generation instructing the second pitch is not maintained at the timing when the standby time has elapsed, the song to be generated The song sound generation device according to claim 1, wherein the sound production pitch of the sound is not corrected.
  4.  前記決定部は、前記待機時間を、前記取得された音節情報が示す音素種別に基づき決定する請求項1~3のいずれか1項に記載の歌唱音生成装置。 The singing sound generating apparatus according to any one of claims 1 to 3, wherein the determination unit determines the waiting time based on a phoneme type indicated by the acquired syllable information.
  5.  前記決定部は、前記待機時間を、前記取得された音節情報が示す音量エンベロープの波形に基づき決定する請求項1~3のいずれか1項に記載の歌唱音生成装置。 The song sound generation device according to any one of claims 1 to 3, wherein the determination unit determines the standby time based on a waveform of a volume envelope indicated by the acquired syllable information.
  6.  前記音節取得部は、複数の音節情報の中から、予め定められた順番で、前記1つの音節を示す音節情報を取得する請求項1~5のいずれか1項に記載の歌唱音生成装置。 The singing voice generating device according to any one of claims 1 to 5, wherein the syllable acquiring unit acquires syllable information indicating the one syllable in a predetermined order from a plurality of syllable information.
  7.  発音する1つの音節を示す音節情報を取得する音節取得ステップと、
     前記音節取得ステップにより取得された音節情報に応じて待機時間を決定する決定ステップと、
     音高を指定した、発音または発音解除の指示を取得する指示取得ステップと、
     前記指示取得ステップにより取得された発音または発音解除の指示に基づいて、前記指示取得ステップにより発音の指示が取得されてから、前記決定ステップにより決定された待機時間の経過後に、単一の発音音高を確定する確定ステップと、
     前記音節取得ステップにより取得された音節情報と前記確定ステップにより確定された発音音高とに基づいて歌唱音を生成する生成ステップと、を有する歌唱音生成方法。
    A syllable acquisition step of acquiring syllable information indicating one syllable to be pronounced;
    A determination step of determining a waiting time according to syllable information acquired by the syllable acquisition step;
    An instruction acquiring step of acquiring an instruction of pronunciation or pronunciation cancellation with a designated pitch;
    A single sounding sound is generated after an elapse of the waiting time determined by the determination step after the instruction of the sound generation is acquired by the instruction acquisition step based on the instruction of the sound generation or the sound generation cancellation acquired by the instruction acquisition step. A determination step to determine the height;
    A generation step of generating a singing sound based on syllable information acquired in the syllable acquiring step and the tone pitch decided in the deciding step.
  8.  発音する1つの音節を示す音節情報を取得する音節取得ステップと、
     前記音節取得ステップにより取得された音節情報に応じて待機時間を決定する決定ステップと、
     音高を指定した、発音または発音解除の指示を取得する指示取得ステップと、
     前記指示取得ステップにより取得された発音または発音解除の指示に基づいて、前記指示取得ステップにより発音の指示が取得されてから、前記決定ステップにより決定された待機時間の経過後に、単一の発音音高を確定する確定ステップと、
     前記音節取得ステップにより取得された音節情報と前記確定ステップにより確定された発音音高とに基づいて歌唱音を生成する生成ステップと、をコンピュータに実行させるプログラム。
     
                      
                      
    A syllable acquisition step of acquiring syllable information indicating one syllable to be pronounced;
    A determination step of determining a waiting time according to syllable information acquired by the syllable acquisition step;
    An instruction acquiring step of acquiring an instruction of pronunciation or pronunciation cancellation with a designated pitch;
    A single sounding sound is generated after an elapse of the waiting time determined by the determination step after the instruction of the sound generation is acquired by the instruction acquisition step based on the instruction of the sound generation or the sound generation cancellation acquired by the instruction acquisition step. A determination step to determine the height;
    A program for causing a computer to execute a generation step of generating a singing sound based on syllable information acquired in the syllable acquisition step and the sound production pitch determined in the determination step.


PCT/JP2017/023786 2017-06-28 2017-06-28 Singing sound generation device, method and program WO2019003350A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
PCT/JP2017/023786 WO2019003350A1 (en) 2017-06-28 2017-06-28 Singing sound generation device, method and program
CN201780091652.2A CN110709922B (en) 2017-06-28 2017-06-28 Singing voice generating device and method, recording medium
JP2019526039A JP6809608B2 (en) 2017-06-28 2017-06-28 Singing sound generator and method, program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2017/023786 WO2019003350A1 (en) 2017-06-28 2017-06-28 Singing sound generation device, method and program

Publications (1)

Publication Number Publication Date
WO2019003350A1 true WO2019003350A1 (en) 2019-01-03

Family

ID=64742886

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2017/023786 WO2019003350A1 (en) 2017-06-28 2017-06-28 Singing sound generation device, method and program

Country Status (3)

Country Link
JP (1) JP6809608B2 (en)
CN (1) CN110709922B (en)
WO (1) WO2019003350A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113593520B (en) * 2021-09-08 2024-05-17 广州虎牙科技有限公司 Singing voice synthesizing method and device, electronic equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04349497A (en) * 1991-05-27 1992-12-03 Yamaha Corp Electronic musical instrument
JP2014098801A (en) * 2012-11-14 2014-05-29 Yamaha Corp Voice synthesizing apparatus
JP2016050985A (en) * 2014-08-29 2016-04-11 ヤマハ株式会社 Performance information editing device
WO2016152717A1 (en) * 2015-03-25 2016-09-29 ヤマハ株式会社 Sound control device, sound control method, and sound control program

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4747434B2 (en) * 2001-04-18 2011-08-17 日本電気株式会社 Speech synthesis method, speech synthesis apparatus, semiconductor device, and speech synthesis program
JP4296290B2 (en) * 2003-10-22 2009-07-15 株式会社豊田中央研究所 Speech recognition apparatus, speech recognition method and program
US8249873B2 (en) * 2005-08-12 2012-08-21 Avaya Inc. Tonal correction of speech
JP6236757B2 (en) * 2012-09-20 2017-11-29 ヤマハ株式会社 Singing composition device and singing composition program
JP6171711B2 (en) * 2013-08-09 2017-08-02 ヤマハ株式会社 Speech analysis apparatus and speech analysis method
JP6208056B2 (en) * 2014-03-17 2017-10-04 株式会社東芝 Pronunciation display device, pronunciation display method, and program
JP2016080827A (en) * 2014-10-15 2016-05-16 ヤマハ株式会社 Phoneme information synthesis device and voice synthesis device
JP6539887B2 (en) * 2014-10-24 2019-07-10 株式会社コルグ Tone evaluation device and program
CN106157979B (en) * 2016-06-24 2019-10-08 广州酷狗计算机科技有限公司 A kind of method and apparatus obtaining voice pitch data
CN106205571A (en) * 2016-06-24 2016-12-07 腾讯科技(深圳)有限公司 A kind for the treatment of method and apparatus of singing voice

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04349497A (en) * 1991-05-27 1992-12-03 Yamaha Corp Electronic musical instrument
JP2014098801A (en) * 2012-11-14 2014-05-29 Yamaha Corp Voice synthesizing apparatus
JP2016050985A (en) * 2014-08-29 2016-04-11 ヤマハ株式会社 Performance information editing device
WO2016152717A1 (en) * 2015-03-25 2016-09-29 ヤマハ株式会社 Sound control device, sound control method, and sound control program

Also Published As

Publication number Publication date
CN110709922B (en) 2023-05-26
CN110709922A (en) 2020-01-17
JP6809608B2 (en) 2021-01-06
JPWO2019003350A1 (en) 2020-01-16

Similar Documents

Publication Publication Date Title
US10002604B2 (en) Voice synthesizing method and voice synthesizing apparatus
JP6705272B2 (en) Sound control device, sound control method, and program
CN107430849B (en) Sound control device, sound control method, and computer-readable recording medium storing sound control program
CN108053814B (en) Speech synthesis system and method for simulating singing voice of user
JP7484952B2 (en) Electronic device, electronic musical instrument, method and program
US20220076651A1 (en) Electronic musical instrument, method, and storage medium
US20220076658A1 (en) Electronic musical instrument, method, and storage medium
JP6809608B2 (en) Singing sound generator and method, program
JP6167503B2 (en) Speech synthesizer
JP2013210501A (en) Synthesis unit registration device, voice synthesis device, and program
JP4180548B2 (en) Karaoke device with vocal range notification function
WO2019239971A1 (en) Information processing method, information processing device and program
CN110720122B (en) Sound generating device and method
JP6299141B2 (en) Musical sound information generating apparatus and musical sound information generating method
WO2016152708A1 (en) Sound control device, sound control method, and sound control program
WO2018198380A1 (en) Song lyric display device and method
JP2018151547A (en) Sound production device and sound production control method
JP6732216B2 (en) Lyrics display device, lyrics display method in lyrics display device, and electronic musical instrument
JP2018151548A (en) Pronunciation device and loop section setting method
WO2019003348A1 (en) Singing sound effect generation device, method and program
WO2019239972A1 (en) Information processing method, information processing device and program
JP5845857B2 (en) Parameter extraction device, speech synthesis system
KR100994340B1 (en) Music contents production device using tts

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17916438

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2019526039

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17916438

Country of ref document: EP

Kind code of ref document: A1