WO2023153033A1 - Information processing method, program, and information processing device - Google Patents

Information processing method, program, and information processing device Download PDF

Info

Publication number
WO2023153033A1
WO2023153033A1 PCT/JP2022/040701 JP2022040701W WO2023153033A1 WO 2023153033 A1 WO2023153033 A1 WO 2023153033A1 JP 2022040701 W JP2022040701 W JP 2022040701W WO 2023153033 A1 WO2023153033 A1 WO 2023153033A1
Authority
WO
WIPO (PCT)
Prior art keywords
performance
musical
musical score
sound
symbol
Prior art date
Application number
PCT/JP2022/040701
Other languages
French (fr)
Japanese (ja)
Inventor
秀一 松本
Original Assignee
ヤマハ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by ヤマハ株式会社 filed Critical ヤマハ株式会社
Publication of WO2023153033A1 publication Critical patent/WO2023153033A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10GREPRESENTATION OF MUSIC; RECORDING MUSIC IN NOTATION FORM; ACCESSORIES FOR MUSIC OR MUSICAL INSTRUMENTS NOT OTHERWISE PROVIDED FOR, e.g. SUPPORTS
    • G10G1/00Means for the representation of music
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems

Definitions

  • an information processing method is implemented by a computer system, and based on musical score data representing a musical score including one or more performance symbols, sounds related to the performance symbols are generated. Generates an acoustic signal representing the
  • a program causes a computer system to function as a generation unit that generates an acoustic signal representing a sound related to one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
  • An information processing apparatus includes a generation unit that generates an acoustic signal representing a sound associated with one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
  • FIG. 4 is a diagram schematically showing processing of a text generation unit 32; FIG. FIG. 4 is a diagram illustrating a musical score; FIG. 10 is a diagram schematically showing read-out timings of read-out text; FIG. 10 is a diagram schematically showing read-out timings of read-out text; FIG. 11 is a diagram illustrating a display screen during reading of a musical score; 4 is a flowchart illustrating a specific procedure of processing for the control device 11 to execute a musical score reading application. 4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30.
  • FIG. FIG. 11 is a diagram illustrating a display screen during execution of a table-of-contents presentation function; FIG. 11 is a block diagram illustrating a functional configuration of a control device 11A in a third embodiment; FIG.
  • the control device 11 is composed of one or more processors that control each element of the information processing device 10 .
  • the control device 11 includes a CPU (Central Processing Unit), an SPU (Sound Processing Unit), a DSP (Digital Signal Processor), an FPGA (Field Programmable Gate Array), or an ASIC (Application Specific Integral). one or more types such as rated circuit) It consists of a processor.
  • the storage device 12 is a single or multiple memories that store the program PG (see FIG. 2) executed by the control device 11 and various data used by the control device 11 .
  • the storage device 12 is composed of a known recording medium such as a magnetic recording medium or a semiconductor recording medium, or a combination of a plurality of types of recording media.
  • a portable recording medium that can be attached to and detached from the information processing device 10, or a recording medium (for example, cloud storage) that can be written or read by the control device 11 via a communication network is used as the storage device 12.
  • the sound collection device 13 detects ambient sounds (air vibrations) and outputs them as acoustic signals.
  • the sound pickup device 13 is, for example, a microphone. Note that the sound collecting device 13, which is separate from the information processing device 10, may be connected to the information processing device 10 by wire or wirelessly.
  • the sound emitting device 14 reproduces the sound represented by the acoustic signal.
  • the sound emitting device 14 is, for example, a speaker or headphones.
  • a D/A converter that converts an acoustic signal from digital to analog and an amplifier that amplifies the acoustic signal are omitted from the drawing for the sake of convenience.
  • the sound emitting device 14, which is separate from the information processing device 10 may be connected to the information processing device 10 by wire or wirelessly.
  • the display device 16 displays images under the control of the control device 11 .
  • various display panels such as a liquid crystal display panel or an organic EL (Electroluminescence) panel are used as the display device 16 .
  • the display device 16, which is separate from the information processing device 10 may be connected to the information processing device 10 by wire or wirelessly.
  • the musical score data SD is stored in the storage device 12 after being distributed to the information processing device 10 via a communication network such as the Internet from, for example, a distribution device (typically a web server).
  • a plurality of score data SD may be stored in the storage device 12 .
  • one score data SD is created corresponding to one piece of music.
  • a musical score is a representation of a piece of music using musical symbols, including performance symbols.
  • musical symbols include musical note symbols, clefs, time signatures, key signatures, and performance symbols.
  • musical note symbols include notes, rests and accidentals that are attached to notes in a musical score.
  • the clef is marked on the left end of the staff and specifies the relationship between the position on the staff and the pitch of the sound.
  • the time signature specifies the number of beats in one measure and the type of note that constitutes one beat.
  • a key signature is a set of accidental symbols for designating the key of a piece of music.
  • Performance symbols are written on the score as a supplement to indicate to the performer the nuances that cannot be expressed with only notes and rests when performing a piece of music.
  • Performance symbols include speed symbols such as adagio and andante, expressions such as affettuoso and agitato, dynamics such as fortissimo and pianissimo, and articulation symbols such as tenuto and staccato (hereinafter referred to as articulation symbols). ”), repeat marks such as da capo and se ⁇ o, decorative marks such as trills and turns, abbreviations such as ottava alta and ottava bassa, and playing style marks indicating specific playing styles for instruments such as pedals and pizzicato.
  • the performance symbols also include finger numbers that designate fingers used when playing the notes written in the musical score.
  • the musical score data SD includes attribute information B for each musical symbol forming the target musical piece.
  • the attribute information B is information that defines musical attributes of each musical symbol, and includes a beat identifier B1 and a symbol identifier B2.
  • the beat identifier B1 is information specifying the temporal position of the musical symbol in the target music. If the music symbol is a musical note symbol or performance symbol, the number of beats from the beginning of the target song to the corresponding music symbol (for example, the beat number obtained by counting an eighth note as one beat) is preferably used as the beat identifier B1.
  • the symbol identifier B2 is information for identifying the type of musical symbol.
  • the symbol identifier B2 includes the note name (note number) and note value.
  • the note name represents the pitch of the note
  • the note value represents the duration of the note on the score.
  • a character string indicating the name of the musical symbol is preferably used as the symbolic identifier B2.
  • the musical score data SD includes musical score image data MD.
  • the musical score image data MD is data representing an image of the musical score of the target musical piece (hereinafter referred to as "score image").
  • an image file for example, a PDF file
  • a musical score image as a plane image in raster format or vector format is suitable as the musical score image data MD.
  • the symbol text data TD is data containing text corresponding to musical symbols written on the musical score.
  • the symbolic text data TD includes symbolic identifiers C1, name texts C2 and semantic texts C3.
  • the symbolic identifier C1 is information for identifying the type of musical symbol, and is information in the same format as the symbolic identifier B2 of the attribute information B.
  • FIG. The symbolic identifier C1 corresponding to the musical note symbol may be only the pitch name.
  • the name text C2 is text indicating the name of the symbol specified by the symbol identifier C1.
  • the symbolic identifier C1 and the name text C2 may be the same character string. If the musical symbol is a note symbol, the name text C2 is "do", "re”, and so on. If the musical symbol is a performance symbol, the name text C2 is "crescendo", "forte”, or the like.
  • the musical score reading application is capable of reading musical scores in multiple languages.
  • the languages that can be selected during reading are Japanese or English. Therefore, the name text C2 includes Japanese text indicating the name of the musical symbol and English text indicating the name of the musical symbol.
  • the meaning text C3 is text indicating the meaning of the symbol specified by the symbol identifier C1. For example, if the name of the musical symbol is "adagio", the semantic text C3 is "slowly”. If the musical symbols are note symbols, the semantic text C3 may not be provided.
  • the semantic text C3 also includes Japanese text indicating the meaning of the musical symbol and English text indicating the meaning of the musical symbol.
  • a word used as a phrase indicating the rest when taking a beat such as "Un” for a quarter rest, "U” for an eighth rest, etc. may be used.
  • FIG. 4 is a block diagram illustrating the functional configuration of the control device 11.
  • the control device 11 generates sound signals representing sounds related to musical symbols forming a musical score (hereinafter referred to as “symbolic sounds”) based on musical score data SD representing the musical score.
  • the musical score represented by the musical score data SD includes one or more performance symbols. Therefore, it can also be said that the control device 11 generates an acoustic signal representing a sound related to the performance symbol based on the musical score data SD representing a musical score including one or more performance symbols.
  • the sound related to the musical performance symbol is, for example, a sound indicating the name of the musical performance symbol or a sound indicating a phrase corresponding to the meaning of the musical performance symbol.
  • the names of the performance symbols correspond to the name texts C2 of the performance symbols
  • the words corresponding to the meanings of the performance symbols correspond to the meaning texts C3 of the performance symbols.
  • the musical score includes musical note symbols in addition to performance symbols. Accordingly, the control device 11 generates acoustic signals representing sounds related to performance symbols and sounds related to musical note symbols.
  • a sound related to a musical note symbol is, for example, a sound indicating the pitch name of the note indicated by the musical note symbol.
  • the acoustic signal is a signal for causing the sound emitting device 14 to reproduce the symbol sound.
  • the control device 11 has a plurality of functions for generating and reproducing sound signals (the instruction receiving unit 30, the text generating unit 32, the voice synthesizing unit 34, the performance The analysis unit 38 and the output control unit 40) are realized.
  • FIG. 5 to 9 are diagrams exemplifying instruction reception screens by the instruction reception unit 30.
  • the instruction receiving unit 30 causes the touch panel T to display a reception screen SC1 for selecting a song to be read aloud, as shown in FIG. 5, for example.
  • the user designates the musical score data SD to be read out by touching the display NA1 to NA5 corresponding to the desired musical score data SD.
  • musical score data "yyy.xml" corresponding to display NA2 is specified.
  • the instruction receiving unit 30 causes the touch panel T to display a reception screen SC2 for selecting a staff notation to be read out of the musical score data SD, as shown in FIG.
  • the reception screen SC2 displays options NB1 and NB2 for designating a staff notation to be read aloud out of the grand staff.
  • the option NB1 designates the reading of the right-hand staff positioned above the grand staff.
  • the option NB2 designates the reading of the staff notation for the left hand located on the lower stage of the grand staff.
  • the user designates the staff notation to be read out by checking the check box CK of at least one of the options NB1 and NB2.
  • the musical score represented by the musical score data SD includes a plurality of performance symbols, and each of the plurality of performance symbols belongs to one of a plurality of classifications.
  • a plurality of classifications correspond to the types of performance symbols shown in FIG.
  • the instruction receiving unit 30 receives selection of at least one of a plurality of classifications of performance symbols.
  • the control device 11 generates acoustic signals for performance symbols belonging to one or more selected categories among the plurality of performance symbols.
  • the instruction receiving unit 30 causes the touch panel T to display a reception screen SC4 for selecting a setting for reading out the musical score data SD, as shown in FIG. 8, for example.
  • Options ND1 and ND2 for designating information to be output when reading a musical score are displayed in the upper area E1 of the reception screen SC4.
  • the option ND1 designates only the reading of musical scores. That is, the option ND1 designates only audio output.
  • the option ND2 designates displaying a musical score image in addition to reading out the musical score. That is, the option ND2 designates output of audio and images.
  • Either option ND1 or ND2 can be selected using a radio button. By touching the radio button corresponding to the option ND1 or the radio button corresponding to the option ND2, the user specifies the information to be output when reading out the musical score.
  • options NE1 to NE4 for specifying the tempo for reading out the score are displayed in the lower region E2 of the reception screen SC4.
  • the option NE1 designates reading at the tempo designated by the musical score.
  • the option NE2 designates reading out all the symbols designated on the reception screen SC3 regardless of the tempo designated by the musical score.
  • Option NE3 designates reading aloud at a tempo synchronized with the performance of the musical score by the user.
  • the user designates an arbitrary tempo. In the illustrated example, the user specifies the reading tempo by specifying the number of beats per minute.
  • the speaking time for one syllable in other words, the number of reading syllables per unit time
  • the speaking time for one syllable in other words, the number of reading syllables per unit time
  • the speaking time for one syllable in other words, the number of reading syllables per unit time
  • the instruction reception unit 30 causes the touch panel T to display a selection instruction reception screen SC5 for further setting, as shown in FIG. 9, for example.
  • Options NF1 to NF2 for designating the language to be used when reading out the musical score are displayed in the upper area E3 of the reception screen SC5.
  • Option NF1 designates reading in Japanese. Reading in Japanese corresponds to, for example, using "do-re-mi-fa-so-la-si" as the note name, using Japanese-like pronunciation when uttering the names of performance symbols, and the like.
  • Choice NF2 designates reading in English.
  • Reading out in English corresponds to, for example, using "C, D, E, F, G, A, B" as note names, using English-like pronunciation when uttering names of performance symbols, and the like.
  • a language other than Japanese and English may be specified on the reception screen SC5.
  • the symbolic text data TD includes the name text C2 and the meaning text C3 of the language.
  • options NG1 to NG2 are displayed for designating the contents of reading out the performance symbols.
  • the option NG1 designates reading out the name of the performance symbol.
  • the name text C2 of the symbol text data TD is read out when the performance symbols are read out.
  • the option NG2 designates the reading of a phrase indicating the meaning of the performance symbol.
  • the meaning text C3 of the symbol text data TD is read out when reading out the performance symbol.
  • the type of voice for reading musical note symbols and the type of voice for reading performance symbols may be designated as different types. Also, if four or more voice types can be specified, for example, the right-hand staff note symbol, the left-hand staff note symbol, the right-hand staff performance symbol, and the left-hand staff symbol It may also be possible to designate the performance symbols to be read out in different voices.
  • the sound emitting device 14 may be set so that the right speaker reads out the staff notation for the right hand, and the left speaker reads out the staff notation for the left hand. Further, when the sound emitting device 14 is a stereo speaker, it may be possible to designate a speaker for outputting the reading sound of musical note symbols and a speaker for outputting the reading sound of performance symbols separately.
  • the user when reading out a chord, the user may be able to select whether to read out each note constituting the chord individually or read out the code name corresponding to the chord.
  • the name text C2 of the symbol text data TD may be stored with the text indicating the pitch name of each note forming the chord
  • the semantic text C3 may be stored with the text indicating the chord name of the chord.
  • the control device 11 starts the acoustic signal generation process. Further, the touch panel T displays, for example, a button for instructing the start of reading out a musical score (hereinafter referred to as "performance start button"). The user presses the performance start button at an appropriate timing to start reading out the musical score.
  • performance start button a button for instructing the start of reading out a musical score
  • the text generation unit 32 shown in FIG. 4 generates text indicating the content of the musical score.
  • FIG. 10 is a diagram schematically showing processing of the text generation unit 32.
  • the text generator 32 reads the score data SD specified on the reception screen SC1 shown in FIG. 5 (S1).
  • the text generation unit 32 classifies the musical score data SD into right-hand data representing a staff notation for the right hand and left-hand data representing a staff notation for the left hand (S2).
  • the data corresponding to the staff notation to be read aloud specified on the reception screen SC2 shown in FIG. 6 is to be processed thereafter.
  • the staff notation data to be read out includes attribute information B of all types of musical symbols (S3).
  • the text generation unit 32 extracts the attribute information B of the symbol to be read aloud specified on the reception screen SC3 shown in FIG. ).
  • the text generation unit 32 collates the symbolic identifier B2 of the extracted attribute information B with the symbolic identifier C1 of the symbolic text data TD, and reads out the name text C2 or the meaning text C3 corresponding to the symbolic identifier C1 (S5). Which of the name text C2 and the meaning text C3 is read depends on which of the options NG1 or NG2 on the reception screen SC5 shown in FIG. 9 is selected. Further, which of the Japanese text and the English text is read depends on which of the options NF1 and NF2 on the reception screen SC5 is selected. In the drawing, these selection contents are described as "designation of reading contents”. The read texts are arranged in the same order as the attribute information B (time series). Through the above processing, a text indicating the content of the musical score (hereinafter referred to as "read-aloud text”) is generated (S6).
  • FIG. 11 is a diagram illustrating musical scores. 12 and 13 are diagrams schematically showing read-out timings of read-out texts.
  • a musical score G shown in FIG. 11 indicates, for example, the first two bars of the musical score data (yyy.xml) specified to be read out on the reception screen SC1 shown in FIG.
  • the musical score G includes a musical score for the right hand and a musical score for the left hand. Based on the tempo information TP, the musical score G is specified as 120 beats per minute, with one quarter note as one beat.
  • FIG. 12 shows right-hand read-out sounds representing right-hand read-out sounds and left-hand read-out sounds representing left-hand read-out sounds.
  • a time axis is shown between the reading sound for the right hand and the reading sound of the musical score for the left hand.
  • One scale (t1) of the time axis is based on the eighth note, which is the shortest note in the musical score G. Based on the tempo of the musical score G described above, one scale (t1) on the time axis is 0.25 seconds.
  • the control device 11 controls when the target point in time at which the piece of music progresses at a speed corresponding to the tempo designated by the tempo information TP reaches the point in time corresponding to the performance symbol. , generates an acoustic signal so that the sound associated with the performance symbol is pronounced.
  • the text generation unit 32 may reduce the text to be read out. For example, if the speech time for one syllable is determined during reading, it is possible to determine whether or not reading is possible at the time the read-aloud text is generated. The text generation unit 32 determines whether or not the read-out text can be read out within the time based on the generated words, the tempo of the music, and the utterance time of one syllable.
  • the text generation unit 32 reads a part of the plurality of performance symbols. The remaining performance symbols may be excluded from reading. Taking the period P1 as an example, the text generation unit 32 does not need to read out one of "mesopiano” or “staccato” and not read out the other. That is, when “mesopiano” and “staccato” are pronounced together, the control device 11 selects either “mesopiano” or “staccato” and reads out the text corresponding to the selected performance symbol.
  • the control device 11 selects either “mesopiano” or “staccato” and reads out the text corresponding to the selected performance symbol.
  • include in “Mezzo piano” is an example of a sound associated with the first performance symbol
  • staccato is an example of a sound associated with the second performance symbol.
  • the instruction receiving unit 30 may allow the user to set the reading priority for each classification of performance symbols.
  • the text generator 32 deletes from the read-aloud text in order from the text of the performance symbols belonging to the lower priority category.
  • the non-verbal sound is included in the read-aloud text instead of the text corresponding to the performance symbol. good too.
  • the user can grasp the rhythm of the musical piece as well as the pitches of the notes indicated in the musical score.
  • FIG. 13 shows only the reading sounds for the right hand and omits the reading sounds for the left hand.
  • symbol Mti (i is an integer from 1 to 9) indicates a metronome sound. The user can grasp the break of the beat by the metronome sound Mti.
  • control device 11 may generate an acoustic signal indicating a sound produced by pronouncing a sound related to the performance symbol regardless of the tempo of the music.
  • the user can comprehend all the symbols of the specified type, and can comprehend the content of the musical score without omission.
  • the option NE4 is a mode in which the user designates an arbitrary tempo by designating the number of beats per minute.
  • the user may designate the progress of reading using the manipulator.
  • the operator may be an operation button displayed on the touch panel T, for example.
  • the operator may be a member of the musical instrument.
  • the musical instrument is a piano, pedals can be used as operators. In this case, for example, when the user steps on the damper pedal once, reading progresses in units of one note symbol or one bar, and when the user steps on the soft pedal once, the reading progresses by one note symbol. , or the reading may be reversed in units of one bar or the like.
  • the voice synthesizing unit 34 shown in FIG. 4 uses the read-out text generated by the text generating unit 32 and the voice data VD to generate an acoustic signal.
  • the speech synthesizer 34 is an example of a generator.
  • the speech synthesizing unit 34 sequentially selects speech segments corresponding to the text to be read out of a plurality of speech segments included in the speech data VD, adjusts the pitch of each speech segment, and then connects them to each other. to generate an acoustic signal.
  • the pitch of the sound related to the musical note symbol in the reading text may be matched with the pitch of the musical note symbol, or may be a predetermined pitch.
  • the performance analysis unit 38 operates only when the user designates reading out at a tempo synchronized with the performance of the musical score (option NE3) on the acceptance screen SC4 shown in FIG. good.
  • the output control unit 40 controls the output of sound based on the acoustic signal and the output of the musical score image based on the musical score image data MD.
  • the output control unit 40 causes the sound emitting device 14 to reproduce the sound represented by the acoustic signal generated by the voice synthesis unit 34.
  • the output control unit 40 outputs the sound represented by the acoustic signal generated by the voice synthesis unit 34 to the sound emitting device 14.
  • the musical score image data MD is displayed on the display device 16 while being reproduced.
  • FIG. 14 is a diagram illustrating a display screen during reading of musical scores.
  • the display of the touch panel T which is the display device 16
  • the display of the touch panel T is switched to the display screen SC6 shown in FIG.
  • a message 601 indicating that the reading sound of the musical score is being reproduced a musical score image 602, a pause button 604, a fast forward button 606, a rewind button 608, a repeat button 610,
  • An end button 612 is displayed.
  • the musical score image 602 is an image displaying the musical score image data MD included in the musical score data SD to be read aloud.
  • a bar 603 indicating the reading position is superimposed on the musical score image 602 and displayed.
  • the output control unit 40 scrolls the musical score image 602 based on the timing label attached to the reading text. At this time, the output control unit 40 adjusts the scrolling speed of the musical score image 602 so that the music symbol being read out and the bar 603 are superimposed. It should be noted that instead of displaying the read-out position with the bar 603, the musical symbols to be read-out may be highlighted.
  • FIG. 14 exemplifies a musical score using staff notation as the musical score image 602
  • the musical score image 602 may be displayed as a piano roll, for example. Further, in the reception screen SC4 shown in FIG. 8, when only the reading of the musical score is specified (option ND1), the musical score image 602 is not displayed.
  • a pause button 604, a fast-forward button 606, a rewind button 608, a review button 610, and an end button 612 accept operations related to reading out the musical score.
  • the output control unit 40 pauses reading out the score.
  • the fast-forward button 606 the output control unit 40 fast-forwards reading of the musical score. For example, when the fast-forward button 606 is touched once, the output control unit 40 changes the readout position to the beginning of the bar next to the bar containing the current readout position.
  • the rewind button 608 is operated, the output control unit 40 rewinds the reading of the musical score.
  • the output control unit 40 selects reading aloud at a tempo synchronized with the performance of the musical score by the user (option NE3). Based on this, adjust the output timing of the reading sound.
  • the output control unit 40 adjusts the output speed of the reading sound so that, for example, on the musical score, the position ahead of the performance position by a predetermined beat is read out.
  • the predetermined beat may be specified by the user.
  • the output control unit 40 may control the music included in the N+1-th measure immediately before the performance of the N-th measure to end.
  • the symbols may be read aloud.
  • Immediately before the performance of the N-th measure ends is, for example, after the last note of the N-th measure is played. This imitates, for example, a teaching method in which a chorus conductor prereads and shows lyrics to be sung next to the chorus members.
  • the generation of the read-out text (S102) and the generation of the acoustic signal (S104) may be performed after the user instructs to read out the score (S106: YES).
  • acoustic signals representing sounds related to performance symbols are generated based on musical score data SD including one or more performance symbols. Therefore, the performance symbols included in the musical score can be comprehended aurally, and even visually impaired people, beginners and small children who are not accustomed to reading musical scores can easily comprehend the musical score.
  • the sound related to the performance symbol is a sound indicating the name of the performance symbol or a sound indicating a phrase corresponding to the meaning of the performance symbol. If the sound associated with the performance symbol is the sound indicating the name of the performance symbol, the description on the musical score can be accurately grasped. In addition, if the sound related to the musical performance symbol is a sound indicating a phrase corresponding to the meaning of the musical performance symbol, even if the user lacks knowledge of the musical performance symbol and cannot understand the meaning of the performance symbol only by the name of the musical performance symbol, Able to grasp the contents indicated by musical scores.
  • sounds corresponding to performance symbols are pronounced at timings corresponding to the tempo of music. This makes it easier for the user to grasp the positions of the performance symbols in the music, thereby improving convenience.
  • an acoustic signal representing the sound related to the musical note symbol is generated. Therefore, the musical note symbols included in the musical score can be comprehended aurally, and the comprehension of the musical score can be further facilitated.
  • a non-verbal notification sound is pronounced as the sound related to the rest.
  • the user can immediately recognize that the sound associated with the rest corresponds to the rest when the sound related to the rest is pronounced.
  • the musical score data SD specified by the user is read out.
  • a list of data names of the musical score data SD such as the reception screen SC1 shown in FIG. 5 is displayed, the user may not be able to identify the musical score data SD corresponding to the desired song.
  • a part of a plurality of musical score data SD is continuously read out so that the user can specify the musical score data SD corresponding to the desired music piece.
  • the function of continuously reading a part of a plurality of musical score data SD is hereinafter referred to as a "table of contents presentation function".
  • FIG. 16 is a diagram illustrating an instruction reception screen by the instruction reception unit 30.
  • the instruction receiving unit 30 when the score reading application is activated, the instruction receiving unit 30 causes the touch panel T to display a menu selection instruction receiving screen SC7 as shown in FIG. 16, for example. Choices NI1 and NI2 are displayed on the acceptance screen SC7.
  • the option NI1 specifies reading out the musical score data SD selected by the user as shown in the first embodiment.
  • the instruction receiving unit 30 displays the receiving screen SC1 shown in FIG. 5, and receives the designation of the musical score data SD to be read out from the user.
  • a part of the musical score is, for example, a part or all of a specific structural section among multiple sections (hereinafter referred to as "structural sections") that divide a piece of music according to its musical meaning.
  • Structural sections are, for example, sections such as an intro, an A melody, a B melody, a chorus, and an outro.
  • the text generation unit 32 generates, for each of the plurality of musical score data SD, a text for reading out, for example, the structure section of the “chorus” of the music.
  • the text generation unit 32 generates, for each of the plurality of musical score data SD, a text for reading out the structural section of the "intro" (predetermined number of bars at the beginning of the musical score) of the music, for example.
  • the musical score data SD is the first musical score data
  • the storage device 12 also stores second musical score data different from the first musical score data.
  • the control device 11 generates a first acoustic signal representing a sound related to the performance symbol and a sound related to the musical note symbol included in a portion of the first musical score corresponding to the first musical score data, and a second acoustic signal corresponding to the second musical score data. Generating a second acoustic signal indicating sounds related to performance symbols and sounds related to musical note symbols included in a portion of the musical score.
  • the control device 11 causes the sound emitting device 14 to sequentially reproduce the first acoustic signal and the second acoustic signal.
  • the first musical score data is musical score data "xxx.xml”
  • the second musical score data is musical score data "yyy.xml”.
  • the control device 11 selects a portion from each of a plurality of musical score data, and sequentially reproduces sounds related to performance symbols and musical note symbols included in the selected portion.
  • the user can easily grasp which musical score of each piece of musical score data SD corresponds to, and quickly select the desired musical score data SD from among the plurality of musical score data SD. can do.
  • the information processing device 10 reads out the musical score data SD.
  • the information processing apparatus 10 assists the user in making the sound of the performance closer to the sound shown in the score.
  • FIG. 18 is a block diagram illustrating the functional configuration of the control device 11A in the third embodiment.
  • the control device 11A includes a performance evaluation section 42 in addition to the configuration of the control device 11 (see FIG. 4) according to the first embodiment.
  • the performance evaluation section 42 evaluates the performance of the musical instrument by the user based on the analysis result of the performance analysis section 38 .
  • the performance analysis unit 38 analyzed the performance position of the musical instrument by the user.
  • the performance analysis unit 38 analyzes the volume of the performance sound of the musical instrument in addition to the analysis of the performance position.
  • the performance evaluation unit 42 evaluates whether the user's performance conforms to the musical symbols of the score. More specifically, the performance evaluation unit 42 detects the difference between the performance sound, which is the sound of the musical composition played by the user, and the sound indicated by the musical symbols included in the musical score representing the musical composition, and determines that the difference is determined in advance. Determine whether or not it is out of the allowable range.
  • the user sets the allowable range of the difference, for example, based on his or her performance skill level. Generally, it is considered that the allowable difference becomes smaller as the skill level of the user increases. If there is a portion where the difference is out of the allowable range, the text generation unit 32 generates text pointing out the portion. Specifically, for example, "Right hand, 2nd measure, 'Fa, re, me' was changed to 'Fa, mi, le'.” In addition, it generates a text that reads out the pitch and duration of the performance performed by the user. Such text is referred to as "supporting text".
  • the performance evaluation unit 42 evaluates the performance by detecting the difference between the volume of the sound played by the user and the volume of the performance along the dynamic symbol. Further, the performance evaluation section 42 detects the difference between the duration of the sound played by the user when the performance symbol is an articulation symbol, and the duration when the performance is performed along the articulation symbol. by doing. For example, the smaller the difference, the performance evaluation unit 42 evaluates that the performance is performed along the performance symbols of the musical score, that is, the performance skill is high.
  • the user sets the allowable range of the difference, for example, based on his or her performance skill level. If there is a portion where the difference is out of the allowable range, the text generation unit 32 generates support text pointing out the portion. Specifically, for example, "Right hand, 1st measure, 'staccato, mi, staccato, fa', staccato momentum is weak," etc. Generate supporting text to indicate that the performance symbols were not reflected.
  • the speech synthesis unit 34 uses the supporting text and the speech data VD to generate an acoustic signal.
  • the text indicating the pitch of the musical note symbol may be read aloud with a voice of a pitch corresponding to the pitch.
  • the output control unit 40 causes the sound emitting device 14 to reproduce sound based on the acoustic signal.
  • the performance sound of the user may be recorded, and the part of the recorded performance sound corresponding to the point pointed out in the support text may be reproduced together with the support text read aloud.
  • the performance symbols may always be read aloud regardless of the presence or absence of differences. In this case, for example, if the difference between the performance symbols and the performance is large, the read-out voice is increased (the greater the difference, the louder the read-out voice is) so that the user can check whether the performance follows the performance symbols. can be grasped.
  • the control device 11 acquires the performance sound, which is the sound of the music played by the user, and detects the difference between the sound indicated by the musical symbols included in the musical score representing the music and the performance sound. do.
  • the control device 11 When the difference is out of a predetermined allowable range, the control device 11 generates an acoustic signal representing a sound related to the musical symbol included in the portion of the musical score corresponding to the location where the difference occurs.
  • the user can grasp the difference between his/her own performance and the content indicated by the musical score, and can efficiently master the performance of the musical piece indicated by the musical score.
  • the control device 11 indicates to the user the position on the musical score of the portion where the difference occurs by reading out the musical symbol of the portion where the difference occurs.
  • the user can intuitively grasp the location on the musical score where the difference occurs, compared to, for example, simply reading out the position (bar number, etc.) in the musical score mechanically.
  • the content of the user's performance is verbalized.
  • the control device 11 reads aloud the pitch and value of the performance performed by the user. This allows the user to objectively grasp the details of his/her own error.
  • the speech synthesizing unit 34 performs segment-connected speech synthesis, but the method of speech synthesis is not limited to the above examples.
  • statistical model type speech synthesis using statistical models such as deep neural networks or HMMs (Hidden Markov Models) may be used.
  • the information processing device 10 may be implemented by a server device that communicates with an information device such as a smart phone or a tablet terminal.
  • the information processing device 10 receives designation of musical score data SD from the information device, and generates an acoustic signal by speech synthesis processing using the designated musical score data SD.
  • the information processing device 10 transmits an acoustic signal generated by speech synthesis processing to the information device.
  • the information device reproduces the acoustic signal.
  • the functions of the information processing device 10 include the control device 11 as described above. It is realized by cooperation of one or more constituent processors and the program PG stored in the storage device 12 .
  • the manner in which the user looks at the items displayed on the touch panel T and touches the touch panel T when performing various settings and instructions in the musical score reading application has been described.
  • the presentation of information (such as selection items in settings) to the user may be performed by reading aloud.
  • the input from the user to the information processing device 10 may be performed by voice input.
  • voice input when a visually impaired person uses a musical score reading application, the use of voice is effective.
  • the above program can be provided in a form stored in a computer-readable recording medium and installed in the computer.
  • the recording medium is, for example, a non-transitory recording medium, and an optical recording medium (optical disc) such as a CD-ROM is a good example.
  • a recording medium for storing the program in the distribution device corresponds to the non-transitory recording medium described above.
  • An information processing apparatus is realized by a computer system, and generates an acoustic signal representing a sound related to the performance symbols based on musical score data representing a musical score including one or more performance symbols. do. Therefore, the performance symbols included in the musical score can be grasped by hearing, and even visually handicapped people, beginners who are not accustomed to reading musical scores, and small children can easily grasp the musical score.
  • the sound related to the performance symbol is a sound indicating the name of the performance symbol or a sound indicating a phrase corresponding to the meaning of the performance symbol.
  • the sound associated with the performance symbol is the sound indicating the name of the performance symbol
  • the sound associated with the performance symbol is a sound indicating a word or phrase corresponding to the meaning of the performance symbol, even if the user lacks knowledge of the performance symbol and cannot understand the meaning of the performance symbol only by the name of the performance symbol, the musical score can be used. It is possible to grasp the contents indicated by .
  • the musical score data includes tempo information specifying a tempo of a piece of music indicated by the musical score, and in generating the acoustic signal, the tempo specified by the tempo information.
  • the acoustic signal is generated such that when the target point in time at which the music progresses at the speed corresponding to reaches the point in time corresponding to the performance symbol, the sound associated with the performance symbol is pronounced.
  • sounds corresponding to performance symbols are produced at timings corresponding to the tempo of music. Therefore, it becomes easier for the user to grasp the positions of the performance symbols in the music, and convenience can be improved.
  • the one or more performance symbols include a first performance symbol and a second performance symbol, and in generating the acoustic signal, When the sound related to the first performance symbol and the sound related to the second performance symbol are superimposed, selecting either the first performance symbol or the second performance symbol, generating said acoustic information representing a sound;
  • the first performance symbol and the second performance symbol are superimposed, either the first performance symbol or the second performance symbol is selected.
  • Generates acoustic information Therefore, the sound related to the first performance symbol and the sound related to the second performance symbol do not overlap, and the audibility of the sound related to the performance symbol can be improved.
  • the musical score data includes tempo information specifying the tempo of the music indicated by the musical score, and in generating the acoustic signal, the musical performance symbols are used regardless of the tempo of the music. Generating an acoustic signal indicative of the sound produced.
  • sounds corresponding to performance symbols are produced regardless of the tempo of music. Therefore, the sounds related to the performance symbols are not over-pronounced, and the audibility of the sounds related to the performance symbols can be improved.
  • the one or more performance symbols are a plurality of performance symbols, and each of the plurality of performance symbols belongs to one of a plurality of classifications. , receiving a selection of at least one of the plurality of categories, and generating the acoustic signal for a musical performance symbol belonging to the one or more categories related to the selection among the plurality of musical performance symbols. Generate.
  • acoustic signals are generated for performance symbols belonging to the selected classification. Therefore, it is possible to selectively produce sounds related to the performance symbols required by the user, thereby improving convenience.
  • the musical score includes musical note symbols in addition to the performance symbols, and generating the acoustic signal includes sounds associated with the performance symbols and generating said acoustic signal representing a sound associated with a musical note symbol.
  • acoustic signals representing sounds related to musical note symbols are generated. Therefore, the musical note symbols included in the musical score can be comprehended aurally, and the comprehension of the musical score can be further facilitated.
  • an acoustic signal representing a non-verbal notification sound is generated as the sound related to the rest.
  • non-verbal notification sounds are used as the sounds related to rests. Therefore, when a sound related to a rest is pronounced, the user can immediately recognize that the sound corresponds to the rest.
  • the musical score data is first musical score data
  • the generation of the acoustic signal is included in a portion of the first musical score corresponding to the first musical score data.
  • generating a second acoustic signal indicative of the sound associated with the musical note symbol and further comprising causing a sound emitting device to sequentially reproduce the first acoustic signal and the second acoustic signal.
  • a portion is selected from each of a plurality of musical score data, and sounds associated with performance symbols and musical note symbols included in the selected portion are sequentially reproduced. Therefore, the user can easily grasp the musical score of which musical piece each of the plurality of musical score data corresponds to, and can quickly select the desired musical score data from the plurality of musical score data.
  • a program according to one aspect (a tenth aspect) of the present disclosure functions as a generation unit that generates an acoustic signal representing a sound related to one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
  • An information processing apparatus includes a generation unit that generates an acoustic signal representing a sound associated with one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
  • Braille sheet music requires about three times as much paper space to write the same content as regular sheet music, and it takes time to read. For this reason, for example, when a user forgets the title of a piece of music and wants to find the desired score from the contents of the score, the user has to spend time reading a plurality of scores, which is inconvenient. .
  • An information processing apparatus is implemented by a computer system, and includes a first musical symbol representing a sound related to the musical symbol included in a portion of a first musical score corresponding to first musical score data including one or more musical symbols. generating an acoustic signal, and generating a second acoustic signal indicating a sound associated with the musical symbol included in a portion of a second musical score that includes one or more of the musical symbols and corresponds to second musical score data that is different from the first musical score data; and causing a sound emitting device to sequentially reproduce the first acoustic signal and the second acoustic signal.
  • An information processing apparatus is realized by a computer system, acquires a performance sound that is a sound of a musical piece played by a user, and obtains a sound indicated by a musical symbol included in a musical score indicating the musical piece, Detecting a difference from the performance sound, and generating an acoustic signal indicating a sound related to a musical symbol included in a portion of the musical score corresponding to the portion where the difference occurs when the difference is out of a predetermined allowable range. do.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Auxiliary Devices For Music (AREA)

Abstract

An information processing device 10 generates, on the basis of score data SD representing a score including at least one performance mark, an acoustic signal representing a sound relating to the performance mark.

Description

情報処理方法、プログラム、および情報処理装置Information processing method, program, and information processing device
 本開示は、楽譜の内容の把握を支援する技術に関する。 The present disclosure relates to technology for assisting in grasping the content of musical scores.
 従来、視覚障碍者が楽譜の内容を把握できるようにするために、点字楽譜が利用されている。例えば下記特許文献1には、コンピュータを用いて楽譜データを点訳し、点訳された楽譜データを点字タイプライタに打ち出して点字楽譜を生成する楽譜自動点訳システムが開示されている。 Conventionally, Braille musical scores have been used to enable visually impaired people to understand the content of musical scores. For example, Japanese Patent Laid-Open No. 2002-200000 discloses an automatic score translation system that uses a computer to transcribe score data into Braille and outputs the transcribed score data to a Braille typewriter to generate a Braille score.
特開昭60-119594号公報JP-A-60-119594
 点字楽譜の理解には専門知識が必要であり、初学者や小さな子供には理解が難しいという課題がある。また、通常楽譜と点字楽譜とを比較すると、同じ内容の譜面を表記するのに、点字楽譜は通常楽譜の約3倍の紙面が必要であり、読み取りに時間がかかるという課題がある。以上の事情を考慮して、本開示のひとつの態様は、楽譜の内容を音によって提示することを目的とする。 Expert knowledge is required to understand Braille music scores, and there is a problem that it is difficult for beginners and small children to understand. In addition, when comparing a normal musical score and a Braille musical score, the Braille musical score requires about three times as much paper space as the normal musical score to write the same musical score, and there is a problem that it takes time to read. In consideration of the above circumstances, one aspect of the present disclosure aims at presenting the contents of a musical score by sound.
 以上の課題を解決するために、本開示のひとつの態様に係る情報処理方法は、コンピュータシステムにより実現され、1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する。 In order to solve the above problems, an information processing method according to one aspect of the present disclosure is implemented by a computer system, and based on musical score data representing a musical score including one or more performance symbols, sounds related to the performance symbols are generated. Generates an acoustic signal representing the
 本開示のひとつの態様に係るプログラムは、1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する生成部としてコンピュータシステムを機能させる。 A program according to one aspect of the present disclosure causes a computer system to function as a generation unit that generates an acoustic signal representing a sound related to one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
 本開示のひとつの態様に係る情報処理装置は、1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する生成部を備える。 An information processing apparatus according to one aspect of the present disclosure includes a generation unit that generates an acoustic signal representing a sound associated with one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
第1実施形態に係る情報処理装置10の構成を例示するブロック図である。1 is a block diagram illustrating the configuration of an information processing apparatus 10 according to a first embodiment; FIG. 記憶装置12が記憶するデータの構成を例示する図である。4 is a diagram illustrating a configuration of data stored in a storage device 12; FIG. 音楽記号の種類を示す図である。It is a figure which shows the kind of musical symbol. 制御装置11の機能的な構成を例示するブロック図である。3 is a block diagram illustrating the functional configuration of the control device 11; FIG. 指示受付部30による指示の受付画面を例示する図である。4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30. FIG. 指示受付部30による指示の受付画面を例示する図である。4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30. FIG. 指示受付部30による指示の受付画面を例示する図である。4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30. FIG. 指示受付部30による指示の受付画面を例示する図である。4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30. FIG. 指示受付部30による指示の受付画面を例示する図である。4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30. FIG. テキスト生成部32の処理を模式的に示す図である。4 is a diagram schematically showing processing of a text generation unit 32; FIG. 楽譜を例示する図である。FIG. 4 is a diagram illustrating a musical score; 読み上げテキストの読み上げタイミングを模式的に示す図である。FIG. 10 is a diagram schematically showing read-out timings of read-out text; 読み上げテキストの読み上げタイミングを模式的に示す図である。FIG. 10 is a diagram schematically showing read-out timings of read-out text; 楽譜の読み上げ中における表示画面を例示する図である。FIG. 11 is a diagram illustrating a display screen during reading of a musical score; 制御装置11が楽譜読み上げアプリケーションを実行する処理の具体的な手順を例示するフローチャートである。4 is a flowchart illustrating a specific procedure of processing for the control device 11 to execute a musical score reading application. 指示受付部30による指示の受付画面を例示する図である。4 is a diagram exemplifying an instruction reception screen by an instruction reception unit 30. FIG. 目次提示機能の実行中の表示画面を例示する図である。FIG. 11 is a diagram illustrating a display screen during execution of a table-of-contents presentation function; 第3実施形態における制御装置11Aの機能的な構成を例示するブロック図である。FIG. 11 is a block diagram illustrating a functional configuration of a control device 11A in a third embodiment; FIG.
A:第1実施形態
 図1は、第1実施形態に係る情報処理装置10の構成を例示するブロック図である。情報処理装置10は、制御装置11と、記憶装置12と、収音装置13と、放音装置14と、操作装置15と、表示装置16とを備えるコンピュータシステムである。情報処理装置10は、例えばスマートフォン、タブレット端末またはパーソナルコンピュータ等の情報端末により実現される。本実施形態では、情報処理装置10はスマートフォンであるものとする。なお、情報処理装置10は、単体の装置で実現されるほか、相互に別体で構成された複数の装置(例えばクライアントサーバシステム)でも実現される。
A: First Embodiment FIG. 1 is a block diagram illustrating the configuration of an information processing apparatus 10 according to the first embodiment. The information processing device 10 is a computer system that includes a control device 11 , a storage device 12 , a sound collection device 13 , a sound emission device 14 , an operation device 15 and a display device 16 . The information processing device 10 is realized by an information terminal such as a smart phone, a tablet terminal, or a personal computer, for example. In this embodiment, the information processing device 10 is assumed to be a smart phone. The information processing apparatus 10 may be implemented as a single device, or may be implemented as a plurality of devices configured separately from each other (for example, a client-server system).
 制御装置11は、情報処理装置10の各要素を制御する単数または複数のプロセッサで構成される。例えば、制御装置11は、CPU(Central Processing Unit)、SPU(Sound Processing Unit)、DSP(Digital Signal Processor)、FPGA(Field Programmable Gate Array)、またはASIC(Application Specific Integrated Circuit)等の1種類以上のプロセッサにより構成される。 The control device 11 is composed of one or more processors that control each element of the information processing device 10 . For example, the control device 11 includes a CPU (Central Processing Unit), an SPU (Sound Processing Unit), a DSP (Digital Signal Processor), an FPGA (Field Programmable Gate Array), or an ASIC (Application Specific Integral). one or more types such as rated circuit) It consists of a processor.
 記憶装置12は、制御装置11が実行するプログラムPG(図2参照)と、制御装置11が使用する各種のデータとを記憶する単数または複数のメモリである。記憶装置12は、例えば磁気記録媒体もしくは半導体記録媒体等の公知の記録媒体、または、複数種の記録媒体の組合せで構成される。なお、情報処理装置10に対して着脱される可搬型の記録媒体、または通信網を介して制御装置11が書込または読出を実行可能な記録媒体(例えばクラウドストレージ)を、記憶装置12として利用してもよい。 The storage device 12 is a single or multiple memories that store the program PG (see FIG. 2) executed by the control device 11 and various data used by the control device 11 . The storage device 12 is composed of a known recording medium such as a magnetic recording medium or a semiconductor recording medium, or a combination of a plurality of types of recording media. A portable recording medium that can be attached to and detached from the information processing device 10, or a recording medium (for example, cloud storage) that can be written or read by the control device 11 via a communication network is used as the storage device 12. You may
 収音装置13は、周囲の音(空気振動)を検出して音響信号として出力する。収音装置13は、例えばマイクロホンである。なお、情報処理装置10とは別体の収音装置13を、情報処理装置10に対して有線または無線により接続してもよい。 The sound collection device 13 detects ambient sounds (air vibrations) and outputs them as acoustic signals. The sound pickup device 13 is, for example, a microphone. Note that the sound collecting device 13, which is separate from the information processing device 10, may be connected to the information processing device 10 by wire or wirelessly.
 放音装置14は、音響信号が表す音を再生する。放音装置14は、例えばスピーカまたはヘッドホンである。なお、音響信号をデジタルからアナログに変換するD/A変換器と、音響信号を増幅する増幅器とは、便宜的に図示が省略されている。また、情報処理装置10とは別体の放音装置14を、情報処理装置10に対して有線または無線により接続してもよい。 The sound emitting device 14 reproduces the sound represented by the acoustic signal. The sound emitting device 14 is, for example, a speaker or headphones. A D/A converter that converts an acoustic signal from digital to analog and an amplifier that amplifies the acoustic signal are omitted from the drawing for the sake of convenience. Further, the sound emitting device 14, which is separate from the information processing device 10, may be connected to the information processing device 10 by wire or wirelessly.
 操作装置15は、利用者からの指示を受け付ける入力機器である。操作装置15は、例えば、利用者が操作する操作子、または、利用者による接触を検知するタッチパネルTである。本実施形態では、操作装置15としてタッチパネルTを用いるものとする。この場合、タッチパネルTは、操作装置15と後述する表示装置16を兼ねている。なお、情報処理装置10とは別体の操作装置15(例えばマウスまたはキーボード)を、情報処理装置10に対して有線または無線により接続してもよい。 The operation device 15 is an input device that receives instructions from the user. The operation device 15 is, for example, an operator operated by a user or a touch panel T that detects contact by the user. In this embodiment, the touch panel T shall be used as the operating device 15. FIG. In this case, the touch panel T serves as both the operation device 15 and the display device 16, which will be described later. An operation device 15 (for example, a mouse or a keyboard) separate from the information processing device 10 may be connected to the information processing device 10 by wire or wirelessly.
 表示装置16は、制御装置11による制御のもとで画像を表示する。例えば液晶表示パネルまたは有機EL(Electroluminescence)パネル等の各種の表示パネルが表示装置16として利用される。なお、情報処理装置10とは別体の表示装置16を、情報処理装置10に対して有線または無線により接続してもよい。 The display device 16 displays images under the control of the control device 11 . For example, various display panels such as a liquid crystal display panel or an organic EL (Electroluminescence) panel are used as the display device 16 . Note that the display device 16, which is separate from the information processing device 10, may be connected to the information processing device 10 by wire or wirelessly.
 図2は、記憶装置12が記憶するデータの構成を例示する図である。図3は、音楽記号の種類を示す図である。図2に示すように、記憶装置12は、制御装置11が実行するプログラムPGと、音声データVDと、楽譜データSDと、記号テキストデータTDを記憶する。本実施形態において、プログラムPGは、楽譜読み上げアプリケーションを実行するためのプログラムである。楽譜読み上げアプリケーションは、楽譜データSDに対応する楽譜に記載された各種の情報に関する音を示す音響信号を生成し、当該音響信号を再生するアプリケーションである。より詳細には、楽譜読み上げアプリケーションは、音楽記号に対応するテキストを読み上げることにより、楽譜の内容を利用者が聴覚を用いて把握可能とする。以下、本実施形態において、楽譜の読み上げとは、楽譜に含まれる音楽記号に関する音を示す音響信号を生成し、当該音響信号を再生することに対応する。 FIG. 2 is a diagram illustrating the configuration of data stored in the storage device 12. As shown in FIG. FIG. 3 is a diagram showing types of musical symbols. As shown in FIG. 2, the storage device 12 stores a program PG executed by the control device 11, voice data VD, score data SD, and symbol text data TD. In this embodiment, the program PG is a program for executing a musical score reading application. The score read-aloud application is an application that generates an acoustic signal indicating sounds related to various information written on the score corresponding to the score data SD, and reproduces the acoustic signal. More specifically, the score-to-speech application reads aloud the text corresponding to the musical symbols so that the user can perceive the contents of the score using the sense of hearing. Hereinafter, in the present embodiment, reading out a musical score corresponds to generating an acoustic signal indicating sounds related to musical symbols included in the musical score and reproducing the acoustic signal.
 音声データVDは、楽譜を読み上げる合成音声を生成するためのデータである。音声データVDは、複数の音声素片を含む音声合成用ライブラリである。各音声素片は、言語的な意味の最小単位である音素単体(例えば母音または子音)、または複数の音素を連結した音素連鎖である。本実施形態では、音声データVDは、男性の声を示す男声音声データと、女性の声を示す女声音声データとを含む。また、本実施形態では、音声データVDは、日本語を発音する日本語音声データと、英語を発音する英語音声データとを含む。すなわち、音声データVDは、2種類の性別と2種類の言語との組み合わせ、少なくとも4種類を含む。複数の種類の音声データVDを用いることにより、例えば楽譜のパート毎または楽譜中の記号の種類に応じて読み上げ音声の種類を変えることができ、利便性を向上させることができる。 The voice data VD is data for generating synthesized voice that reads out the musical score. The speech data VD is a speech synthesis library containing a plurality of speech segments. Each phonetic segment is a single phoneme (for example, a vowel or a consonant), which is the minimum unit of linguistic meaning, or a phoneme chain in which a plurality of phonemes are connected. In this embodiment, the voice data VD includes male voice data representing a male voice and female voice data representing a female voice. In addition, in this embodiment, the voice data VD includes Japanese voice data for pronouncing Japanese and English voice data for pronouncing English. That is, the voice data VD includes at least four combinations of two genders and two languages. By using a plurality of types of voice data VD, it is possible to change the type of reading voice for each part of a musical score or according to the type of symbols in the musical score, thereby improving convenience.
 楽譜データSDは、楽曲の楽譜を表すデータである。楽譜データSDは、例えばウェブサーバ等の配信装置(図示略)からネットワークを介して配信されることにより、または、楽譜データSDを記録した記録媒体が店舗で販売されることにより、流通する。楽譜データSDは、利用者が健常者であるか、聴覚障碍者であるか等に関わらず取得可能な一般的なデータである。本実施形態において、楽譜データSDは、楽曲の楽譜の内容を特定のデータ記述言語で記述する。具体的には、楽譜データSDは、音楽記号等の楽譜の要素が論理的な情報として表現された楽譜表現用のファイル(例えばMusicXML形式のファイル)である。 The musical score data SD is data representing the musical score of a musical piece. The musical score data SD is distributed, for example, by being distributed via a network from a distribution device (not shown) such as a web server, or by selling recording media recording the musical score data SD at stores. The musical score data SD is general data that can be obtained regardless of whether the user is a healthy person or a hearing-impaired person. In this embodiment, the musical score data SD describes the content of the musical score of a piece of music in a specific data description language. Specifically, the musical score data SD is a file for expressing musical scores (for example, a file in MusicXML format) in which elements of musical scores such as musical symbols are expressed as logical information.
 楽譜データSDは、例えば配信装置(典型的にはウェブサーバ)からインターネット等の通信網を介して情報処理装置10に配信された上で記憶装置12に格納される。記憶装置12に記憶される楽譜データSDは、複数であってもよい。一般的には、1つの楽曲に対応して1つの楽譜データSDが作成される。 The musical score data SD is stored in the storage device 12 after being distributed to the information processing device 10 via a communication network such as the Internet from, for example, a distribution device (typically a web server). A plurality of score data SD may be stored in the storage device 12 . Generally, one score data SD is created corresponding to one piece of music.
 楽譜は、演奏記号を含む音楽記号によって楽曲を書き表したものである。図3に示すように、音楽記号は、音符記号、音部記号、拍子記号、調号、および演奏記号を含む。音符記号は、音符、休符および楽譜中の音符に付加される臨時記号を含む。音部記号は、五線の左端に記され五線上の位置と音の高さとの関係を指定する。拍子記号は、一小節内における拍の数および一拍とする音符の種類を指定する。調号は、楽曲の調を指定するための変化記号のセットである。 A musical score is a representation of a piece of music using musical symbols, including performance symbols. As shown in FIG. 3, musical symbols include musical note symbols, clefs, time signatures, key signatures, and performance symbols. Musical note symbols include notes, rests and accidentals that are attached to notes in a musical score. The clef is marked on the left end of the staff and specifies the relationship between the position on the staff and the pitch of the sound. The time signature specifies the number of beats in one measure and the type of note that constitutes one beat. A key signature is a set of accidental symbols for designating the key of a piece of music.
 演奏記号は、楽曲の演奏に際し、音符や休符だけでは表現できないニュアンスを演奏者に指示するために、補助的に楽譜に記される。演奏記号には、アダージョ、アンダンテ等の速度記号、アッフェットゥオーソ、アジタート等の発想記号、フォルテシモ、ピアニッシモ等の強弱記号、テヌート、スタッカート等のアーティキュレーションを示す記号(以下「アーティキュレーション記号」という)、ダ・カーポ、セーニョ等の反復記号、トリル、ターン等の装飾記号、オッターヴァアルタ、オッターヴァバッサ等の省略記号、ペダル、ピチカート等の楽器特有の奏法を示す奏法記号等が含まれる。また、本実施形態では、楽譜に記載された音符を演奏する際に用いる指を指定する指番号も、演奏記号に含めるものとする。  Performance symbols are written on the score as a supplement to indicate to the performer the nuances that cannot be expressed with only notes and rests when performing a piece of music. Performance symbols include speed symbols such as adagio and andante, expressions such as affettuoso and agitato, dynamics such as fortissimo and pianissimo, and articulation symbols such as tenuto and staccato (hereinafter referred to as articulation symbols). ”), repeat marks such as da capo and seño, decorative marks such as trills and turns, abbreviations such as ottava alta and ottava bassa, and playing style marks indicating specific playing styles for instruments such as pedals and pizzicato. be In the present embodiment, the performance symbols also include finger numbers that designate fingers used when playing the notes written in the musical score.
 図2に示すように、楽譜データSDは、対象楽曲を構成する音楽記号毎に属性情報Bを包含する。属性情報Bは、それぞれの音楽記号の音楽的な属性を規定する情報であり、拍識別子B1と記号識別子B2とを含む。拍識別子B1は、対象楽曲における当該音楽記号の時間的な位置を指定する情報である。音楽記号が音符記号または演奏記号の場合、対象楽曲の先頭から当該音楽記号までの拍数(例えば8分音符を1拍として計数した拍番号)が拍識別子B1として好適に利用される。記号識別子B2は、音楽記号の種類を識別するための情報である。例えば音楽記号が音符記号の場合、記号識別子B2は、音名(ノートナンバー)と音価とを含む。音名は音符の音高を表し、音価は楽譜上での音符の継続長を表す。また、音楽記号が音符記号以外である場合には、音楽記号の名称を示す文字列が記号識別子B2として好適に利用される。 As shown in FIG. 2, the musical score data SD includes attribute information B for each musical symbol forming the target musical piece. The attribute information B is information that defines musical attributes of each musical symbol, and includes a beat identifier B1 and a symbol identifier B2. The beat identifier B1 is information specifying the temporal position of the musical symbol in the target music. If the music symbol is a musical note symbol or performance symbol, the number of beats from the beginning of the target song to the corresponding music symbol (for example, the beat number obtained by counting an eighth note as one beat) is preferably used as the beat identifier B1. The symbol identifier B2 is information for identifying the type of musical symbol. For example, if the music symbol is a note symbol, the symbol identifier B2 includes the note name (note number) and note value. The note name represents the pitch of the note, and the note value represents the duration of the note on the score. Also, when the musical symbol is other than the note symbol, a character string indicating the name of the musical symbol is preferably used as the symbolic identifier B2.
 また、楽譜データSDは、楽譜が示す楽曲のテンポを指定するテンポ情報TPを含む。テンポ情報TPは、例えば1分間(単位時間)当たりの拍数、および1拍とする音符の種類を含む。 The musical score data SD also includes tempo information TP that specifies the tempo of the music indicated by the musical score. The tempo information TP includes, for example, the number of beats per minute (unit time) and the type of note that constitutes one beat.
 また、楽譜データSDは、楽譜画像データMDを含む。楽譜画像データMDは、対象楽曲の楽譜の画像(以下「楽譜画像」という)を表すデータである。具体的には、例えばラスタ形式またはベクタ形式の平面画像として楽譜画像を表現する画像ファイル(例えばPDFファイル)が、楽譜画像データMDとして好適である。 Also, the musical score data SD includes musical score image data MD. The musical score image data MD is data representing an image of the musical score of the target musical piece (hereinafter referred to as "score image"). Specifically, for example, an image file (for example, a PDF file) representing a musical score image as a plane image in raster format or vector format is suitable as the musical score image data MD.
 記号テキストデータTDは、楽譜に記載された音楽記号に対応するテキストを含むデータである。記号テキストデータTDは、記号識別子C1と、名称テキストC2と、意味テキストC3とを含む。記号識別子C1は、音楽記号の種類を識別するための情報であり、属性情報Bの記号識別子B2と同一形式の情報である。音符記号に対応する記号識別子C1は、音名のみであってもよい。 The symbol text data TD is data containing text corresponding to musical symbols written on the musical score. The symbolic text data TD includes symbolic identifiers C1, name texts C2 and semantic texts C3. The symbolic identifier C1 is information for identifying the type of musical symbol, and is information in the same format as the symbolic identifier B2 of the attribute information B. FIG. The symbolic identifier C1 corresponding to the musical note symbol may be only the pitch name.
 名称テキストC2は、記号識別子C1で特定される記号の名称を示すテキストである。記号識別子C1と名称テキストC2とが同一の文字列であってもよい。音楽記号が音符記号である場合、名称テキストC2は「ド」、「レ」等である。音楽記号が演奏記号である場合、名称テキストC2は「クレッシェンド」、「フォルテ」等である。ここで、本実施形態では、楽譜読み上げアプリケーションは、多言語による楽譜の読み上げが可能であるものとする。一例として、読み上げ時に選択可能な言語を、日本語または英語とする。このため、名称テキストC2は、音楽記号の名称を示す日本語のテキストと、音楽記号の名称を示す英語のテキストとを含む。 The name text C2 is text indicating the name of the symbol specified by the symbol identifier C1. The symbolic identifier C1 and the name text C2 may be the same character string. If the musical symbol is a note symbol, the name text C2 is "do", "re", and so on. If the musical symbol is a performance symbol, the name text C2 is "crescendo", "forte", or the like. Here, in the present embodiment, it is assumed that the musical score reading application is capable of reading musical scores in multiple languages. As an example, the languages that can be selected during reading are Japanese or English. Therefore, the name text C2 includes Japanese text indicating the name of the musical symbol and English text indicating the name of the musical symbol.
 意味テキストC3は、記号識別子C1で特定される記号の意味を示すテキストである。例えば、音楽記号の名称が「アダージョ」である場合、意味テキストC3は「ゆるやかに」である。音楽記号が音符記号である場合、意味テキストC3は設けられなくてもよい。意味テキストC3も、音楽記号の意味を示す日本語のテキストと、音楽記号の意味を示す英語のテキストとを含む。 The meaning text C3 is text indicating the meaning of the symbol specified by the symbol identifier C1. For example, if the name of the musical symbol is "adagio", the semantic text C3 is "slowly". If the musical symbols are note symbols, the semantic text C3 may not be provided. The semantic text C3 also includes Japanese text indicating the meaning of the musical symbol and English text indicating the meaning of the musical symbol.
 なお、休符に対応する名称テキストC2として、例えば非言語の通知音を用いてもよい。この場合、音響信号の生成においては、休符に関する音として、非言語の通知音を表す音響信号を生成する。非言語の通知音とは、言語的な意味を有さない音であり、例えばメトロノーム音、クリック音、ビープ音等を含む。休符に関する音として、非言語の通知音を用いることにより、利用者は、通知音が発音された際に休符に対応する音であることを即座に把握することができる。本実施形態では、休符に対応する名称テキストC2として、クリック音を用いる。4分休符に対応するクリック音は「カチッ」、8分音符に対応するクリック音は「カッ」である。また、休符に対応するクリック音は、「カチッ」、「カッ」に限らず、例えば4分休符は「ウン」、8分休符は「ウ」などとしてもよい。 A non-verbal notification sound, for example, may be used as the name text C2 corresponding to the rest. In this case, in the generation of the acoustic signal, an acoustic signal representing a non-verbal notification sound is generated as the sound related to the rest. A non-verbal notification sound is a sound that has no linguistic meaning, and includes, for example, a metronome sound, a click sound, a beep sound, and the like. By using a non-verbal notification sound as the sound related to the rest, the user can immediately recognize that the notification sound corresponds to the rest when the notification sound is pronounced. In this embodiment, a click sound is used as the name text C2 corresponding to the rest. A click sound corresponding to a quarter note is "click", and a click sound corresponding to an eighth note is "ka". Also, the click sound corresponding to the rest is not limited to "click" and "ka".
 また、休符に対応する名称テキストC2として、例えば4分休符であれば「ウン」、8分休符であれば「ウ」など、拍を取る際に休符を示す語句として用いられる語を用いてもよい。 Also, as the name text C2 corresponding to the rest, a word used as a phrase indicating the rest when taking a beat, such as "Un" for a quarter rest, "U" for an eighth rest, etc. may be used.
 図4は、制御装置11の機能的な構成を例示するブロック図である。制御装置11は、楽譜を表す楽譜データSDに基づいて、楽譜を構成する音楽記号に関する音(以下「記号音」という)を表す音響信号を生成する。楽譜データSDが表す楽譜には、1以上の演奏記号が含まれている。よって、制御装置11は、1以上の演奏記号を含む楽譜を表す楽譜データSDに基づいて、演奏記号に関する音を表す音響信号を生成する、ともいえる。演奏記号に関する音は、例えば演奏記号の名称を示す音、または、演奏記号の意味に対応する語句を示す音である。本実施形態において、演奏記号の名称は演奏記号の名称テキストC2に対応し、演奏記号の意味に対応する語句は演奏記号の意味テキストC3に対応する。また、本実施形態では、楽譜は、演奏記号に加え、音符記号を含む。よって、制御装置11は、演奏記号に関する音と、音符記号に関する音とを表す音響信号を生成する。音符記号に関する音は、例えば音符記号が示す音符の音名を示す音である。 FIG. 4 is a block diagram illustrating the functional configuration of the control device 11. As shown in FIG. The control device 11 generates sound signals representing sounds related to musical symbols forming a musical score (hereinafter referred to as “symbolic sounds”) based on musical score data SD representing the musical score. The musical score represented by the musical score data SD includes one or more performance symbols. Therefore, it can also be said that the control device 11 generates an acoustic signal representing a sound related to the performance symbol based on the musical score data SD representing a musical score including one or more performance symbols. The sound related to the musical performance symbol is, for example, a sound indicating the name of the musical performance symbol or a sound indicating a phrase corresponding to the meaning of the musical performance symbol. In this embodiment, the names of the performance symbols correspond to the name texts C2 of the performance symbols, and the words corresponding to the meanings of the performance symbols correspond to the meaning texts C3 of the performance symbols. Further, in this embodiment, the musical score includes musical note symbols in addition to performance symbols. Accordingly, the control device 11 generates acoustic signals representing sounds related to performance symbols and sounds related to musical note symbols. A sound related to a musical note symbol is, for example, a sound indicating the pitch name of the note indicated by the musical note symbol.
 音響信号は、放音装置14に記号音を再生させるための信号である。制御装置11は、記憶装置12に記憶されたプログラムPGを実行することで、音響信号を生成し、再生するための複数の機能(指示受付部30、テキスト生成部32、音声合成部34、演奏解析部38、出力制御部40)を実現する。 The acoustic signal is a signal for causing the sound emitting device 14 to reproduce the symbol sound. By executing the program PG stored in the storage device 12, the control device 11 has a plurality of functions for generating and reproducing sound signals (the instruction receiving unit 30, the text generating unit 32, the voice synthesizing unit 34, the performance The analysis unit 38 and the output control unit 40) are realized.
 指示受付部30は、操作装置15に対する利用者からの指示を受け付ける。指示受付部30は、例えばタッチパネルTに利用者からの指示の受付画面を表示する。利用者は、タッチパネルTに表示された受付画面に対してタッチ操作を行うことにより、指示を入力する。 The instruction receiving unit 30 receives instructions from the user to the operation device 15. The instruction receiving unit 30 displays a screen for receiving instructions from the user on the touch panel T, for example. The user inputs an instruction by performing a touch operation on the reception screen displayed on the touch panel T. FIG.
 図5~図9は、指示受付部30による指示の受付画面を例示する図である。楽譜読み上げアプリケーションが起動されると、指示受付部30は、例えば図5に示すように、読み上げ対象とする楽曲の選択指示の受付画面SC1をタッチパネルTに表示させる。受付画面SC1には、例えば記憶装置12に記憶された楽譜データSDのデータ名を示す表示NA1~NA5が表示される。利用者は、所望の楽譜データSDに対応する表示NA1~NA5をタッチすることにより、読み上げ対象とする楽譜データSDを指定する。図5の例では、表示NA2に対応する楽譜データ「yyy.xml」が指定されている。この状態で利用者がOKボタンBTをタッチすると、楽譜データ「yyy.xml」の指定が確定する。以降の受付画面においても、OKボタンBTのタッチによって、利用者の選択指示が確定する。なお、データ名を示す表示NA1~NA5に代えて、楽譜データSDに対応する楽曲のタイトルが表示されてもよい。 5 to 9 are diagrams exemplifying instruction reception screens by the instruction reception unit 30. FIG. When the score reading application is activated, the instruction receiving unit 30 causes the touch panel T to display a reception screen SC1 for selecting a song to be read aloud, as shown in FIG. 5, for example. Indications NA1 to NA5 indicating the data names of the musical score data SD stored in the storage device 12, for example, are displayed on the acceptance screen SC1. The user designates the musical score data SD to be read out by touching the display NA1 to NA5 corresponding to the desired musical score data SD. In the example of FIG. 5, musical score data "yyy.xml" corresponding to display NA2 is specified. When the user touches the OK button BT in this state, the specification of the musical score data "yyy.xml" is confirmed. In the following reception screens as well, the user's selection instruction is confirmed by touching the OK button BT. It should be noted that instead of the displays NA1 to NA5 indicating the data names, titles of songs corresponding to the musical score data SD may be displayed.
 楽譜データSDが指定されると、指示受付部30は、例えば図6に示すように、楽譜データSDのうち読み上げ対象とする五線譜の選択指示の受付画面SC2をタッチパネルTに表示させる。受付画面SC2には、大譜表のうち読み上げ対象とする五線譜を指定するための選択肢NB1およびNB2が表示される。選択肢NB1は、大譜表の上段に位置する右手用の五線譜の読み上げを指定する。選択肢NB2は、大譜表の下段に位置する左手用の五線譜の読み上げを指定する。利用者は、選択肢NB1またはNB2の少なくともいずれかのチェックボックスCKにチェックを入れることにより、読み上げ対象の五線譜を指定する。 When the musical score data SD is specified, the instruction receiving unit 30 causes the touch panel T to display a reception screen SC2 for selecting a staff notation to be read out of the musical score data SD, as shown in FIG. The reception screen SC2 displays options NB1 and NB2 for designating a staff notation to be read aloud out of the grand staff. The option NB1 designates the reading of the right-hand staff positioned above the grand staff. The option NB2 designates the reading of the staff notation for the left hand located on the lower stage of the grand staff. The user designates the staff notation to be read out by checking the check box CK of at least one of the options NB1 and NB2.
 読み上げ対象の五線譜が指定されると、指示受付部30は、例えば図7に示すように、読み上げ対象とする記号の種類の選択指示の受付画面SC3をタッチパネルTに表示させる。受付画面SC3には、読み上げ対象とする記号の種類を指定するための選択肢NC1~NC11が表示される。選択肢NC1は、音符記号の読み上げを指定する。選択肢NC2は、演奏記号の読み上げを指定する。なお、図3に示す音楽記号のうち、音部記号、拍子記号および調号については、1つの楽譜内で変化することが稀であるため、本実施形態では継続的な読み上げ対象としていない。一方で、音部記号、拍子記号および調号についても、音符記号および演奏記号同様、読み上げ対象とするか否かを利用者が指示できてもよい。 When the staff notation to be read aloud is designated, the instruction receiving unit 30 causes the touch panel T to display a reception screen SC3 for selecting the type of symbol to be read aloud, as shown in FIG. 7, for example. Options NC1 to NC11 for designating the type of symbol to be read aloud are displayed on the acceptance screen SC3. Choice NC1 designates reading of musical note symbols. Option NC2 designates the reading of performance symbols. Of the music symbols shown in FIG. 3, the clef, time signature, and key signature are not subject to continuous reading in this embodiment because they rarely change within one score. On the other hand, the user may be able to specify whether or not to read aloud the clef, time signature, and key signature as well as the musical note symbol and performance symbol.
 また、演奏記号については、読み上げ対象とする記号の種類を更に詳細に指定可能である。選択肢NC3は、速度記号の読み上げを指定する。選択肢NC4は、発想記号の読み上げを指定する。選択肢NC5は、強弱記号の読み上げを指定する。選択肢NC6は、アーティキュレーション記号の読み上げを指定する。選択肢NC7は、反復記号の読み上げを指定する。選択肢NC8は、装飾記号の読み上げを指定する。選択肢NC9は、省略記号の読み上げを指定する。選択肢NC10は、奏法記号の読み上げを指定する。選択肢NC11は、指番号の読み上げを指定する。 Also, for performance symbols, it is possible to specify in more detail the types of symbols to be read aloud. Choice NC3 designates reading of the speed symbol. Option NC4 designates the reading of expression symbols. Option NC5 designates reading of dynamic symbols. Choice NC6 specifies the reading of articulation symbols. Choice NC7 specifies the reading of repeat symbols. Option NC8 designates reading of decorative symbols. Choice NC9 designates reading of ellipsis. Option NC10 designates reading of rendition style symbols. Option NC11 designates the reading of finger numbers.
 すなわち、楽譜データSDで表される楽譜は、複数の演奏記号を含み、複数の演奏記号の各々は、複数の分類のいずれかに属する。複数の分類とは、図3に示す演奏記号の種類に対応する。指示受付部30は、演奏記号の複数の分類のうち少なくとも1つの分類の選択を受け付ける。制御装置11は、複数の演奏記号のうち選択に係る1以上の分類に属する演奏記号について、音響信号を生成する。 That is, the musical score represented by the musical score data SD includes a plurality of performance symbols, and each of the plurality of performance symbols belongs to one of a plurality of classifications. A plurality of classifications correspond to the types of performance symbols shown in FIG. The instruction receiving unit 30 receives selection of at least one of a plurality of classifications of performance symbols. The control device 11 generates acoustic signals for performance symbols belonging to one or more selected categories among the plurality of performance symbols.
 読み上げ対象の項目が指定されると、指示受付部30は、例えば図8に示すように、楽譜データSDを読み上げる際の設定の選択指示の受付画面SC4をタッチパネルTに表示させる。受付画面SC4のうち、上部の領域E1には、楽譜の読み上げ時に出力する情報を指定するための選択肢ND1およびND2が表示される。選択肢ND1は、楽譜の読み上げのみを指定する。すなわち、選択肢ND1は、音声の出力のみを指定する。選択肢ND2は、楽譜の読み上げに加えて楽譜画像を表示することを指定する。すなわち、選択肢ND2は、音声および画像を出力することを指定する。選択肢ND1およびND2は、ラジオボタンを用いていずれかを選択可能となっている。利用者は、選択肢ND1に対応するラジオボタンまたは選択肢ND2に対応するラジオボタンをタッチすることにより、楽譜の読み上げ時に出力する情報を指定する。 When an item to be read aloud is designated, the instruction receiving unit 30 causes the touch panel T to display a reception screen SC4 for selecting a setting for reading out the musical score data SD, as shown in FIG. 8, for example. Options ND1 and ND2 for designating information to be output when reading a musical score are displayed in the upper area E1 of the reception screen SC4. The option ND1 designates only the reading of musical scores. That is, the option ND1 designates only audio output. The option ND2 designates displaying a musical score image in addition to reading out the musical score. That is, the option ND2 designates output of audio and images. Either option ND1 or ND2 can be selected using a radio button. By touching the radio button corresponding to the option ND1 or the radio button corresponding to the option ND2, the user specifies the information to be output when reading out the musical score.
 また、受付画面SC4のうち、下部の領域E2には、楽譜の読み上げのテンポを指定するための選択肢NE1~NE4が表示される。選択肢NE1は、楽譜で指定されたテンポでの読み上げを指定する。選択肢NE1を選択した場合、読み上げテンポと読み上げ音節数との関係で、受付画面SC3で指定された記号の全てを読み上げられない場合がある。この場合には、読み上げられる記号が適宜減らされる。選択肢NE2は、楽譜で指定されたテンポに関わらず、受付画面SC3で指定された記号の全てを読み上げることを指定する。選択肢NE3は、利用者による楽譜の演奏と同期したテンポでの読み上げを指定する。選択肢NE4は、利用者が任意のテンポを指定する。図の例では、利用者は、1分間当たりの拍数を指定することにより、読み上げのテンポを指定する。 Also, options NE1 to NE4 for specifying the tempo for reading out the score are displayed in the lower region E2 of the reception screen SC4. The option NE1 designates reading at the tempo designated by the musical score. When the option NE1 is selected, it may not be possible to read out all the symbols specified on the reception screen SC3 due to the relationship between the readout tempo and the number of readout syllables. In this case, the symbols to be read are reduced accordingly. The option NE2 designates reading out all the symbols designated on the reception screen SC3 regardless of the tempo designated by the musical score. Option NE3 designates reading aloud at a tempo synchronized with the performance of the musical score by the user. For option NE4, the user designates an arbitrary tempo. In the illustrated example, the user specifies the reading tempo by specifying the number of beats per minute.
 なお、読み上げテンポの指定と関連して、例えば読み上げ時における1音節の発話時間(言い換えると、単位時間当たりの読み上げ音節数)を設定できてもよい。例えば選択肢NE1が選択された場合、1音節の発話時間が短いほど受付画面SC3で指定された記号をより多く読み上げることができる。また、選択肢NE2が選択された場合、1音節の発話時間が短いほど、より短時間に読み上げを完了することができる。 In addition, in connection with specifying the reading tempo, for example, it may be possible to set the speaking time for one syllable (in other words, the number of reading syllables per unit time) during reading. For example, when the option NE1 is selected, the shorter the utterance time of one syllable, the more symbols specified on the reception screen SC3 can be read out. Further, when the option NE2 is selected, the shorter the utterance time of one syllable, the shorter the reading can be completed.
 受付画面SC4のOKボタンBTがタッチされると、指示受付部30は、例えば図9に示すように、更に他の設定に関する選択指示の受付画面SC5をタッチパネルTに表示させる。受付画面SC5のうち、上部の領域E3には、楽譜の読み上げる際に用いる言語を指定するための選択肢NF1~NF2が表示される。選択肢NF1は、日本語での読み上げを指定する。日本語での読み上げとは、例えば音名として「ド・レ・ミ・ファ・ソ・ラ・シ」を用いる、演奏記号の名称の発話に際して日本語的な発音を用いる、等に対応する。選択肢NF2は、英語での読み上げを指定する。英語での読み上げとは、例えば音名として「C・D・E・F・G・A・B」を用いる、演奏記号の名称の発話に際して英語的な発音を用いる、等に対応する。なお、受付画面SC5において、日本語および英語以外の言語が指定できてもよい。この場合、記号テキストデータTDには、当該言語の名称テキストC2および意味テキストC3が含まれる。 When the OK button BT on the reception screen SC4 is touched, the instruction reception unit 30 causes the touch panel T to display a selection instruction reception screen SC5 for further setting, as shown in FIG. 9, for example. Options NF1 to NF2 for designating the language to be used when reading out the musical score are displayed in the upper area E3 of the reception screen SC5. Option NF1 designates reading in Japanese. Reading in Japanese corresponds to, for example, using "do-re-mi-fa-so-la-si" as the note name, using Japanese-like pronunciation when uttering the names of performance symbols, and the like. Choice NF2 designates reading in English. Reading out in English corresponds to, for example, using "C, D, E, F, G, A, B" as note names, using English-like pronunciation when uttering names of performance symbols, and the like. A language other than Japanese and English may be specified on the reception screen SC5. In this case, the symbolic text data TD includes the name text C2 and the meaning text C3 of the language.
 また、受付画面SC5のうち、中部の領域E4には、演奏記号を読み上げる際の内容を指定するための選択肢NG1~NG2が表示される。選択肢NG1は、演奏記号の名称の読み上げを指定する。選択肢NG1が指定された場合、演奏記号の読み上げ時には、記号テキストデータTDのうち名称テキストC2が読み上げられる。選択肢NG2は、演奏記号の意味を示す語句の読み上げを指定する。選択肢NG2が指定された場合、演奏記号の読み上げ時には、記号テキストデータTDのうち意味テキストC3が読み上げられる。 Also, in the central area E4 of the reception screen SC5, options NG1 to NG2 are displayed for designating the contents of reading out the performance symbols. The option NG1 designates reading out the name of the performance symbol. When the option NG1 is specified, the name text C2 of the symbol text data TD is read out when the performance symbols are read out. The option NG2 designates the reading of a phrase indicating the meaning of the performance symbol. When the option NG2 is specified, the meaning text C3 of the symbol text data TD is read out when reading out the performance symbol.
 また、受付画面SC5のうち、下部の領域E5には、読み上げ音声の種類を指定するための選択肢NH1~NH2が表示される。例えば受付画面SC2で右手用の五線譜および左手用の五線譜の両方を読み上げ対象の五線譜として指定した場合、同時に複数の音符記号を示すテキストが読み上げられることがある。本実施形態では、利用者が読み上げられたテキストを識別しやすくするために、右手用の五線譜を読み上げる音声と、左手用の五線譜を読み上げる音声とを異なる種類に指定できるようにしている。すなわち、指示受付部30は、楽曲の複数のパートの各々について音声の種類を個別に設定可能である。本実施形態では、音声の種類として、男性の声である男声と、女性の声である女声とが指定できる。選択肢NH1は、右手用の五線譜を読み上げる音声を、男声または女声のいずれかに指定する。選択肢NH2は、左手用の五線譜を読み上げる音声を、男声または女声のいずれかに指定する。 In addition, options NH1 to NH2 for specifying the type of reading voice are displayed in the lower region E5 of the reception screen SC5. For example, if both the staff notation for the right hand and the staff notation for the left hand are designated as the staff notation to be read out on the reception screen SC2, texts indicating a plurality of note symbols may be read out at the same time. In this embodiment, in order to make it easier for the user to identify the read-out text, it is possible to specify different types of voices for reading the staff notation for the right hand and for reading the staff notation for the left hand. That is, the instruction receiving unit 30 can individually set the type of sound for each of the multiple parts of the music. In this embodiment, as the type of voice, male voice and female voice can be specified. The option NH1 designates either a male voice or a female voice as the voice for reading out the staff notation for the right hand. Option NH2 designates either a male voice or a female voice for reading the staff notation for the left hand.
 なお、例えば音符記号を読み上げる音声の種類と、演奏記号を読み上げる音声の種類とを異なる種類に指定できてもよい。また、音声の種類が4以上指定可能な場合、例えば右手用の五線譜上の音符記号と、左手用の五線譜上の音符記号と、右手用の五線譜上の演奏記号と、左手用の五線譜上の演奏記号とを、それぞれの異なる音声で読み上げるように指定できてもよい。 It should be noted that, for example, the type of voice for reading musical note symbols and the type of voice for reading performance symbols may be designated as different types. Also, if four or more voice types can be specified, for example, the right-hand staff note symbol, the left-hand staff note symbol, the right-hand staff performance symbol, and the left-hand staff symbol It may also be possible to designate the performance symbols to be read out in different voices.
 また、放音装置14がステレオスピーカである場合、右手用の五線譜を読み上げる音声を右側のスピーカから、左手用の五線譜を読み上げる音声を左側のスピーカから、それぞれ出力するように設定できてもよい。また、放音装置14がステレオスピーカである場合、音符記号の読み上げ音を出力するスピーカと、演奏記号の読み上げ音を出力するスピーカとが別になるように指定できてもよい。 Also, if the sound emitting device 14 is a stereo speaker, it may be set so that the right speaker reads out the staff notation for the right hand, and the left speaker reads out the staff notation for the left hand. Further, when the sound emitting device 14 is a stereo speaker, it may be possible to designate a speaker for outputting the reading sound of musical note symbols and a speaker for outputting the reading sound of performance symbols separately.
 この他、例えば、和音を読み上げる際に、和音を構成する各音符を個々に読み上げるか、和音に対応するコードネームを読み上げるかを利用者が選択できてもよい。この場合、例えば記号テキストデータTDの名称テキストC2に和音を構成する各音符の音名を示すテキストを、意味テキストC3に和音のコードネームを示すテキストを、それぞれ記憶するようにしてもよい。 In addition, for example, when reading out a chord, the user may be able to select whether to read out each note constituting the chord individually or read out the code name corresponding to the chord. In this case, for example, the name text C2 of the symbol text data TD may be stored with the text indicating the pitch name of each note forming the chord, and the semantic text C3 may be stored with the text indicating the chord name of the chord.
 これらの設定が完了し、図9のOKボタンBTが押下されると、制御装置11は、音響信号の生成処理を開始する。また、タッチパネルTには、例えば楽譜の読み上げ開始を指示するボタン(以下「演奏開始ボタン」という)が表示される。利用者は、適宜のタイミングで演奏開始ボタンを押下して、楽譜の読み上げを開始させる。 When these settings are completed and the OK button BT in FIG. 9 is pressed, the control device 11 starts the acoustic signal generation process. Further, the touch panel T displays, for example, a button for instructing the start of reading out a musical score (hereinafter referred to as "performance start button"). The user presses the performance start button at an appropriate timing to start reading out the musical score.
 図4に示すテキスト生成部32は、楽譜の内容を示すテキストを生成する。図10は、テキスト生成部32の処理を模式的に示す図である。テキスト生成部32は、図5に示す受付画面SC1で指定された楽譜データSDを読み込む(S1)。テキスト生成部32は、楽譜データSDを、右手用の五線譜を示す右手データと、左手用の五線譜を示す左手データとに分類する(S2)。右手データおよび左手データのうち、図6に示す受付画面SC2で指定された読み上げ対象の五線譜に対応するデータが、以降の処理対象となる。 The text generation unit 32 shown in FIG. 4 generates text indicating the content of the musical score. FIG. 10 is a diagram schematically showing processing of the text generation unit 32. As shown in FIG. The text generator 32 reads the score data SD specified on the reception screen SC1 shown in FIG. 5 (S1). The text generation unit 32 classifies the musical score data SD into right-hand data representing a staff notation for the right hand and left-hand data representing a staff notation for the left hand (S2). Of the right hand data and left hand data, the data corresponding to the staff notation to be read aloud specified on the reception screen SC2 shown in FIG. 6 is to be processed thereafter.
 読み上げ対象の五線譜のデータには、全ての種類の音楽記号の属性情報Bが含まれる(S3)。テキスト生成部32は、図7に示す受付画面SC3で指定された読み上げ対象の記号の属性情報Bを、読み上げ対象の五線譜のデータから抽出し、拍識別子B1に基づいて時系列に配列する(S4)。 The staff notation data to be read out includes attribute information B of all types of musical symbols (S3). The text generation unit 32 extracts the attribute information B of the symbol to be read aloud specified on the reception screen SC3 shown in FIG. ).
 テキスト生成部32は、抽出した属性情報Bの記号識別子B2と、記号テキストデータTDの記号識別子C1とを照合し、記号識別子C1に対応する名称テキストC2または意味テキストC3を読み出す(S5)。名称テキストC2または意味テキストC3のいずれを読み出すかは、図9に示す受付画面SC5の選択肢NG1またはNG2のいずれが選択されたかによって決まる。また、日本語のテキストまたは英語のテキストのいずれを読み出すかは、受付画面SC5の選択肢NF1またはNF2のいずれが選択されたかによって決まる。図上において、これらの選択内容を「読み上げ内容の指定」と表記する。読み出されたテキストは、属性情報Bと同じ順番(時系列)で配置される。以上の処理により、楽譜の内容を示すテキスト(以下「読み上げテキスト」という)が生成される(S6)。 The text generation unit 32 collates the symbolic identifier B2 of the extracted attribute information B with the symbolic identifier C1 of the symbolic text data TD, and reads out the name text C2 or the meaning text C3 corresponding to the symbolic identifier C1 (S5). Which of the name text C2 and the meaning text C3 is read depends on which of the options NG1 or NG2 on the reception screen SC5 shown in FIG. 9 is selected. Further, which of the Japanese text and the English text is read depends on which of the options NF1 and NF2 on the reception screen SC5 is selected. In the drawing, these selection contents are described as "designation of reading contents". The read texts are arranged in the same order as the attribute information B (time series). Through the above processing, a text indicating the content of the musical score (hereinafter referred to as "read-aloud text") is generated (S6).
 また、テキスト生成部32は、読み上げテキストに対してタイミングラベルを付加する(S7)。タイミングラベルは、読み上げテキストの読み上げタイミングを特定する情報である。ここで、同じ内容の読み上げテキストであっても、図8に示す受付画面SC4における楽譜の読み上げのテンポの指定に応じて、読み上げの速度は異なる。よって、テキスト生成部32は、読み上げテキストに対して、楽譜の読み上げのテンポの設定に応じたタイミングラベルを付与する。 Also, the text generation unit 32 adds a timing label to the read-out text (S7). The timing label is information specifying the reading timing of the reading text. Here, even if the read-aloud text has the same content, the read-aloud speed differs according to the designated tempo of the read-aloud of the musical score on the reception screen SC4 shown in FIG. Therefore, the text generation unit 32 gives the read-out text a timing label corresponding to the setting of the read-aloud tempo of the musical score.
 図11は、楽譜を例示する図である。また、図12および図13は、読み上げテキストの読み上げタイミングを模式的に示す図である。図11に示す楽譜Gは、例えば図5に示す受付画面SC1で読み上げ対象として指定した楽譜データ(yyy.xml)の冒頭の2小節を示す。楽譜Gは、右手用の楽譜と左手用の楽譜とを含む。楽譜Gは、テンポ情報TPに基づき、4分音符を1拍とし、1分間当たり120拍と指定されている。 FIG. 11 is a diagram illustrating musical scores. 12 and 13 are diagrams schematically showing read-out timings of read-out texts. A musical score G shown in FIG. 11 indicates, for example, the first two bars of the musical score data (yyy.xml) specified to be read out on the reception screen SC1 shown in FIG. The musical score G includes a musical score for the right hand and a musical score for the left hand. Based on the tempo information TP, the musical score G is specified as 120 beats per minute, with one quarter note as one beat.
 例えば、図8に示す受付画面SC4において、楽曲のテンポでの読み上げ(選択肢NE1)が指定された場合、図12に示すタイミングでの読み上げを行うようにタイミングラベルが付加される。図12には、右手用の楽譜の読み上げ音を示す右手用読み上げ音と、左手用の楽譜の読み上げ音を示す左手用読み上げ音とが示されている。右手用読み上げ音と左手用の楽譜の読み上げ音との間に、時間軸を示す。時間軸の1目盛り(t1)は、楽譜G内で最も短い音符である8分音符を基準としている。上述した楽譜Gのテンポに基づくと、時間軸の1目盛り(t1)は0.25秒となる。 For example, in the reception screen SC4 shown in FIG. 8, when reading aloud at the tempo of the music (option NE1) is specified, a timing label is added so as to read aloud at the timing shown in FIG. FIG. 12 shows right-hand read-out sounds representing right-hand read-out sounds and left-hand read-out sounds representing left-hand read-out sounds. A time axis is shown between the reading sound for the right hand and the reading sound of the musical score for the left hand. One scale (t1) of the time axis is based on the eighth note, which is the shortest note in the musical score G. Based on the tempo of the musical score G described above, one scale (t1) on the time axis is 0.25 seconds.
 右手用読み上げ音について説明する。読み上げが開始されると、まず0.25秒の間(期間P1)に、「メゾピアノ」と「スタッカート」と「ミ」とが読み上げられる。「メゾピアノ」と「スタッカート」と「ミ」の読み上げ順序は任意である。次に、0.25秒の間(期間P2)に、「スタッカート」「ファ」が読み上げられる。期間P1と期間P2を比較すると、期間P1の方が単位時間当たりの読み上げ音節数が多いので、期間P1の方が期間P2よりも読み上げ速度を早くする必要がある。次に、0.5秒の間(期間P3)に、「ソ」が読み上げられる。期間P2と期間P3を比較すると、期間P3の方が単位時間当たりの読み上げ音節数が少ないので、期間P3の方が期間P2よりも読み上げ速度が遅くなる。次に、0.5秒の間(期間P4)に、「ミ」が読み上げられる。期間P3と期間P4を比較すると、両者の単位時間当たりの読み上げ音節数は同一なので、期間P3と期間P4の読み上げ速度は略同一となる。次に、0.5秒の間(期間P5)に、4分休符を示すクリック音(「カチッ」)が読み上げられる。 I will explain the reading sound for the right hand. When the reading is started, "mesopiano", "staccato" and "mi" are read out for 0.25 seconds (period P1). The reading order of "mesopiano", "staccato" and "mi" is arbitrary. Next, "Staccato" and "Fa" are read aloud for 0.25 seconds (period P2). Comparing the period P1 and the period P2, the number of syllables to be read out per unit time is larger in the period P1, so the reading speed in the period P1 needs to be faster than that in the period P2. Next, for 0.5 seconds (period P3), "So" is read out. When the period P2 and the period P3 are compared, since the number of syllables to be read out per unit time is smaller in the period P3, the reading speed in the period P3 is lower than that in the period P2. Next, "mi" is read aloud for 0.5 seconds (period P4). When the period P3 and the period P4 are compared, since the number of syllables to be read out per unit time is the same in both periods, the reading speed in the period P3 and the period P4 is substantially the same. Next, for 0.5 seconds (period P5), a click sound (“click”) indicating a quarter note rest is read out.
 すなわち、楽曲のテンポでの読み上げが指定された場合、制御装置11は、テンポ情報TPが指定するテンポに対応する速度で楽曲を進行する目標時点が、演奏記号に対応する時点に到達したときに、当該演奏記号に関する音が発音されるように音響信号を生成する。 That is, when reading aloud at the tempo of a piece of music is designated, the control device 11 controls when the target point in time at which the piece of music progresses at a speed corresponding to the tempo designated by the tempo information TP reaches the point in time corresponding to the performance symbol. , generates an acoustic signal so that the sound associated with the performance symbol is pronounced.
 なお、例えば期間P1のように、単位時間当たりの読み上げ音節数が多く、利用者の理解が困難な可能性がある場合、テキスト生成部32は、読み上げ対象のテキストを減少させてもよい。例えば、読み上げ時における1音節の発話時間が決められている場合には、読み上げテキストが生成された時点で、読み上げの可否を判断可能である。テキスト生成部32は、読み上げテキストを生成語、楽曲のテンポと、1音節の発話時間とに基づいて、読み上げテキストの読み上げが時間内に行えるか否かを判断する。 It should be noted that, for example, when the number of syllables to be read out per unit time is large and the user's understanding may be difficult, as in period P1, the text generation unit 32 may reduce the text to be read out. For example, if the speech time for one syllable is determined during reading, it is possible to determine whether or not reading is possible at the time the read-aloud text is generated. The text generation unit 32 determines whether or not the read-out text can be read out within the time based on the generated words, the tempo of the music, and the utterance time of one syllable.
 テキスト生成部32は、読み上げが時間内に行えないと判断した場合、演奏記号に対応するテキストを読み上げテキストから削除し、音符記号の読み上げのみを行ってもよい。または、テキスト生成部32は、読み上げが時間内に行えないと判断した場合、それぞれの記号の読み上げ音を重ねて発音してもよい。期間P1を例にすると、「メゾピアノ」と「スタッカート」と「ミ」とを重ねて発音してもよい。 If the text generation unit 32 determines that the reading cannot be performed within the time, it may delete the text corresponding to the performance symbol from the reading text and only read the musical note symbol. Alternatively, when the text generation unit 32 determines that the reading cannot be performed within the time, the text generating unit 32 may superimpose the reading sound of each symbol. Taking the period P1 as an example, "mesopiano", "staccato" and "mi" may be pronounced together.
 また、複数の演奏記号を読み上げ対象に含む期間(例えば図12の期間P1)において、読み上げが時間内に行えないと判断した場合、テキスト生成部32は、複数の演奏記号のうち一部を読み上げ対象とし、残りの演奏記号は読み上げ対象から外してもよい。期間P1を例にすると、テキスト生成部32は、例えば「メゾピアノ」または「スタッカート」の一方を読み上げ対象とし、他方を読み上げ対象としなくてもよい。すなわち、制御装置11は、「メゾピアノ」と「スタッカート」とが重ねて発音される場合に、「メゾピアノ」または「スタッカート」のいずれかを選択して、選択した演奏記号に対応するテキストを読み上げテキストに含める。「メゾピアノ」は第1の演奏記号に関する音の一例であり、「スタッカート」は第2の演奏記号に関する音の一例である。 Further, when it is determined that reading cannot be performed within the time period during which a plurality of performance symbols are to be read aloud (for example, period P1 in FIG. 12), the text generation unit 32 reads a part of the plurality of performance symbols. The remaining performance symbols may be excluded from reading. Taking the period P1 as an example, the text generation unit 32 does not need to read out one of "mesopiano" or "staccato" and not read out the other. That is, when "mesopiano" and "staccato" are pronounced together, the control device 11 selects either "mesopiano" or "staccato" and reads out the text corresponding to the selected performance symbol. include in "Mezzo piano" is an example of a sound associated with the first performance symbol, and "staccato" is an example of a sound associated with the second performance symbol.
 なお、指示受付部30は、演奏記号の分類別に読み上げの優先度を利用者に設定させてもよい。この場合、テキスト生成部32は、優先度の低い分類に属する演奏記号のテキストから順に、読み上げテキストから削除する。 It should be noted that the instruction receiving unit 30 may allow the user to set the reading priority for each classification of performance symbols. In this case, the text generator 32 deletes from the read-aloud text in order from the text of the performance symbols belonging to the lower priority category.
 または、例えばそれぞれの演奏記号に対応する非言語音を予め定めておき、読み上げが時間内に行えないと判断した場合、演奏記号に対応するテキストに代えて、非言語音を読み上げテキストに入れてもよい。 Alternatively, for example, if a non-verbal sound corresponding to each performance symbol is determined in advance and it is determined that the reading cannot be performed within the time limit, the non-verbal sound is included in the read-aloud text instead of the text corresponding to the performance symbol. good too.
 左手用読み上げ音について説明する。左手用楽譜には、3つのピッチクラスからなる三和音が示されている。読み上げが開始されると、1小節目の和音が読み上げられる。1小節目の和音の読み上げは、2秒間継続される。本実施形態では、和音が「ドミソ」と発話されるのではなく、「ド」と「ミ」と「ソ」が、それぞれ独立した音として読み上げられる。このとき、「ド」と「ミ」と「ソ」の読み上げが同時に開始されると、利用者が音の識別を行えない可能性がある。よって、図12に示すように、「ド」と「ミ」と「ソ」の読み上げ開始タイミングをわずかにずらして開始してもよい。わずかに、とは、例えば楽譜内で最も短い音符に対応する時間より短い時間であってもよい。楽譜Gの場合には、8分音符に対応する0.25秒より短い時間が、わずかに、に対応する。 I will explain the reading sound for the left hand. The left-hand score shows triads consisting of three pitch classes. When the reading is started, the chord of the first measure is read. The reading of the chords of the first measure continues for two seconds. In this embodiment, the chord is not pronounced as "domiso", but "do", "mi", and "so" are read out as independent sounds. At this time, if reading out of "do", "mi" and "so" is started at the same time, the user may not be able to distinguish the sounds. Therefore, as shown in FIG. 12, reading start timings of "do", "mi" and "so" may be slightly shifted. Slightly may be, for example, a time less than the time corresponding to the shortest note in the musical score. In the case of the musical score G, the time less than 0.25 seconds corresponding to the eighth note corresponds slightly to .
 このように、楽曲のテンポに合わせて楽譜を読み上げることによって、利用者は楽譜に示された音符の音高とともに、楽曲のリズムを把握することができる。 In this way, by reading out the musical score according to the tempo of the musical piece, the user can grasp the rhythm of the musical piece as well as the pitches of the notes indicated in the musical score.
 また、例えば、図8に示す受付画面SC4において、全項目読み上げ(選択肢NE2)が指定された場合、図13に示すタイミングでの読み上げを行うようにタイミングラベルが付加される。全項目読み上げにおいては、単位時間当たりの読み上げ音節数を一定とする。図13には、右手用読み上げ音のみを示し、左手用読み上げ音は省略している。図13において、符号Mti(iは1から9の整数)は、メトロノーム音を示す。利用者は、メトロノーム音Mtiにより、拍の区切りを把握することができる。 Also, for example, when reading all items (option NE2) is specified on the reception screen SC4 shown in FIG. 8, a timing label is added so that reading is performed at the timing shown in FIG. In reading out all items, the number of syllables read out per unit time is fixed. FIG. 13 shows only the reading sounds for the right hand and omits the reading sounds for the left hand. In FIG. 13, symbol Mti (i is an integer from 1 to 9) indicates a metronome sound. The user can grasp the break of the beat by the metronome sound Mti.
 読み上げが開始されると、まず第1小節の読み上げが行われることを示す語句「第1小節」が読み上げられる。次に、メトロノーム音Mt1が発音され、その後、「メゾピアノ」、「スタッカート」、「ミ」、「スタッカート」、「ファ」が読み上げられる。次に、メトロノーム音Mt2が発音され、「ソ」が発音される。以降も同様に、メトロノーム音Mtiの間に、音符記号の音高を示すテキストが読み上げられる。なお、第2小節の最後の音符は2分音符であり、2拍分の長さがある。この場合、メトロノーム音Mt7の後、「ミ」の読み上げが「ミー」と伸ばして継続され、メトロノーム音Mt8が「ミー」と重ねて発音される。その後、「ミー」の読み上げが終了してからメトロノーム音Mt9が発音される。 When the reading starts, the phrase "first bar" is read aloud, indicating that the first bar will be read. Next, a metronome sound Mt1 is pronounced, after which "mesopiano", "staccato", "mi", "staccato" and "fa" are read aloud. Next, the metronome sound Mt2 is pronounced, and "so" is pronounced. Likewise, the text indicating the pitch of the musical note symbol is read aloud between the metronome sounds Mti. Note that the last note of the second bar is a half note and has a length of two beats. In this case, after the metronome sound Mt7, the reading of "mi" is extended to "me" and continued, and the metronome sound Mt8 is superimposed with "me". After that, the metronome sound Mt9 is pronounced after the reading of "Me" is finished.
 左手用読み上げ音については、右手用読み上げ音と連動させるのが好ましい。例えばメトロノーム音Mt1が発音された後、語句「第2小節」が読み上げられる直前まで「ドー」、「ミー」、「ソー」の読み上げが継続される。「ドー」、「ミー」、「ソー」の読み上げ開始タイミングは、上述のようにわずかにずらされていてもよい。その後、メトロノーム音Mt5が発音された後、メトロノーム音Mt9が発音される直前までは、「シー」、「レー」、「ソー」の読み上げが継続される。 It is preferable to link the reading sound for the left hand with the reading sound for the right hand. For example, after the metronome sound Mt1 is pronounced, reading of "do", "me", and "so" is continued until immediately before the phrase "second bar" is read. The reading start timings of "do", "me", and "so" may be slightly shifted as described above. After that, after the metronome sound Mt5 is pronounced, reading out of "shi", "re", and "so" is continued until immediately before the metronome sound Mt9 is pronounced.
 このように、制御装置11は、楽曲のテンポに関わらず演奏記号に関する音を発音した音を示す音響信号を生成してもよい。これにより、利用者は指定した種類の記号を全て把握することができ、楽譜の内容を漏れなく把握することができる。 In this way, the control device 11 may generate an acoustic signal indicating a sound produced by pronouncing a sound related to the performance symbol regardless of the tempo of the music. As a result, the user can comprehend all the symbols of the specified type, and can comprehend the content of the musical score without omission.
 なお、図8に示す受付画面SC4において、利用者による楽譜の演奏と同期したテンポでの読み上げ(選択肢NE3)が指定された場合、読み上げタイミングは事前に予測できないため、テキスト生成部32によるタイミングラベルの付加は行われなくてもよい。また利用者により任意のテンポが指定された場合(選択肢NE4)、上述した楽曲のテンポでの読み上げ(選択肢NE1)の説明における楽曲のテンポを、利用者が指定したテンポに読み替えて、同様の処理を行えばよい。 Note that if the user designates reading at a tempo synchronized with the performance of the musical score by the user (option NE3) on the reception screen SC4 shown in FIG. may not be added. If the user designates an arbitrary tempo (choice NE4), the tempo of the music in the explanation of reading out at the tempo of the music (option NE1) is replaced with the tempo designated by the user, and the same process is performed. should be done.
 なお、図8に示す受付画面SC4において、選択肢NE4は、1分間当たりの拍数を指定することにより、利用者が任意のテンポを指定する態様であった。これに限らず、例えば利用者が操作子を用いて読み上げの進行を指定するようにしてもよい。操作子は、例えばタッチパネルTに表示された操作ボタンであってもよい。また、例えば情報処理装置10と利用者が演奏する楽器とが接続されている場合には、操作子は、楽器の部材であってもよい。例えば楽器がピアノである場合、操作子としてペダルを用いることができる。この場合、例えば利用者がダンパーペダルを1回踏むと、1つの音符記号分、または1小節分等の単位で読み上げが進行され、利用者がソフトペダルを1回踏むと、1つの音符記号分、または1小節分等の単位で読み上げが後退される、などとしてもよい。 In addition, in the reception screen SC4 shown in FIG. 8, the option NE4 is a mode in which the user designates an arbitrary tempo by designating the number of beats per minute. Alternatively, for example, the user may designate the progress of reading using the manipulator. The operator may be an operation button displayed on the touch panel T, for example. Further, for example, when the information processing apparatus 10 is connected to a musical instrument played by the user, the operator may be a member of the musical instrument. For example, if the musical instrument is a piano, pedals can be used as operators. In this case, for example, when the user steps on the damper pedal once, reading progresses in units of one note symbol or one bar, and when the user steps on the soft pedal once, the reading progresses by one note symbol. , or the reading may be reversed in units of one bar or the like.
 図4に示す音声合成部34は、テキスト生成部32が生成した読み上げテキストと、音声データVDを用いて、音響信号を生成する。音声合成部34は、生成部の一例である。音声合成部34は、音声データVDに含まれる複数の音声素片のうち読み上げテキストに対応する音声素片を順次に選択し、各音声素片の音高を調整してから相互に接続することで、音響信号を生成する。読み上げテキストのうち音符記号に関する音の音高は、当該音符記号の音高と一致させてもよいし、予め定められた所定の音高であってもよい。音声合成部34が生成した音響信号が放音装置14に供給されることで、楽譜を示す音が放音装置14から再生される。 The voice synthesizing unit 34 shown in FIG. 4 uses the read-out text generated by the text generating unit 32 and the voice data VD to generate an acoustic signal. The speech synthesizer 34 is an example of a generator. The speech synthesizing unit 34 sequentially selects speech segments corresponding to the text to be read out of a plurality of speech segments included in the speech data VD, adjusts the pitch of each speech segment, and then connects them to each other. to generate an acoustic signal. The pitch of the sound related to the musical note symbol in the reading text may be matched with the pitch of the musical note symbol, or may be a predetermined pitch. By supplying the sound signal generated by the voice synthesizing unit 34 to the sound emitting device 14 , the sound indicating the musical score is reproduced from the sound emitting device 14 .
 演奏解析部38は、利用者による楽器の演奏を解析する。演奏解析部38は、例えば、楽曲のうち利用者が楽器により演奏している位置(演奏位置)を解析する。演奏解析部38は、例えば収音装置13により楽器の演奏音を収音し、演奏音の音高および継続時間を解析する。演奏解析部38は、解析した演奏音の音高と、楽譜データSD上の音符の音高とを照合し、時間軸上の複数の時点の各々において、楽曲内の演奏位置を順次に解析する。 The performance analysis unit 38 analyzes the performance of the musical instrument by the user. The performance analysis unit 38 analyzes, for example, a position (playing position) in a piece of music where the user is playing the musical instrument. The performance analysis unit 38, for example, picks up the performance sound of the musical instrument using the sound collection device 13, and analyzes the pitch and duration of the performance sound. The performance analysis unit 38 compares the pitches of the analyzed performance sounds with the pitches of the notes on the score data SD, and sequentially analyzes the performance positions in the music at each of a plurality of points in time on the time axis. .
 また、例えば楽器が電子楽器である場合、演奏解析部38は、楽器に対する操作状態を示す演奏情報を電子楽器から取得してもよい。操作状態とは、例えば電子楽器が電子ピアノである場合には、押下されている鍵の識別子および押圧力などである。この場合、演奏解析部38は、演奏情報を用いて、各時点における演奏位置を楽譜上にマッピングする。 Further, for example, if the musical instrument is an electronic musical instrument, the performance analysis unit 38 may acquire performance information indicating the operating state of the musical instrument from the electronic musical instrument. For example, if the electronic musical instrument is an electronic piano, the operating state is the identifier of the pressed key and the pressing force. In this case, the performance analysis section 38 uses the performance information to map the performance position at each point on the musical score.
 なお、第1実施形態では、演奏解析部38は、図8に示す受付画面SC4において、利用者による楽譜の演奏と同期したテンポでの読み上げ(選択肢NE3)が指定された場合にのみ動作すればよい。 In the first embodiment, the performance analysis unit 38 operates only when the user designates reading out at a tempo synchronized with the performance of the musical score (option NE3) on the acceptance screen SC4 shown in FIG. good.
 出力制御部40は、音響信号に基づく音の出力、および楽譜画像データMDに基づく楽譜画像の出力を制御する。例えば図8に示す受付画面SC4において、楽譜の読み上げのみが指定された場合(選択肢ND1)、出力制御部40は、音声合成部34により生成された音響信号が表す音を放音装置14に再生させる。また、受付画面SC4において、楽譜の読み上げと楽譜画像の表示が指定された場合(選択肢ND2)、出力制御部40は、音声合成部34により生成された音響信号が表す音を放音装置14に再生させるとともに、楽譜画像データMDを表示装置16に表示させる。 The output control unit 40 controls the output of sound based on the acoustic signal and the output of the musical score image based on the musical score image data MD. For example, in the reception screen SC4 shown in FIG. 8, when only the reading of the musical score is specified (option ND1), the output control unit 40 causes the sound emitting device 14 to reproduce the sound represented by the acoustic signal generated by the voice synthesis unit 34. Let Further, when reading out the score and displaying the score image are designated on the reception screen SC4 (option ND2), the output control unit 40 outputs the sound represented by the acoustic signal generated by the voice synthesis unit 34 to the sound emitting device 14. The musical score image data MD is displayed on the display device 16 while being reproduced.
 図14は、楽譜の読み上げ中における表示画面を例示する図である。例えば図9に示す受付画面SC5への選択指示の入力が終了した後、タッチパネルTに表示される演奏開始ボタンを利用者がタッチすると、音響信号が表す音である楽譜の読み上げ音が放音装置14から出力されるとともに、表示装置16であるタッチパネルTの表示が図14に示す表示画面SC6に切り替わる。表示画面SC6には、楽譜の読み上げ音が再生中であることを示すメッセージ601と、楽譜画像602と、一時停止ボタン604と、早送りボタン606と、巻き戻しボタン608と、聞き直しボタン610と、終了ボタン612とが表示される。 FIG. 14 is a diagram illustrating a display screen during reading of musical scores. For example, when the user touches a performance start button displayed on the touch panel T after finishing inputting a selection instruction to the reception screen SC5 shown in FIG. 14, the display of the touch panel T, which is the display device 16, is switched to the display screen SC6 shown in FIG. On the display screen SC6, a message 601 indicating that the reading sound of the musical score is being reproduced, a musical score image 602, a pause button 604, a fast forward button 606, a rewind button 608, a repeat button 610, An end button 612 is displayed.
 楽譜画像602は、読み上げ対象の楽譜データSDに含まれる楽譜画像データMDを表示した画像である。楽譜画像602には、読み上げ位置を示すバー603が重畳して表示される。出力制御部40は、読み上げテキストに付されたタイミングラベルに基づいて、楽譜画像602をスクロールさせる。この時、出力制御部40は、読み上げられている音楽記号とバー603とが重畳するように、楽譜画像602のスクロール速度を調整する。なお、読み上げ位置をバー603で表示するのに代えて、読み上げ対象となる音楽記号をハイライト表示してもよい。 The musical score image 602 is an image displaying the musical score image data MD included in the musical score data SD to be read aloud. A bar 603 indicating the reading position is superimposed on the musical score image 602 and displayed. The output control unit 40 scrolls the musical score image 602 based on the timing label attached to the reading text. At this time, the output control unit 40 adjusts the scrolling speed of the musical score image 602 so that the music symbol being read out and the bar 603 are superimposed. It should be noted that instead of displaying the read-out position with the bar 603, the musical symbols to be read-out may be highlighted.
 図14では楽譜画像602として五線譜を用いた楽譜を例示したが、これに限らず、例えば楽譜画像602としてピアノロールが表示されてもよい。また、図8に示す受付画面SC4において、楽譜の読み上げのみが指定された場合(選択肢ND1)には、楽譜画像602は表示されない。 Although FIG. 14 exemplifies a musical score using staff notation as the musical score image 602, the musical score image 602 may be displayed as a piano roll, for example. Further, in the reception screen SC4 shown in FIG. 8, when only the reading of the musical score is specified (option ND1), the musical score image 602 is not displayed.
 一時停止ボタン604、早送りボタン606、巻き戻しボタン608、聞き直しボタン610および終了ボタン612は、楽譜の読み上げに関する操作を受け付ける。一時停止ボタン604が操作された場合、出力制御部40は、楽譜の読み上げを一時停止する。早送りボタン606が操作された場合、出力制御部40は、楽譜の読み上げを早送りする。例えば早送りボタン606が1回タッチされた場合、出力制御部40は、現在の読み上げ位置を含む小節の次の小節の先頭に、読み上げ位置を変更する。巻き戻しボタン608が操作された場合、出力制御部40は、楽譜の読み上げを巻き戻す。例えば巻き戻しボタン608が1回タッチされた場合、出力制御部40は、現在の読み上げ位置を含む小節の冒頭に、読み上げ位置を変更する。聞き直しボタン610が操作された場合、出力制御部40は、読み上げ中の楽譜を最初から読み上げし直す。言い換えると、出力制御部40は、読み上げ中の楽譜の第1小節の先頭に読み上げ位置を変更する。終了ボタン612が操作されると、出力制御部40は、読み上げ中の楽譜の読み上げを終了する。 A pause button 604, a fast-forward button 606, a rewind button 608, a review button 610, and an end button 612 accept operations related to reading out the musical score. When the pause button 604 is operated, the output control unit 40 pauses reading out the score. When the fast-forward button 606 is operated, the output control unit 40 fast-forwards reading of the musical score. For example, when the fast-forward button 606 is touched once, the output control unit 40 changes the readout position to the beginning of the bar next to the bar containing the current readout position. When the rewind button 608 is operated, the output control unit 40 rewinds the reading of the musical score. For example, when the rewind button 608 is touched once, the output control unit 40 changes the readout position to the beginning of the bar containing the current readout position. When the re-listen button 610 is operated, the output control section 40 re-reads the score being read aloud from the beginning. In other words, the output control unit 40 changes the reading position to the beginning of the first bar of the musical score being read. When the end button 612 is operated, the output control unit 40 ends the reading of the score being read.
 なお、楽譜の読み上げ開始位置を利用者が指定できるようにしてもよい。例えば、出力制御部40は、読み上げ開始の指示を待機している間に、タッチパネルTに演奏開始ボタンと楽譜画像を表示する。図14を援用して説明すると、表示画面SC6におけるメッセージ601に代えて、読み上げ開始ボタンが表示される。利用者は、楽譜画像602をスクロールして、楽譜画像602上の読み上げを開始したい位置とバー603とが重なるようにする。この状態で読み上げ開始ボタンがタッチされると、楽譜画像602上のバー603と重なる位置から読み上げが開始される。または、例えば読み上げを開始したい小節の番号を指定することにより、楽譜の読み上げ開始位置を指定するようにしてもよい。 It should be noted that the user may be able to specify the reading start position of the musical score. For example, the output control unit 40 displays a performance start button and a musical score image on the touch panel T while waiting for an instruction to start reading. Referring to FIG. 14, instead of the message 601 on the display screen SC6, a reading start button is displayed. The user scrolls the musical score image 602 so that the position on the musical score image 602 where he/she wants to start reading aloud overlaps with the bar 603 . When the read-out start button is touched in this state, reading-out is started from a position overlapping the bar 603 on the musical score image 602 . Alternatively, the reading start position of the musical score may be specified by specifying the number of the bar at which reading is to be started, for example.
 また、出力制御部40は、図8に示す受付画面SC4において、利用者による楽譜の演奏と同期したテンポでの読み上げ(選択肢NE3)が指定された場合、演奏解析部38が解析した演奏位置に基づいて、読み上げ音の出力タイミングを調整する。出力制御部40は、例えば楽譜上において、演奏位置に対して所定拍進んだ位置が読み上げられるように、読み上げ音の出力速度を調整する。上記所定拍は、利用者が指定できてもよい。 8, the output control unit 40 selects reading aloud at a tempo synchronized with the performance of the musical score by the user (option NE3). Based on this, adjust the output timing of the reading sound. The output control unit 40 adjusts the output speed of the reading sound so that, for example, on the musical score, the position ahead of the performance position by a predetermined beat is read out. The predetermined beat may be specified by the user.
 また、他の例として、出力制御部40は、例えば演奏位置が第N小節(Nは1以上の整数)の場合において、第N小節の演奏が終了する直前に、第N+1小節に含まれる音楽記号を読み上げるようにしてもよい。第N小節の演奏が終了する直前とは、例えば第N小節の末尾の音符が演奏された後である。これは、例えば合唱の指揮者が、合唱のメンバーに対して次に歌う歌詞を先読みして示す教示法を模している。 As another example, when the performance position is the N-th measure (N is an integer equal to or greater than 1), the output control unit 40 may control the music included in the N+1-th measure immediately before the performance of the N-th measure to end. The symbols may be read aloud. Immediately before the performance of the N-th measure ends is, for example, after the last note of the N-th measure is played. This imitates, for example, a teaching method in which a chorus conductor prereads and shows lyrics to be sung next to the chorus members.
 図15は、制御装置11が楽譜読み上げアプリケーションを実行する処理の具体的な手順を例示するフローチャートである。例えば操作装置15に対する利用者からの指示を契機として楽譜読み上げアプリケーションが起動される。 FIG. 15 is a flow chart illustrating a specific procedure of processing for the control device 11 to execute the musical score reading application. For example, an instruction from the user to the operation device 15 triggers the start of the musical score read-aloud application.
 楽譜読み上げアプリケーションが起動されると、制御装置11(指示受付部30)は、図5~図9に示す受付画面SC1~SC5を表示して、楽譜の読み上げに関する各種の指定を利用者から受け付ける(S100)。各種の指定とは、例えば読み上げの対象とする楽譜データSDの指定、読み上げ対象とする記号の種類の指定、読み上げ言語の指定等である。 When the score reading application is started, the control device 11 (instruction receiving unit 30) displays reception screens SC1 to SC5 shown in FIGS. S100). The various designations include, for example, designation of musical score data SD to be read aloud, designation of types of symbols to be read aloud, designation of a reading language, and the like.
 制御装置11(テキスト生成部32)は、S100で受け付けた各種の指定の内容と、楽譜データSDと、記号テキストデータTDとに基づいて、読み上げテキストを生成する(S102)。制御装置11(音声合成部34)は、読み上げテキストと音声データVDとを用いて、音声合成により読み上げテキストを読み上げた音響信号を生成する(S104)。制御装置11(出力制御部40)は、利用者から楽譜の読み上げの指示があるまで待機する(S106:NO)。制御装置11(出力制御部40)は、利用者から楽譜の読み上げの指示があると(S106:YES)、放音装置14から音響信号を再生し(S108)、本フローチャートによる処理を終了する。 The control device 11 (text generation unit 32) generates reading text based on the contents of various designations received in S100, the musical score data SD, and the symbol text data TD (S102). The control device 11 (speech synthesizing unit 34) uses the read-out text and the voice data VD to generate an acoustic signal that reads out the read-out text by voice synthesis (S104). The control device 11 (output control unit 40) waits until the user gives an instruction to read out the musical score (S106: NO). When the user instructs to read out the musical score (S106: YES), the control device 11 (output control unit 40) reproduces the sound signal from the sound emitting device 14 (S108), and ends the processing according to this flowchart.
 なお、読み上げテキストの生成(S102)および音響信号の生成(S104)は、利用者からの楽譜の読み上げの指示(S106:YES)の後に行ってもよい。 It should be noted that the generation of the read-out text (S102) and the generation of the acoustic signal (S104) may be performed after the user instructs to read out the score (S106: YES).
 以上の通り、第1実施形態においては、1以上の演奏記号を含む楽譜データSDに基づいて、演奏記号に関する音を表す音響信号が生成される。したがって、楽譜に含まれる演奏記号を聴覚により把握することができ、例えば視覚障碍者、または読譜に慣れていない初学者や小さい子供においても、楽譜の把握を容易とすることができる。 As described above, in the first embodiment, acoustic signals representing sounds related to performance symbols are generated based on musical score data SD including one or more performance symbols. Therefore, the performance symbols included in the musical score can be comprehended aurally, and even visually impaired people, beginners and small children who are not accustomed to reading musical scores can easily comprehend the musical score.
 また、第1実施形態においては、演奏記号に関する音は、演奏記号の名称を示す音、または、演奏記号の意味に対応する語句を示す音である。演奏記号に関する音が、演奏記号の名称を示す音である場合には、楽譜上の記述を正確に把握することができる。また、演奏記号に関する音が、演奏記号の意味に対応する語句を示す音である場合には、利用者が演奏記号の知識に乏しく、演奏記号の名称のみではその意味が理解できない場合においても、楽譜が示す内容を把握することができる。 In addition, in the first embodiment, the sound related to the performance symbol is a sound indicating the name of the performance symbol or a sound indicating a phrase corresponding to the meaning of the performance symbol. If the sound associated with the performance symbol is the sound indicating the name of the performance symbol, the description on the musical score can be accurately grasped. In addition, if the sound related to the musical performance symbol is a sound indicating a phrase corresponding to the meaning of the musical performance symbol, even if the user lacks knowledge of the musical performance symbol and cannot understand the meaning of the performance symbol only by the name of the musical performance symbol, Able to grasp the contents indicated by musical scores.
 また、第1実施形態においては、楽曲のテンポに対応するタイミングで演奏記号に対応する音を発音する。これにより、利用者は、楽曲内における演奏記号の位置を把握しやすくなり、利便性を向上させることができる。 Also, in the first embodiment, sounds corresponding to performance symbols are pronounced at timings corresponding to the tempo of music. This makes it easier for the user to grasp the positions of the performance symbols in the music, thereby improving convenience.
 また、第1実施形態においては、第1の演奏記号に関する音と第2の演奏記号に関する音が重ねて発音される場合に、第1の演奏記号または第2の演奏記号のいずれかを選択して音響信号を生成する。これにより、第1の演奏記号に関する音と第2の演奏記号に関する音が重ねて発音されることがなく、演奏記号に関する音の聞き取り易さを向上させることができる。 Further, in the first embodiment, when a sound related to the first performance symbol and a sound related to the second performance symbol are superimposed, either the first performance symbol or the second performance symbol is selected. to generate an acoustic signal. As a result, the sound related to the first performance symbol and the sound related to the second performance symbol do not overlap, and the audibility of the sound related to the performance symbol can be improved.
 また、第1実施形態においては、楽曲のテンポに関わらず演奏記号に対応する音が発音される。これにより、演奏記号に関する音が重ねて発音されることがなく、演奏記号に関する音の聞き取り易さを向上させることができる。 Also, in the first embodiment, sounds corresponding to performance symbols are pronounced regardless of the tempo of the music. As a result, the sounds related to the performance symbols are not over-pronounced, and the audibility of the sounds related to the performance symbols can be improved.
 また、第1実施形態においては、複数の分類の中から選択された分類に属する演奏記号について音響信号を生成する。したがって、利用者が必要とする演奏記号に関する音を選択的に発音させることができ、利便性を向上させることができる。 Also, in the first embodiment, acoustic signals are generated for performance symbols belonging to a category selected from a plurality of categories. Therefore, it is possible to selectively produce sounds related to performance symbols required by the user, thereby improving convenience.
 また、第1実施形態においては、演奏記号に関する音に加え、音符記号に関する音を表す音響信号を生成する。したがって、楽譜に含まれる音符記号を聴覚により把握することができ、楽譜の把握を更に容易とすることができる。 Also, in the first embodiment, in addition to the sound related to the performance symbol, an acoustic signal representing the sound related to the musical note symbol is generated. Therefore, the musical note symbols included in the musical score can be comprehended aurally, and the comprehension of the musical score can be further facilitated.
 また、第1実施形態においては、休符に関する音として、非言語の通知音が発音される。これにより、利用者は、休符に関する音が発音された際に休符に対応する音であることを即座に把握することができる。 Also, in the first embodiment, a non-verbal notification sound is pronounced as the sound related to the rest. As a result, the user can immediately recognize that the sound associated with the rest corresponds to the rest when the sound related to the rest is pronounced.
B:第2実施形態
 第2実施形態を説明する。なお、以下に例示する各態様において機能が第1実施形態と同様である要素については、第1実施形態の説明と同様の符号を流用して各々の詳細な説明を適宜に省略する。
B: Second Embodiment A second embodiment will be described. In each aspect illustrated below, elements having the same functions as those of the first embodiment are denoted by the same reference numerals as in the description of the first embodiment, and detailed descriptions thereof are appropriately omitted.
 第1実施形態では、情報処理装置10に記憶された複数の楽譜データSDのうち、利用者が指定した楽譜データSDの読み上げを行った。一方で、例えば図5に示す受付画面SC1のような楽譜データSDのデータ名の一覧が表示されても、利用者が所望の楽曲に対応する楽譜データSDを識別できない場合がある。第2実施形態では、複数の楽譜データSDの一部分を連続的に読み上げ、利用者が所望の楽曲に対応する楽譜データSDを特定できるようにする。複数の楽譜データSDの一部分を連続的に読み上げる機能を、以下「目次提示機能」と称する。 In the first embodiment, among the plurality of musical score data SD stored in the information processing device 10, the musical score data SD specified by the user is read out. On the other hand, even if a list of data names of the musical score data SD such as the reception screen SC1 shown in FIG. 5 is displayed, the user may not be able to identify the musical score data SD corresponding to the desired song. In the second embodiment, a part of a plurality of musical score data SD is continuously read out so that the user can specify the musical score data SD corresponding to the desired music piece. The function of continuously reading a part of a plurality of musical score data SD is hereinafter referred to as a "table of contents presentation function".
 図16は、指示受付部30による指示の受付画面を例示する図である。第2実施形態において楽譜読み上げアプリケーションが起動されると、指示受付部30は、例えば図16に示すようなメニュー選択指示の受付画面SC7をタッチパネルTに表示させる。受付画面SC7には、選択肢NI1およびNI2が表示される。選択肢NI1は、第1実施形態に示すような、利用者が選択した楽譜データSDの読み上げを指定する。選択肢NI1がタッチされた場合、指示受付部30は、図5に示す受付画面SC1を表示させ、利用者から読み上げ対象の楽譜データSDの指定を受け付ける。 FIG. 16 is a diagram illustrating an instruction reception screen by the instruction reception unit 30. FIG. In the second embodiment, when the score reading application is activated, the instruction receiving unit 30 causes the touch panel T to display a menu selection instruction receiving screen SC7 as shown in FIG. 16, for example. Choices NI1 and NI2 are displayed on the acceptance screen SC7. The option NI1 specifies reading out the musical score data SD selected by the user as shown in the first embodiment. When the option NI1 is touched, the instruction receiving unit 30 displays the receiving screen SC1 shown in FIG. 5, and receives the designation of the musical score data SD to be read out from the user.
 選択肢NI2は、目次提示機能の実行を指定する。選択肢NI2では、目次提示機能が「メロディ目次」と称されている。選択肢NI2が選択されると、テキスト生成部32は、記憶装置12に記憶された複数の楽譜データSDのそれぞれについて、楽譜データSDが表す楽譜の一部分を読み上げるための読み上げテキストを生成する。楽譜の一部分には、例えば演奏記号と音符記号とが含まれる。 Option NI2 designates execution of the table of contents presentation function. In option NI2, the table of contents presentation function is called "melody table of contents". When the option NI2 is selected, the text generator 32 generates read-out text for reading a part of the musical score represented by the musical score data SD for each of the plurality of musical score data SD stored in the storage device 12. A portion of the musical score includes, for example, performance symbols and musical note symbols.
 楽譜の一部分は、例えば、楽曲を音楽的な意味に応じて区画した複数の区間(以下「構造区間」という)のうち特定の構造区間の一部または全部である。構造区間は、例えば、イントロ、Aメロ、Bメロ、サビおよびアウトロ等の区間である。具体的には、テキスト生成部32は、例えば楽曲のうち「サビ」の構造区間を読上げるためのテキストを複数の楽譜データSDの各々について生成する。または、テキスト生成部32は、例えば楽曲のうち「イントロ」(楽譜の冒頭の所定数の小節)の構造区間を読上げるためのテキストを複数の楽譜データSDの各々について生成する。 A part of the musical score is, for example, a part or all of a specific structural section among multiple sections (hereinafter referred to as "structural sections") that divide a piece of music according to its musical meaning. Structural sections are, for example, sections such as an intro, an A melody, a B melody, a chorus, and an outro. Specifically, the text generation unit 32 generates, for each of the plurality of musical score data SD, a text for reading out, for example, the structure section of the “chorus” of the music. Alternatively, the text generation unit 32 generates, for each of the plurality of musical score data SD, a text for reading out the structural section of the "intro" (predetermined number of bars at the beginning of the musical score) of the music, for example.
 なお、楽譜中における位置情報(例えば小節番号)と構造区間との対応関係を示す情報を楽譜データSDに含めておいてもよい。また、目次提示機能の実行時においても、第1実施形態で指示受付部30が行った各種の指示の受け付け(図6~図9参照)を行ってもよい。音声合成部34は、各楽譜データSDから生成された読み上げテキストと、音声データVDを用いて、音響信号を生成する。出力制御部40は、音響信号に基づく音を放音装置14に再生させる。 It should be noted that the musical score data SD may include information indicating the correspondence relationship between position information (for example, bar numbers) in the musical score and structure sections. Further, various instructions (see FIGS. 6 to 9) performed by the instruction receiving unit 30 in the first embodiment may be received even when the table of contents presentation function is executed. The voice synthesizing unit 34 generates an acoustic signal by using the reading text generated from each musical score data SD and the voice data VD. The output control unit 40 causes the sound emitting device 14 to reproduce sound based on the acoustic signal.
 図17は、目次提示機能の実行中の表示画面を例示する図である。表示画面SC8には、記憶装置12に記憶された楽譜データSDのデータ名を示す表示NA1~NA5が表示される。表示NA1~NA5は縦方向に配置されており、表示NA1で示される楽譜データ「xxx.xml」から順に、楽譜の一部分が読み上げられる。楽譜データ「xxx.xml」の読み上げが終了すると、表示NA2で示される楽譜データ「yyy.xml」の読み上げが開始される。読み上げ中の楽譜データに対応する表示(図17においては表示NA2)は、他の表示と異なる背景色で表示されてもよい。 FIG. 17 is a diagram exemplifying the display screen during execution of the table of contents presentation function. Displays NA1 to NA5 indicating the data names of the musical score data SD stored in the storage device 12 are displayed on the display screen SC8. The displays NA1 to NA5 are arranged in the vertical direction, and a part of the musical score is read out in order from the musical score data "xxx.xml" indicated by the display NA1. When the reading of the musical score data "xxx.xml" is completed, the reading of the musical score data "yyy.xml" indicated by display NA2 is started. The display corresponding to the musical score data being read (display NA2 in FIG. 17) may be displayed in a background color different from that of other displays.
 なお、図17の表示画面SC8において、記憶装置12に記憶された複数の楽譜データSDのうち、目次提示機能において読み上げの対象とする楽譜データSDを利用者が選択できるようにしてもよい。また、目次提示機能における楽譜データSDの読み上げの順序を、利用者が指定できるようにしてもよい。 It should be noted that the display screen SC8 of FIG. 17 may allow the user to select the musical score data SD to be read aloud by the table of contents presentation function from among the plurality of musical score data SD stored in the storage device 12. Further, the user may be allowed to specify the order of reading out the musical score data SD in the table of contents presentation function.
 また、表示画面SC8には、図14に示す表示画面SC6と同様の、一時停止ボタン604と、早送りボタン606と、巻き戻しボタン608と、聞き直しボタン610と、終了ボタン612とが表示される。利用者は、目次提示機能の実行中においても、一時停止ボタン604、早送りボタン606、巻き戻しボタン608、聞き直しボタン610および終了ボタン612を用いて、楽譜の読み上げに関する操作を行うことができる。 Further, the display screen SC8 displays a pause button 604, a fast forward button 606, a rewind button 608, a repeat button 610, and an end button 612 similar to the display screen SC6 shown in FIG. . The user can use the pause button 604, the fast-forward button 606, the rewind button 608, the re-listen button 610, and the end button 612 to perform operations related to reading out the musical score even while the table of contents presentation function is being executed.
 すなわち、第2実施形態において、楽譜データSDは第1楽譜データであり、記憶装置12には、第1楽譜データと異なる第2楽譜データも記憶されている。制御装置11は、第1楽譜データに対応する第1楽譜の一部分に含まれる演奏記号に関する音と、音符記号に関する音とを示す第1音響信号の生成と、第2楽譜データに対応する第2楽譜の一部分に含まれる演奏記号に関する音と、音符記号に関する音とを示す第2音響信号の生成とを行う。また、制御装置11は、第1音響信号と、第2音響信号とを放音装置14に順次に再生させる。例えば、第1楽譜データは楽譜データ「xxx.xml」であり、第2楽譜データは楽譜データ「yyy.xml」である。 That is, in the second embodiment, the musical score data SD is the first musical score data, and the storage device 12 also stores second musical score data different from the first musical score data. The control device 11 generates a first acoustic signal representing a sound related to the performance symbol and a sound related to the musical note symbol included in a portion of the first musical score corresponding to the first musical score data, and a second acoustic signal corresponding to the second musical score data. Generating a second acoustic signal indicating sounds related to performance symbols and sounds related to musical note symbols included in a portion of the musical score. Further, the control device 11 causes the sound emitting device 14 to sequentially reproduce the first acoustic signal and the second acoustic signal. For example, the first musical score data is musical score data "xxx.xml", and the second musical score data is musical score data "yyy.xml".
 第2実施形態によれば、制御装置11は、複数の楽譜データから各々一部分を選択し、選択された一部分に含まれる演奏記号および音符記号に関する音を順次再生する。これにより、利用者は、複数の楽譜データSDが、それぞれどの楽曲の楽譜に対応するものかを容易に把握することができ、複数の楽譜データSDの中から所望の楽譜データSDを迅速に選択することができる。 According to the second embodiment, the control device 11 selects a portion from each of a plurality of musical score data, and sequentially reproduces sounds related to performance symbols and musical note symbols included in the selected portion. As a result, the user can easily grasp which musical score of each piece of musical score data SD corresponds to, and quickly select the desired musical score data SD from among the plurality of musical score data SD. can do.
C:第3実施形態
 第3実施形態を説明する。なお、以下に例示する各態様において機能が第1実施形態と同様である要素については、第1実施形態の説明と同様の符号を流用して各々の詳細な説明を適宜に省略する。
C: Third Embodiment A third embodiment will be described. In each aspect illustrated below, elements having the same functions as those of the first embodiment are denoted by the same reference numerals as in the description of the first embodiment, and detailed descriptions thereof are appropriately omitted.
 第1実施形態では、情報処理装置10は、楽譜データSDの読み上げを行った。第3実施形態では、情報処理装置10は、楽譜データSDの読み上げに加えて、利用者の演奏音が楽譜に示された音に近づくように支援する。 In the first embodiment, the information processing device 10 reads out the musical score data SD. In the third embodiment, in addition to reading out the score data SD, the information processing apparatus 10 assists the user in making the sound of the performance closer to the sound shown in the score.
 図18は、第3実施形態における制御装置11Aの機能的な構成を例示するブロック図である。制御装置11Aは、第1実施形態に係る制御装置11(図4参照)の構成に加えて、演奏評価部42を備える。演奏評価部42は、演奏解析部38による解析結果に基づいて、利用者による楽器の演奏を評価する。ここで、第1実施形態では、演奏解析部38は、利用者による楽器の演奏位置を解析した。第3実施形態では、演奏解析部38は、演奏位置の解析に加えて、楽器の演奏音の大きさを解析する。 FIG. 18 is a block diagram illustrating the functional configuration of the control device 11A in the third embodiment. The control device 11A includes a performance evaluation section 42 in addition to the configuration of the control device 11 (see FIG. 4) according to the first embodiment. The performance evaluation section 42 evaluates the performance of the musical instrument by the user based on the analysis result of the performance analysis section 38 . Here, in the first embodiment, the performance analysis unit 38 analyzed the performance position of the musical instrument by the user. In the third embodiment, the performance analysis unit 38 analyzes the volume of the performance sound of the musical instrument in addition to the analysis of the performance position.
 演奏評価部42は、利用者による演奏が、楽譜の音楽記号に沿っているかを評価する。より詳細には、演奏評価部42は、利用者により楽曲を演奏した音である演奏音と、楽曲を示す楽譜に含まれる音楽記号が示す音との差分を検出し、差分が予め定められた許容範囲から外れているか否かを判定する。 The performance evaluation unit 42 evaluates whether the user's performance conforms to the musical symbols of the score. More specifically, the performance evaluation unit 42 detects the difference between the performance sound, which is the sound of the musical composition played by the user, and the sound indicated by the musical symbols included in the musical score representing the musical composition, and determines that the difference is determined in advance. Determine whether or not it is out of the allowable range.
 例えば音符記号に沿った演奏か否かの評価は、利用者による演奏音の音高と楽譜上の音符の音高との差分、および、演奏音における音の継続長と楽譜上の音符の音価との差分を検出することによって行う。演奏評価部42は、例えば上記差分が小さいほど、楽譜の音符記号に沿って演奏が行われている、すなわち演奏の技術が高いと評価する。 For example, the evaluation of whether or not the performance follows the musical notation is based on the difference between the pitch of the sound played by the user and the pitch of the note on the score, and the duration of the sound in the performance and the sound of the note on the score. by detecting the difference between the For example, the smaller the difference, the performance evaluation unit 42 evaluates that the performance is being performed along the musical note symbols of the musical score, that is, the performance skill is high.
 利用者は、例えば自身の演奏の熟練度に基づいて、差分の許容範囲を設定しておく。一般には、熟練度が高い利用者ほど、許容できる差分は小さくなると考えられる。差分が許容範囲から外れた箇所がある場合、テキスト生成部32は、当該箇所を指摘するテキストを生成する。具体的には、例えば「右手、第2小節、『ファ、レ、ミー』が、『ファ、ミ、レー』となっていました」など、楽譜に記された正しい音高および音価を読み上げるとともに、利用者が行った演奏における音高および音価を読み上げるテキストを生成する。このようなテキストを「支援テキスト」という。 The user sets the allowable range of the difference, for example, based on his or her performance skill level. Generally, it is considered that the allowable difference becomes smaller as the skill level of the user increases. If there is a portion where the difference is out of the allowable range, the text generation unit 32 generates text pointing out the portion. Specifically, for example, "Right hand, 2nd measure, 'Fa, re, me' was changed to 'Fa, mi, le'." In addition, it generates a text that reads out the pitch and duration of the performance performed by the user. Such text is referred to as "supporting text".
 また、演奏記号に沿った演奏か否かは、演奏記号毎に判断される。演奏評価部42は、例えば演奏記号が強弱記号の場合には利用者による演奏音の音量と強弱記号に沿った演奏を行った場合の音量との差分を検出することによって演奏を評価する。また、演奏評価部42は、例えば演奏記号がアーティキュレーション記号の場合には利用者による演奏音の継続長とアーティキュレーション記号に沿った演奏を行った場合の継続長との差分を検出することによって行う。演奏評価部42は、例えば上記差分が小さいほど、楽譜の演奏記号に沿って演奏が行われている、すなわち演奏の技術が高いと評価する。 Also, whether or not the performance follows the performance symbols is determined for each performance symbol. For example, when the performance symbol is a dynamic symbol, the performance evaluation unit 42 evaluates the performance by detecting the difference between the volume of the sound played by the user and the volume of the performance along the dynamic symbol. Further, the performance evaluation section 42 detects the difference between the duration of the sound played by the user when the performance symbol is an articulation symbol, and the duration when the performance is performed along the articulation symbol. by doing. For example, the smaller the difference, the performance evaluation unit 42 evaluates that the performance is performed along the performance symbols of the musical score, that is, the performance skill is high.
 利用者は、例えば自身の演奏の熟練度に基づいて、差分の許容範囲を設定しておく。差分が許容範囲から外れた箇所がある場合、テキスト生成部32は、当該箇所を指摘する支援テキストを生成する。具体的には、例えば「右手、第1小節、『スタッカート、ミ、スタッカート、ファ』、スタッカートの弾みが弱いです」など、楽譜に記された演奏記号を読み上げるとともに、利用者が行った演奏が演奏記号を反映していなかったことを示す支援テキストを生成する。 The user sets the allowable range of the difference, for example, based on his or her performance skill level. If there is a portion where the difference is out of the allowable range, the text generation unit 32 generates support text pointing out the portion. Specifically, for example, "Right hand, 1st measure, 'staccato, mi, staccato, fa', staccato momentum is weak," etc. Generate supporting text to indicate that the performance symbols were not reflected.
 音声合成部34は、支援テキストと、音声データVDを用いて、音響信号を生成する。支援テキストのうち、音符記号の音高を示すテキストは、当該音高に対応する高さの音声で読み上げてもよい。出力制御部40は、音響信号に基づく音を放音装置14に再生させる。 The speech synthesis unit 34 uses the supporting text and the speech data VD to generate an acoustic signal. Of the supporting text, the text indicating the pitch of the musical note symbol may be read aloud with a voice of a pitch corresponding to the pitch. The output control unit 40 causes the sound emitting device 14 to reproduce sound based on the acoustic signal.
 音響信号の再生は、利用者による当該楽曲の演奏が終了してから行ってもよいし、演奏の途中で行ってもよい。演奏の途中で行う場合、出力制御部40は、例えば上記許容範囲から外れた差分が生じたら即座に支援テキストを再生してもよい。この場合、支援テキストの再生後、出力制御部40は、上記許容範囲から外れた差分が生じた箇所(支援テキストにより指摘された箇所)の演奏を再度行うように促す音声を再生してもよい。利用者が支援テキストにより指摘された箇所の演奏を行うと、演奏評価部42は、演奏が楽譜の音楽記号に沿っているかを評価し、上記の処理を繰り返す。これにより、利用者が苦手な箇所(楽譜の音楽記号に沿った演奏をしにくい箇所)の反復練習を促すことができ、利用者は、楽譜に表された楽曲の演奏を効率的に習得することができる。 The audio signal may be played after the user has finished playing the music, or may be played during the performance. When performing during the performance, the output control unit 40 may immediately reproduce the support text when the difference is out of the allowable range, for example. In this case, after reproducing the support text, the output control unit 40 may reproduce a voice prompting the player to replay the part where the difference is out of the allowable range (the part pointed out by the support text). . When the user plays the part pointed out by the support text, the performance evaluation unit 42 evaluates whether the performance follows the musical symbols of the score and repeats the above process. As a result, it is possible to encourage the user to repeatedly practice the parts that the user is not good at (the parts that are difficult to play according to the musical symbols of the score), so that the user can efficiently master the performance of the music represented by the score. be able to.
 なお、例えば利用者の演奏音を録音しておき、録音した演奏音のうち支援テキストにおいて指摘された箇所に対応する部分を、支援テキストの読み上げとともに再生するようにしてもよい。 It should be noted that, for example, the performance sound of the user may be recorded, and the part of the recorded performance sound corresponding to the point pointed out in the support text may be reproduced together with the support text read aloud.
 また、支援テキストを生成する際に、演奏記号については、差分の有無に関わらず、常時読み上げるようにしてもよい。この場合、例えば演奏記号と演奏との差分が大きい場合には、読み上げ音声を大きくする(差分が大きいほど読み上げ音声を大きくする)ことにより、演奏記号に沿った演奏ができているかを、利用者が把握できるようにしてもよい。 Also, when generating the support text, the performance symbols may always be read aloud regardless of the presence or absence of differences. In this case, for example, if the difference between the performance symbols and the performance is large, the read-out voice is increased (the greater the difference, the louder the read-out voice is) so that the user can check whether the performance follows the performance symbols. can be grasped.
 すなわち、第3実施形態において、制御装置11は、利用者により楽曲を演奏した音である演奏音を取得し、楽曲を示す楽譜に含まれる音楽記号が示す音と、演奏音との差分を検出する。制御装置11は、差分が予め定められた許容範囲から外れた場合、差分が生じた箇所に対応する楽譜の一部分に含まれる音楽記号に関する音を示す音響信号を生成する。 That is, in the third embodiment, the control device 11 acquires the performance sound, which is the sound of the music played by the user, and detects the difference between the sound indicated by the musical symbols included in the musical score representing the music and the performance sound. do. When the difference is out of a predetermined allowable range, the control device 11 generates an acoustic signal representing a sound related to the musical symbol included in the portion of the musical score corresponding to the location where the difference occurs.
 第3実施形態によれば、利用者は、自身の演奏と楽譜が示す内容との差分を把握することができ、楽譜が示す楽曲の演奏を効率的に習得することができる。具体的には、制御装置11は、上記差分が生じた箇所の音楽記号を読み上げることにより、差分が生じた箇所の楽譜上の位置を利用者に示す。これにより、利用者は、例えば単に楽譜内の位置(小節番号など)が機械的に読み上げられるのと比較して、差分が生じている楽譜上の箇所を直感的に把握することができる。また、第3実施形態では、利用者の演奏の内容を言語化する。例えば、制御装置11は、利用者が行った演奏における音高および音価を読み上げる。これにより、利用者は、自身の誤りの内容を客観的に把握することができる。 According to the third embodiment, the user can grasp the difference between his/her own performance and the content indicated by the musical score, and can efficiently master the performance of the musical piece indicated by the musical score. Specifically, the control device 11 indicates to the user the position on the musical score of the portion where the difference occurs by reading out the musical symbol of the portion where the difference occurs. As a result, the user can intuitively grasp the location on the musical score where the difference occurs, compared to, for example, simply reading out the position (bar number, etc.) in the musical score mechanically. Further, in the third embodiment, the content of the user's performance is verbalized. For example, the control device 11 reads aloud the pitch and value of the performance performed by the user. This allows the user to objectively grasp the details of his/her own error.
D:変形例
 以上に例示した各態様に付加される具体的な変形の態様を以下に例示する。以下の例示から任意に選択された複数の態様を、相互に矛盾しない範囲で適宜に併合してもよい。
D: Modifications Examples of specific modifications added to the above-exemplified embodiments are given below. A plurality of aspects arbitrarily selected from the following examples may be combined as appropriate within a mutually consistent range.
(1)前述の各形態においては、音声合成部34は、素片接続型の音声合成を行ったが、音声合成の方法は以上の例示に限定されない。例えば、深層ニューラルネットワークまたはHMM(Hidden Markov Model)等の統計モデルを利用した統計モデル型の音声合成が利用されてもよい。 (1) In each of the above-described embodiments, the speech synthesizing unit 34 performs segment-connected speech synthesis, but the method of speech synthesis is not limited to the above examples. For example, statistical model type speech synthesis using statistical models such as deep neural networks or HMMs (Hidden Markov Models) may be used.
(2)前述の各形態においては、楽譜データSDを用いて楽譜に含まれる記号に関する音を読み上げた。楽譜データSDの利用方法はこれに限らず、楽譜データSDを用いて演奏音を再生してもよい。具体的には、例えば楽譜データSDが表す楽譜の左手用の五線譜の演奏音を情報処理装置10が再生するとともに、右手用の五線譜内の音楽記号の読み上げを行ってもよい。利用者は、音楽記号の読み上げ音を聞きながら右手での演奏の練習を行う。左手用の五線譜の演奏音が再生されることにより、利用者は、右手での演奏のタイミングや楽曲のハーモニーを効率的に習得することができる。 (2) In each of the above embodiments, the musical score data SD is used to read out the sounds associated with the symbols included in the musical score. The method of using the score data SD is not limited to this, and performance sounds may be reproduced using the score data SD. Specifically, for example, the information processing apparatus 10 may reproduce the performance sound of the left-hand staff notation of the musical score represented by the score data SD, and read out the music symbols in the right-hand staff. The user practices playing with the right hand while listening to the reading sound of the musical symbols. By reproducing the performance sound of the staff notation for the left hand, the user can efficiently master the timing of performance with the right hand and the harmony of the music.
(3)例えばスマートフォンまたはタブレット端末等の情報装置と通信するサーバ装置により情報処理装置10が実現されてもよい。例えば、情報処理装置10は、情報装置から楽譜データSDの指定を受け付け、指定された楽譜データSDを用いた音声合成処理により音響信号を生成する。情報処理装置10は、音声合成処理により生成した音響信号を情報装置に送信する。情報装置では音響信号が再生される。 (3) The information processing device 10 may be implemented by a server device that communicates with an information device such as a smart phone or a tablet terminal. For example, the information processing device 10 receives designation of musical score data SD from the information device, and generates an acoustic signal by speech synthesis processing using the designated musical score data SD. The information processing device 10 transmits an acoustic signal generated by speech synthesis processing to the information device. The information device reproduces the acoustic signal.
(4)情報処理装置10の機能(指示受付部30、テキスト生成部32、音声合成部34、演奏解析部38、出力制御部40、演奏評価部42)は、前述の通り、制御装置11を構成する単数または複数のプロセッサと、記憶装置12に記憶されたプログラムPGとの協働により実現される。 (4) The functions of the information processing device 10 (the instruction receiving unit 30, the text generating unit 32, the speech synthesizing unit 34, the performance analyzing unit 38, the output control unit 40, and the performance evaluating unit 42) include the control device 11 as described above. It is realized by cooperation of one or more constituent processors and the program PG stored in the storage device 12 .
(5)前述の各形態においては、楽譜読み上げアプリケーションにおける各種設定や指示に際して、タッチパネルTに表示された項目を利用者が目視して、タッチパネルTに対するタッチ操作を行う態様について説明した。これに限らず、例えば利用者に対する情報(設定における選択項目など)の提示が音声での読み上げによって行われてもよい。また、情報処理装置10に対する利用者からの入力が、音声入力によって行われてもよい。特に、楽譜読み上げアプリケーションを視覚障碍者が利用する場合には、音声を用いた態様が有効である。 (5) In each of the above-described embodiments, the manner in which the user looks at the items displayed on the touch panel T and touches the touch panel T when performing various settings and instructions in the musical score reading application has been described. For example, the presentation of information (such as selection items in settings) to the user may be performed by reading aloud. Further, the input from the user to the information processing device 10 may be performed by voice input. In particular, when a visually impaired person uses a musical score reading application, the use of voice is effective.
 以上のプログラムは、コンピュータが読取可能な記録媒体に格納された形態で提供されてコンピュータにインストールされ得る。記録媒体は、例えば非一過性(non-transitory)の記録媒体であり、CD-ROM等の光学式記録媒体(光ディスク)が好例であるが、半導体記録媒体または磁気記録媒体等の公知の任意の形式の記録媒体も包含される。なお、非一過性の記録媒体とは、一過性の伝搬信号(transitory,
 propagating signal)を除く任意の記録媒体を含み、揮発性の記録媒体も除外されない。また、配信装置が通信網を介してプログラムを配信する構成では、当該配信装置においてプログラムを記憶する記録媒体が、前述の非一過性の記録媒体に相当する。
The above program can be provided in a form stored in a computer-readable recording medium and installed in the computer. The recording medium is, for example, a non-transitory recording medium, and an optical recording medium (optical disc) such as a CD-ROM is a good example. Also included are recording media in the form of Note that the non-transient recording medium means a transient propagation signal (transitory,
volatile recording media are not excluded. Also, in a configuration in which a distribution device distributes a program via a communication network, a recording medium for storing the program in the distribution device corresponds to the non-transitory recording medium described above.
E:付記
 以上に例示した形態から、例えば以下の構成が把握される。
E: Supplementary Note The following configurations, for example, can be grasped from the above-exemplified forms.
 本開示のひとつの態様(態様1)に係る情報処理装置は、コンピュータシステムにより実現され、1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する。したがって、楽譜に含まれる演奏記号を聴覚により把握することができ、例えば視覚障碍者、読譜に慣れていない初学者や小さい子供においても、楽譜の把握を容易とすることができる。 An information processing apparatus according to one aspect (aspect 1) of the present disclosure is realized by a computer system, and generates an acoustic signal representing a sound related to the performance symbols based on musical score data representing a musical score including one or more performance symbols. do. Therefore, the performance symbols included in the musical score can be grasped by hearing, and even visually handicapped people, beginners who are not accustomed to reading musical scores, and small children can easily grasp the musical score.
 態様1の具体例(態様2)において、前記演奏記号に関する音は、前記演奏記号の名称を示す音、または、前記演奏記号の意味に対応する語句を示す音である。以上の態様においては、演奏記号に関する音が、演奏記号の名称を示す音である場合には、楽譜上の記述を正確に把握することができる。また、演奏記号に関する音が、演奏記号の意味に対応する語句を示す音である場合には、ユーザが演奏記号の知識に乏しく、演奏記号の名称のみではその意味が理解できない場合においても、楽譜が示す内容を把握することができる。 In the specific example of aspect 1 (aspect 2), the sound related to the performance symbol is a sound indicating the name of the performance symbol or a sound indicating a phrase corresponding to the meaning of the performance symbol. In the above aspect, when the sound associated with the performance symbol is the sound indicating the name of the performance symbol, it is possible to accurately grasp the description on the musical score. In addition, when the sound associated with the performance symbol is a sound indicating a word or phrase corresponding to the meaning of the performance symbol, even if the user lacks knowledge of the performance symbol and cannot understand the meaning of the performance symbol only by the name of the performance symbol, the musical score can be used. It is possible to grasp the contents indicated by .
 態様1または態様2の具体例(態様3)において、前記楽譜データは、前記楽譜が示す楽曲のテンポを指定するテンポ情報を含み、前記音響信号の生成においては、前記テンポ情報が指定する前記テンポに対応する速度で前記楽曲を進行する目標時点が、前記演奏記号に対応する時点に到達したときに、当該演奏記号に関する音が発音されるように、前記音響信号を生成する。以上の態様においては、楽曲のテンポに対応するタイミングで演奏記号に対応する音が発音される。したがって、ユーザは、楽曲内における演奏記号の位置を把握しやすくなり、利便性を向上させることができる。 In a specific example of Aspect 1 or Aspect 2 (Aspect 3), the musical score data includes tempo information specifying a tempo of a piece of music indicated by the musical score, and in generating the acoustic signal, the tempo specified by the tempo information. The acoustic signal is generated such that when the target point in time at which the music progresses at the speed corresponding to reaches the point in time corresponding to the performance symbol, the sound associated with the performance symbol is pronounced. In the above mode, sounds corresponding to performance symbols are produced at timings corresponding to the tempo of music. Therefore, it becomes easier for the user to grasp the positions of the performance symbols in the music, and convenience can be improved.
 態様1から態様3のいずれか1つの具体例(態様4)において、前記1以上の演奏記号は、第1の演奏記号と第2の演奏記号とを含み、前記音響信号の生成においては、前記第1の演奏記号に関する音と前記第2の演奏記号に関する音とが重ねて発音される場合に、第1の演奏記号または第2の演奏記号のいずれかを選択して、選択した演奏記号に関する音を表す前記音響情報を生成する。以上の態様においては、第1の演奏記号に関する音と第2の演奏記号に関する音が重ねて発音される場合に、前記第1の演奏記号または前記第2の演奏記号のいずれかを選択して音響情報を生成する。したがって、第1の演奏記号に関する音と第2の演奏記号に関する音が重ねて発音されることがなく、演奏記号に関する音の聞き取り易さを向上させることができる。 In a specific example of any one of Aspects 1 to 3 (Aspect 4), the one or more performance symbols include a first performance symbol and a second performance symbol, and in generating the acoustic signal, When the sound related to the first performance symbol and the sound related to the second performance symbol are superimposed, selecting either the first performance symbol or the second performance symbol, generating said acoustic information representing a sound; In the above aspect, when a sound related to the first performance symbol and a sound related to the second performance symbol are superimposed, either the first performance symbol or the second performance symbol is selected. Generates acoustic information. Therefore, the sound related to the first performance symbol and the sound related to the second performance symbol do not overlap, and the audibility of the sound related to the performance symbol can be improved.
 態様1の具体例(態様5)において、前記楽譜データは、前記楽譜が示す楽曲のテンポを指定するテンポ情報を含み、前記音響信号の生成においては、前記楽曲のテンポに関わらず前記演奏記号に関する音を発音した音を示す音響信号を生成する。以上の態様においては、楽曲のテンポに関わらず演奏記号に対応する音が発音される。したがって、演奏記号に関する音が重ねて発音されることがなく、演奏記号に関する音の聞き取り易さを向上させることができる。 In the specific example of Aspect 1 (Aspect 5), the musical score data includes tempo information specifying the tempo of the music indicated by the musical score, and in generating the acoustic signal, the musical performance symbols are used regardless of the tempo of the music. Generating an acoustic signal indicative of the sound produced. In the above mode, sounds corresponding to performance symbols are produced regardless of the tempo of music. Therefore, the sounds related to the performance symbols are not over-pronounced, and the audibility of the sounds related to the performance symbols can be improved.
 態様1から態様5のいずれか1つの具体例(態様6)において、前記1以上の演奏記号は、複数の演奏記号であり、前記複数の演奏記号の各々は、複数の分類のいずれかに属し、前記複数の分類のうち少なくとも1つの分類の選択を受け付け、前記音響信号の生成においては、前記複数の演奏記号のうち前記選択に係る前記1以上の分類に属する演奏記号について、前記音響信号を生成する。以上の態様においては、選択された分類に属する演奏記号について音響信号を生成する。したがって、ユーザが必要とする演奏記号に関する音を選択的に発音させることができ、利便性を向上させることができる。 In a specific example of any one of Aspects 1 to 5 (Aspect 6), the one or more performance symbols are a plurality of performance symbols, and each of the plurality of performance symbols belongs to one of a plurality of classifications. , receiving a selection of at least one of the plurality of categories, and generating the acoustic signal for a musical performance symbol belonging to the one or more categories related to the selection among the plurality of musical performance symbols. Generate. In the above aspect, acoustic signals are generated for performance symbols belonging to the selected classification. Therefore, it is possible to selectively produce sounds related to the performance symbols required by the user, thereby improving convenience.
 態様1から態様6のいずれか1つの具体例(態様7)において、前記楽譜は、前記演奏記号に加え、音符記号を含み、前記音響信号を生成することは、前記演奏記号に関する音と、前記音符記号に関する音とを表す前記音響信号を生成することを含む。以上の態様においては、演奏記号に関する音に加え、音符記号に関する音を表す音響信号を生成する。したがって、楽譜に含まれる音符記号を聴覚により把握することができ、楽譜の把握を更に容易とすることができる。 In a specific example of any one of Aspects 1 to 6 (Aspect 7), the musical score includes musical note symbols in addition to the performance symbols, and generating the acoustic signal includes sounds associated with the performance symbols and generating said acoustic signal representing a sound associated with a musical note symbol. In the above aspect, in addition to the sounds related to performance symbols, acoustic signals representing sounds related to musical note symbols are generated. Therefore, the musical note symbols included in the musical score can be comprehended aurally, and the comprehension of the musical score can be further facilitated.
 態様7の具体例(態様8)において、前記音響信号の生成においては、休符に関する音として、非言語の通知音を表す音響信号を生成する。以上の態様においては、休符に関する音として、非言語の通知音を用いる。したがって、ユーザは、休符に関する音が発音された際に休符に対応する音であることを即座に把握することができる。 In the specific example of aspect 7 (aspect 8), in generating the acoustic signal, an acoustic signal representing a non-verbal notification sound is generated as the sound related to the rest. In the above embodiments, non-verbal notification sounds are used as the sounds related to rests. Therefore, when a sound related to a rest is pronounced, the user can immediately recognize that the sound corresponds to the rest.
 態様7または態様8の具体例(態様9)において、前記楽譜データは、第1楽譜データであり、前記音響信号の生成は、前記第1楽譜データに対応する第1楽譜の一部分に含まれる前記演奏記号に関する音と、前記音符記号に関する音とを示す第1音響信号の生成と、前記第1楽譜データと異なる第2楽譜データに対応する第2楽譜の一部分に含まれる前記演奏記号に関する音と、前記音符記号に関する音とを示す第2音響信号の生成とを含み、前記第1音響信号と、前記第2音響信号とを放音装置に順次に再生させることと、を更に含む。以上の態様においては、複数の楽譜データから各々一部分を選択し、選択された一部分に含まれる演奏記号および音符記号に関する音を順次再生する。したがって、ユーザは、複数の楽譜データが、それぞれどの楽曲の楽譜に対応するものかを容易に把握することができ、複数の楽譜データの中から所望の楽譜データを迅速に選択することができる。 In a specific example of Aspect 7 or Aspect 8 (Aspect 9), the musical score data is first musical score data, and the generation of the acoustic signal is included in a portion of the first musical score corresponding to the first musical score data. generating a first acoustic signal indicating a sound related to a performance symbol and a sound related to the musical note symbol; , generating a second acoustic signal indicative of the sound associated with the musical note symbol, and further comprising causing a sound emitting device to sequentially reproduce the first acoustic signal and the second acoustic signal. In the above aspect, a portion is selected from each of a plurality of musical score data, and sounds associated with performance symbols and musical note symbols included in the selected portion are sequentially reproduced. Therefore, the user can easily grasp the musical score of which musical piece each of the plurality of musical score data corresponds to, and can quickly select the desired musical score data from the plurality of musical score data.
 本開示のひとつの態様(態様10)に係るプログラムは、1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する生成部としてコンピュータシステムを機能させる。 A program according to one aspect (a tenth aspect) of the present disclosure functions as a generation unit that generates an acoustic signal representing a sound related to one or more performance symbols based on musical score data representing a musical score including one or more performance symbols. Let
 本開示のひとつの態様(態様11)に係る情報処理装置は、1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する生成部を備える。 An information processing apparatus according to one aspect (aspect 11) of the present disclosure includes a generation unit that generates an acoustic signal representing a sound associated with one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
 ここで、点字楽譜は通常楽譜と比べて、同じ内容の譜面を表記するのに約3倍の紙面が必要であり、読み取りに時間がかかる。このため、例えば利用者が楽曲の曲名を忘れてしまい、楽譜の内容から所望の楽譜を探したい場合に、複数の楽譜を時間をかけて読み取らなくてはならず、不便であるという課題がある。 Here, Braille sheet music requires about three times as much paper space to write the same content as regular sheet music, and it takes time to read. For this reason, for example, when a user forgets the title of a piece of music and wants to find the desired score from the contents of the score, the user has to spend time reading a plurality of scores, which is inconvenient. .
 本開示のひとつの態様に係る情報処理装置は、コンピュータシステムにより実現され、1以上の音楽記号を含む第1楽譜データに対応する第1楽譜の一部分に含まれる前記音楽記号に関する音を示す第1音響信号を生成し、1以上の前記音楽記号を含み前記第1楽譜データと異なる第2楽譜データに対応する第2楽譜の一部分に含まれる前記音楽記号に関する音を示す第2音響信号を生成し、前記第1音響信号と、前記第2音響信号とを放音装置に順次に再生させる。 An information processing apparatus according to one aspect of the present disclosure is implemented by a computer system, and includes a first musical symbol representing a sound related to the musical symbol included in a portion of a first musical score corresponding to first musical score data including one or more musical symbols. generating an acoustic signal, and generating a second acoustic signal indicating a sound associated with the musical symbol included in a portion of a second musical score that includes one or more of the musical symbols and corresponds to second musical score data that is different from the first musical score data; and causing a sound emitting device to sequentially reproduce the first acoustic signal and the second acoustic signal.
 また、楽器の教習は、楽譜に記載された音楽記号を正しく読み取り、当該音楽記号が示す音を正確に演奏することを目的とする。一方で、楽譜に記載された音楽記号を正しく読み取れているか、および音楽記号が示す音を正確に演奏できているかは、演奏者自身で識別することができない場合がある。このため、一般に演奏者は指導者に指導を依頼することにより教習を受ける。しかしながら、指導者を常時側に置いて演奏することは、現実的ではなく、演奏に対するフィードバックを受けられる機会は限られている。 In addition, the purpose of musical instrument training is to correctly read the musical symbols written on the score and to play the sounds indicated by the musical symbols. On the other hand, there are cases in which the player himself/herself cannot identify whether the musical symbols written on the musical score are correctly read and whether the sounds indicated by the musical symbols are correctly played. For this reason, performers generally receive training by requesting guidance from instructors. However, it is not realistic to have an instructor by your side all the time, and opportunities to receive feedback on your performance are limited.
 本開示のひとつの態様に係る情報処理装置は、コンピュータシステムにより実現され、利用者により楽曲を演奏した音である演奏音を取得し、前記楽曲を示す楽譜に含まれる音楽記号が示す音と、前記演奏音との差分を検出し、前記差分が予め定められた許容範囲から外れた場合、前記差分が生じた箇所に対応する前記楽譜の一部分に含まれる音楽記号に関する音を示す音響信号を生成する。 An information processing apparatus according to one aspect of the present disclosure is realized by a computer system, acquires a performance sound that is a sound of a musical piece played by a user, and obtains a sound indicated by a musical symbol included in a musical score indicating the musical piece, Detecting a difference from the performance sound, and generating an acoustic signal indicating a sound related to a musical symbol included in a portion of the musical score corresponding to the portion where the difference occurs when the difference is out of a predetermined allowable range. do.
10…情報処理装置、11,11A…制御装置、12…記憶装置、13…収音装置、14…放音装置、15…操作装置、16…表示装置、30…指示受付部、32…テキスト生成部、34…音声合成部、38…演奏解析部、40…出力制御部、42…演奏評価部、PG…プログラム、SD…楽譜データ、T…タッチパネル、TD…記号テキストデータ、VD…音声データ。 DESCRIPTION OF SYMBOLS 10... Information processing apparatus 11, 11A... Control apparatus 12... Storage apparatus 13... Sound collection apparatus 14... Sound emission apparatus 15... Operation apparatus 16... Display apparatus 30... Instruction reception part 32... Text generation Section 34...Speech synthesis section 38...Performance analysis section 40...Output control section 42...Performance evaluation section PG...Program SD...Score data T...Touch panel TD...Symbol text data VD...Voice data.

Claims (11)

  1.  1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する
     コンピュータシステムにより実現される情報処理方法。
    An information processing method implemented by a computer system, comprising generating an acoustic signal representing a sound associated with one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
  2.  前記演奏記号に関する音は、前記演奏記号の名称を示す音、または、前記演奏記号の意味に対応する語句を示す音である
     請求項1記載の情報処理方法。
    2. The information processing method according to claim 1, wherein the sound associated with the musical performance symbol is a sound indicating the name of the musical performance symbol or a sound indicating a phrase corresponding to the meaning of the musical performance symbol.
  3.  前記楽譜データは、前記楽譜が示す楽曲のテンポを指定するテンポ情報を含み、
     前記音響信号の生成においては、前記テンポ情報が指定する前記テンポに対応する速度で前記楽曲を進行する目標時点が、前記演奏記号に対応する時点に到達したときに、当該演奏記号に関する音が発音されるように、前記音響信号を生成する
     請求項1または2記載の情報処理方法。
    the musical score data includes tempo information specifying the tempo of the music indicated by the musical score;
    In the generation of the acoustic signal, when a target point in time at which the music piece progresses at a speed corresponding to the tempo specified by the tempo information reaches a point in time corresponding to the performance symbol, a sound associated with the performance symbol is produced. 3. The information processing method according to claim 1 or 2, wherein the acoustic signal is generated so as to be
  4.  前記1以上の演奏記号は、第1の演奏記号と第2の演奏記号とを含み、
     前記音響信号の生成においては、前記第1の演奏記号に関する音と前記第2の演奏記号に関する音とが重ねて発音される場合に、前記第1の演奏記号または前記第2の演奏記号のいずれかを選択して、選択した演奏記号に関する音を表す前記音響信号を生成する
     請求項1から3のいずれか1項記載の情報処理方法。
    the one or more performance symbols include a first performance symbol and a second performance symbol;
    In the generation of the acoustic signal, when a sound related to the first performance symbol and a sound related to the second performance symbol are superimposed, either the first performance symbol or the second performance symbol is generated. 4. The information processing method according to any one of claims 1 to 3, wherein the acoustic signal representing the sound associated with the selected performance symbol is generated by selecting one of:
  5.  前記楽譜データは、前記楽譜が示す楽曲のテンポを指定するテンポ情報を含み、
     前記音響信号の生成においては、前記楽曲のテンポに関わらず前記演奏記号に関する音を発音した音を示す音響信号を生成する
     請求項1記載の情報処理方法。
    the musical score data includes tempo information specifying the tempo of the music indicated by the musical score;
    2. The information processing method according to claim 1, wherein in the generation of the acoustic signal, the acoustic signal representing the sound of the sound associated with the performance symbol is generated regardless of the tempo of the music piece.
  6.  前記1以上の演奏記号は、複数の演奏記号であり、
     前記複数の演奏記号の各々は、複数の分類のいずれかに属し、
     前記複数の分類のうち少なくとも1つの分類の選択を受け付け、
     前記音響信号の生成においては、前記複数の演奏記号のうち前記選択に係る前記1以上の分類に属する演奏記号について、前記音響信号を生成する
     請求項1から5のいずれか1項記載の情報処理方法。
    The one or more performance symbols are a plurality of performance symbols,
    each of the plurality of performance symbols belongs to one of a plurality of classifications,
    Receiving selection of at least one classification from the plurality of classifications;
    6. The information processing according to any one of claims 1 to 5, wherein in generating the acoustic signal, the acoustic signal is generated for a musical performance symbol belonging to the one or more classifications related to the selection among the plurality of musical performance symbols. Method.
  7.  前記楽譜は、前記演奏記号に加え、音符記号を含み、
     前記音響信号を生成することは、前記演奏記号に関する音と、前記音符記号に関する音とを表す前記音響信号を生成することを含む
     請求項1から6のいずれか1項記載の情報処理方法。
    The musical score includes musical note symbols in addition to the performance symbols,
    7. The information processing method according to any one of claims 1 to 6, wherein generating the acoustic signal includes generating the acoustic signal representing a sound related to the performance symbol and a sound related to the musical note symbol.
  8.  前記音響信号の生成においては、休符に関する音として、非言語の通知音を表す音響信号を生成する
     請求項7記載の情報処理方法。
    8. The information processing method according to claim 7, wherein in the generation of the acoustic signal, an acoustic signal representing a non-verbal notification sound is generated as the sound related to the rest.
  9.  前記楽譜データは、第1楽譜データであり、
     前記音響信号の生成は、
     前記第1楽譜データに対応する第1楽譜の一部分に含まれる前記演奏記号に関する音と、前記音符記号に関する音とを示す第1音響信号の生成と、
     前記第1楽譜データと異なる第2楽譜データに対応する第2楽譜の一部分に含まれる前記演奏記号に関する音と、前記音符記号に関する音とを示す第2音響信号の生成とを含み、
     前記第1音響信号と、前記第2音響信号とを放音装置に順次に再生させることと、を更に含む
     請求項7または8記載の情報処理方法。
    The musical score data is first musical score data,
    Generating the acoustic signal includes:
    generating a first acoustic signal indicating a sound related to the performance symbol and a sound related to the musical note symbol included in a portion of the first musical score corresponding to the first musical score data;
    generating a second acoustic signal indicating a sound related to the performance symbol and a sound related to the musical note symbol included in a portion of a second musical score corresponding to second musical score data different from the first musical score data;
    9. The information processing method according to claim 7, further comprising causing a sound emitting device to sequentially reproduce the first acoustic signal and the second acoustic signal.
  10.  1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する生成部
     としてコンピュータシステムを機能させるプログラム。
    A program that causes a computer system to function as a generation unit that generates an acoustic signal representing a sound associated with one or more performance symbols, based on musical score data representing a musical score containing one or more performance symbols.
  11.  1以上の演奏記号を含む楽譜を表す楽譜データに基づいて、前記演奏記号に関する音を表す音響信号を生成する生成部
     を備える情報処理装置。
    An information processing apparatus comprising: a generating unit configured to generate an acoustic signal representing a sound associated with one or more performance symbols based on musical score data representing a musical score including one or more performance symbols.
PCT/JP2022/040701 2022-02-10 2022-10-31 Information processing method, program, and information processing device WO2023153033A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-019211 2022-02-10
JP2022019211A JP2023116866A (en) 2022-02-10 2022-02-10 Information processing method, program, and information processing device

Publications (1)

Publication Number Publication Date
WO2023153033A1 true WO2023153033A1 (en) 2023-08-17

Family

ID=87564095

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2022/040701 WO2023153033A1 (en) 2022-02-10 2022-10-31 Information processing method, program, and information processing device

Country Status (2)

Country Link
JP (1) JP2023116866A (en)
WO (1) WO2023153033A1 (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000214848A (en) * 1999-01-21 2000-08-04 Yamaha Corp Performance support device, performance support method, and recording medium with performance support program recorded therein

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000214848A (en) * 1999-01-21 2000-08-04 Yamaha Corp Performance support device, performance support method, and recording medium with performance support program recorded therein

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SHINYA YAMADA, TOSHIYUKI GOTOH, NAOYOSHI TAMURA: "Automated Generation System of DAISY Content with Braill and Oral from Digital Music Scores", IEICE TECHNICAL REPORT, WIT, IEICE, JP, vol. 109, no. 358 (WIT2009-71), 1 January 2010 (2010-01-01), JP, pages 19 - 24, XP009548320 *

Also Published As

Publication number Publication date
JP2023116866A (en) 2023-08-23

Similar Documents

Publication Publication Date Title
JP5821824B2 (en) Speech synthesizer
US7754955B2 (en) Virtual reality composer platform system
KR20180063163A (en) Automated music composition and creation machines, systems and processes employing musical experience descriptors based on language and / or graphic icons
JP5949607B2 (en) Speech synthesizer
JP2008015195A (en) Musical piece practice support device
JP6728754B2 (en) Pronunciation device, pronunciation method and pronunciation program
WO2015194423A1 (en) Controller and system for voice generation based on characters
JP2008026622A (en) Evaluation apparatus
US20220076658A1 (en) Electronic musical instrument, method, and storage medium
US20220076651A1 (en) Electronic musical instrument, method, and storage medium
KR20170120234A (en) System for providing music synchronized with syllable of english words
JP2023015302A (en) Electronic apparatus, electronic musical instrument, method and program
JP2006259471A (en) Singing practice system and program for singing practice system
JP4929604B2 (en) Song data input program
JP6589356B2 (en) Display control device, electronic musical instrument, and program
JP4038836B2 (en) Karaoke equipment
JP2009169103A (en) Practice support device
WO2023153033A1 (en) Information processing method, program, and information processing device
JP6044284B2 (en) Speech synthesizer
US20220044662A1 (en) Audio Information Playback Method, Audio Information Playback Device, Audio Information Generation Method and Audio Information Generation Device
JP2022065554A (en) Method for synthesizing voice and program
JP4501874B2 (en) Music practice device
WO2022190502A1 (en) Sound generation device, control method therefor, program, and electronic musical instrument
JP2002182675A (en) Speech synthesizer, vocal data former and singing apparatus
JP5953743B2 (en) Speech synthesis apparatus and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22926049

Country of ref document: EP

Kind code of ref document: A1