WO2022202199A1 - Code estimation device, training device, code estimation method, and training method - Google Patents
Code estimation device, training device, code estimation method, and training method Download PDFInfo
- Publication number
- WO2022202199A1 WO2022202199A1 PCT/JP2022/009233 JP2022009233W WO2022202199A1 WO 2022202199 A1 WO2022202199 A1 WO 2022202199A1 JP 2022009233 W JP2022009233 W JP 2022009233W WO 2022202199 A1 WO2022202199 A1 WO 2022202199A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- series data
- string
- information
- time
- chord
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 25
- 238000010801 machine learning Methods 0.000 description 15
- 238000010276 construction Methods 0.000 description 13
- 238000003860 storage Methods 0.000 description 13
- 238000010586 diagram Methods 0.000 description 9
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000003825 pressing Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000003864 performance function Effects 0.000 description 1
- 239000011435 rock Substances 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10G—REPRESENTATION OF MUSIC; RECORDING MUSIC IN NOTATION FORM; ACCESSORIES FOR MUSIC OR MUSICAL INSTRUMENTS NOT OTHERWISE PROVIDED FOR, e.g. SUPPORTS
- G10G3/00—Recording music in notation form, e.g. recording the mechanical operation of a musical instrument
- G10G3/04—Recording music in notation form, e.g. recording the mechanical operation of a musical instrument using electrical means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/36—Accompaniment arrangements
- G10H1/38—Chord
Definitions
- the present invention relates to a chord estimation device and method for estimating chords for playing a musical instrument, and a training device and method for constructing a chord estimation device.
- a code is estimated for each specific section. For example, one chord is estimated per bar. If it is possible to perform chord estimation with a higher degree of freedom from given notes, it is expected that it will be possible to more appropriately support the production of musical scores with chords.
- the purpose of the present invention is to perform chord estimation with a high degree of freedom based on musical note strings.
- a chord estimation apparatus uses a receiving unit that receives time-series data including a string of notes composed of a plurality of notes, and a trained model to generate a code string corresponding to the string of notes based on the time-series data. and an estimating unit for estimating the code string information to be indicated.
- a training apparatus includes a first acquisition unit that acquires input time-series data including a reference note string composed of a plurality of notes, and output code string information that indicates a code string corresponding to the reference note string.
- a chord estimation method is executed by a computer, accepts time series data including a string of notes, uses a trained model, and corresponds to the string of notes based on the time series data. Estimate code string information that indicates the code string to be used.
- a training method is executed by a computer, acquires input time-series data including a reference note string consisting of a plurality of notes, and outputs code string information indicating a code string corresponding to the reference note string. acquire and build a trained model that has learned the input/output relationship between the input time-series data and the output code string information.
- chord estimation with a high degree of freedom can be performed based on a string of musical notes.
- FIG. 1 is a block diagram showing the configuration of a processing system including a chord estimation device and a training device according to one embodiment of the present invention.
- FIG. 2 is a diagram showing an example of input time-series data included in training data.
- FIG. 3 is a diagram showing an example of output code string information included in training data.
- FIG. 4 is a block diagram showing the configuration of the training device and chord estimation device.
- FIG. 5 shows an example of an arranged musical score displayed on the display unit.
- FIG. 6 is a flowchart showing an example of training processing.
- FIG. 7 is a flowchart showing an example of chord estimation processing.
- FIG. 8 is a diagram showing a modified example of output code string information included in training data.
- FIG. 1 is a block diagram showing the configuration of a processing system including a chord estimation device and a training device according to one embodiment of the present invention.
- the processing system 100 includes a RAM (Random Access Memory) 110, a ROM (Read Only Memory) 120, a CPU (Central Processing Unit) 130, a storage section 140, an operation section 150 and a display section 160. .
- RAM Random Access Memory
- ROM Read Only Memory
- CPU Central Processing Unit
- the processing system 100 is implemented by a computer such as a personal computer, tablet terminal, or smart phone.
- the processing system 100 may be realized by cooperative operation of a plurality of computers connected by a communication path such as Ethernet, or may be realized by an electronic musical instrument such as an electronic piano having performance functions.
- RAM 110 , ROM 120 , CPU 130 , storage section 140 , operation section 150 and display section 160 are connected to the bus 170 .
- RAM 110 , ROM 120 and CPU 130 constitute training device 10 and chord estimation device 20 .
- training device 10 and chord estimation device 20 are configured by common processing system 100 in this embodiment, they may be configured by separate processing systems.
- the RAM 110 consists of, for example, a volatile memory, and is used as a work area for the CPU 130.
- the ROM 120 is, for example, a non-volatile memory and stores a training program and a code estimation program.
- CPU 130 performs a training process by executing a training program stored in ROM 120 on RAM 110 . Further, the CPU 130 performs code estimation processing by executing a code estimation program stored in the ROM 120 on the RAM 110 . Details of the training process and the code estimation process will be described later.
- the training program or code estimation program may be stored in the storage unit 140 instead of the ROM 120.
- the training program or code estimation program may be provided in a form stored in a computer-readable storage medium and installed in ROM 120 or storage unit 140 .
- a training program or code estimation program distributed from a server (including a cloud server) on the network is installed in the ROM 120 or the storage unit 140.
- the storage unit 140 includes a storage medium such as a hard disk, an optical disk, a magnetic disk, or a memory card, and stores a trained model M and a plurality of training data D.
- the trained model M or each piece of training data D may not be stored in the storage unit 140, but may be stored in a computer-readable storage medium.
- the trained model M or respective training data D may be stored on a server on that network.
- the trained model M is a machine learning model that has been trained to present chord strings to be referred to when the user of the chord estimation device 20 (hereinafter referred to as a performer) plays a piece of music. be.
- a trained model M is constructed using a plurality of training data D.
- a user of the training device 10 can generate the training data D by operating the operation unit 150 .
- the training data D is data created based on the musical knowledge or musical sense of the reference performer.
- the reference performer has a relatively high level of skill in playing the piece of music.
- a reference performer may be the performer's mentor or teacher in the performance of the musical composition.
- the training data D indicates a set of input time-series data and output code string information.
- the input time-series data indicates a reference note string consisting of a plurality of notes.
- the input time-series data is data that forms a melody or accompaniment sound with a plurality of notes.
- the input time-series data may be image data representing images of musical scores.
- the output code string information is data in which codes corresponding to the reference note string are arranged in time series. A code string corresponding to the reference note string is provided by the reference performer.
- Figs. 2 and 3 are diagrams showing an example of each training data D.
- Figs. The example in FIG. 2 shows input time-series data including a reference note string consisting of a plurality of notes.
- the example in FIG. 3 shows output code string information indicating a code string corresponding to the reference note string.
- the input time-series data has a metrical structure and additional information in addition to the reference note string.
- the input time-series data A shown in FIG. 2 is data obtained by extracting data for the first two bars of a song. In the input time-series data A, bars are separated by "bar", and beats are separated by "beat". In this way, the input time-series data A has a metrical structure with the "bar” and "beat” information.
- Elements A1 to A37 indicate the reference note string of the first bar. That is, the elements A1 to A37 are separated into bars by the "bar” before the element A1 and the "bar” after the element A37. In addition, it is divided into beats by "beat” after elements A8, A18, and A26.
- the element A0 is additional information.
- additional information for example, key information, genre information, difficulty level information, and the like are used.
- key information is added by the Key element.
- the key information is information specifying the key of the music represented by the reference note string.
- the numerical value following Key is the numerical value that designates the key.
- Genre information is information that designates the genre of music represented by the reference note string.
- genre information for example, genres such as rock, pops, and jazz are specified. By designating genre information as additional information, a reference note string and a code string corresponding to the genre are machine-learned.
- the difficulty level information is information indicating the difficulty level of the musical score indicated by the reference note string.
- difficulty level information is information indicating the difficulty level of the musical score indicated by the reference note string.
- a code string corresponding to the difficulty level of the reference note string and score is machine-learned. For example, in the case of a score with a low difficulty level, machine learning is performed while interpolating notes from a small number of tones. In the case of a score with a high degree of difficulty, machine learning is performed while selecting notes that form chords from an excessive number of tones.
- elements other than the element A0, "bar” and “beat” correspond to the reference note string.
- Elements A1 to A37 indicate the reference note string of the first measure.
- the element A0 is placed at the beginning of the input time-series data A, that is, before the reference note string (elements A1 to A37), but it may be placed at any position in the input time-series data A.
- elements A1 to A37 in the reference note string, “L” means left hand, “R” means right hand, and the number following “L” or “R” means scale. Also, “on” and “off” mean key depression and key release, respectively. Also, “wait” means waiting, and the number following "wait” means the length of time.
- elements A1-A5 indicate pressing the keys of scale 77 and 74 with the right hand simultaneously while simultaneously pressing the keys of scale 53 and 46 with the left hand, followed by holding for 11 units of time.
- elements A6 to A8 indicate that the left hand keys of scale 53 and scale 46 are released at the same time and then held for 1 unit of time. Then, after maintaining for one unit time, elements A9 to A11 indicate that the left hand presses scale 53 and scale 46 again, and then waits for five unit time.
- the output code string information B shown in FIG. 3 indicates a code string corresponding to the reference note string included in the input time-series data A.
- Code strings corresponding to elements A1 to A37 of input time-series data A are represented by elements B1 to B3 and elements B4 to B6. That is, the elements B1 to B6 indicate the code string corresponding to the first bar of the input time-series data A.
- FIG. In the output code string information B bars are also separated by “bar” and beats by "beat”. A range delimited by "bar” before the element B1 and "bar” after the element B6 corresponds to the first bar.
- one code is indicated by three elements.
- Elements B1 to B3 define the chord of the first beat of the first bar.
- Elements B4 to B6 define the chord of the fourth beat of the first bar.
- Elements B7 to B9 define the chord of the first beat of the fourth bar.
- the first element (B1, B4, B7) represents basic code information.
- the basic chord information (chord) indicates a numerical value from 1 to 24 that designates the type of major chord and minor chord for each of the 12 tones (C, C#, D, D#, . . . A, A#, B).
- the second element (B2, B5, B8) of the three elements indicating the chord indicates chord type information.
- the chord type information indicates a numerical value designating the type of tension chord.
- the third element (B3, B6, B9) represents chord root information.
- the chord root information (root) indicates a numerical value designating the root note of the on-chord.
- FIG. 4 is a block diagram showing the configuration of the training device 10 and the chord estimation device 20.
- the training device 10 includes a first acquisition unit 11, a second acquisition unit 12, and a construction unit 13 as functional units.
- the functional units of the training device 10 are implemented by the CPU 130 of FIG. 1 executing the training program. At least part of the functional units of the training device 10 may be realized by hardware such as an electronic circuit.
- the first acquisition unit 11 acquires the input time-series data A from each training data D stored in the storage unit 140 or the like.
- the second acquisition unit 12 acquires output code string information B from each training data D.
- the construction unit 13 uses the input time-series data A acquired by the first acquisition unit 11 as an input element, and the output code string information B acquired by the second acquisition unit 12 as an output element. perform machine learning to By repeating machine learning for a plurality of training data D, the construction unit 13 constructs a trained model M indicating the input/output relationship between the input time-series data A and the output code string information B.
- the building unit 13 builds the trained model M by training the Transformer, but the embodiment is not limited to this.
- the construction unit 13 may construct the trained model M by training a machine learning model of another method that handles time series.
- the trained model M constructed by the construction unit 13 is stored in the storage unit 140, for example.
- the trained model M constructed by the construction unit 13 may be stored in a server or the like on the network.
- the code estimation device 20 includes a reception unit 21, an estimation unit 22, and a generation unit 23 as functional units.
- the functional units of the code estimation device 20 are implemented by the CPU 130 of FIG. 1 executing the code estimation program. At least part of the functional units of the code estimation device 20 may be realized by hardware such as an electronic circuit.
- the reception unit 21 receives time-series data including a string of notes made up of a plurality of notes.
- the performer can give image data representing an image of the musical score to the reception unit 21 as time-series data.
- the performer can generate time-series data by operating the operation unit 150 and provide it to the reception unit 21 .
- the time-series data has the same configuration as the input time-series data A in FIG. In other words, time-series data has a metrical structure and additional information in addition to a string of musical notes.
- the estimation unit 22 estimates code string information using the trained model M stored in the storage unit 140 or the like.
- the code string information indicates a code string corresponding to the note string accepted by the accepting unit 21, and is estimated based on the note string and additional information. Since the time-series data has the same configuration as the input time-series data A, the code string information has the same configuration as the output code string information B.
- the generation unit 23 generates score information based on the note sequence of the time-series data received by the reception unit 21 and the code string information estimated by the estimation unit 22 .
- the musical score information is information on an arranged musical score for a piano, and is data in which chord information is added to a staff notation.
- the musical score information is MIDI data to which code string information is added.
- the display unit 160 displays the musical score with chords based on the musical score information generated by the generating unit 23 .
- FIG. 5 shows an example of a musical score with chords displayed on the display unit 160.
- the code-attached musical score indicates that the code string information estimated by the estimating section 22 corresponds to each note of the note string accepted by the accepting section 21 .
- FIG. 6 is a flowchart showing an example of training processing by the training apparatus 10 of FIG.
- the training process in FIG. 6 is performed by CPU 130 in FIG. 1 executing a training program.
- the first acquisition unit 11 acquires the input time-series data A from each training data D (step S1).
- the second acquisition unit 12 acquires the output code string information B from each training data D (step S2). Either of steps S1 and S2 may be performed first, or may be performed simultaneously.
- the construction unit 13 performs machine learning using the input time-series data A obtained in step S1 as an input element and the output code string information B obtained in step S2 as an output element. (Step S3). Subsequently, the construction unit 13 determines whether or not sufficient machine learning has been performed (step S4). If the machine learning is insufficient, the construction unit 13 returns to step S3. Steps S3 and S4 are repeated while changing the parameters until sufficient machine learning is performed. The number of iterations of machine learning changes according to quality conditions that the trained model M to be constructed should satisfy.
- the construction unit 13 saves the input/output relationship between the input time-series data A and the output code string information B learned by the machine learning in step S3 as a trained model M (step S5). This completes the training process.
- FIG. 7 is a flowchart showing an example of chord estimation processing by the chord estimation device 20 of FIG.
- the chord estimation process in FIG. 7 is performed by CPU 130 in FIG. 1 executing a chord estimation program.
- the receiving unit 21 receives time-series data (step S11).
- the estimation unit 22 estimates code string information from the time-series data received in step S11 using the trained model M saved in step S5 of the training process (step S12).
- code string information including one or a plurality of code strings is estimated from the note strings included in the time-series data
- chord estimation is performed with a high degree of freedom.
- the chord change timing is also estimated in the course of time, more appropriate chord estimation is performed.
- the time-series data does not contain information that serves as a chord change delimiter, but the estimation unit 22 performs chord estimation including chord change timing.
- the generation unit 23 After that, the generation unit 23 generates score information based on the note string of the time-series data received in step S11 and the code string information estimated in step S12 (step S13). A score with chords may be displayed on the display unit 160 based on the generated score information. This completes the chord estimation process.
- the chord estimation apparatus 20 includes the receiving unit 21 that receives time-series data including a string of notes composed of a plurality of notes, and the trained model M: and an estimating unit 22 for estimating code string information indicating a code string corresponding to the musical note string.
- the trained model M is used to estimate appropriate code string information from the temporal flow of multiple notes in the time-series data. This makes it possible to present a coded musical score based on time-series data including a string of notes. Since one or more chord strings are estimated from the note string, chord estimation is performed with a high degree of freedom.
- the trained model M learns the input/output relationship between the input time-series data A including a reference note string consisting of a plurality of notes and the output code string information B indicating a code string corresponding to each note in the reference note string. It may be a machine learning model that In this case, code string information can be easily estimated from time-series data.
- the estimation unit 22 may also estimate the chord change timing in the code string. As a result, more appropriate chord estimation corresponding to the note string is performed.
- the input time-series data A may include genre information specifying the genre of music represented by the reference note string.
- the time-series data may also include genre information that designates the genre of music represented by a string of musical notes.
- the estimation unit 22 may estimate the code string information based on the time-series data including the genre information. Thus, chord estimation suitable for the genre of music is performed.
- the input time-series data A may include key information that specifies the key of the music represented by the reference note string.
- the time-series data may also include key information that specifies the key of music represented by a string of notes.
- the estimating section 22 may estimate the code string information based on the time-series data including the key information. This provides a chord estimation that is appropriate for the key of the music.
- the input time-series data A may include difficulty level information specifying the difficulty level of the musical score indicated by the reference note string.
- the time-series data may also include difficulty level information that designates the difficulty level of the musical score indicated by the note string.
- the estimation unit 22 may estimate the code string information based on the time-series data including the difficulty level information. As a result, appropriate chord estimation is performed according to the difficulty level of the musical score indicated by the note string.
- the chord estimating device 20 may further include a generation unit 23 that generates musical score information indicating a chorded musical score to which code string information is added so as to correspond to each note of the musical note string.
- the training apparatus 10 includes a first acquisition unit 11 that acquires input time-series data A including a reference note string composed of a plurality of notes, and an output code string that indicates a code string corresponding to the reference note string.
- a second acquisition unit 12 that acquires information B, and a construction unit 13 that constructs a trained model M that has learned the input/output relationship between the input time-series data A and the output code string information B.
- a trained model M that has learned the input/output relationship between the input time-series data A and the output code string information B can be easily constructed.
- the input time-series data A includes additional information, and the time-series data includes additional information, but the embodiment is not limited to this.
- the input time-series data A only needs to include the reference note string, and does not have to include additional information.
- the time-series data may include musical note sequences and may not include additional information.
- the input time-series data A has "bar” and "beat” information as the metrical structure, but the embodiment is not limited to this.
- the input time-series data A may not have a metrical structure.
- FIG. 8 is a diagram showing an example of output code string information B prepared for input time-series data A having no metrical structure. As shown in FIG. 8, the output code string information B does not have a metrical structure consisting of "bar" and "beat” information.
- the construction unit 13 may construct different trained models M according to the type of additional information, or may construct one trained model M.
- the input time-series data A may include, as additional information, a plurality of information out of key information, genre information, and difficulty level information.
- the code estimation device 20 includes the generator 23, but the embodiment is not limited to this.
- the player can create a musical score with chords by transcribing the chord string information estimated by the estimating section 22 to a desired musical score. Therefore, the code estimation device 20 does not have to include the generator 23 .
- the training data D is trained to estimate chord string information when performing on the piano, but the embodiment is not limited to this.
- the training data D may be trained to estimate chord string information when performing with other musical instruments such as guitars and drums.
- the user of the chord estimation device 20 is a performer.
- the machine learning by the training device 10 may be performed in advance by the staff of the musical score production company.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Electrophonic Musical Instruments (AREA)
Abstract
Description
以下、本発明の実施の形態に係るコード推定装置、訓練装置、コード推定方法および訓練方法について図面を用いて詳細に説明する。図1は、本発明の一実施の形態に係るコード推定装置および訓練装置を含む処理システムの構成を示すブロック図である。図1に示すように、処理システム100は、RAM(ランダムアクセスメモリ)110、ROM(リードオンリメモリ)120、CPU(中央演算処理装置)130、記憶部140、操作部150および表示部160を備える。 (1) Configuration of Processing System Hereinafter, a chord estimation device, a training device, a chord estimation method, and a training method according to embodiments of the present invention will be described in detail with reference to the drawings. FIG. 1 is a block diagram showing the configuration of a processing system including a chord estimation device and a training device according to one embodiment of the present invention. As shown in FIG. 1, the
訓練済モデルMは、コード推定装置20の使用者(以下、演奏者と呼ぶ。)が楽曲を演奏するときに参照するコード列を提示するために訓練された機械学習モデルである。訓練済モデルMは、複数の訓練データDを用いて構築される。訓練装置10の使用者は、操作部150を操作することにより、訓練データDを生成することができる。訓練データDは、参照演奏者の音楽的知識または音楽的センス等に基づいて作成されたデータである。参照演奏者は、楽曲の演奏に関して比較的高い技量を有する。参照演奏者は、楽曲の演奏における演奏者の指導者または師であってもよい。 (2) Training Data The trained model M is a machine learning model that has been trained to present chord strings to be referred to when the user of the chord estimation device 20 (hereinafter referred to as a performer) plays a piece of music. be. A trained model M is constructed using a plurality of training data D. A user of the
図4は、訓練装置10およびコード推定装置20の構成を示すブロック図である。図4に示すように、訓練装置10は、機能部として、第1の取得部11、第2の取得部12および構築部13を含む。図1のCPU130が訓練プログラムを実行することにより、訓練装置10の機能部が実現される。訓練装置10の機能部の少なくとも一部は、電子回路等のハードウエアにより実現されてもよい。 (3) Training Device and Chord Estimation Device FIG. 4 is a block diagram showing the configuration of the
図6は、図4の訓練装置10による訓練処理の一例を示すフローチャートである。図6の訓練処理は、図1のCPU130が訓練プログラムを実行することにより行われる。まず、第1の取得部11は、各訓練データDから入力時系列データAを取得する(ステップS1)。また、第2の取得部12は、各訓練データDから出力コード列情報Bを取得する(ステップS2)。ステップS1,S2は、いずれが先に実行されてもよいし、同時に実行されてもよい。 (4) Training Processing and Chord Estimation Processing FIG. 6 is a flowchart showing an example of training processing by the
以上説明したように、本実施の形態に係るコード推定装置20は、複数の音符からなる音符列を含む時系列データを受け付ける受付部21と、訓練済モデルMを用いて、音符列に対応するコード列を示すコード列情報を推定する推定部22とを備える。この構成によれば、訓練済モデルMを用いて、時系列データにおける複数の音符の時間的流れから適切なコード列情報が推定される。これにより、音符列を含む時系列データに基づいてコード付き楽譜を提示することができる。音符列からは1つ、または複数のコード列が推定されるので、自由度の高いコード推定が行われる。 (5) Effect of the Embodiment As described above, the
上記実施の形態において、入力時系列データAは付加情報を含み、時系列データは付加情報を含むが、実施の形態はこれに限定されない。入力時系列データAは、参照音符列を含めばよく、付加情報を含まなくてもよい。同様に、時系列データは、音符列を含めばよく、付加情報を含まなくてもよい。 (6) Other Embodiments In the above embodiment, the input time-series data A includes additional information, and the time-series data includes additional information, but the embodiment is not limited to this. The input time-series data A only needs to include the reference note string, and does not have to include additional information. Similarly, the time-series data may include musical note sequences and may not include additional information.
Claims (16)
- 複数の音符からなる音符列を含む時系列データを受け付ける受付部と、
訓練済モデルを用い、前記時系列データに基づいて、前記音符列に対応するコード列を示すコード列情報を推定する推定部と、を備えるコード推定装置。 a reception unit that receives time-series data including a string of notes made up of a plurality of notes;
an estimating unit for estimating code string information indicating a code string corresponding to the musical note string based on the time-series data using a trained model. - 前記訓練済モデルは、複数の音符からなる参照音符列を含む入力時系列データと、前記参照音符列に対応するコード列を示す出力コード列情報との間の入出力関係を習得したモデルである、請求項1に記載のコード推定装置。 The trained model is a model that has learned the input/output relationship between input time-series data including a reference note string consisting of a plurality of notes and output code string information indicating a code string corresponding to the reference note string. 2. A code estimation apparatus according to claim 1.
- 前記推定部は、前記コード列におけるコードチェンジのタイミングについても推定する、請求項1または請求項2に記載のコード推定装置。 3. The chord estimating apparatus according to claim 1, wherein said estimating unit also estimates timing of chord change in said code string.
- 前記入力時系列データは、前記参照音符列で表現される音楽のジャンルを指定するジャンル情報を含み、
前記時系列データは、前記音符列で表現される音楽のジャンルを指定するジャンル情報を含み、
前記推定部は、ジャンル情報を含む前記時系列データに基づいて、前記コード列情報を推定する、請求項2に記載のコード推定装置。 the input time-series data includes genre information specifying a genre of music represented by the reference note sequence;
The time-series data includes genre information specifying the genre of music represented by the musical note sequence,
3. The chord estimation device according to claim 2, wherein said estimation unit estimates said code string information based on said time series data including genre information. - 前記入力時系列データは、前記参照音符列で表現される音楽の調を指定する調情報を含み、
前記時系列データは、前記音符列で表現される音楽の調を指定する調情報を含み、
前記推定部は、調情報を含む前記時系列データに基づいて、前記コード列情報を推定する、請求項2に記載のコード推定装置。 The input time-series data includes key information specifying the key of music represented by the reference note string,
The time-series data includes key information specifying the key of the music represented by the string of notes,
3. The chord estimation apparatus according to claim 2, wherein said estimation unit estimates said chord string information based on said time-series data including key information. - 前記入力時系列データは、前記参照音符列で示される楽譜の難易度を指定する難易度情報を含み、
前記時系列データは、前記音符列で示される楽譜の難易度を指定する難易度情報を含み、
前記推定部は、難易度情報を含む前記時系列データに基づいて、前記コード列情報を推定する、請求項2に記載のコード推定装置。 The input time-series data includes difficulty level information specifying the difficulty level of the musical score indicated by the reference note string,
The time-series data includes difficulty level information that specifies the difficulty level of the musical score indicated by the string of notes,
3. The chord estimation device according to claim 2, wherein said estimation unit estimates said code string information based on said time-series data including difficulty level information. - 前記音符列の各音符に対応するように前記コード列情報が付されたコード付き楽譜を示す楽譜情報を生成する生成部をさらに備える、請求項1~6のいずれか一項に記載のコード推定装置。 7. The chord estimation according to any one of claims 1 to 6, further comprising a generating unit that generates score information indicating a chorded score to which the code string information is attached so as to correspond to each note of the note string. Device.
- 複数の音符からなる参照音符列を含む入力時系列データを取得する第1の取得部と、
前記参照音符列に対応するコード列を示す出力コード列情報を取得する第2の取得部と、
前記入力時系列データと前記出力コード列情報との間の入出力関係を習得した訓練済モデルを構築する構築部と、を備える訓練装置。 a first acquisition unit that acquires input time-series data including a reference note string consisting of a plurality of notes;
a second acquisition unit for acquiring output code string information indicating a code string corresponding to the reference note string;
A training device, comprising: a building unit that builds a trained model that has learned the input/output relationship between the input time-series data and the output code string information. - 複数の音符からなる音符列を含む時系列データを受け付け、
訓練済モデルを用い、前記時系列データに基づいて、前記音符列に対応するコード列を示すコード列情報を推定する、コンピュータにより実行されるコード推定方法。 Accepts time-series data containing note strings consisting of multiple notes,
A computer-implemented chord estimation method for estimating chord string information indicating a chord string corresponding to said string of notes based on said time-series data using a trained model. - 前記訓練済モデルは、複数の音符からなる参照音符列を含む入力時系列データと、前記参照音符列に対応するコード列を示す出力コード列情報との間の入出力関係を習得したモデルである、請求項9に記載のコンピュータにより実行されるコード推定方法。 The trained model is a model that has learned the input/output relationship between input time-series data including a reference note string consisting of a plurality of notes and output code string information indicating a code string corresponding to the reference note string. 10. The computer implemented code estimation method of claim 9.
- 前記推定することは、前記コード列におけるコードチェンジのタイミングについても推定する、請求項9または請求項10に記載のコンピュータにより実行されるコード推定方法。 11. A computer-implemented chord estimation method according to claim 9 or 10, wherein said estimating also estimates the timing of chord changes in said chord sequence.
- 前記入力時系列データは、前記参照音符列で表現される音楽のジャンルを指定するジャンル情報を含み、
前記時系列データは、前記音符列で表現される音楽のジャンルを指定するジャンル情報を含み、
前記推定することは、ジャンル情報を含む前記時系列データに基づいて、前記コード列情報を推定する、請求項10に記載のコンピュータにより実行されるコード推定方法。 the input time-series data includes genre information specifying a genre of music represented by the reference note sequence;
The time-series data includes genre information specifying the genre of music represented by the musical note sequence,
11. The computer-implemented method of estimating chords of claim 10, wherein the estimating estimates the code string information based on the time series data including genre information. - 前記入力時系列データは、前記参照音符列で表現される音楽の調を指定する調情報を含み、
前記時系列データは、前記音符列で表現される音楽の調を指定する調情報を含み、
前記推定することは、調情報を含む前記時系列データに基づいて、前記コード列情報を推定する、請求項10に記載のコンピュータにより実行されるコード推定方法。 The input time-series data includes key information specifying the key of music represented by the reference note string,
The time-series data includes key information specifying the key of the music represented by the string of notes,
11. The computer-implemented chord estimation method of claim 10, wherein said estimating estimates said chord string information based on said time series data including key information. - 前記入力時系列データは、前記参照音符列で示される楽譜の難易度を指定する難易度情報を含み、
前記時系列データは、前記音符列で示される楽譜の難易度を指定する難易度情報を含み、
前記推定することは、難易度情報を含む前記時系列データに基づいて、前記コード列情報を推定する、請求項10に記載のコンピュータにより実行されるコード推定方法。 The input time-series data includes difficulty level information specifying the difficulty level of the musical score indicated by the reference note string,
The time-series data includes difficulty level information that specifies the difficulty level of the musical score indicated by the string of notes,
11. The computer-implemented method of estimating chords of claim 10, wherein the estimating estimates the chord string information based on the time series data including difficulty information. - さらに、前記音符列の各音符に対応するように前記コード列情報が付されたコード付き楽譜を示す楽譜情報を生成する、請求項9~14のいずれか一項に記載のコンピュータにより実行されるコード推定方法。 15. The computer according to any one of claims 9 to 14, further comprising generating musical score information indicating a chorded musical score to which the code string information has been added so as to correspond to each note of the musical note string. Code estimation method.
- 複数の音符からなる参照音符列を含む入力時系列データを取得し、
前記参照音符列に対応するコード列を示す出力コード列情報を取得し、
前記入力時系列データと前記出力コード列情報との間の入出力関係を習得した訓練済モデルを構築する、コンピュータにより実行される訓練方法。 Take input time series data containing a reference note string consisting of multiple notes,
Acquiring output code string information indicating a code string corresponding to the reference note string;
A computer-implemented training method for building a trained model that has learned input-output relationships between said input time series data and said output code string information.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2023508892A JPWO2022202199A1 (en) | 2021-03-26 | 2022-03-03 | |
CN202280023333.9A CN117043852A (en) | 2021-03-26 | 2022-03-03 | Chord estimation device, training device, chord estimation method, and training method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2021-052532 | 2021-03-26 | ||
JP2021052532 | 2021-03-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022202199A1 true WO2022202199A1 (en) | 2022-09-29 |
Family
ID=83396894
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2022/009233 WO2022202199A1 (en) | 2021-03-26 | 2022-03-03 | Code estimation device, training device, code estimation method, and training method |
Country Status (3)
Country | Link |
---|---|
JP (1) | JPWO2022202199A1 (en) |
CN (1) | CN117043852A (en) |
WO (1) | WO2022202199A1 (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015031738A (en) * | 2013-07-31 | 2015-02-16 | 株式会社河合楽器製作所 | Chord progression estimation and detection device and chord progression estimation and detection program |
WO2020145326A1 (en) * | 2019-01-11 | 2020-07-16 | ヤマハ株式会社 | Acoustic analysis method and acoustic analysis device |
-
2022
- 2022-03-03 JP JP2023508892A patent/JPWO2022202199A1/ja active Pending
- 2022-03-03 CN CN202280023333.9A patent/CN117043852A/en active Pending
- 2022-03-03 WO PCT/JP2022/009233 patent/WO2022202199A1/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2015031738A (en) * | 2013-07-31 | 2015-02-16 | 株式会社河合楽器製作所 | Chord progression estimation and detection device and chord progression estimation and detection program |
WO2020145326A1 (en) * | 2019-01-11 | 2020-07-16 | ヤマハ株式会社 | Acoustic analysis method and acoustic analysis device |
Also Published As
Publication number | Publication date |
---|---|
CN117043852A (en) | 2023-11-10 |
JPWO2022202199A1 (en) | 2022-09-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112382257B (en) | Audio processing method, device, equipment and medium | |
CN111630590B (en) | Method for generating music data | |
KR101942814B1 (en) | Method for providing accompaniment based on user humming melody and apparatus for the same | |
EP3489946A1 (en) | Real-time jamming assistance for groups of musicians | |
US7411125B2 (en) | Chord estimation apparatus and method | |
JP2019152716A (en) | Information processing method and information processor | |
JP6760450B2 (en) | Automatic arrangement method | |
JP2012506061A (en) | Analysis method of digital music sound signal | |
JP6565528B2 (en) | Automatic arrangement device and program | |
JP6693176B2 (en) | Lyrics generation device and lyrics generation method | |
Jensen | Evolutionary music composition: A quantitative approach | |
JP6645085B2 (en) | Automatic arrangement device and program | |
WO2022202199A1 (en) | Code estimation device, training device, code estimation method, and training method | |
JP7375302B2 (en) | Acoustic analysis method, acoustic analysis device and program | |
US20220383843A1 (en) | Arrangement generation method, arrangement generation device, and generation program | |
US6984781B2 (en) | Music formulation | |
JP2019109357A (en) | Feature analysis method for music information and its device | |
CN116710998A (en) | Information processing system, electronic musical instrument, information processing method, and program | |
Vargas et al. | Artificial musical pattern generation with genetic algorithms | |
Suthaphan et al. | Music generator for elderly using deep learning | |
WO2022244403A1 (en) | Musical score writing device, training device, musical score writing method and training method | |
WO2022190453A1 (en) | Fingering presentation device, training device, fingering presentation method, and training method | |
KR102490769B1 (en) | Method and device for evaluating ballet movements based on ai using musical elements | |
Akimoto et al. | SketTune: Real-time input assistance for novices to compose music for self-expression | |
WO2020171035A1 (en) | Sound signal synthesis method, generative model training method, sound signal synthesis system, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22774995 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 202280023333.9 Country of ref document: CN Ref document number: 2023508892 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 22774995 Country of ref document: EP Kind code of ref document: A1 |