WO2017195292A1 - Music structure analysis device, method for analyzing music structure, and music structure analysis program - Google Patents

Music structure analysis device, method for analyzing music structure, and music structure analysis program Download PDF

Info

Publication number
WO2017195292A1
WO2017195292A1 PCT/JP2016/063981 JP2016063981W WO2017195292A1 WO 2017195292 A1 WO2017195292 A1 WO 2017195292A1 JP 2016063981 W JP2016063981 W JP 2016063981W WO 2017195292 A1 WO2017195292 A1 WO 2017195292A1
Authority
WO
WIPO (PCT)
Prior art keywords
section
music
development
structure analysis
feature
Prior art date
Application number
PCT/JP2016/063981
Other languages
French (fr)
Japanese (ja)
Inventor
四郎 鈴木
Original Assignee
Pioneer DJ株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pioneer DJ株式会社 filed Critical Pioneer DJ株式会社
Priority to JP2018516262A priority Critical patent/JPWO2017195292A1/en
Priority to EP16901640.9A priority patent/EP3457395A4/en
Priority to PCT/JP2016/063981 priority patent/WO2017195292A1/en
Publication of WO2017195292A1 publication Critical patent/WO2017195292A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/061Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of musical phrases, isolation of musically relevant segments, e.g. musical thumbnail generation, or for temporal structure analysis of a musical piece, e.g. determination of the movement sequence of a musical work
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/076Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction of timing, tempo; Beat detection
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/081Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for automatic key or tonality recognition, e.g. using musical rules or a knowledge base
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/571Chords; Chord sequences
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/121Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods

Definitions

  • the present invention relates to a music structure analysis apparatus, a music structure analysis method, and a music structure analysis program.
  • the music structure is automatically assigned by assigning characteristic sections that characterize the music structure such as so-called Intro, A melody (Verse1), B melody (Verse2), Sabi (Hook), Outro, etc. Analysis techniques are known.
  • Patent Literature 1 discloses a technique for assigning feature sections such as stanzas and refrains to music data by performing similarity determination between segments (feature sections) assigned to music data. Yes.
  • An object of the present invention is to provide a music structure analysis apparatus, a music structure analysis method, and a music structure analysis program capable of easily assigning characteristic sections characterizing music data.
  • the music structure analysis apparatus of the present invention is A music structure analysis apparatus that assigns the feature section to the music data set with the development points of the feature section characterizing the structure of the music data, A position information acquisition unit for acquiring position information of the development point; Based on the position information of the expansion points acquired by the position information acquisition unit, a sound number analysis unit that analyzes the number of sounds with different frequencies for each section between the expansion points; A feature section allocating section that assigns a feature section to a section between other development points based on a section of development points that takes the maximum value of the number of sounds analyzed by the sound number analysis section, It is characterized by having.
  • the music structure analysis method of the present invention includes: A music structure analysis method for assigning a characteristic section to music data in which a development point of a characteristic section characterizing the structure of the music data is set, Obtaining position information of the development point; A procedure for analyzing the number of sounds having different frequencies for each section between the development points based on the acquired position information of the development points; A procedure for assigning a feature section to a section between other development points based on a section between development points taking the maximum value of the number of analyzed sounds; It is characterized by implementing.
  • the music structure analysis program of the present invention causes a computer to function as the music structure analysis apparatus described above.
  • the schematic diagram showing the structure of the acoustic system which concerns on embodiment of this invention.
  • FIG. 1 shows an acoustic control system 1 according to an embodiment of the present invention.
  • the acoustic control system 1 includes two digital players 2, a digital mixer 3, A computer 4 and a speaker 5 are provided.
  • the digital player 2 includes a jog dial 2A, a plurality of operation buttons (not shown), and a display 2B, and the operator of the digital player 2 operates the jog dial 2A or the operation buttons to perform acoustic control information corresponding to the operation. Can be output.
  • the acoustic control information is output to the computer 4 via a USB (Universal Serial Bus) cable 6 capable of bidirectional communication.
  • USB Universal Serial Bus
  • the digital mixer 3 includes an operation switch 3A, a volume adjustment lever 3B, and a left / right switching lever 3C. By operating these switches 3A, levers 3B, 3C, acoustic control information can be output.
  • the acoustic control information is output to the computer 4 via the USB cable 7.
  • the music information processed by the computer 4 is input to the digital mixer 3, and the music information including the input digital signal is converted into an analog signal and output from the speaker 5 through the analog cable 8.
  • the digital player 2 and the digital mixer 3 are connected to each other via a LAN (Local Area Network) cable 9 compliant with the IEEE 1394 standard, and the sound control generated by operating the digital player 2 without using the computer 4. Information can also be output directly to the digital mixer 3 for DJ performance.
  • LAN Local Area Network
  • FIG. 2 shows a functional block diagram of the computer 4 as a music structure analyzing apparatus.
  • the computer 4 includes a position information acquisition unit 11, a sound number analysis unit 12, a bass level analysis unit 13, a ratio calculation unit 14, a feature section allocation unit 15, and a display as a music structure analysis program executed on the arithmetic processing device 10.
  • An information generation unit 16 is provided.
  • the position information acquisition unit 11 acquires the number of measures of the development point set in the music data M1 as the position information of the development point. Specifically, as shown in FIG. 3, the position information acquisition unit 11 has development points P1, P2,..., Pn, Pn + 3,... Set between bars in the music data M1.
  • the position information of Pe-4 in this embodiment, the number of bars is acquired.
  • the expansion point is the frequency analysis of the number of sounds with different frequencies for each measure by FFT, etc., the number of peaks of the sound pressure level is counted, and the other is based on the measure that is the maximum number of sounds in the music data M1.
  • the ratio of the sound of the bars is calculated and set as the inter-bar position where the ratio changes greatly.
  • the expansion points can be set in the computer 4 using the sound number analysis unit 12, but the expansion points P1, P2,..., Pn, by analyzing the number of sounds in advance as in this embodiment.
  • Music data M1 in which Pn + 3,... Pe-4 is set may be used.
  • the setting of the development points P1, P2,..., Pn, Pn + 3,... Pe-4 is not limited to the method described above, and may be performed based on, for example, the similarity of phrases in the music data M1. .
  • the sound number analysis unit 12 detects the signal level of each frequency band for each segment between the development points P1, P2,..., Pn, Pn + 3,. Analyze the number of sounds.
  • the “number of sounds” may be a case where sounds having different frequencies may be counted as different sounds, or a fundamental tone, overtones, etc. may be counted as one with the same scale.
  • the input music data M1 may be music data stored on a hard disk in the computer 4, or music data recorded on a CD, a Blu-ray disc or the like inserted in a slot of the digital player 2. Alternatively, it may be music data that can be downloaded from a network via a communication line.
  • the sound number analysis unit 12 for example, as shown in FIG. 4, for each section between the development points P1, P2,..., Pn, Pn + 3,.
  • Count the number of sounds by counting the number of peaks in the frequency.
  • the analysis of sounds having different frequencies is performed using FFT, but the present invention is not limited to this, and for example, frequency conversion may be performed using discrete cosine transform or discrete Fourier transform.
  • the sound number analysis unit 12 outputs the analysis result to the ratio calculation unit 14.
  • the bass level analysis unit 13 takes into account the average value for each bar section of the low-frequency sound pressure peak level, which is the low-frequency signal level, and develops points P1, P2, ..., Pn, Pn + 3, ... Pe-4. It is provided to allocate a chorus section in the section between.
  • the low sound level analysis unit 13 has a frequency lower than a predetermined frequency in a section between the development points P1, P2,..., Pn, Pn + 3,. Analyzes the low sound pressure peak level. Specifically, the bass level analysis unit 13 acquires bass sound levels such as bass drums and bass, and calculates the average value of the bass sound pressure peak levels in the sections for each measure section as the development point P1, .., Pn, Pn + 3,... Pe-4. The bass level analysis unit 13 outputs the analysis result to the feature section allocation unit 15.
  • the feature interval allocation unit 15 sets the interval between the expansion points P 1, P 2,..., Pn, Pn + 3,.
  • Characteristic sections such as an intro section, an A melody (Verse1) section, a B melody (Verse2) section, a chorus (Hook) section, a C melody (Verse3) section, and an outro section are allocated.
  • the feature section allocating unit 15 searches for a maximum value 1 and a maximum value 2 that are larger than the preceding and following sections based on the analysis result of the sound number analysis unit 12. Then, maximal value 1 and maximal value 2 are set as candidates for the rust section (Hook).
  • the feature section allocating unit 15 obtains the average value of the low frequency sound pressure peak levels of each feature section based on the analysis result of the bass level analyzing unit 13, and the maximum value 1 Then, it is determined whether or not the average value of the low-frequency sound pressure peak level in the section having the maximum value 2 is higher than a predetermined threshold value, and the hook section is allocated.
  • the feature interval allocation unit 15 sets the interval before the maximum value 1 and the maximum value 2 as the A melody interval (Verse1), and the subsequent interval as the B melody interval (Verse2). Further, the subsequent section is allocated to the C melody section (Verse 3) or the like.
  • Which feature section is assigned is determined by whether or not the number of sounds exceeds a predetermined threshold.
  • the predetermined threshold value may be a fixed threshold value that is smaller than the maximum value, or may be a threshold value that is set as a predetermined ratio with respect to the maximum value and varies according to the maximum value.
  • the name of the feature section to be assigned is arbitrary. In FIG. 5, names such as A-Verse and B-Verse may be assigned.
  • the feature section allocating unit 15 assigns intro sections and outro sections in advance. Keep going.
  • the display information generating unit 16 generates the feature section assigned by the feature section assigning unit 15 as display information together with the music data M1. Specifically, as shown in FIG. 6, the characteristic section is displayed as the music data M1 progresses, and the display information for changing the color of the characteristic section as the music data M1 progresses is generated.
  • the display information generated by the display information generation unit 16 is output to the display 2B serving as a display device of the digital player 2, and the DJ performer is performing which characteristic section is being played as the music performance of the music data M1 progresses. Can be confirmed.
  • the position information acquisition unit 11 acquires position information of the development points P1, P2,..., Pn, Pn + 3,... Pe-4 in the music data M1 (step S1).
  • the sound number analysis unit 12 analyzes the number of sounds in the section between the development points P1, P2,..., Pn, Pn + 3,... Pe-4 (step S2).
  • the ratio calculation unit 14 uses the interval between the expansion points Pn and Pn + 3 having the maximum number of sounds as a reference to other expansion points P1, P2,. -4 is calculated (step S3).
  • the feature section assigning unit 15 assigns the intro section to the section from the start point of the music data M1 to the first development point P1 (step S4). Subsequently, the feature section assigning unit 15 assigns an outro section from the last development point Pe-4 of the music data M1 to the end point of the music-multiple M1 (step S5).
  • the feature section allocating unit 15 searches for the maximum value in the section between the development points other than the intro section and the outro section (step S6). The search may start from the section next to the intro section or may start from the section before the outro section. When the section having the maximum value is found, the feature section allocating unit 15 obtains the average value of the low-frequency sound pressure peak levels in the section between the expansion points P1, P2,..., Pn, Pn + 3,. (Procedure S7).
  • the feature section allocating unit 15 determines whether or not the average value of the low frequency sound pressure peak level in the section having the maximum value exceeds a predetermined threshold (step S8). When the value is equal to or less than the predetermined threshold, the feature section allocating unit 15 searches for the next maximum value. When the predetermined threshold value is exceeded, the feature section allocating unit 15 allocates a hook section to the section (step S9). The feature section allocating unit 15 repeats steps S6 to S9 for all sections having the maximum value in the music data M1 (step S10). Steps S8 and S9 are performed in order to improve the detection accuracy of the chorus section, and only the search for the maximum value of the section between the expansion points that takes the maximum value of the number of sounds, A rust section may be allocated.
  • the feature section assigning unit 15 acquires the number of section sounds between other development points before and after the section set as the chorus section (step S11).
  • the feature section allocating unit 15 determines whether or not the number of sounds in the section between other development points exceeds a predetermined threshold (step S12).
  • the feature section allocating unit 15 allocates an A melody (Verse1) section to the section (step S13), and when the ratio is equal to or less than the predetermined threshold, the feature section allocating unit. 15 assigns a B melody (Verse2) section to the section (step S14).
  • the feature section assigning unit 15 repeats until the assignment of the feature section to the section between all the development points P1, P2,..., Pn, Pn + 3,.
  • the feature section assigning unit 15 outputs the assigned result to the display information generating unit 16, and the display information generating unit 16 generates display information based on the assignment result, The generated display information is output to the display 2B of the digital player 2 (step S16).
  • the music data M1 can be easily and quickly allocated. It is possible to assign feature sections.
  • a user performing DJ performance can visually recognize which feature section is currently being played, and therefore performs a higher level DJ performance. be able to.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Auxiliary Devices For Music (AREA)

Abstract

Provided is a music structure analysis device (4) that allocates feature segments that characterize the structure of music data (M1) to music data (M1) in which development points are established in feature segments, wherein the music structure analysis device (4) includes: a position information acquisition unit (11) for acquiring the position information for the development points; a sound number analyzing unit (12) for analyzing a number of sounds with different frequencies for each segment between each of the development points on the basis of the position information for the development points acquired by the position information acquisition unit (11); and a feature segment allocation unit (15) for allocating the feature segments to a segment between other development points on the basis of the segment between development points determined to contain a local maximum value for the number of sounds analyzed by the sound number analyzing unit (12).

Description

楽曲構造解析装置、楽曲構造解析方法および楽曲構造解析プログラムMusic structure analysis apparatus, music structure analysis method, and music structure analysis program
 本発明は、楽曲構造解析装置、楽曲構造解析方法および楽曲構造解析プログラムに関する。 The present invention relates to a music structure analysis apparatus, a music structure analysis method, and a music structure analysis program.
 従来、楽曲データに、いわゆるイントロ(Intro)、Aメロ(Verse1)、Bメロ(Verse2)、サビ(Hook)、アウトロ(Outro)等の楽曲構造を特徴づける特徴区間を割り付け、楽曲構造を自動的に解析する技術が知られている。
 例えば、特許文献1には、楽曲データに割り付けられたセグメント(特徴区間)同士を、セグメント間の類似性判定を行うことにより、スタンザ、リフレイン等の特徴区間を楽曲データに割り付ける技術が開示されている。
Conventionally, the music structure is automatically assigned by assigning characteristic sections that characterize the music structure such as so-called Intro, A melody (Verse1), B melody (Verse2), Sabi (Hook), Outro, etc. Analysis techniques are known.
For example, Patent Literature 1 discloses a technique for assigning feature sections such as stanzas and refrains to music data by performing similarity determination between segments (feature sections) assigned to music data. Yes.
特許4775380号公報Japanese Patent No. 4775380
 しかしながら、前記特許文献1に記載の技術では、特徴区間同士の類似性に基づいて、特徴区間の割り付けを行っているため、サビ、Bメロ等の一部の特徴区間の割り付けをすることができないという課題がある。
 また、割り付けられたとしても、イントロ、Aメロ、Bメロ、サビ、Cメロ、アウトロそれぞれについて、類似性を対比して割り付けなければならず、特徴区間の割り付けが煩雑化するという課題がある。
However, in the technique described in Patent Document 1, since the feature sections are assigned based on the similarity between the feature sections, it is not possible to assign some feature sections such as rust and B melody. There is a problem.
Even if assigned, the intro, A melody, B melody, rust, C melody, and outro must be assigned in comparison with each other, and the assignment of feature sections becomes complicated.
 本発明の目的は、楽曲データを特徴づける特徴区間を、簡単に割り付けることのできる楽曲構造解析装置、楽曲構造解析方法および楽曲構造解析プログラムを提供することにある。 An object of the present invention is to provide a music structure analysis apparatus, a music structure analysis method, and a music structure analysis program capable of easily assigning characteristic sections characterizing music data.
 本発明の楽曲構造解析装置は、
 楽曲データの構造を特徴づける特徴区間の展開ポイントが設定された楽曲データに、前記特徴区間を割り付ける楽曲構造解析装置であって、
 前記展開ポイントの位置情報を取得する位置情報取得部と、
 前記位置情報取得部によって取得された展開ポイントの位置情報に基づいて、それぞれの展開ポイント間の区間毎に、周波数の異なる音の数を解析する音数解析部と、
 前記音数解析部により解析された音の数の極大値をとる展開ポイントの区間に基づいて、他の展開ポイント間の区間に、特徴区間を割り付ける特徴区間割付部と、
を備えていることを特徴とする。
The music structure analysis apparatus of the present invention is
A music structure analysis apparatus that assigns the feature section to the music data set with the development points of the feature section characterizing the structure of the music data,
A position information acquisition unit for acquiring position information of the development point;
Based on the position information of the expansion points acquired by the position information acquisition unit, a sound number analysis unit that analyzes the number of sounds with different frequencies for each section between the expansion points;
A feature section allocating section that assigns a feature section to a section between other development points based on a section of development points that takes the maximum value of the number of sounds analyzed by the sound number analysis section,
It is characterized by having.
 本発明の楽曲構造解析方法は、
 楽曲データの構造を特徴づける特徴区間の展開ポイントが設定された楽曲データに、特徴区間を割り付ける楽曲構造解析方法であって、
 前記展開ポイントの位置情報を取得する手順と、
 取得された展開ポイントの位置情報に基づいて、それぞれの展開ポイント間の区間毎に周波数の異なる音の数を解析する手順と、
 解析された音の数の極大値をとる展開ポイント間の区間に基づいて、他の展開ポイント間の区間に、特徴区間を割り付ける手順と、
を実施することを特徴とする。
The music structure analysis method of the present invention includes:
A music structure analysis method for assigning a characteristic section to music data in which a development point of a characteristic section characterizing the structure of the music data is set,
Obtaining position information of the development point;
A procedure for analyzing the number of sounds having different frequencies for each section between the development points based on the acquired position information of the development points;
A procedure for assigning a feature section to a section between other development points based on a section between development points taking the maximum value of the number of analyzed sounds;
It is characterized by implementing.
 本発明の楽曲構造解析プログラムは、コンピュータを、前述した楽曲構造解析装置として機能させることを特徴とする。 The music structure analysis program of the present invention causes a computer to function as the music structure analysis apparatus described above.
本発明の実施形態に係る音響システムの構造を表す模式図。The schematic diagram showing the structure of the acoustic system which concerns on embodiment of this invention. 前記実施形態における楽曲構造解析装置の構造を表すブロック図。The block diagram showing the structure of the music structure analysis apparatus in the said embodiment. 前記実施形態における展開ポイントを説明するための模式図。The schematic diagram for demonstrating the expansion | deployment point in the said embodiment. 前記実施形態における楽曲データの展開ポイント間の区間を説明するための模式図。The schematic diagram for demonstrating the area between the expansion | deployment points of the music data in the said embodiment. 前記実施形態における特徴区間割付部を説明するためのグラフ。The graph for demonstrating the characteristic area allocation part in the said embodiment. 前記実施形態における表示情報生成部により生成された表示情報を説明するための模式図。The schematic diagram for demonstrating the display information produced | generated by the display information production | generation part in the said embodiment. 前記実施形態における作用を説明するためのフローチャート。The flowchart for demonstrating the effect | action in the said embodiment.
 [1]音響制御システムの全体構成
 図1には、本発明の実施形態に係る音響制御システム1が示されており、音響制御システム1は、2台のデジタルプレーヤー2と、デジタルミキサー3と、コンピュータ4と、スピーカー5とを備える。
 デジタルプレーヤー2は、ジョグダイヤル2Aと、図示しない複数の操作ボタンと、ディスプレイ2Bとを備え、デジタルプレーヤー2の操作者は、ジョグダイヤル2Aや、操作ボタンを操作することにより、操作に応じた音響制御情報を出力することができる。音響制御情報は、双方向通信可能なUSB(Universal Serial Bus)ケーブル6を介してコンピュータ4に出力される。
[1] Overall Configuration of Acoustic Control System FIG. 1 shows an acoustic control system 1 according to an embodiment of the present invention. The acoustic control system 1 includes two digital players 2, a digital mixer 3, A computer 4 and a speaker 5 are provided.
The digital player 2 includes a jog dial 2A, a plurality of operation buttons (not shown), and a display 2B, and the operator of the digital player 2 operates the jog dial 2A or the operation buttons to perform acoustic control information corresponding to the operation. Can be output. The acoustic control information is output to the computer 4 via a USB (Universal Serial Bus) cable 6 capable of bidirectional communication.
 デジタルミキサー3は、操作スイッチ3Aと、音量調整レバー3Bと、左右切替レバー3Cとを備え、これらのスイッチ3A、レバー3B、3Cを操作することにより、音響制御情報を出力することができる。音響制御情報は、USBケーブル7を介してコンピュータ4に出力される。また、デジタルミキサー3には、コンピュータ4で処理された楽曲情報が入力され、入力されたデジタル信号からなる楽曲情報は、アナログ信号に変換され、アナログケーブル8を介して、スピーカー5から音声出力される。
 また、デジタルプレーヤー2およびデジタルミキサー3は、IEEE1394規格に準拠したLAN(Local Area Network)ケーブル9を介して互いに接続され、コンピュータ4を使わなくとも、デジタルプレーヤー2を操作して生成された音響制御情報を、直接デジタルミキサー3に出力して、DJパフォーマンスを行うこともできる。
The digital mixer 3 includes an operation switch 3A, a volume adjustment lever 3B, and a left / right switching lever 3C. By operating these switches 3A, levers 3B, 3C, acoustic control information can be output. The acoustic control information is output to the computer 4 via the USB cable 7. The music information processed by the computer 4 is input to the digital mixer 3, and the music information including the input digital signal is converted into an analog signal and output from the speaker 5 through the analog cable 8. The
The digital player 2 and the digital mixer 3 are connected to each other via a LAN (Local Area Network) cable 9 compliant with the IEEE 1394 standard, and the sound control generated by operating the digital player 2 without using the computer 4. Information can also be output directly to the digital mixer 3 for DJ performance.
 [2]コンピュータ4の機能ブロック構成
 図2には、楽曲構造解析装置としてのコンピュータ4の機能ブロック図が示されている。コンピュータ4は、演算処理装置10上で実行される楽曲構造解析プログラムとしての
位置情報取得部11、音数解析部12、低音レベル解析部13、比率演算部14、特徴区間割付部15、および表示情報生成部16を備える。
[2] Functional Block Configuration of Computer 4 FIG. 2 shows a functional block diagram of the computer 4 as a music structure analyzing apparatus. The computer 4 includes a position information acquisition unit 11, a sound number analysis unit 12, a bass level analysis unit 13, a ratio calculation unit 14, a feature section allocation unit 15, and a display as a music structure analysis program executed on the arithmetic processing device 10. An information generation unit 16 is provided.
 位置情報取得部11は、楽曲データM1中に設定された展開ポイントの小節数を、展開ポイントの位置情報として取得する。具体的には、図3に示されるように、位置情報取得部11は、楽曲データM1中における、小節と小節の間に設定された展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4の位置情報、本実施形態では小節数を取得する。
 展開ポイントは、小節毎に周波数の異なる音の数を、FFT等により周波数解析し、音圧レベルのピーク本数をカウントし、楽曲データM1中の最大の音の数となる小節を基準として、他の小節の音の比率を演算し、比率が大きく変化する小節間位置として設定される。展開ポイントは、音数解析部12を利用して、コンピュータ4内で設定することもできるが、本実施形態のように、予め音の数を解析して展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4が設定された楽曲データM1を利用してもよい。
 また、展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4の設定は、前述の方法に限られず、たとえば、楽曲データM1内のフレーズの類似性に基づいて、行ってもよい。
The position information acquisition unit 11 acquires the number of measures of the development point set in the music data M1 as the position information of the development point. Specifically, as shown in FIG. 3, the position information acquisition unit 11 has development points P1, P2,..., Pn, Pn + 3,... Set between bars in the music data M1. The position information of Pe-4, in this embodiment, the number of bars is acquired.
The expansion point is the frequency analysis of the number of sounds with different frequencies for each measure by FFT, etc., the number of peaks of the sound pressure level is counted, and the other is based on the measure that is the maximum number of sounds in the music data M1. The ratio of the sound of the bars is calculated and set as the inter-bar position where the ratio changes greatly. The expansion points can be set in the computer 4 using the sound number analysis unit 12, but the expansion points P1, P2,..., Pn, by analyzing the number of sounds in advance as in this embodiment. Music data M1 in which Pn + 3,... Pe-4 is set may be used.
Further, the setting of the development points P1, P2,..., Pn, Pn + 3,... Pe-4 is not limited to the method described above, and may be performed based on, for example, the similarity of phrases in the music data M1. .
 音数解析部12は、入力される楽曲データM1について、それぞれの展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間毎に、各周波数帯域の信号レベルを検出し、音の数を解析する。ここで、”音の数”とは、周波数の異なる音を異なる音として数えてもよく、基音、倍音等を同じ音階として1つに数える場合であってもよい。なお、入力される楽曲データM1は、コンピュータ4内のハードディスクに保存された楽曲データであってもよく、デジタルプレーヤー2のスロットに挿入されたCD、ブルーレイディスク等に記録された楽曲データであってもよく、通信回線を介してネットワーク上からダウンロードされ得る楽曲データであってもよい。 The sound number analysis unit 12 detects the signal level of each frequency band for each segment between the development points P1, P2,..., Pn, Pn + 3,. Analyze the number of sounds. Here, the “number of sounds” may be a case where sounds having different frequencies may be counted as different sounds, or a fundamental tone, overtones, etc. may be counted as one with the same scale. Note that the input music data M1 may be music data stored on a hard disk in the computer 4, or music data recorded on a CD, a Blu-ray disc or the like inserted in a slot of the digital player 2. Alternatively, it may be music data that can be downloaded from a network via a communication line.
 具体的には、音数解析部12は、たとえば、図4に示されるように、楽曲データM1の展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間毎に、周波数のピーク本数を数えることにより、音の数をカウントする。なお、本実施形態ではFFTを用いて周波数の異なる音の解析をおこなっているが、本発明はこれに限られず、例えば、離散コサイン変換、離散フーリエ変換を用いて周波数変換を行ってもよい。
 音数解析部12は、解析結果を比率演算部14に出力する。
Specifically, the sound number analysis unit 12, for example, as shown in FIG. 4, for each section between the development points P1, P2,..., Pn, Pn + 3,. Count the number of sounds by counting the number of peaks in the frequency. In the present embodiment, the analysis of sounds having different frequencies is performed using FFT, but the present invention is not limited to this, and for example, frequency conversion may be performed using discrete cosine transform or discrete Fourier transform.
The sound number analysis unit 12 outputs the analysis result to the ratio calculation unit 14.
 低音レベル解析部13は、低域の信号レベルとなる低域音圧ピークレベルの小節区間毎の平均値を加味して、展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間に、サビ区間を割り付けるために設けられている。
 低音レベル解析部13は、楽曲データM1の展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間の所定の周波数よりも低域の周波数、たとえば、100Hz以下の低音の低域音圧ピークレベルを解析する。具体的には、低音レベル解析部13は、バスドラム、ベース等の低域音圧レベルを取得して、小節区間毎の区間内の低域音圧ピークレベルの平均値を、展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間の信号レベルとする。
 低音レベル解析部13は、解析結果を特徴区間割付部15に出力する。
The bass level analysis unit 13 takes into account the average value for each bar section of the low-frequency sound pressure peak level, which is the low-frequency signal level, and develops points P1, P2, ..., Pn, Pn + 3, ... Pe-4. It is provided to allocate a chorus section in the section between.
The low sound level analysis unit 13 has a frequency lower than a predetermined frequency in a section between the development points P1, P2,..., Pn, Pn + 3,. Analyzes the low sound pressure peak level. Specifically, the bass level analysis unit 13 acquires bass sound levels such as bass drums and bass, and calculates the average value of the bass sound pressure peak levels in the sections for each measure section as the development point P1, .., Pn, Pn + 3,... Pe-4.
The bass level analysis unit 13 outputs the analysis result to the feature section allocation unit 15.
 比率演算部14は、音数解析部12の解析結果に基づいて、最大の音の数となる展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間を基準として、他の区間の音の数の比率を演算する。具体的には、比率演算部14は、図4に示されるように、音の数の最も大きなFFT2を与えるPn番目の区間を特定し、n番目の小節目における音の数Nmaxを取得し、下記式(1)に基づいて、他の小節の比率Rnを演算する。
 Rn=Nn/Nmax・・・(1)
 他の区間の比率は、小節を表す矩形部分内の数字のように、最大となるn番目の展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間に対する比率として演算される。比率演算部14は、演算結果を特徴区間割付部15に出力する。
Based on the analysis result of the sound number analysis unit 12, the ratio calculation unit 14 uses the interval between the expansion points P1, P2,..., Pn, Pn + 3,. Calculate the ratio of the number of sounds in other sections. Specifically, as shown in FIG. 4, the ratio calculation unit 14 specifies the Pn-th section that gives the FFT 2 having the largest number of sounds, acquires the number Nmax of sounds in the n-th bar, Based on the following formula (1), the ratio Rn of other measures is calculated.
Rn = Nn / Nmax (1)
The ratio of the other sections is calculated as the ratio to the section between the nth expansion points P1, P2,..., Pn, Pn + 3,. Is done. The ratio calculation unit 14 outputs the calculation result to the feature interval allocation unit 15.
 特徴区間割付部15は、比率演算部14の演算結果、および低音レベル解析部13の解析結果に基づいて、展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間にイントロ(Intro)区間、Aメロ(Verse1)区間、Bメロ(Verse2)区間、サビ(Hook)区間、Cメロ(Verse3)区間、アウトロ(Outro)区間等の特徴区間を割り付ける。
 具体的には、特徴区間割付部15は、図5のグラフG1に示されるように、音数解析部12の解析結果に基づいて、前後の区間よりも大きな極大値1、極大値2を探索し、極大値1、極大値2をサビ区間(Hook)の候補とする。
 特徴区間割付部15は、図5のグラフG2に示されるように、低音レベル解析部13の解析結果に基づいて、各特徴区間の低域音圧ピークレベルの平均値を取得し、極大値1、極大値2をとる区間における低域音圧ピークレベルの平均値が、所定の閾値よりも上であるか否かを判定し、サビ(Hook)区間を割り付ける。
Based on the calculation result of the ratio calculation unit 14 and the analysis result of the bass level analysis unit 13, the feature interval allocation unit 15 sets the interval between the expansion points P 1, P 2,..., Pn, Pn + 3,. Characteristic sections such as an intro section, an A melody (Verse1) section, a B melody (Verse2) section, a chorus (Hook) section, a C melody (Verse3) section, and an outro section are allocated.
Specifically, as shown in the graph G1 of FIG. 5, the feature section allocating unit 15 searches for a maximum value 1 and a maximum value 2 that are larger than the preceding and following sections based on the analysis result of the sound number analysis unit 12. Then, maximal value 1 and maximal value 2 are set as candidates for the rust section (Hook).
As shown in the graph G2 of FIG. 5, the feature section allocating unit 15 obtains the average value of the low frequency sound pressure peak levels of each feature section based on the analysis result of the bass level analyzing unit 13, and the maximum value 1 Then, it is determined whether or not the average value of the low-frequency sound pressure peak level in the section having the maximum value 2 is higher than a predetermined threshold value, and the hook section is allocated.
 次に、特徴区間割付部15は、図5のグラフG3に示されるように、極大値1、極大値2の前の区間をAメロ区間(Verse1)、後ろの区間をBメロ区間(Verse2)、さらにその後ろの区間をCメロ区間(Verse3)等に割り付ける。どちらの特徴区間を割り付けるかは、音の数が、所定の閾値を超えるか否かで決定する。所定の閾値は、極大値よりも小さな固定された閾値としてもよく、極大値に対する所定の割合として設定された、極大値に応じて変動する閾値としてもよい。
 割り付ける特徴区間の名称の付け方は任意であり、図5において、A-Verse、B-Verse等の名称を付与してもよい。また、イントロ区間は楽曲データM1の最初の区間であり、アウトロ区間は楽曲データM1の最後の区間であることは明らかであるから、特徴区間割付部15は、予めイントロ区間、アウトロ区間の割り付けを行っておく。
Next, as shown in the graph G3 of FIG. 5, the feature interval allocation unit 15 sets the interval before the maximum value 1 and the maximum value 2 as the A melody interval (Verse1), and the subsequent interval as the B melody interval (Verse2). Further, the subsequent section is allocated to the C melody section (Verse 3) or the like. Which feature section is assigned is determined by whether or not the number of sounds exceeds a predetermined threshold. The predetermined threshold value may be a fixed threshold value that is smaller than the maximum value, or may be a threshold value that is set as a predetermined ratio with respect to the maximum value and varies according to the maximum value.
The name of the feature section to be assigned is arbitrary. In FIG. 5, names such as A-Verse and B-Verse may be assigned. In addition, since the intro section is the first section of the music data M1 and the outro section is the last section of the music data M1, the feature section allocating unit 15 assigns intro sections and outro sections in advance. Keep going.
 表示情報生成部16は、特徴区間割付部15で割り付けられた特徴区間を楽曲データM1とともに、表示情報として生成する。具体的には、図6に示されるように、楽曲データM1の進行とともに、特徴区間を表示し、楽曲データM1の進行とともに、特徴区間の色を変更させていく表示情報を生成する。
 表示情報生成部16により生成された表示情報は、デジタルプレーヤー2の表示装置となるディスプレイ2Bに出力され、DJパフォーマーは、楽曲データM1の楽曲の演奏進行とともに、どこの特徴区間を演奏しているかを確認することができる。
The display information generating unit 16 generates the feature section assigned by the feature section assigning unit 15 as display information together with the music data M1. Specifically, as shown in FIG. 6, the characteristic section is displayed as the music data M1 progresses, and the display information for changing the color of the characteristic section as the music data M1 progresses is generated.
The display information generated by the display information generation unit 16 is output to the display 2B serving as a display device of the digital player 2, and the DJ performer is performing which characteristic section is being played as the music performance of the music data M1 progresses. Can be confirmed.
 [3]実施形態の作用および効果
 次に、本実施形態の作用となる楽曲構造解析方法について、図7に示されるフローチャートに基づいて説明する。
 位置情報取得部11は、楽曲データM1中の展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4の位置情報を取得する(手順S1)。
 次に、音数解析部12は、展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間の音の数の解析を行う(手順S2)。
 比率演算部14は、音数解析部12の解析結果に基づいて、音の数の最大値をとる展開ポイントPn、Pn+3間の区間を基準として、他の展開ポイントP1、P2、…Pe-4間の区間の比率を演算する(手順S3)。
[3] Operation and Effect of Embodiment Next, a music structure analysis method which is the operation of the present embodiment will be described based on the flowchart shown in FIG.
The position information acquisition unit 11 acquires position information of the development points P1, P2,..., Pn, Pn + 3,... Pe-4 in the music data M1 (step S1).
Next, the sound number analysis unit 12 analyzes the number of sounds in the section between the development points P1, P2,..., Pn, Pn + 3,... Pe-4 (step S2).
Based on the analysis result of the sound number analysis unit 12, the ratio calculation unit 14 uses the interval between the expansion points Pn and Pn + 3 having the maximum number of sounds as a reference to other expansion points P1, P2,. -4 is calculated (step S3).
 特徴区間割付部15は、楽曲データM1の開始時点から最初の展開ポイントP1までの区間に、イントロ区間を割り付ける(手順S4)。
 続けて、特徴区間割付部15は、楽曲データM1の最後の展開ポイントPe-4から楽曲-多M1の終了時点に、アウトロ区間を割り付ける(手順S5)。
 特徴区間割付部15は、イントロ区間およびアウトロ区間以外の展開ポイント間の区間について、極大値を探索する(手順S6)。探索は、イントロ区間の次の区間から始めてもよく、アウトロ区間の前の区間から始めてもよい。
 特徴区間割付部15は、極大値をとる区間が探索されたら、展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間の低域音圧ピークレベルの平均値を取得する(手順S7)。
The feature section assigning unit 15 assigns the intro section to the section from the start point of the music data M1 to the first development point P1 (step S4).
Subsequently, the feature section assigning unit 15 assigns an outro section from the last development point Pe-4 of the music data M1 to the end point of the music-multiple M1 (step S5).
The feature section allocating unit 15 searches for the maximum value in the section between the development points other than the intro section and the outro section (step S6). The search may start from the section next to the intro section or may start from the section before the outro section.
When the section having the maximum value is found, the feature section allocating unit 15 obtains the average value of the low-frequency sound pressure peak levels in the section between the expansion points P1, P2,..., Pn, Pn + 3,. (Procedure S7).
 特徴区間割付部15は、極大値をとる区間の低域音圧ピークレベルの平均値が所定の閾値を超えるか否かを判定する(手順S8)。
 所定の閾値以下である場合、特徴区間割付部15は、次の極大値の探索を行う。
 所定の閾値を超える場合、特徴区間割付部15は、当該区間にサビ(Hook)区間を割り付ける(手順S9)。
 特徴区間割付部15は、楽曲データM1中のすべての極大値をとる区間について、手順S6から手順S9を繰り返す(手順S10)。
 なお、手順S8および手順S9は、サビ区間の検出精度を向上させるために行っており、音の数の極大値をとる展開ポイント間の区間の極大値の探索のみで、展開ポイント間の区間に、サビ区間を割り付けてもよい。
The feature section allocating unit 15 determines whether or not the average value of the low frequency sound pressure peak level in the section having the maximum value exceeds a predetermined threshold (step S8).
When the value is equal to or less than the predetermined threshold, the feature section allocating unit 15 searches for the next maximum value.
When the predetermined threshold value is exceeded, the feature section allocating unit 15 allocates a hook section to the section (step S9).
The feature section allocating unit 15 repeats steps S6 to S9 for all sections having the maximum value in the music data M1 (step S10).
Steps S8 and S9 are performed in order to improve the detection accuracy of the chorus section, and only the search for the maximum value of the section between the expansion points that takes the maximum value of the number of sounds, A rust section may be allocated.
 すべてのサビ区間の割り付けが終了したら、特徴区間割付部15は、サビ区間に設定された区間の前後の他の展開ポイント間の区間音の数を取得する(手順S11)。
 特徴区間割付部15は、他の展開ポイント間の区間の音の数が所定の閾値を超えているか否かを判定する(手順S12)。
When all the chorus sections have been assigned, the feature section assigning unit 15 acquires the number of section sounds between other development points before and after the section set as the chorus section (step S11).
The feature section allocating unit 15 determines whether or not the number of sounds in the section between other development points exceeds a predetermined threshold (step S12).
 音の数の比率が所定の閾値を超えている場合、特徴区間割付部15は、その区間にAメロ(Verse1)区間を割り付け(手順S13)、所定の閾値以下である場合、特徴区間割付部15は、その区間にBメロ(Verse2)区間を割り付ける(手順S14)。
 特徴区間割付部15は、すべての展開ポイントP1、P2、…、Pn、Pn+3、…Pe-4間の区間への特徴区間の割り付けが終わるまで繰り返す(手順S15)。
 すべての区間に特徴区間が割り付けられたら、特徴区間割付部15は、割り付けた結果を表示情報生成部16に出力し、表示情報生成部16は、割り付け結果に基づいて、表示情報を生成し、生成した表示情報をデジタルプレーヤー2のディスプレイ2Bに出力する(手順S16)。
When the ratio of the number of sounds exceeds a predetermined threshold, the feature section allocating unit 15 allocates an A melody (Verse1) section to the section (step S13), and when the ratio is equal to or less than the predetermined threshold, the feature section allocating unit. 15 assigns a B melody (Verse2) section to the section (step S14).
The feature section assigning unit 15 repeats until the assignment of the feature section to the section between all the development points P1, P2,..., Pn, Pn + 3,.
When feature sections are assigned to all sections, the feature section assigning unit 15 outputs the assigned result to the display information generating unit 16, and the display information generating unit 16 generates display information based on the assignment result, The generated display information is output to the display 2B of the digital player 2 (step S16).
 このような本実施形態によれば、音数解析部12による音の数の解析を行うだけで、全ての特徴区間の割り付けを行うことができるため、簡単にかつ迅速に楽曲データM1に対して特徴区間の割り付けを行うことができる。
 また、表示情報生成部16からディスプレイ2Bに表示情報を出力することにより、DJパフォーマンスを行うユーザーは、現在どの特徴区間が演奏されているかを視認することができるため、より高度なDJパフォーマンスを行うことができる。
According to the present embodiment, since all the feature sections can be allocated simply by analyzing the number of sounds by the sound number analysis unit 12, the music data M1 can be easily and quickly allocated. It is possible to assign feature sections.
In addition, by outputting display information from the display information generation unit 16 to the display 2B, a user performing DJ performance can visually recognize which feature section is currently being played, and therefore performs a higher level DJ performance. be able to.
 1…音響制御システム、2…デジタルプレーヤー、2A…ジョグダイヤル、2B…ディスプレイ、3…デジタルミキサー、3A…操作スイッチ、3B…音量調整レバー、3C…左右切替レバー、4…コンピュータ、5…スピーカー、6…ケーブル、7…USBケーブル、8…アナログケーブル、9…ケーブル、10…演算処理装置、11…位置情報取得部、12…音数解析部、13…低音レベル解析部、14…比率演算部、15…特徴区間割付部、16…表示情報生成部、G1…グラフ、G2…グラフ、G3…グラフ、M1…楽曲データ、P1、P2、Pn、Pn+3、Pe-4…展開ポイント。
 
DESCRIPTION OF SYMBOLS 1 ... Sound control system, 2 ... Digital player, 2A ... Jog dial, 2B ... Display, 3 ... Digital mixer, 3A ... Operation switch, 3B ... Volume control lever, 3C ... Left / right switching lever, 4 ... Computer, 5 ... Speaker, 6 ... Cable, 7 ... USB cable, 8 ... Analog cable, 9 ... Cable, 10 ... Arithmetic processing device, 11 ... Position information acquisition unit, 12 ... Sound number analysis unit, 13 ... Bass level analysis unit, 14 ... Ratio calculation unit, 15 ... Characteristic section allocation unit, 16 ... Display information generation unit, G1 ... Graph, G2 ... Graph, G3 ... Graph, M1 ... Music data, P1, P2, Pn, Pn + 3, Pe-4 ... Expansion point.

Claims (9)

  1.  楽曲データの構造を特徴づける特徴区間の展開ポイントが設定された楽曲データに、前記特徴区間を割り付ける楽曲構造解析装置であって、
     前記展開ポイントの位置情報を取得する位置情報取得部と、
     前記位置情報取得部によって取得された展開ポイントの位置情報に基づいて、それぞれの展開ポイント間の区間毎に、周波数の異なる音の数を解析する音数解析部と、
     前記音数解析部により解析された、音の数の極大値をとる展開ポイント間の区間に基づいて、他の展開ポイント間の区間に、前記特徴区間を割り付ける特徴区間割付部と、
    を備えていることを特徴とする楽曲構造解析装置。
    A music structure analysis apparatus that assigns the feature section to the music data set with the development points of the feature section characterizing the structure of the music data,
    A position information acquisition unit for acquiring position information of the development point;
    Based on the position information of the expansion points acquired by the position information acquisition unit, a sound number analysis unit that analyzes the number of sounds with different frequencies for each section between the expansion points;
    Based on the section between the development points that takes the maximum value of the number of sounds analyzed by the sound number analysis section, a feature section allocating section that assigns the feature section to a section between other development points;
    A music structure analyzing apparatus comprising:
  2.  請求項1に記載の楽曲構造解析装置において、
     前記位置情報取得部で取得された展開ポイントの位置情報に基づいて、展開ポイント間の区間毎に、所定の周波数よりも低域の周波数の音の信号レベルを解析する低音レベル解析部を備え、
     前記特徴区間割付部は、前記低音レベル解析部により解析された低域の信号レベルの極大値をとる展開ポイント間の区間も加味して、前記特徴区間を割り付けることを特徴とする楽曲構造解析装置。
    In the music structure analysis apparatus according to claim 1,
    Based on the position information of the development points acquired by the position information acquisition unit, for each section between the development points, comprising a bass level analysis unit for analyzing the signal level of the sound of the frequency lower than the predetermined frequency,
    The feature section allocating unit assigns the feature section in consideration of a section between development points taking a maximum value of a low-frequency signal level analyzed by the bass level analyzing section. .
  3.  請求項1または請求項2に記載の楽曲構造解析装置において、
     前記特徴区間割付部は、音の数の極大値をとる小節を含む展開ポイント間に、サビ(Hook)区間を割り付けることを特徴とする楽曲構造解析装置。
    In the music structure analysis apparatus according to claim 1 or 2,
    The musical piece structure analyzing apparatus characterized in that the characteristic section allocating section allocates a hook section between development points including a measure having a maximum value of the number of sounds.
  4.  請求項3に記載の楽曲構造解析装置において、
     前記特徴区間割付部は、他の展開ポイント間の区間の音の数が、固定された閾値を超えるか否かに基づいて、前記他の展開ポイント間の区間に、特徴区間の割り付けを行うことを特徴とする楽曲構造解析装置。
    In the music structure analysis apparatus according to claim 3,
    The feature section allocation unit assigns feature sections to sections between the other development points based on whether the number of sounds in the section between other development points exceeds a fixed threshold value. A music structure analysis apparatus characterized by
  5.  請求項3に記載の楽曲構造解析装置において、
     前記特徴区間割付部は、他の展開ポイント間の区間の音の数が、前記極大値に応じて変動する閾値を超えるか否かに基づいて、前記他の展開ポイント間の区間に、特徴区間の割り付けを行うことを特徴とする楽曲構造解析装置。
    In the music structure analysis apparatus according to claim 3,
    The feature section allocating unit determines whether or not the number of sounds in the section between the other development points exceeds the threshold that varies according to the maximum value, in the section between the other development points. The music structure analysis apparatus characterized by performing the allocation of music.
  6.  請求項1から請求項5のいずれか一項に記載の楽曲構造解析装置において、
     前記音数解析部は、周波数変換を行い、各周波数帯域の信号レベルにより音の数を解析することを特徴とする楽曲構造解析装置。
    In the music structure analysis apparatus according to any one of claims 1 to 5,
    The music number analysis unit performs frequency conversion and analyzes the number of sounds according to the signal level of each frequency band.
  7.  請求項1から請求項6のいずれか一項に記載の楽曲構造解析装置において、
     前記特徴区間割付部によって割り付けられた特徴区間を、楽曲データとともに表示装置上に表示情報を生成する表示情報生成部を備えていることを特徴とする楽曲構造解析装置。
    In the music structure analysis apparatus according to any one of claims 1 to 6,
    A music structure analyzing apparatus, comprising: a display information generating unit configured to generate display information on the display device together with music data on the characteristic sections allocated by the characteristic section allocating unit.
  8.  楽曲データの構造を特徴づける特徴区間の展開ポイントが設定された楽曲データに、特徴区間を割り付ける楽曲構造解析方法であって、
     前記展開ポイントの位置情報を取得する手順と、
     取得された展開ポイントの位置情報に基づいて、それぞれの展開ポイント間の区間毎に、周波数の異なる音の数を解析する手順と、
     解析された音の数の極大値をとる展開ポイント間の区間に基づいて、他の展開ポイント区間に、特徴区間を割り付ける手順と、
    を実施することを特徴とする楽曲構造解析方法。
    A music structure analysis method for assigning a characteristic section to music data in which a development point of a characteristic section characterizing the structure of the music data is set,
    Obtaining position information of the development point;
    A procedure for analyzing the number of sounds having different frequencies for each section between the development points based on the acquired location information of the development points;
    A procedure for assigning a feature section to another development point section based on a section between development points taking the maximum value of the analyzed number of sounds;
    The music structure analysis method characterized by implementing.
  9.  コンピュータを、請求項1から請求項7のいずれか一項に記載の楽曲構造解析装置として機能させることを特徴とする楽曲構造解析プログラム。
     
    A music structure analysis program for causing a computer to function as the music structure analysis apparatus according to any one of claims 1 to 7.
PCT/JP2016/063981 2016-05-11 2016-05-11 Music structure analysis device, method for analyzing music structure, and music structure analysis program WO2017195292A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2018516262A JPWO2017195292A1 (en) 2016-05-11 2016-05-11 Music structure analysis apparatus, music structure analysis method, and music structure analysis program
EP16901640.9A EP3457395A4 (en) 2016-05-11 2016-05-11 Music structure analysis device, method for analyzing music structure, and music structure analysis program
PCT/JP2016/063981 WO2017195292A1 (en) 2016-05-11 2016-05-11 Music structure analysis device, method for analyzing music structure, and music structure analysis program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2016/063981 WO2017195292A1 (en) 2016-05-11 2016-05-11 Music structure analysis device, method for analyzing music structure, and music structure analysis program

Publications (1)

Publication Number Publication Date
WO2017195292A1 true WO2017195292A1 (en) 2017-11-16

Family

ID=60266426

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2016/063981 WO2017195292A1 (en) 2016-05-11 2016-05-11 Music structure analysis device, method for analyzing music structure, and music structure analysis program

Country Status (3)

Country Link
EP (1) EP3457395A4 (en)
JP (1) JPWO2017195292A1 (en)
WO (1) WO2017195292A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022070396A1 (en) * 2020-10-02 2022-04-07 AlphaTheta株式会社 Music analysis device, music analysis method, and program
WO2023054237A1 (en) * 2021-09-30 2023-04-06 パイオニア株式会社 Sound effect output device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4775380B2 (en) 2004-09-28 2011-09-21 ソニー株式会社 Apparatus and method for grouping time segments of music
JP2012194387A (en) * 2011-03-16 2012-10-11 Yamaha Corp Intonation determination device

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4926737A (en) * 1987-04-08 1990-05-22 Casio Computer Co., Ltd. Automatic composer using input motif information
EP2088518A1 (en) * 2007-12-17 2009-08-12 Sony Corporation Method for music structure analysis

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4775380B2 (en) 2004-09-28 2011-09-21 ソニー株式会社 Apparatus and method for grouping time segments of music
JP2012194387A (en) * 2011-03-16 2012-10-11 Yamaha Corp Intonation determination device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3457395A4 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022070396A1 (en) * 2020-10-02 2022-04-07 AlphaTheta株式会社 Music analysis device, music analysis method, and program
WO2023054237A1 (en) * 2021-09-30 2023-04-06 パイオニア株式会社 Sound effect output device

Also Published As

Publication number Publication date
EP3457395A4 (en) 2019-10-30
EP3457395A1 (en) 2019-03-20
JPWO2017195292A1 (en) 2019-03-07

Similar Documents

Publication Publication Date Title
US11915725B2 (en) Post-processing of audio recordings
US9117429B2 (en) Input interface for generating control signals by acoustic gestures
JP4650662B2 (en) Signal processing apparatus, signal processing method, program, and recording medium
JP4645241B2 (en) Voice processing apparatus and program
KR20090130833A (en) System and method for automatically producing haptic events from a digital audio file
WO2017057530A1 (en) Audio processing device and audio processing method
JPH04195196A (en) Midi chord forming device
WO2017195292A1 (en) Music structure analysis device, method for analyzing music structure, and music structure analysis program
US9087503B2 (en) Sampling device and sampling method
JP6281211B2 (en) Acoustic signal alignment apparatus, alignment method, and computer program
CN108369800B (en) Sound processing device
JP2015018112A (en) Musical sound generating apparatus, musical sound generating method, and program
WO2017135350A1 (en) Recording medium, acoustic processing device, and acoustic processing method
JP6625202B2 (en) Music structure analysis device, music structure analysis method, and music structure analysis program
Stöter et al. Unison Source Separation.
CN114724583A (en) Music fragment positioning method, device, equipment and storage medium
US10424279B2 (en) Performance apparatus, performance method, recording medium, and electronic musical instrument
JP2015087436A (en) Voice sound processing device, control method and program for voice sound processing device
JP6357772B2 (en) Electronic musical instrument, program and pronunciation pitch selection method
WO2024034115A1 (en) Audio signal processing device, audio signal processing method, and program
JP7176114B2 (en) MUSIC ANALYSIS DEVICE, PROGRAM AND MUSIC ANALYSIS METHOD
JP5151603B2 (en) Electronic musical instruments
WO2024034116A1 (en) Audio data processing device, audio data processing method, and program
JP5375869B2 (en) Music playback device, music playback method and program
JP4238807B2 (en) Sound source waveform data determination device

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2018516262

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16901640

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2016901640

Country of ref document: EP

Effective date: 20181211