CN103761965B - A kind of sorting technique of instrument signal - Google Patents

A kind of sorting technique of instrument signal Download PDF

Info

Publication number
CN103761965B
CN103761965B CN201410008533.4A CN201410008533A CN103761965B CN 103761965 B CN103761965 B CN 103761965B CN 201410008533 A CN201410008533 A CN 201410008533A CN 103761965 B CN103761965 B CN 103761965B
Authority
CN
China
Prior art keywords
tau
matrix
phase space
musical instrument
dimension
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201410008533.4A
Other languages
Chinese (zh)
Other versions
CN103761965A (en
Inventor
郭一娜
王志社
郅逍遥
王晓梅
李临生
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Taiyuan University of Science and Technology
Original Assignee
Taiyuan University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Taiyuan University of Science and Technology filed Critical Taiyuan University of Science and Technology
Priority to CN201410008533.4A priority Critical patent/CN103761965B/en
Publication of CN103761965A publication Critical patent/CN103761965A/en
Application granted granted Critical
Publication of CN103761965B publication Critical patent/CN103761965B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Complex Calculations (AREA)

Abstract

A sorting technique for instrument signal, belongs to electronic information technical field, comprises phase space reconfiguration module, principal component analysis module, characteristic extracting module and flexible Neural Tree module. Be characterised in that implementation step is that the time series that different musical instrument sample signals are produced is carried out phase space reconfiguration, remove redundancy by principal component analysis, reach dimensionality reduction object, again by analyzing various musical instrument characteristics, adopt probability density function to portray the difference of each musical instrument in phase space, finally, adopt flexible Neural Tree model to classify as grader, it can effectively solve artificial neural network structure's high dependency problem, can make the classification accuracy of single musical instrument be up to 98.7%.

Description

Classification method of musical instrument signals
Technical Field
The invention belongs to the technical field of electronic information, and particularly relates to a method for classifying musical instrument signals.
Background
Music is an indispensable part in life of people, however, the mode of knowing music information can only be used for acquiring interesting contents from the music information through real-time audition. With the development of science and technology, the information that people can contact is rapidly increasing, so that it is not easy to obtain interesting music in mass data through real-time audition. In recent years, music data analysis and retrieval have become a research hotspot at home and abroad.
Most of music is expressed by depending on the performance of musical instruments, and the same song is performed by different musical instruments, so that different enjoyment is brought to the hearing of people. In the field of computer identification, with the continuous development of signal analysis and data mining technologies, various intercrossed technologies such as time-frequency analysis and neural networks are gradually applied to musical instrument classification.
Efficient feature extraction and classifier selection have been key steps in instrument classification. Mel cepstral coefficient features are commonly applied in the field of speech signals, but also magnify the heteroscedasticity in the classification of music genres and instrument signals. MPEG-7, as a standardized framework for the purpose of audio feature extraction and description, is indispensable in feature extraction of music signals, and is particularly more effective for similar music signals. In addition to numerous audio features, a wide range of classifiers have been extensively studied in audio classification, such as neural networks, gaussian mixture models, hidden markov models, bayesian decision classifiers, and support vector machines. Some researchers use a single-channel voice separation method to separate two or more musical instruments possibly through Gaussian modeling, others use amplitude modulation-frequency modulation characteristics to analyze and identify musical instrument signals (the accuracy rate can reach 70%), others use Mel cepstrum coefficient characteristics, constant q transformation and autocorrelation functions to extract characteristics, and a Bayesian decision classifier is combined to enable the identification accuracy rate to reach 79% -84%. However, in the existing research results, the classification accuracy of some audio signals is generally low, which indicates that the above audio characteristics do not well delineate the differences between all the signals, and the parameter selection and operation processes of most classifiers have the disadvantages of complexity and complexity.
Disclosure of Invention
The invention aims to provide a method for classifying musical instrument signals, which can effectively overcome the defects in the prior art.
The invention is realized in such a way that various musical instrument signals are subjected to feature extraction by adopting a phase space reconstruction module, a principal component analysis module, a feature extraction module and a flexible neural tree module, and then are classified by a classifier, and the method is characterized by comprising the following implementation steps of:
the method comprises the following steps of firstly, carrying out phase space reconstruction on collected instrument sample signals, mainly determining the delay time and the embedding dimension of the sample signals, and comprising the following steps:
let a one-dimensional time sequence of a certain instrument signal sample be x ═ x (x)1,x2,...,xK)TSetting the delay time as tau, establishing an m-dimensional embedding space Y, mapping x into the embedding space, and then reconstructing a vector of a phase space as:
Y = ( y 1 , y 2 , ... , y n ) = x ( 1 ) x ( 2 ) ... x ( n ) x ( 1 + τ ) x ( 2 + τ ) ... x ( n + τ ) . . . . . . . . . x ( 1 + ( m - 1 ) τ ) x ( 2 + ( m - 1 ) τ ) ... ... x ( n + ( m - 1 ) τ ) - - - ( 1 )
wherein K is the length of the audio time series x; the number N of phase vectors in phase space is K- (m-1) τ (N is 1, 2.., N),
(1) determining the optimal delay time tau by the average mutual information method
And (3) selecting an average mutual information method to determine the optimal delay time tau:
I ( τ ) = Σ n = 1 N P ( x n , x n + τ ) log 2 P ( x n , x n + τ ) P ( x n ) P ( x n + τ ) - - - ( 2 )
wherein P (x)n,xn+τ) To reconstruct the joint distribution probability in the graph, P (x)n),P(xn+τ) Selecting tau of a first local minimum value of I (tau) as an optimal delay time for the edge distribution probability;
(2) determining embedding dimension m by false neighbor method
For arbitrary phasor y in m-dimensional phasor spacen(i)={x(i),x(i+τ),...,x[i+(m-1)τ]All have a nearest neighbor y within a certain distancen NN(i) And the distance between the two points is:
Dm(i)=||yn(i)-yn NN(i)||(3)
Dm(i) is a small quantity, the distance between these two phase points changes when the embedding dimension of the phase space increases from the m dimension to the m +1 dimension, in which case the adjacent distance is Dm+1(i) Then, there are:
D2 m+1(i)=D2 m(i)+||x(i+τm)-xNN(i+τm)||(4)
judging which points are false neighbors by giving a certain threshold value, and enabling
ΔD m ( i , m ) = D m + 1 2 ( i ) - D m 2 ( i ) D m 2 ( i ) = | x ( i + τ m ) - x N N ( i + τ m ) | D m ( i ) - - - ( 5 )
Wherein Δ Dm(i, m) represents the amount of change in the distance of the nearest neighbor point in m dimensions with respect to the time in m +1 dimensions,
if Δ Dm(i, m) is more than or equal to 10%, the adjacent points are false, whether the adjacent points of each phasor in the m-dimensional space are false is judged in sequence, the proportion of the false adjacent points in all phasors is calculated, when the proportion of the false adjacent points is less than 5%, the phasor space track is considered to be completely opened, and m at the moment is the optimal embedding dimension;
(3) phase space reconstruction
After determining the delay time τ and the embedding dimension m, the original one-dimensional time sequence x may be determined as (x)1,x2,...,xK)TDelayed reconstruction to obtain an embedded vector xm=(xt,xt-τ,...,xt-(m-1)τ)TThese vectors define a point in the m-dimensional embedding space Y, and the respective delay amounts at different time points produce a series of such points in the m-dimensional space, thereby forming a phase space reconstruction matrix a of order m × n reflecting the original system information;
and secondly, performing principal component analysis on the reconstructed musical instrument sample signal, and eliminating redundant information to achieve the effect of reducing the dimension:
R=E(AAT) Where A is m × n order matrix obtained after phase space reconstruction, R is autocorrelation matrix of m components, V is m × m order eigenvector matrix of R, column vector is orthogonal normalized eigenvector of R, Λ is characteristic diagonal matrix of R, and lambda is V Λi(i ═ 1, 2.. times, m) is the element on the ith diagonal; constructing m new independent variables Y ═ VTA,Y={y1,y2,...,ym}TTo λiAfter the (i ═ 1, 2.. multidot.m) is arranged in a descending order, the eigenvectors corresponding to the previous p larger eigenvalues are taken to obtain a vector matrix B of p × n order after dimension reduction, wherein p is more than or equal to 3;
thirdly, extracting the characteristics of the matrix B obtained after the principal component analysis processing, wherein the extracted characteristic values of the probability density function are as follows:
let us assume that the sample data region G is a tiny cube with x as the center and a side length of h, and first define a kernel function:
the probability density function is then found to be:
p ( x ) = 1 N Σ n = 1 N 1 h m ( x - x n h ) - - - ( 7 )
n is the number of sample points, N is 1, 2., N, m is the data dimension, k (u) is a defined kernel function, thereby obtaining an m × N-order matrix, and after sampling the matrix line by line, forming an m × c-order feature set H, wherein c is less than N;
and fourthly, training a neural network by adopting a flexible neural tree model as a classifier of the musical instrument recognition module:
the calculation flow adopted in the flexible neural tree model is as follows: firstly, setting an initial structure of a flexible neural tree, and then optimizing and training a fuzzy classification model according to an obtained feature set H and the set initial structure, wherein the model comprises two types of instructions: the flexible neuron instruction is used for connecting a non-leaf node of a tree structure with a subtree, the terminal instruction is each input characteristic, and a function instruction set and a terminal instruction set are represented as follows:
S=F∪T={+2,+3,...,+N}∪{z1,z2,...zn}(8)
whereini(2, 3.. An, N) represents a non-leaf node instruction havingiAn input is namelyiAn input variable; z is a radical of1,z2,...znRepresenting a leaf node instruction, the leaf node having no input variable;
adopting a Gaussian function as a flexible excitation function:
f ( j i , b i , z ) = e - ( z - j i b i ) 2 - - - ( 9 )
wherein,jiand biAre adjustable parameters that are randomly generated during the creation of the flexible neural tree. The fitness function is another problem that needs to be determined in advance in addition to the instruction set, and the root mean square error is selected as the fitness function.
The invention has the advantages and positive effects that: the invention adopts a new musical instrument classification method to classify musical instrument signals, and the method discloses the specific attribute of each musical instrument by carrying out phase space reconstruction on time sequences generated by different musical instruments; redundant information is removed through principal component analysis, and the purpose of reducing the dimension is achieved; by analyzing the characteristics of various instruments, the difference of the various instruments in a phase space is described by adopting a probability density function; and finally, a flexible neural tree model is used as a classifier for classification, so that the problem of high dependency of an artificial neural network structure is effectively solved. The method has the advantages that the parameter selection and the operation process are simple, the highest classification accuracy of a single musical instrument can reach 98.7% through experimental results, and the lowest root mean square error of the classifier can reach 0.045570.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention;
FIGS. 2 and 3 are time domain waveforms of signals used in the experiment, and FIG. 2 shows a sample signal x with a circle number1-1Fig. 3 is a sample signal x of a piano2-1Time domain waveform diagrams of (a);
FIGS. 4 and 5 are respectively a sample x of a circle number determined by mutual information1-1And piano sample x2-1The delay time τ of (d);
FIGS. 6 and 7 are respectively a sample x of the circle number determined by the pseudo-neighbor method1-1And piano sample x2-1The embedding dimension m;
FIGS. 8 and 9 are respectively a sample x of the circle number1-1And piano sample x2-1Reconstructing a picture in phase space in three-dimensional space;
FIGS. 10 and 11 are respectively a sample x of the circle number1-1And piano sample x2-1The eigenvalues of the probability density function of (1);
FIG. 12 is the final trained flexible neural tree model;
FIGS. 13 and 14 are respectively a sample x of the circle number1-1And piano sample x2-1The predicted output and the target output of the training samples and the test samples.
In the figure: amplitude, N the number of sample points, I (τ) the average mutual information content, τ the delay time, m the embedding dimension, Δ DmThe method comprises the steps of representing a false neighbor rate, PDF representing a probability density function value, time representing time, value of forecast and target output, Traindata being a training sample, Testdata representing a test sample, forecast representing a predicted value, and target representing a target value.
Detailed Description
Matlab7.0 and Visualstudio2008 are selected as software platforms, and the design of the scheme is realized through programming. The implementation process is to select two common instrument signals: round horn, piano as the subject of experiment. For each instrument signal, 10 sets of measured sample data were selected, the sample data length being 2000 points. The first 8 groups were used as training samples and the other 2 groups were used as test samples.
Taking one sample of two instrument signals as an example for explanation, the specific operation steps are as follows:
firstly, carrying out phase space reconstruction processing on a sample signal, then carrying out principal component analysis to reduce redundancy, extracting principal elements, then entering a characteristic extraction part, finally carrying out musical instrument identification, and training a neural network by adopting a flexible neural tree to achieve the identification purpose.
First, the specific processing steps of phase space reconstruction in this embodiment are as follows:
(1) determination of the optimum delay time tau
Selecting an average mutual information method to determine the optimal delay time tau, as follows:
I ( τ ) = Σ n = 1 N P ( x n , x n + τ ) log 2 P ( x n , x n + τ ) P ( x n ) P ( x n + τ ) - - - ( 10 )
the optimum delay time is chosen to be τ of the first local minimum of the average mutual information quantity I (τ). FIGS. 4 and 5 are respectively a sample x of the circle number determined by this method1-1And piano sample x2-1Is delayed by a delay time τ, τ1-1=3,τ2-1=6
(2) Determination of the optimal embedding dimension m
And (3) determining an embedding dimension m by selecting a false neighbor point method, and judging which points are false neighbors according to the following formula:
ΔD m ( i , m ) = D m + 1 2 ( i ) - D m 2 ( i ) D m 2 ( i ) = | | y n + τ m ( i + τ m ) - y n + τ m N N ( i + τ m ) | | D m ( i ) - - - ( 11 )
if Δ Dm(i, m) ≧ 10%, this indicates that the neighbor is spurious, where τ1-1=3,τ2-16. Separately calculate x1-1And x2-1When the proportion of the false neighbors is less than 5%, the phase space trajectory is considered to be completely opened, and m at the moment is the optimal embedding dimension. FIGS. 6 and 7 are respectively a sample x of the circle numbers determined by this method1-1And piano sample x2-1M, m of the embedding dimension1-1=m2-1=5。
(3) Phase space reconstruction
Determining x1-1And x2-1Is delayed by a time τ1-1=3,τ2-16 and embedding dimension m1-1=m2-1The embedded vector x is obtained from their respective time series delayed reconstruction5=(xt,xt-3,xt-6,xt-9,xt-12)TAnd x5=(xt,xt-6,xt-12,xt-18,xt-24)TThese areThe vectors each define a point in a respective 5-dimensional embedding space, and the respective delay amounts at different points in time produce a series of such points in the 5-dimensional space, thereby forming two 5-dimensional phase spaces reflecting the original system information. In order to visually see the reconstructed image, three dimensions are selected for imaging, and x is shown in fig. 8 and 9 respectively1-1And x2-1The image is reconstructed in phase space in three-dimensional space.
Second, the specific processing steps of principal component analysis in this embodiment are as follows:
R=E(AAT) And RV ═ V Λ where a is x1-1A 5 × 1991 matrix obtained after phase space reconstruction, wherein R is an autocorrelation matrix of 5-dimensional reconstruction variables, V is a 5 × 5 th order eigenvector matrix of R, column vectors of the matrix are orthogonal normalized eigenvectors of R, Λ is an eigendiagonal matrix of R and lambdai(i ═ 1, 2.., 5) is the element on the ith diagonal; constructing 5 new irrelevant variables Y ═ VTX,Y={y1,y2,...,y5}TFor Y ═ Y1,y2,...,y5}TAfter arranging according to descending order, the eigenvectors corresponding to the previous 3 larger eigenvalues are taken to obtain a reduced-dimension 3 × 1991 order vector matrix B1
By the same token, x is obtained2-1Vector matrix B after principal component analysis dimensionality reduction2
Step three, the specific processing steps of the feature extraction part in this embodiment are as follows:
a matrix B obtained by reducing the dimension after the principal component analysis1And B2Extracting the characteristic value of the probability density function:
p ( x ) = 1 N Σ n = 1 N 1 h m k ( x - x n h ) - - - ( 12 )
n1991, m 3, H0.2, and two feature sets H of 3 × 10 are formed by sampling1-1And H2-1. FIGS. 10 and 11 are respectively a sample x of the circle number1-1And piano sample x2-1A figure of eigenvalues of the probability density function of (2).
And obtaining the feature set H of the rest 9 groups of sample data of the round number and the piano according to the same method1-iAnd H2-i,i=2,...,10。
In the fourth step, in this embodiment, a flexible neural tree is used as a classifier to identify musical instruments, and is mainly implemented on a visual studio2008 software platform, and the specific steps are as follows:
circle number data H to be used as training sample1-iThe feature sets of (i ═ 1,. and 8) are placed in sequence in the file of text. Firstly, setting an initial structure of a flexible neural tree, and then optimizing and training a fuzzy classification model according to a feature set in a train. And then inputting the test sample signal into the trained flexible neural tree model for performance verification. Similarly, for piano data H2-i(i 1.., 8) were trained and tested.
Fig. 12 shows the final trained flexible neural tree model. Wherein the function instruction set is F { +2,+3,+4And the terminal instruction set is as follows: t is={x1,x2,x3,x4,x5,x6,x7}. Selecting root mean square error RMSE of 0.1 as a threshold value to judge whether the classification result is correct, and when the output RMSE value is larger than 0.1, identifying the RMSE as other musical instruments; otherwise, the identification is considered to be correct. The recognition rate of the i-th instrument is as follows:
R i = n N × 100 %
wherein i is a sample category, N is the number of samples identified as i-type instruments, and N is the total number of samples of i-type instruments. The effect of this embodiment is that the sample x of the circle number1-1And piano sample x2-1The recognition rates of (a) and (b) are 84.9% and 98.7%, respectively, the mean root mean square error is 0.088685 and 0.045570, respectively, and fig. 13 and 14 are the circle number sample x, respectively1-1And piano sample x2-1The predicted output of the test sample is substantially consistent with the target output, although partially deviated, but generally very close. Experiments have shown that the present invention is very effective for the classification of instrument signals.

Claims (1)

1. A method for classifying musical instrument signals comprises the steps of carrying out feature extraction on various musical instrument signals by adopting a phase space reconstruction module, a principal component analysis module, a feature extraction module and a flexible neural tree module, and then classifying the musical instrument signals by a classifier, and is characterized by comprising the following implementation steps of:
the method comprises the following steps of firstly, carrying out phase space reconstruction on collected instrument sample signals, mainly determining the delay time and the embedding dimension of the sample signals, and comprising the following steps:
let a one-dimensional time sequence of a certain instrument signal sample be x ═ x (x)1,x2,...,xK)TSetting the delay time as tau, establishing an m-dimensional embedding space Y, mapping x into the embedding space, and then reconstructing a vector of a phase space as:
Y = ( y 1 , y 2 , ... , y n ) = x ( 1 ) x ( 2 ) ... x ( n ) x ( 1 + τ ) x ( 2 + τ ) ... x ( n + τ ) . . . . . . . . . x ( 1 + ( m - 1 ) τ ) x ( 2 + ( m - 1 ) τ ) ... x ( n + ( m - 1 ) τ ) - - - ( 1 )
wherein K is the length of the audio time series x; the number N of phase vectors in phase space is K- (m-1) τ (N is 1, 2.., N),
(1) determining the optimal delay time tau by the average mutual information method
And (3) selecting an average mutual information method to determine the optimal delay time tau:
I ( τ ) = Σ n = 1 N P ( x n , x n + τ ) log 2 P ( x n , x n + τ ) P ( x n ) P ( x n + τ ) - - - ( 2 )
wherein P (x)n,xn+τ) To reconstruct the joint distribution probability in the graph, P (x)n),P(xn+τ) Selecting tau of a first local minimum value of I (tau) as an optimal delay time for the edge distribution probability;
(2) determining embedding dimension m by false neighbor method
For arbitrary phasor y in m-dimensional phasor spacen(i)={x(i),x(i+τ),...,x[i+(m-1)τ]All have a nearest neighbor y within a certain distancen NN(i) And the distance between the two points is:
D m ( i ) = | | y n ( i ) - y n N N ( i ) | | - - - ( 3 )
Dm(i) is a small quantity, the distance between these two phase points changes when the embedding dimension of the phase space increases from the m dimension to the m +1 dimension, in which case the adjacent distance is Dm+1(i) Then, there are:
D m + 1 2 ( i ) = D m 2 ( i ) + | | x ( i + τ m ) - x N N ( i + τ m ) | | - - - ( 4 )
judging which points are false neighbors by giving a certain threshold value, and enabling
ΔD m ( i , m ) = D m + 1 2 ( i ) - D m 2 ( i ) D m 2 ( i ) = | | y n + τ m ( i + τ m ) - y n + τ m N N ( i + τ m ) | | D m ( i ) - - - ( 5 )
Wherein Δ Dm(i, m) represents the amount of change in the distance of the nearest neighbor point in m dimensions with respect to the time in m +1 dimensions,
if Δ Dm(i, m) is more than or equal to 10%, the adjacent points are false, whether the adjacent points of each phasor in the m-dimensional space are false is judged in sequence, the proportion of the false adjacent points in all phasors is calculated, when the proportion of the false adjacent points is less than 5%, the phasor space track is considered to be completely opened, and m at the moment is the optimal embedding dimension;
(3) phase space reconstruction
After determining the delay time τ and the embedding dimension m, the original one-dimensional time sequence x may be determined as (x)1,x2,...,xK)TDelayed reconstruction to obtain an embedded vector xm=(xt,xt-τ,...,xt-(m-1)τ)TThese vectors define a point in the m-dimensional embedding space Y, and the respective delay amounts at different points in time produce a series of such points in the m-dimensional space, thereby constituting a signal reflecting the original system informationA phase space reconstruction matrix A of order m × n;
and secondly, performing principal component analysis on the reconstructed musical instrument sample signal, and eliminating redundant information to achieve the effect of reducing the dimension:
R=E(AAT) Where A is m × n order matrix obtained after phase space reconstruction, R is autocorrelation matrix of m components, V is m × m order eigenvector matrix of R, column vector is orthogonal normalized eigenvector of R, Λ is characteristic diagonal matrix of R, and lambda is V Λi(i ═ 1, 2.. times, m) is the element on the ith diagonal; constructing m new independent variables Y ═ VTA,Y={y1,y2,...,ym}TTo λiAfter the (i ═ 1, 2.. multidot.m) is arranged in a descending order, the eigenvectors corresponding to the previous p larger eigenvalues are taken to obtain a vector matrix B of p × n order after dimension reduction, wherein p is more than or equal to 3;
thirdly, extracting the characteristics of the matrix B obtained after the principal component analysis processing, wherein the extracted characteristic values of the probability density function are as follows:
let us assume that the sample data region G is a tiny cube with x as the center and a side length of h, and first define a kernel function:
the probability density function is then found to be:
p ( x ) = 1 N Σ n = 1 N 1 h m k ( x - x n h ) - - - ( 7 )
n is the number of sample points, N is 1, 2., N, m is the data dimension, k (u) is a defined kernel function, thereby obtaining an m × N-order matrix, and after sampling the matrix line by line, forming an m × c-order feature set H, wherein c is less than N;
and fourthly, training a neural network by adopting a flexible neural tree model as a classifier of the musical instrument recognition module:
the calculation flow adopted in the flexible neural tree model is as follows: firstly, setting an initial structure of a flexible neural tree, and then optimizing and training a fuzzy classification model according to an obtained feature set H and the set initial structure, wherein the model comprises two types of instructions: the flexible neuron instruction is used for connecting a non-leaf node of a tree structure with a subtree, the terminal instruction is each input characteristic, and a function instruction set and a terminal instruction set are represented as follows:
S=F∪T={+2,+3,…,+N}∪{z1,z2,…,zn}(8)
whereini(2, 3.. N) represents a non-leaf node instruction, the non-leaf node having i inputs, i.e., i input variables; z is a radical of1,z2,...znRepresenting a leaf node instruction, the leaf node having no input variable;
adopting a Gaussian function as a flexible excitation function:
f ( j i , b i , z ) = e - ( z - j i b i ) 2 - - - ( 9 )
wherein j isiAnd biIs an adjustable parameter randomly generated in the process of creating the flexible neural tree, a fitness function is another problem needing to be determined in advance besides an instruction set, and a root mean square error is selected as the fitness function.
CN201410008533.4A 2014-01-09 2014-01-09 A kind of sorting technique of instrument signal Expired - Fee Related CN103761965B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410008533.4A CN103761965B (en) 2014-01-09 2014-01-09 A kind of sorting technique of instrument signal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410008533.4A CN103761965B (en) 2014-01-09 2014-01-09 A kind of sorting technique of instrument signal

Publications (2)

Publication Number Publication Date
CN103761965A CN103761965A (en) 2014-04-30
CN103761965B true CN103761965B (en) 2016-05-25

Family

ID=50529193

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410008533.4A Expired - Fee Related CN103761965B (en) 2014-01-09 2014-01-09 A kind of sorting technique of instrument signal

Country Status (1)

Country Link
CN (1) CN103761965B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105096987B (en) * 2015-06-01 2019-01-15 努比亚技术有限公司 A kind of processing method and terminal of audio data
CN105070301B (en) * 2015-07-14 2018-11-27 福州大学 A variety of particular instrument idetified separation methods in the separation of single channel music voice
CN108962279A (en) * 2018-07-05 2018-12-07 平安科技(深圳)有限公司 New Method for Instrument Recognition and device, electronic equipment, the storage medium of audio data
CN110310666B (en) * 2019-06-27 2021-07-23 成都潜在人工智能科技有限公司 Musical instrument identification method and system based on SE convolutional network
CN110657882B (en) * 2019-09-23 2021-07-27 暨南大学 Bridge real-time safety state monitoring method utilizing single-measuring-point response
CN111461201B (en) * 2020-03-30 2023-09-19 重庆大学 Sensor data classification method based on phase space reconstruction
CN111681674B (en) * 2020-06-01 2024-03-08 中国人民大学 Musical instrument type identification method and system based on naive Bayesian model
CN114724533A (en) * 2022-02-25 2022-07-08 杭州小伴熊科技有限公司 Musical composition extraction method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101092228B1 (en) * 2009-12-21 2011-12-12 세종대학교산학협력단 System and method for recognizing instrument to classify signal source

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
A study on traditional Malay musical instruments sounds classification system;Norhalina Senann etc;《Proceedings of the 11th International Conference on Information Integration and Web-based Applications & Services》;20091231;729-733页 *
Instrument Identification in Monophonic Music Using Spectral Information;Mizuki Ihara etc;《2007 IEEE International Symposium on Signal Processing and Information Technology》;20071231;595-599页 *
基于声学特征的乐器识别研究;林玉志;《中国优秀硕士学位论文全文数据库 信息科技辑》;20130115(第1期);全文 *
基于声学特征的乐器识别综述;邓见光 等;《东莞理工学院学报》;20120630;第19卷(第3期);58-64页 *
基于音频的数字媒体中乐器识别方法研究;杭小兵;《中国优秀硕士学位论文全文数据库 信息科技辑》;20111215(第S1期);全文 *

Also Published As

Publication number Publication date
CN103761965A (en) 2014-04-30

Similar Documents

Publication Publication Date Title
CN103761965B (en) A kind of sorting technique of instrument signal
Lidy et al. CQT-based Convolutional Neural Networks for Audio Scene Classification.
CN111754988B (en) Sound scene classification method based on attention mechanism and double-path depth residual error network
CN112562741B (en) Singing voice detection method based on dot product self-attention convolution neural network
KR102198273B1 (en) Machine learning based voice data analysis method, device and program
CN109408660B (en) Music automatic classification method based on audio features
Massoudi et al. Urban sound classification using CNN
CN103544963A (en) Voice emotion recognition method based on core semi-supervised discrimination and analysis
CN110633725A (en) Method and device for training classification model and classification method and device
CN111400540B (en) Singing voice detection method based on extrusion and excitation residual error network
CN112750442B (en) Crested mill population ecological system monitoring system with wavelet transformation and method thereof
CN114237046B (en) Partial discharge pattern recognition method based on SIFT data feature extraction algorithm and BP neural network model
Ozkok et al. Convolutional neural network analysis of recurrence plots for high resolution melting classification
Benamer et al. Database for arabic speech commands recognition
Naranjo-Alcazar et al. On the performance of residual block design alternatives in convolutional neural networks for end-to-end audio classification
CN112735442B (en) Wetland ecology monitoring system with audio separation voiceprint recognition function and audio separation method thereof
CN105006231A (en) Distributed large population speaker recognition method based on fuzzy clustering decision tree
CN117976006A (en) Audio processing method, device, computer equipment and storage medium
CN114048770B (en) Automatic detection method and system for digital audio deletion and insertion tampering operation
CN115472179A (en) Automatic detection method and system for digital audio deletion and insertion tampering operation
CN112687280B (en) Biodiversity monitoring system with frequency spectrum-time space interface
JP2002062892A (en) Acoustic classifying device
Hebbar et al. A Comparison of Audio Preprocessing Techniques and Deep Learning Algorithms for Raga Recognition
CN107492384B (en) Voice emotion recognition method based on fuzzy nearest neighbor algorithm
DURDAG et al. A New Genre Classification with the Colors of Music

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160525

Termination date: 20190109