CN102760444A - Support vector machine based classification method of base-band time-domain voice-frequency signal - Google Patents

Support vector machine based classification method of base-band time-domain voice-frequency signal Download PDF

Info

Publication number
CN102760444A
CN102760444A CN2012101250857A CN201210125085A CN102760444A CN 102760444 A CN102760444 A CN 102760444A CN 2012101250857 A CN2012101250857 A CN 2012101250857A CN 201210125085 A CN201210125085 A CN 201210125085A CN 102760444 A CN102760444 A CN 102760444A
Authority
CN
China
Prior art keywords
signal
band time
subsequence
zero
mean
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2012101250857A
Other languages
Chinese (zh)
Other versions
CN102760444B (en
Inventor
刘一民
李元新
孟华东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201210125085.7A priority Critical patent/CN102760444B/en
Publication of CN102760444A publication Critical patent/CN102760444A/en
Application granted granted Critical
Publication of CN102760444B publication Critical patent/CN102760444B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Complex Calculations (AREA)

Abstract

The invention relates to a support vector machine based classification method of base-band time-domain voice-frequency signals, comprising the following steps of: firstly segmenting a base-band time-domain voice-frequency signal sequence to obtain initial segmented subsequences; then respectively subtracting respective mean value from each initial segmented subsequence to obtain zero-mean-value segmented subsequences; then carrying out windowing treatment on each zero-mean-value segmented subsequence, respectively carrying out Fourier transformation treatment on results to obtain the spectrum amplitudes of the zero-mean-value segmented subsequences, and respectively solving the standard difference of each spectrum amplitude to obtain a characteristic quantity; sequentially combining the zero-mean-value segmented subsequences into a long subsequence according to an order; then calculating a normalized autocorrelation matrix of the long subsequence, and carrying out singular value decomposition on the normalized autocorrelation matrix to obtain a demarcation point of a subspace; then calculating the signal to noise ratio parameter of an other characteristic quantity; and finally sending an input vector composed of the two characteristic quantities into a trained SVM (Support Vector Machine) classifier to identify the classification of base-band time-domain voice-frequency signals and distinguish a voice signal and a noise signal.

Description

Base band time-domain audio signal sorting technique based on SVM
Technical field
The invention belongs to the signal processing technology field, be specifically related to a kind of base band time-domain audio signal sorting technique based on SVM.
Background technology
The present invention is applied in the radio detection system; Handled signal is the base band time-domain audio signal after the demodulation; The voice signal that signal possibly polluted by noise in various degree also possibly be pure noise signal, wherein noise all with white noise for leading and be mixed with a small amount of coloured noise; Utilize the principle of SVM to make up a kind of sorter, signal type is carried out simple and effective discriminating classification.
Following article and patent documentation have covered the main background technology in this field basically.In order to explain out the evolution of technology, let their time sequencings arrange, and introduce the main contribution of document one by one.
1.S.Gokhun?Tanyer,Hamza?ozer,“Voice?Activity?Detection?in?Nonstationary?Gaussian?Noise”,Proceedings?of?ICSP,1620-1623,1998.
Sound end detects (Voice Activity Detection; VAD) be meant the process that from noise, screens out voice; Article has proposed the method for energy threshold method, zero-crossing rate method, least square cycle estimator and adaptive energy thresholding; Wherein energy threshold method and zero-crossing rate method are applicable to Signal-to-Noise (signal to noise ratio more; SNR) under the condition with higher, false-alarm is very high when signal to noise ratio (S/N ratio) is low, and least square cycle estimator can periodically cause detecting failure owing to the noise non-stationary comprises.Article also proposes several different methods is integrated into the strategy that the lang tone signal detects simultaneously.
2.C.J.C.Burges,“A?Tutorial?on?Support?Vector?Machines?for?Pattern?Recognition”,Data?Mining?and?Knowledge?Discovery,vol.2,no.2,pp.121-167,1998.
Having introduced ultimate principle and the conclusion of SVM in detail derives; The method of SVM is that the optimal classification lineoid under the linear separability situation puts forward; Its basic thought may be summarized to be at first and through nonlinear transformation the input space is transformed to a higher dimensional space, in this new space, asks for highest possible priority classification lineoid then." largest interval " and " with data projection to higher dimensional space more " is its key concept, and SVM constitutes two quasi-mode sorters on the ordinary meaning.But mostly this article is the SVM ultimate principle is carried out the proof of the derivation of equation, is not given in prompting and guidance that the voice signal context of detection is used.
3.S.Gokhun?Tanyer,Hamza?ozer,“Voice?Activity?Detection?in?Nonstationary?Noise”,IEEE?Trans.Speech?Audio?Process.,vol.8,no.4,pp.478-481,Jul.2001
Propose the sound end detecting method of adaptive energy thresholding and provide implementation strategy, wherein be applied to method of geometry signal calculated SNR, reduced dependence the noise signal prior imformation.But the method for estimation of this SNR receives the influence of signal cumulative distribution, can not fully learn noise signal information, and selection of parameter is comparatively difficult with adjustment, and SNR estimates at deviation under the situation of noise non-stationary.
4.Quanwei?Cai,Ping?Wei,Xianci?Xiao,“A?Digital?Modulation?Recognition?Method”,Proceedings?of?ICASSP,2004,pp?863–866
Signal SNR estimation principle and method based on SVD have been proposed, simple, the performance of this method is not inquired into, do not provide the Choosing computing parameters method yet.
5.Cheol-Sun?Park,Won?Jang,Sun-Phil?Nah.and?Dae?Young?Kim,“Automatic?Modulation?Recognition?using?Support?Vector?Machine?in?Software?Radio?Applications”,in?Proc.9th?IEEE?ICACT,Feb.2007,pp.9-12
Proposition is based on the method for the signal Modulation Mode Recognition of SVM, with the power spectrum density maximal value γ of the normalization center symmetry instantaneous amplitude of signal Max, the symmetry nonlinear component absolute value of the center in the signal strong component instantaneous phase standard deviation sigma Ap, the symmetry nonlinear component of the center in the signal strong component instantaneous phase standard deviation sigma Dp, receive the standard deviation sigma of the normalization center symmetry instantaneous amplitude absolute value of signal AaAnd the standard deviation sigma of normalization instantaneous frequency absolute value in the signal strong component AfInput obtains the result as characteristic quantity, even under the situation of the low SNR of signal, also obtain classification results exactly.
Summary of the invention
In order to overcome the deficiency of above-mentioned prior art; The object of the present invention is to provide a kind of base band time-domain audio signal sorting technique based on SVM; The base band time-domain audio signal is handled; Extract characteristic quantity as the input of sorter obtaining discriminating classification results to signal type, thereby voice signal and noise signal are classified.
To achieve these goals, the technical scheme of the present invention's employing is:
Base band time-domain audio signal sorting technique based on SVM comprises the steps:
Step 1: with total length is the base band time-domain audio signal sequence s={s (1) of N, s (2) ..., s (N) } be divided into the K section, every segment length is L, obtains the initial fragment subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , S wherein i(m)=s ((i-1) L+m) (i=1,2 ..., K, m=1,2 ..., L), each initial fragment subsequence deducts average separately respectively then, can get zero-mean segmentation subsequence x 1 = { x 1 ( 1 ) , x 1 ( 2 ) , . . . , x 1 ( L ) } x 2 = { x 2 ( 1 ) , x 2 ( 2 ) , . . . , x 2 ( L ) } . . . x K = { x K ( 1 ) , x K ( 2 ) , . . . , x K ( L ) } , Wherein x i ( m ) = s i ( m ) - 1 L Σ j = 1 L s i ( j ) ;
Step 2: each zero-mean segmentation subsequence is carried out windowing process, obtain the result and do x 1 ′ = x 1 w T x 2 ′ = x 2 w T . . . x K ′ = x K w T , Wherein w is a Hanning window;
Step 3: the result after the windowing process is carried out Fourier transform respectively handle, the spectrum amplitude sequence that obtains the zero-mean segmentation subsequence after each windowing does f 1 = | FFT ( x 1 ′ ) | = { f 1 ( 1 ) , f 1 ( 2 ) , . . . , f 1 ( M ) } f 2 = | FFT ( x 2 ′ ) | = { f 2 ( 1 ) , f 2 ( 2 ) , . . . , f 2 ( M ) } . . . f K = | FFT ( x K ′ ) | = { f K ( 1 ) , f K ( 2 ) , . . . , f K ( M ) } , Wherein M is the length of spectrum amplitude sequence;
Step 4: the standard deviation d={d (1) that obtains each spectrum amplitude respectively; D (2); ...; D (K) }; Wherein
Figure BDA0000157114150000041
obtains the mean value of all standard deviations then; Obtain a characteristic quantity of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation
Step 5: with each zero-mean segmentation subsequence x 1, x 2..., x KBe combined into long sequence x, i.e. an x={x successively according to order 1, x 2..., x K}={ x (1), x (2) ..., x (N) }, calculate the normalized autocorrelation matrix of this sequence then, the result does Wherein
Figure BDA0000157114150000044
Q is the dimension of autocorrelation matrix, and span is [50,90];
Step 6: autocorrelation matrix R is carried out svd, obtain R=V Λ V H, wherein
Λ=diag (λ 1, λ 2..., λ Q) Q * Q=diag (γ 1+ σ 2..., γ p+ σ 2, σ 2..., σ 2) Q * Q, and γ 1>=γ 2>=...>=γ pThereby, obtain the separation p of subspace;
Step 7: another characteristic quantity according to
Figure BDA0000157114150000045
Figure BDA0000157114150000046
calculates this base band time-domain audio signal sequence is designated as signal to noise ratio (S/N ratio) parameter
Step 8: with two characteristic quantities of this base band time-domain audio signal sequence; Be that spectrum amplitude standard deviation D and signal to noise ratio (S/N ratio) parameter constitute input vector; Send in the svm classifier device of having trained; Thereby identify the kind of this base band time-domain audio signal, distinguish voice signal and noise signal.
Above-mentioned subspace separation p can be obtained by following method: by
Figure BDA0000157114150000049
Wherein
Figure BDA00001571141500000410
Be to the rounding downwards of autocorrelation matrix dimension result, calculate a last T+1 eigenvalue Q-T, λ Q-T+1..., λ QAverage, all are greater than 1.5E then λEigenwert in the maximum following p that is designated as, i.e. p={i| λ i>1.5E λ, λ I+1<1.5E λ.
Above-mentioned is the base band time-domain audio signal sequence s={s (1) of N with total length, s (2) ..., s (N) } be divided into the K section, every pairing period should be not more than 20ms.
Compared with prior art, the present invention more obtains the prior imformation of treating category signal through the mode of training, chooses suitable input feature vector amount and can effectively obtain classification results rapidly.In order to reflect the difference of voice signal and noise signal, select signal SNR parameter and signal spectrum amplitude standard deviation input feature vector amount as sorter, not only make things convenient for calculating but also can be good at realizing the discriminating and the classification of signal.The present invention can detect effectively and differentiate voice signal and noise signal; Two input feature vector amount Signal-to-Noise parameters choosing and signal spectrum amplitude standard deviation are calculated the difference that simply can effectively reflect two kinds of signals again, even under the lower situation of signal to noise ratio (S/N ratio), also can guarantee higher classification accuracy rate.The present invention is applicable to real time signal processing, is easy to realize, can perform well in the radio application.
Description of drawings
Fig. 1 is a process flow diagram of the present invention.
Fig. 2 is the probability density distribution figure of input feature vector amount when being the Signal-to-Noise parameter.
Fig. 3 is the probability density distribution figure of input feature vector amount when being the spectrum amplitude standard deviation.
Fig. 4 is a svm classifier device working result synoptic diagram.
Embodiment
Below in conjunction with accompanying drawing and embodiment the present invention is explained further details.
The present invention is based on SVM principle design sorter; Through base band time-domain audio signal series processing is extracted characteristic quantity; It is sent into the sorter of trained as input, thereby identify the type of sound signal, voice signal and noise signal are correctly classified.
As shown in Figure 1, performing step is following:
Step 1: owing to what will handle is the base band time-domain audio signal sequence of having passed through demodulation, at first tackles signal and carries out pre-service, so that extract the characteristic quantity of abundant reflected signal characteristic.
With total length is the base band time-domain audio signal sequence s={s (1) of N, s (2) ..., s (N) } evenly be divided into the K section, every segment length is L, every pairing period should be not more than 20ms.
Obtain the initial fragment subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , S wherein i(m)=s ((i-1) L+m) (i=1,2 ..., K, m=1,2 ..., L), each initial fragment subsequence deducts separately average respectively removing DC component then, thereby can get zero-mean segmentation subsequence x 1 = { x 1 ( 1 ) , x 1 ( 2 ) , . . . , x 1 ( L ) } x 2 = { x 2 ( 1 ) , x 2 ( 2 ) , . . . , x 2 ( L ) } . . . x K = { x K ( 1 ) , x K ( 2 ) , . . . , x K ( L ) } , Wherein x i ( m ) = s i ( m ) - 1 L Σ j = 1 L s i ( j ) .
Step 2: for reduce to the segmentation subsequence carry out frequency domain when handling secondary lobe select for use Hanning window that each zero-mean segmentation subsequence is carried out windowing process to result's influence.Result after the windowing does x 1 ′ = x 1 w T x 2 ′ = x 2 w T . . . x K ′ = x K w T , Wherein w is the Hanning window sequence.
Step 3: the result after the windowing process is carried out Fourier transform respectively handle, the spectrum amplitude sequence that obtains the zero-mean segmentation subsequence after each windowing does f 1 = | FFT ( x 1 ′ ) | = { f 1 ( 1 ) , f 1 ( 2 ) , . . . , f 1 ( M ) } f 2 = | FFT ( x 2 ′ ) | = { f 2 ( 1 ) , f 2 ( 2 ) , . . . , f 2 ( M ) } . . . f K = | FFT ( x K ′ ) | = { f K ( 1 ) , f K ( 2 ) , . . . , f K ( M ) } , Wherein counting of FFT should be 2 power exponent 2 greater than 2~4 times of sub-sequence length a, M is the length of spectrum amplitude sequence.
Step 4: utilize the no inclined to one side estimated form
Figure BDA0000157114150000066
of standard deviation to obtain the standard deviation d={d (1) of the spectrum amplitude of each segmentation subsequence respectively; D (2); ...; D (K) }; Obtain the mean value of all standard deviations then, just obtain a characteristic quantity of this time-domain audio signal sequence, i.e. the spectrum amplitude standard deviation D = 1 K Σ i = 1 K d ( i ) .
The Signal-to-Noise parameter is as shown in Figure 2, and wherein horizontal ordinate is the span of Signal-to-Noise parameter, and ordinate is a probability density; The probability density function of spectrum amplitude standard deviation is as shown in Figure 3, and wherein horizontal ordinate is the span of spectrum amplitude standard deviation, and ordinate is a probability density.As can be seen from the figure the characteristic quantity of noise signal distributes comparatively concentrated; Therefore single characteristic quantity can reflect the difference of voice signal and noise signal to a certain extent; But can not two types of signals be distinguished completely effectively; Both could realize correct signal classification, therefore continuation execution following steps as the input quantity of sorter so need associating.
Step 5: then audio signal sequence is handled and obtained the another one characteristic quantity.At first with each zero-mean segmentation subsequence x 1, x 2..., x KBe combined into a Chief Signal Boatswain sequence x successively according to order, promptly obtain x={x 1, x 2..., x K}={ x (1), x (2) ..., x (N) }, calculate the normalized autocorrelation matrix of this sequence then, the result does
Figure BDA0000157114150000071
Wherein
Figure BDA0000157114150000072
And Q is the dimension of autocorrelation matrix, and span is [50,90], and value is 70 among the present invention.
Step 6: autocorrelation matrix R is carried out SVD decompose, obtain R=V Λ V HSuppose that voice signal and noise signal are separate, R=R x+ R n=V (Λ x+ Λ n) V H=V Λ V H, R wherein x, R nIt is respectively the autocorrelation matrix of voice signal and noise signal.
Can know Λ by the SVD decomposition x=diag (γ 1, γ 2..., γ p, 0 ..., 0) Q * Q, γ 1>=γ 2>=...>=γ p,
Λ n=diag(σ 2,σ 2,…,σ 2) Q×Q
Λ=diag(λ 1,λ 2,…,λ Q) Q×Q=diag(γ 12,…,γ p2,σ 2,…,σ 2) Q×Q
Through Wherein
Figure BDA0000157114150000074
Be to the rounding downwards of autocorrelation matrix dimension result, calculate a last T+1 eigenvalue Q-T, λ Q-T+1..., λ QAverage, search all then greater than 1.5E λEigenwert in the maximum following separation p that is designated as, i.e. p={i| λ i>1.5E λ, λ I+1<1.5E λ.
Step 7: calculate another characteristic quantity of this base band time-domain audio signal sequence according to
Figure BDA0000157114150000075
Figure BDA0000157114150000076
, promptly signal to noise ratio (S/N ratio) parameter
Figure BDA0000157114150000077
can reflect the state of signal-to-noise of signal to a certain extent.
Step 8: with two characteristic quantities of this base band time-domain audio signal sequence; Be that spectrum amplitude standard deviation D and signal to noise ratio (S/N ratio) parameter
Figure BDA0000157114150000081
constitute input vector; Send in the svm classifier device of trained; Just can obtain the classification results of this base band time-domain audio signal, distinguish voice signal and noise signal.
The sorter working result of carrying out this step is as shown in Figure 4; Wherein "+" is the phonic signal character amount; " * " is the noise signal characteristic quantity; Two category feature amounts can be confirmed this base band time-domain audio signal sorter distinguishing signal type and classification correctly effectively based on SVM thus by correct isolation in the space.

Claims (3)

1. based on the base band time-domain audio signal sorting technique of SVM, it is characterized in that, comprise the steps:
Step 1: with total length is the base band time-domain audio signal sequence s={s (1) of N, s (2) ..., s (N) } be divided into the K section, every segment length is L, obtains the initial fragment subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , S wherein i(m)=s ((i-1) L+m) (i=1,2 ..., K, m=1,2 ..., L), each initial fragment subsequence deducts average separately respectively then, can get zero-mean segmentation subsequence x 1 = { x 1 ( 1 ) , x 1 ( 2 ) , . . . , x 1 ( L ) } x 2 = { x 2 ( 1 ) , x 2 ( 2 ) , . . . , x 2 ( L ) } . . . x K = { x K ( 1 ) , x K ( 2 ) , . . . , x K ( L ) } , Wherein
x i ( m ) = s i ( m ) - 1 L Σ j = 1 L s i ( j ) ;
Step 2: each zero-mean segmentation subsequence is carried out windowing process, obtain the result and do x 1 ′ = x 1 w T x 2 ′ = x 2 w T . . . x K ′ = x K w T , Wherein w is a Hanning window;
Step 3: the result after the windowing process is carried out Fourier transform respectively handle, the spectrum amplitude sequence that obtains the zero-mean segmentation subsequence after each windowing does f 1 = | FFT ( x 1 ′ ) | = { f 1 ( 1 ) , f 1 ( 2 ) , . . . , f 1 ( M ) } f 2 = | FFT ( x 2 ′ ) | = { f 2 ( 1 ) , f 2 ( 2 ) , . . . , f 2 ( M ) } . . . f K = | FFT ( x K ′ ) | = { f K ( 1 ) , f K ( 2 ) , . . . , f K ( M ) } , Wherein M is the length of spectrum amplitude sequence;
Step 4: the standard deviation d={d (1) that obtains each spectrum amplitude respectively; D (2); ...; D (K) }; Wherein
Figure FDA0000157114140000016
obtains the mean value of all standard deviations then; Obtain a characteristic quantity of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation
Figure FDA0000157114140000017
Step 5: with each zero-mean segmentation subsequence x 1, x 2..., x KBe combined into long sequence x, i.e. an x={x successively according to order 1, x 2..., x K}={ x (1), x (2) ..., x (N) }, calculate the normalized autocorrelation matrix of this sequence then, the result does
Figure FDA0000157114140000021
Wherein
Figure FDA0000157114140000022
Q is the dimension of autocorrelation matrix, and span is [50,90];
Step 6: autocorrelation matrix R is carried out svd, obtain R=V Λ V H, wherein
Λ=diag (λ 1, λ 2..., λ Q) Q * Q=diag (γ 1+ σ 2..., γ p+ σ 2, σ 2..., σ 2) Q * Q, and γ 1>=γ 2>=...>=γ pThereby, obtain the separation p of subspace;
Step 7: another characteristic quantity according to
Figure FDA0000157114140000023
Figure FDA0000157114140000024
calculates this base band time-domain audio signal sequence is designated as signal to noise ratio (S/N ratio) parameter
Figure FDA0000157114140000025
Step 8: with two characteristic quantities of this base band time-domain audio signal sequence; Be that spectrum amplitude standard deviation D and signal to noise ratio (S/N ratio) parameter
Figure FDA0000157114140000026
constitute input vector; Send in the svm classifier device of having trained; Thereby identify the kind of this base band time-domain audio signal, distinguish voice signal and noise signal.
2. according to the said signal sorting technique of claim 1, it is characterized in that subspace separation p can be obtained by following method: by
Figure FDA0000157114140000027
Wherein Be to the rounding downwards of autocorrelation matrix dimension result, calculate a last T+1 eigenvalue Q-T, λ Q-T+1..., λ QAverage, all are greater than 1.5E then λEigenwert in the maximum following p that is designated as, i.e. p={i| λ i>1.5E λ, λ I+1<1.5E λ.
3. according to the said signal sorting technique of claim 1, it is characterized in that be divided into the K section in the said step 1, every section time corresponding is not more than 20ms.
CN201210125085.7A 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal Expired - Fee Related CN102760444B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210125085.7A CN102760444B (en) 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210125085.7A CN102760444B (en) 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal

Publications (2)

Publication Number Publication Date
CN102760444A true CN102760444A (en) 2012-10-31
CN102760444B CN102760444B (en) 2014-06-11

Family

ID=47054885

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210125085.7A Expired - Fee Related CN102760444B (en) 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal

Country Status (1)

Country Link
CN (1) CN102760444B (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104240697A (en) * 2013-06-24 2014-12-24 浙江大华技术股份有限公司 Audio data feature extraction method and device
CN104409073A (en) * 2014-11-04 2015-03-11 贵阳供电局 Substation equipment sound and voice identification method
CN104732970A (en) * 2013-12-20 2015-06-24 中国科学院声学研究所 Ship radiation noise recognition method based on comprehensive features
CN104751856A (en) * 2013-12-31 2015-07-01 中国移动通信集团公司 Voice sentence recognizing method and device
CN105743756A (en) * 2016-01-20 2016-07-06 灵芯微电子科技(苏州)有限公司 Frame detection method based on Adaboost algorithm in Wi-Fi system
CN105976822A (en) * 2016-07-12 2016-09-28 西北工业大学 Audio signal extraction method and apparatus based on parameterization supergain beam former
CN106789764A (en) * 2016-11-18 2017-05-31 杭州电子科技大学 The transform domain quadratic estimate method of the denoising of joint Weighted Threshold and balanced judgement
CN107682109A (en) * 2017-10-11 2018-02-09 北京航空航天大学 A kind of interference signal classifying identification method suitable for UAV Communication system
CN108877783A (en) * 2018-07-05 2018-11-23 腾讯音乐娱乐科技(深圳)有限公司 The method and apparatus for determining the audio types of audio data
CN109448389A (en) * 2018-11-23 2019-03-08 西安联丰迅声信息科技有限责任公司 A kind of vehicle whistle intelligent detecting method
CN112466322A (en) * 2020-11-27 2021-03-09 华侨大学 Electromechanical device noise signal feature extraction method
CN113759208A (en) * 2021-06-02 2021-12-07 青岛鼎信通讯股份有限公司 Abnormal waveform identification method based on fault indicator
CN117150224A (en) * 2023-10-30 2023-12-01 宜兴启明星物联技术有限公司 User behavior data storage analysis method based on Internet of things

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080059156A1 (en) * 2006-08-30 2008-03-06 International Business Machines Corporation Method and apparatus for processing speech data
US7505902B2 (en) * 2004-07-28 2009-03-17 University Of Maryland Discrimination of components of audio signals based on multiscale spectro-temporal modulations
CN101529929A (en) * 2006-09-05 2009-09-09 Gn瑞声达A/S A hearing aid with histogram based sound environment classification
US20100008641A1 (en) * 2008-06-24 2010-01-14 Sony Corporation Electronic apparatus, video content editing method, and program
JP2011034342A (en) * 2009-07-31 2011-02-17 Fujifilm Corp Image processing apparatus and method, data processing device and method, and program

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7505902B2 (en) * 2004-07-28 2009-03-17 University Of Maryland Discrimination of components of audio signals based on multiscale spectro-temporal modulations
US20080059156A1 (en) * 2006-08-30 2008-03-06 International Business Machines Corporation Method and apparatus for processing speech data
CN101529929A (en) * 2006-09-05 2009-09-09 Gn瑞声达A/S A hearing aid with histogram based sound environment classification
US20100008641A1 (en) * 2008-06-24 2010-01-14 Sony Corporation Electronic apparatus, video content editing method, and program
JP2011034342A (en) * 2009-07-31 2011-02-17 Fujifilm Corp Image processing apparatus and method, data processing device and method, and program

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHEOL-SUN PARK ET AL: "Automatic Modulation Recognition using Support Vector Machine in Software Radio Applications", 《THE 9TH INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY》 *
JUN-HO CHOI ET AL: "Automatic Modulation Recognition of Digital Signals using Wavelet Features and SVM", 《INTERNATIONAL CONFERENCE ON ADVANCED COMMUNICATION TECHNOLOGY, 2008. ICACT 2008. 10TH》 *

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104240697A (en) * 2013-06-24 2014-12-24 浙江大华技术股份有限公司 Audio data feature extraction method and device
CN104732970B (en) * 2013-12-20 2018-12-04 中国科学院声学研究所 A kind of ship-radiated noise recognition methods based on comprehensive characteristics
CN104732970A (en) * 2013-12-20 2015-06-24 中国科学院声学研究所 Ship radiation noise recognition method based on comprehensive features
CN104751856A (en) * 2013-12-31 2015-07-01 中国移动通信集团公司 Voice sentence recognizing method and device
CN104751856B (en) * 2013-12-31 2017-12-22 中国移动通信集团公司 A kind of speech sentences recognition methods and device
CN104409073A (en) * 2014-11-04 2015-03-11 贵阳供电局 Substation equipment sound and voice identification method
CN105743756A (en) * 2016-01-20 2016-07-06 灵芯微电子科技(苏州)有限公司 Frame detection method based on Adaboost algorithm in Wi-Fi system
CN105743756B (en) * 2016-01-20 2019-03-12 中科威发半导体(苏州)有限公司 Frame detection method based on adaboost algorithm in WiFi system
CN105976822A (en) * 2016-07-12 2016-09-28 西北工业大学 Audio signal extraction method and apparatus based on parameterization supergain beam former
CN105976822B (en) * 2016-07-12 2019-12-03 西北工业大学 Audio signal extracting method and device based on parametrization supergain beamforming device
CN106789764A (en) * 2016-11-18 2017-05-31 杭州电子科技大学 The transform domain quadratic estimate method of the denoising of joint Weighted Threshold and balanced judgement
CN106789764B (en) * 2016-11-18 2019-07-16 杭州电子科技大学 Joint Weighted Threshold denoises and the transform domain quadratic estimate method of balanced judgement
CN107682109A (en) * 2017-10-11 2018-02-09 北京航空航天大学 A kind of interference signal classifying identification method suitable for UAV Communication system
CN108877783A (en) * 2018-07-05 2018-11-23 腾讯音乐娱乐科技(深圳)有限公司 The method and apparatus for determining the audio types of audio data
CN109448389A (en) * 2018-11-23 2019-03-08 西安联丰迅声信息科技有限责任公司 A kind of vehicle whistle intelligent detecting method
CN112466322A (en) * 2020-11-27 2021-03-09 华侨大学 Electromechanical device noise signal feature extraction method
CN112466322B (en) * 2020-11-27 2023-06-20 华侨大学 Noise signal feature extraction method for electromechanical equipment
CN113759208A (en) * 2021-06-02 2021-12-07 青岛鼎信通讯股份有限公司 Abnormal waveform identification method based on fault indicator
CN117150224A (en) * 2023-10-30 2023-12-01 宜兴启明星物联技术有限公司 User behavior data storage analysis method based on Internet of things
CN117150224B (en) * 2023-10-30 2024-01-26 宜兴启明星物联技术有限公司 User behavior data storage analysis method based on Internet of things

Also Published As

Publication number Publication date
CN102760444B (en) 2014-06-11

Similar Documents

Publication Publication Date Title
CN102760444B (en) Support vector machine based classification method of base-band time-domain voice-frequency signal
CN108600135B (en) Method for identifying signal modulation mode
CN106330385B (en) A kind of interference type recognition methods
CN108231067A (en) Sound scenery recognition methods based on convolutional neural networks and random forest classification
CN102610227A (en) Sound signal processing apparatus, sound signal processing method, and program
CN105261367B (en) A kind of method for distinguishing speek person
Miao et al. Underwater acoustic signal classification based on sparse time–frequency representation and deep learning
WO2021008000A1 (en) Voice wakeup method and apparatus, electronic device and storage medium
CN105393305A (en) Method for processing acoustic signal
CN104135327A (en) Spectrum sensing method based on support vector machine
CN101764786A (en) MQAM signal recognition method based on clustering algorithm
CN106357575A (en) Multi-parameter jointly-estimated interference type identification method
CN106098079A (en) Method and device for extracting audio signal
CN111832462A (en) Frequency hopping signal detection and parameter estimation method based on deep neural network
Kong et al. Radar waveform recognition using Fourier-based synchrosqueezing transform and CNN
Yu et al. Discriminative training for multiple observation likelihood ratio based voice activity detection
CN113472390A (en) Frequency hopping signal parameter estimation method based on deep learning
CN112394324A (en) Microphone array-based remote sound source positioning method and system
Wang et al. Binary neural networks for wireless interference identification
CN111325143A (en) Underwater target identification method under unbalanced data set condition
Tran et al. GAN-based data augmentation for UWB NLOS identification using machine learning
CN105743756B (en) Frame detection method based on adaboost algorithm in WiFi system
Yin et al. Co-channel multi-signal modulation classification based on convolution neural network
Wang et al. A new method of automatic modulation recognition based on dimension reduction
CN104467995B (en) Blind primary user detection based on HMM and transmission level recognition methods

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20140611

Termination date: 20210425

CF01 Termination of patent right due to non-payment of annual fee