CN102760444B - Support vector machine based classification method of base-band time-domain voice-frequency signal - Google Patents

Support vector machine based classification method of base-band time-domain voice-frequency signal Download PDF

Info

Publication number
CN102760444B
CN102760444B CN201210125085.7A CN201210125085A CN102760444B CN 102760444 B CN102760444 B CN 102760444B CN 201210125085 A CN201210125085 A CN 201210125085A CN 102760444 B CN102760444 B CN 102760444B
Authority
CN
China
Prior art keywords
signal
band time
sigma
subsequence
zero
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201210125085.7A
Other languages
Chinese (zh)
Other versions
CN102760444A (en
Inventor
刘一民
李元新
孟华东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201210125085.7A priority Critical patent/CN102760444B/en
Publication of CN102760444A publication Critical patent/CN102760444A/en
Application granted granted Critical
Publication of CN102760444B publication Critical patent/CN102760444B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Compression, Expansion, Code Conversion, And Decoders (AREA)
  • Complex Calculations (AREA)

Abstract

The invention relates to a support vector machine based classification method of base-band time-domain voice-frequency signals, comprising the following steps of: firstly segmenting a base-band time-domain voice-frequency signal sequence to obtain initial segmented subsequences; then respectively subtracting respective mean value from each initial segmented subsequence to obtain zero-mean-value segmented subsequences; then carrying out windowing treatment on each zero-mean-value segmented subsequence, respectively carrying out Fourier transformation treatment on results to obtain the spectrum amplitudes of the zero-mean-value segmented subsequences, and respectively solving the standard difference of each spectrum amplitude to obtain a characteristic quantity; sequentially combining the zero-mean-value segmented subsequences into a long subsequence according to an order; then calculating a normalized autocorrelation matrix of the long subsequence, and carrying out singular value decomposition on the normalized autocorrelation matrix to obtain a demarcation point of a subspace; then calculating the signal to noise ratio parameter of an other characteristic quantity; and finally sending an input vector composed of the two characteristic quantities into a trained SVM (Support Vector Machine) classifier to identify the classification of base-band time-domain voice-frequency signals and distinguish a voice signal and a noise signal.

Description

Base band time-domain audio signal sorting technique based on Support Vector Machine
Technical field
The invention belongs to signal processing technology field, be specifically related to a kind of base band time-domain audio signal sorting technique based on Support Vector Machine.
Background technology
The present invention is applied in radio detection system, handled signal is the base band time-domain audio signal after demodulation, the voice signal that signal may be polluted in various degree by noise, also may be pure noise signal, wherein noise all with white noise for leading and be mixed with a small amount of coloured noise, utilize the principle of SVM to build a kind of sorter, signal type is carried out to simple and effective discriminating classification.
Following article and patent documentation, covered the main background technology in this field substantially.In order to explain out the evolution of technology, allow their time sequencings arrange, and introduce one by one the main contributions of document.
1.S.Gokhun?Tanyer,Hamza?ozer,“Voice?Activity?Detection?in?Nonstationary?Gaussian?Noise”,Proceedings?of?ICSP,1620-1623,1998.
Sound end detects (Voice Activity Detection, VAD) refer to the process that screens out voice from noise, article has proposed the method for energy threshold method, zero-crossing rate method, least square cycle estimator and adaptive energy thresholding, wherein energy threshold method and zero-crossing rate method are applicable to Signal-to-Noise (signal to noi se ratio more, SNR) in higher situation, in the time that signal to noise ratio (S/N ratio) is lower, false-alarm is very high, and least square cycle estimator can cause detecting unsuccessfully because noise non-stationary comprises periodically.Article also proposes several different methods to be integrated into the strategy of row Speech signal detection simultaneously.
2.C.J.C.Burges,“A?Tutorial?on?Support?Vector?Machines?for?Pattern?Recognition”,Data?Mining?and?Knowledge?Discovery,vol.2,no.2,pp.121-167,1998.
The ultimate principle and the conclusion that describe SVM in detail are derived, the method of SVM is that the optimal classification lineoid from linear separability situation puts forward, first its basic thought may be summarized to be, by nonlinear transformation, the input space is transformed to a higher dimensional space, then in this new space, asks for highest possible priority classification lineoid." largest interval " and " by extremely more higher dimensional space of data projection " is its key concept, and SVM forms two quasi-mode sorters on ordinary meaning.But mostly this article is SVM ultimate principle to carry out the proof of the derivation of equation, be not given in prompting and the guidance of the application of Speech signal detection aspect.
3.S.Gokhun?Tanyer,Hamza?ozer,“Voice?Activity?Detection?in?Nonstationary?Noise”,IEEE?Trans.Speech?Audio?Process.,vol.8,no.4,pp.478-481,Jul.2001
Propose the sound end detecting method of adaptive energy thresholding and provide implementation strategy, being wherein applied to method of geometry and calculating signal SNR, having reduced the dependence to noise signal prior imformation.But the method for estimation of this SNR is subject to the impact of signal cumulative distribution, can not fully learn noise signal information, comparatively difficulty is chosen and adjusted to parameter, and the in the situation that of noise non-stationary, SNR estimates at deviation.
4.Quanwei?Cai,Ping?Wei,Xianci?Xiao,“A?Digital?Modulation?Recognition?Method”,Proceedings?of?ICASSP,2004,pp?863–866
Signal SNR estimation principle and method based on SVD have been proposed, simple, the performance of the method is not inquired into, do not provide the choosing method of calculating parameter yet.
5.Cheol-Sun?Park,Won?Jang,Sun-Phil?Nah.and?Dae?Young?Kim,“Automatic?Modulation?Recognition?using?Support?Vector?Machine?in?Software?Radio?Applications”,in?Proc.9th?IEEE?ICACT,Feb.2007,pp.9-12
The method of signal Modulation Mode Recognition based on SVM is proposed, with the power spectrum density maximal value γ of the normalization Central Symmetry instantaneous amplitude of signal max, Central Symmetry nonlinear component absolute value in signal strong component instantaneous phase standard deviation sigma ap, Central Symmetry nonlinear component in signal strong component instantaneous phase standard deviation sigma dp, receive the standard deviation sigma of normalization Central Symmetry instantaneous amplitude absolute value of signal aaand the standard deviation sigma of normalization instantaneous frequency absolute value in signal strong component afas characteristic quantity, input obtains result, even if also obtain classification results exactly in the situation that of the low SNR of signal.
Summary of the invention
In order to overcome above-mentioned the deficiencies in the prior art, the object of the present invention is to provide a kind of base band time-domain audio signal sorting technique based on Support Vector Machine, base band time-domain audio signal is processed, extract characteristic quantity as the input of sorter to obtain the discriminating classification results to signal type, thereby by voice signal and noise signal classification.
To achieve these goals, the technical solution used in the present invention is:
Base band time-domain audio signal sorting technique based on Support Vector Machine, comprises the steps:
Step 1: the base band time-domain audio signal sequence s={s (1) that is N by total length, s (2) ..., s (N) } be divided into K section, every segment length is L, obtains initial fragment subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , Wherein s i(m)=s ((i-1) L+m) (i=1,2 ..., K, m=1,2 ..., L), then each initial fragment subsequence deducts respectively average separately, can obtain zero-mean segmentation subsequence x 1 = { x 1 ( 1 ) , x 1 ( 2 ) , . . . , x 1 ( L ) } x 2 = { x 2 ( 1 ) , x 2 ( 2 ) , . . . , x 2 ( L ) } . . . x K = { x K ( 1 ) , x 2 ( K ) , . . . , x K ( L ) } , Wherein x i ( m ) = s i ( m ) - 1 L Σ j = 1 L s i ( j ) ;
Step 2: each zero-mean segmentation subsequence is carried out to windowing process, obtain result and be x 1 ′ = x 1 w T x 2 ′ = x 2 w T . . . x K ′ = x K w T , Wherein w is Hanning window;
Step 3: the result after windowing process is carried out respectively to Fourier transform processing, and the spectrum amplitude sequence that obtains the zero-mean segmentation subsequence after each windowing is f 1 = | FFT ( x 1 ′ ) | = { f 1 ( 1 ) , f 1 ( 2 ) , . . . , f 1 ( M ) } f 2 = | FFT ( x 2 ′ ) | = { f 2 ( 1 ) , f 2 ( 2 ) , . . . , f 2 ( M ) } . . . f K = | FFT ( x K ′ ) | = { f K ( 1 ) , f K ( 2 ) , . . . , f K ( M ) } , Wherein M is the length of spectrum amplitude sequence;
Step 4: obtain respectively the standard deviation d={d (1) of each spectrum amplitude, d (2) ..., d (K) }, wherein d ( i ) = 1 M - 1 Σ j = 1 M ( f i ( j ) - 1 M Σ l = 1 M f i ( l ) ) 2 , Then obtain the mean value of all standard deviations, obtain a characteristic quantity of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation
Figure GDA00001853396400042
Step 5: by each zero-mean segmentation subsequence x 1, x 2..., x kbe combined into successively long sequence x, i.e. an x={x according to order 1, x 2..., x k}={ x (1), x (2) ..., x (N) }, then calculate the normalized autocorrelation matrix of this sequence, result is wherein r i = 1 Σ j = 1 N x ( j ) 2 Σ l = 1 N - i x ( l ) x ( l + i ) , Q is the dimension of autocorrelation matrix, and span is [50,90];
Step 6: autocorrelation matrix R is carried out to svd, obtain R=V Λ V h, wherein Λ=diag (λ 1, λ 2..., λ q) q × Q=diag (γ 1+ σ 2..., γ p+ σ 2, σ 2..., σ 2) q × Q, and γ 1>=γ 2>=...>=γ pthereby, obtain the separation p of subspace;
Step 7: according to σ 2 ^ = 1 Q - p Σ i = p + 1 Q λ i , SNR ^ = 10 log Σ i = 1 p λ i - p × σ 2 ^ Q × σ 2 ^ Another characteristic quantity that calculates this base band time-domain audio signal sequence, is designated as signal to noise ratio (S/N ratio) parameter
Figure GDA00001853396400047
Step 8: by two characteristic quantities of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation D and signal to noise ratio (S/N ratio) parameter
Figure GDA00001853396400048
formation input vector, sends in the svm classifier device of having trained, thereby identifies the kind of this base band time-domain audio signal, distinguishes voice signal and noise signal.
Above-mentioned subspace separation p can obtain by the following method: by
Figure GDA00001853396400049
wherein
Figure GDA000018533964000410
be to the rounding downwards of autocorrelation matrix dimension result, calculate a last T+1 eigenvalue λ q-T, λ q-T+1..., λ qaverage, then all 1.5E that are greater than λeigenwert in be designated as p, i.e. p={i| λ under maximum i>1.5E λ, λ i+1<1.5E λ.
The above-mentioned base band time-domain audio signal sequence s={s (1) that is N by total length, s (2) ..., s (N) } be divided into K section, every corresponding period should be not more than 20ms.
Compared with prior art, the present invention more obtains the prior imformation of signal to be sorted by the mode of training, choose suitable input feature vector amount and can effectively obtain rapidly classification results.In order to reflect the difference of voice signal and noise signal, select signal SNR parameter and the signal spectrum amplitude standard deviation input feature vector amount as sorter, not only facilitate calculating but also can be good at realizing discriminating and the classification of signal.The present invention is detection and identification voice signal and noise signal effectively, two input feature vector amount Signal-to-Noise parameters choosing and signal spectrum amplitude standard deviation are calculated the difference that simply can effectively reflect again two kinds of signals, even if also can guarantee higher classification accuracy rate in the situation that signal to noise ratio (S/N ratio) is lower.The present invention is applicable to real time signal processing, is easy to realize, and can perform well in radio application.
Accompanying drawing explanation
Fig. 1 is process flow diagram of the present invention.
Fig. 2 is the probability density distribution figure of input feature vector amount while being Signal-to-Noise parameter.
Fig. 3 is the probability density distribution figure of input feature vector amount while being spectrum amplitude standard deviation.
Fig. 4 is svm classifier device working result schematic diagram.
Embodiment
Below in conjunction with drawings and Examples, the present invention is described in further details.
The present invention is based on SVM principle design sorter, by base band time-domain audio signal series processing is extracted to characteristic quantity, set it as input and send into the complete sorter of training, thereby identify the type of sound signal, voice signal and noise signal are correctly classified.
As shown in Figure 1, performing step is as follows:
Step 1: due to be processed be the base band time-domain audio signal sequence of having passed through demodulation, first tackle signal and carry out pre-service, so that extract the characteristic quantity of abundant reflected signal feature.
The base band time-domain audio signal sequence s={s (1) that is N by total length, s (2) ..., s (N) } be evenly divided into K section, every segment length is L, every corresponding period should be not more than 20ms.
Obtain initial fragment subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , Wherein s i(m)=s ((i-1) L+m) (i=1,2 ..., K, m=1,2 ..., L), then each initial fragment subsequence deducts respectively average separately to remove DC component, thereby can obtain zero-mean segmentation subsequence x 1 = { x 1 ( 1 ) , x 1 ( 2 ) , . . . , x 1 ( L ) } x 2 = { x 2 ( 1 ) , x 2 ( 2 ) , . . . , x 2 ( L ) } . . . x K = { x K ( 1 ) , x 2 ( K ) , . . . , x K ( L ) } , Wherein x i ( m ) = s i ( m ) - 1 L &Sigma; j = 1 L s i ( j ) .
Step 2: the impact of secondary lobe on result while segmentation subsequence being carried out to frequency domain processing in order to reduce, select Hanning window to carry out windowing process to each zero-mean segmentation subsequence.Result after windowing is x 1 &prime; = x 1 w T x 2 &prime; = x 2 w T . . . x K &prime; = x K w T , Wherein w is Hanning window sequence.
Step 3: the result after windowing process is carried out respectively to Fourier transform processing, and the spectrum amplitude sequence that obtains the zero-mean segmentation subsequence after each windowing is f 1 = | FFT ( x 1 &prime; ) | = { f 1 ( 1 ) , f 1 ( 2 ) , . . . , f 1 ( M ) } f 2 = | FFT ( x 2 &prime; ) | = { f 2 ( 1 ) , f 2 ( 2 ) , . . . , f 2 ( M ) } . . . f K = | FFT ( x K &prime; ) | = { f K ( 1 ) , f K ( 2 ) , . . . , f K ( M ) } ,
Wherein counting of FFT should be 2 the power exponent 2 that is greater than 2~4 times of sub-sequence length a, M is the length of spectrum amplitude sequence.
Step 4: utilize standard deviation without inclined to one side estimated form d ( i ) = 1 M - 1 &Sigma; j = 1 M ( f i ( j ) - 1 M &Sigma; l = 1 M f i ( l ) ) 2 Obtain respectively the standard deviation d={d (1) of the spectrum amplitude of each segmentation subsequence, d (2) ..., d (K) }, then obtain the mean value of all standard deviations, just obtain a characteristic quantity of this time-domain audio signal sequence, i.e. spectrum amplitude standard deviation D = 1 K &Sigma; i = 1 K d ( i ) .
As shown in Figure 2, wherein horizontal ordinate is the span of Signal-to-Noise parameter to Signal-to-Noise parameter, and ordinate is probability density; As shown in Figure 3, wherein horizontal ordinate is the span of spectrum amplitude standard deviation to the probability density function of spectrum amplitude standard deviation, and ordinate is probability density.As can be seen from the figure the characteristic quantity of noise signal distributes comparatively concentrated, therefore single characteristic quantity can reflect the difference of voice signal and noise signal to a certain extent, but two class signals can not be distinguished completely effectively, could realize correct signal classification so need to combine both as the input quantity of sorter, therefore continue to carry out following steps.
Step 5: then audio signal sequence is processed and obtained another one characteristic quantity.First by each zero-mean segmentation subsequence x 1, x 2..., x kbe combined into successively a Chief Signal Boatswain sequence x according to order, obtain x={x 1, x 2..., x k}={ x (1), x (2) ..., x (N) }, then calculate the normalized autocorrelation matrix of this sequence, result is
Figure GDA00001853396400071
wherein r i = 1 &Sigma; j = 1 N x ( j ) 2 &Sigma; l = 1 N - i x ( l ) x ( l + i ) , And Q is the dimension of autocorrelation matrix, and span is [50,90], in the present invention, value is 70.
Step 6: autocorrelation matrix R is carried out to SVD decomposition, obtain R=V Λ V h.Suppose that voice signal and noise signal are separate, R=R x+ R n=V (Λ x+ Λ n) V h=V Λ V h, wherein R x, R nit is respectively the autocorrelation matrix of voice signal and noise signal.
Decompose known Λ by SVD x=diag (γ 1, γ 2..., γ p, 0 ..., 0) q × Q, γ 1>=γ 2>=...>=γ p,
Λ n=diag(σ 22,…,σ 2) Q×Q
Λ=diag(λ 12,…,λ Q) Q×Q=diag(γ 12,…,γ p22,…,σ 2) Q×Q
Pass through
Figure GDA00001853396400073
wherein
Figure GDA00001853396400074
be to the rounding downwards of autocorrelation matrix dimension result, calculate a last T+1 eigenvalue λ q-T, λ q-T+1..., λ qaverage, then search all 1.5E of being greater than λeigenwert in be designated as separation p, i.e. p={i| λ under maximum i>1.5E λ, λ i+1<1.5E λ.
Step 7: according to &sigma; 2 ^ = 1 Q - p &Sigma; i = p + 1 Q &lambda; i , SNR ^ = 10 log &Sigma; i = 1 p &lambda; i - p &times; &sigma; 2 ^ Q &times; &sigma; 2 ^ Calculate another characteristic quantity of this base band time-domain audio signal sequence, i.e. signal to noise ratio (S/N ratio) parameter
Figure GDA00001853396400077
can reflect to a certain extent the state of signal-to-noise of signal.
Step 8: by two characteristic quantities of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation D and signal to noise ratio (S/N ratio) parameter
Figure GDA00001853396400081
form input vector, send in the complete svm classifier device of training, just can obtain the classification results of this base band time-domain audio signal, distinguish voice signal and noise signal.
Carry out the sorter working result of this step as shown in Figure 4, wherein "+" is phonic signal character amount, " * " is noise signal characteristic quantity, in space, two category feature amounts can correctly be isolated, and confirm thus distinguishing signal type and the correctly classification effectively of this base band time-domain audio signal sorter based on SVM.

Claims (2)

1. the base band time-domain audio signal sorting technique based on Support Vector Machine, is characterized in that, comprises the steps:
Step 1: the base band time-domain audio signal sequence s={s (1) that is N by total length, s (2) ..., s (N) } be divided into K section, every segment length is L, obtains initial fragment subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , wherein s i(m)=s ((i-1) L+m) (i=1,2 ..., K, m=1,2 ..., L), then each initial fragment subsequence deducts respectively average separately, can obtain zero-mean segmentation subsequence s 1 = { s 1 ( 1 ) , s 1 ( 2 ) , . . . , s 1 ( L ) } s 2 = { s 2 ( 1 ) , s 2 ( 2 ) , . . . , s 2 ( L ) } . . . s K = { s K ( 1 ) , s K ( 2 ) , . . . , s K ( L ) } , wherein x i ( m ) = s i ( m ) - 1 L &Sigma; j = 1 l s i ( j ) ;
Step 2: each zero-mean segmentation subsequence is carried out to windowing process, obtain result and be x 1 &prime; = x 1 w T x 2 &prime; = x 2 w T . . . x K &prime; = x K w T , wherein w is Hanning window;
Step 3: the result after windowing process is carried out respectively to Fourier transform processing, and the spectrum amplitude sequence that obtains the zero-mean segmentation subsequence after each windowing is f 1 = | FFT ( x 1 &prime; ) | = { f 1 ( 1 ) , f 1 ( 2 ) , . . . , f 1 ( M ) } f 2 = | FFT ( x 1 &prime; ) | = { f 2 ( 1 ) , f 2 ( 2 ) , . . . , f 2 ( M ) } . . . f K = | FFT ( x K &prime; ) | = { f K ( 1 ) , f K ( 2 ) , . . . , f K ( M ) } , Wherein M is the length of spectrum amplitude sequence;
Step 4: obtain respectively the standard deviation d={d (1) of each spectrum amplitude, d (2) ..., d (K) }, wherein
Figure FDA00003268392200016
, then obtain the mean value of all standard deviations, obtain a characteristic quantity of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation
Figure FDA00003268392200017
Step 5: by each zero-mean segmentation subsequence x 1, x 2..., x kbe combined into successively long sequence x, i.e. an x={x according to order 1, x 2..., x k}={ x (1), x (2) ..., x (N) }, then calculate the normalized autocorrelation matrix of this sequence, result is
Figure FDA00003268392200021
, wherein r i = 1 &Sigma; j = 1 N x ( j ) 2 &Sigma; l = 1 N - i x ( l ) x ( l + i ) , Q is the dimension of autocorrelation matrix, span is [50,90];
Step 6: autocorrelation matrix R is carried out to svd, obtain R=V Λ V h, wherein Λ=diag (λ 1, λ 2..., λ q) q × Q=diag (γ 1+ σ 2..., γ p+ σ 2, σ 2..., σ 2) q × Q, and γ 1>=γ 2>=...>=γ p, by
Figure FDA00003268392200023
, wherein
Figure FDA00003268392200028
be to the rounding downwards of autocorrelation matrix dimension result, calculate a last T+1 eigenvalue λ q-T, λ q-T+1..., λ qaverage, then all 1.5E that are greater than λeigenwert in be designated as p, i.e. p={i| λ under maximum i>1.5E λ, λ i+1<1.5E λ, thereby obtain the separation p of subspace;
Step 7: according to &sigma; 2 ^ 1 Q - p &Sigma; i = p + 1 Q &lambda; i , SNR ^ = 10 log &Sigma; i = 1 p &lambda; i - p &times; &sigma; 2 ^ Q &times; &sigma; 2 ^ Another characteristic quantity that calculates this base band time-domain audio signal sequence, is designated as signal to noise ratio (S/N ratio) parameter
Figure FDA00003268392200026
Step 8: by two characteristic quantities of this base band time-domain audio signal sequence, i.e. spectrum amplitude standard deviation D and signal to noise ratio (S/N ratio) parameter
Figure FDA00003268392200027
, formation input vector, sends in the svm classifier device of having trained, thereby identifies the kind of this base band time-domain audio signal, distinguishes voice signal and noise signal.
2. signal sorting technique according to claim 1, is characterized in that, in described step 1, be divided into K section, every corresponding period is not more than 20ms.
CN201210125085.7A 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal Expired - Fee Related CN102760444B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210125085.7A CN102760444B (en) 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210125085.7A CN102760444B (en) 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal

Publications (2)

Publication Number Publication Date
CN102760444A CN102760444A (en) 2012-10-31
CN102760444B true CN102760444B (en) 2014-06-11

Family

ID=47054885

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210125085.7A Expired - Fee Related CN102760444B (en) 2012-04-25 2012-04-25 Support vector machine based classification method of base-band time-domain voice-frequency signal

Country Status (1)

Country Link
CN (1) CN102760444B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104240697A (en) * 2013-06-24 2014-12-24 浙江大华技术股份有限公司 Audio data feature extraction method and device
CN104732970B (en) * 2013-12-20 2018-12-04 中国科学院声学研究所 A kind of ship-radiated noise recognition methods based on comprehensive characteristics
CN104751856B (en) * 2013-12-31 2017-12-22 中国移动通信集团公司 A kind of speech sentences recognition methods and device
CN104409073A (en) * 2014-11-04 2015-03-11 贵阳供电局 Substation equipment sound and voice identification method
CN105743756B (en) * 2016-01-20 2019-03-12 中科威发半导体(苏州)有限公司 Frame detection method based on adaboost algorithm in WiFi system
CN105976822B (en) * 2016-07-12 2019-12-03 西北工业大学 Audio signal extracting method and device based on parametrization supergain beamforming device
CN106789764B (en) * 2016-11-18 2019-07-16 杭州电子科技大学 Joint Weighted Threshold denoises and the transform domain quadratic estimate method of balanced judgement
CN107682109B (en) * 2017-10-11 2019-07-30 北京航空航天大学 A kind of interference signal classifying identification method suitable for UAV Communication system
CN108877783B (en) * 2018-07-05 2021-08-31 腾讯音乐娱乐科技(深圳)有限公司 Method and apparatus for determining audio type of audio data
CN109448389B (en) * 2018-11-23 2021-09-10 西安联丰迅声信息科技有限责任公司 Intelligent detection method for automobile whistling
CN112466322B (en) * 2020-11-27 2023-06-20 华侨大学 Noise signal feature extraction method for electromechanical equipment
CN113759208A (en) * 2021-06-02 2021-12-07 青岛鼎信通讯股份有限公司 Abnormal waveform identification method based on fault indicator
CN117150224B (en) * 2023-10-30 2024-01-26 宜兴启明星物联技术有限公司 User behavior data storage analysis method based on Internet of things

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7505902B2 (en) * 2004-07-28 2009-03-17 University Of Maryland Discrimination of components of audio signals based on multiscale spectro-temporal modulations
CN101529929A (en) * 2006-09-05 2009-09-09 Gn瑞声达A/S A hearing aid with histogram based sound environment classification
JP2011034342A (en) * 2009-07-31 2011-02-17 Fujifilm Corp Image processing apparatus and method, data processing device and method, and program

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101136199B (en) * 2006-08-30 2011-09-07 纽昂斯通讯公司 Voice data processing method and equipment
JP4488091B2 (en) * 2008-06-24 2010-06-23 ソニー株式会社 Electronic device, video content editing method and program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7505902B2 (en) * 2004-07-28 2009-03-17 University Of Maryland Discrimination of components of audio signals based on multiscale spectro-temporal modulations
CN101529929A (en) * 2006-09-05 2009-09-09 Gn瑞声达A/S A hearing aid with histogram based sound environment classification
JP2011034342A (en) * 2009-07-31 2011-02-17 Fujifilm Corp Image processing apparatus and method, data processing device and method, and program

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Automatic Modulation Recognition of Digital Signals using Wavelet Features and SVM;Jun-Ho Choi et al;《International Conference on Advanced Communication Technology, 2008. ICACT 2008. 10th》;20080220;387-390 *
Automatic Modulation Recognition using Support Vector Machine in Software Radio Applications;Cheol-Sun Park et al;《The 9th International Conference on Advanced Communication Technology》;20070214;9-12 *
Cheol-Sun Park et al.Automatic Modulation Recognition using Support Vector Machine in Software Radio Applications.《The 9th International Conference on Advanced Communication Technology》.2007,9-12.
Jun-Ho Choi et al.Automatic Modulation Recognition of Digital Signals using Wavelet Features and SVM.《International Conference on Advanced Communication Technology, 2008. ICACT 2008. 10th》.2008,387-390.

Also Published As

Publication number Publication date
CN102760444A (en) 2012-10-31

Similar Documents

Publication Publication Date Title
CN102760444B (en) Support vector machine based classification method of base-band time-domain voice-frequency signal
CN106330385B (en) A kind of interference type recognition methods
CN102610227A (en) Sound signal processing apparatus, sound signal processing method, and program
CN108600135A (en) A kind of recognition methods of signal modulation mode
US20210399817A1 (en) Wireless channel scenario identification method and system
CN103646649A (en) High-efficiency voice detecting method
CN105261367A (en) Identification method of speaker
CN105393305A (en) Method for processing acoustic signal
Miao et al. Underwater acoustic signal classification based on sparse time–frequency representation and deep learning
Sun et al. Cluster guide particle swarm optimization (CGPSO) for underdetermined blind source separation with advanced conditions
CN102968990A (en) Speaker identifying method and system
CN111832462A (en) Frequency hopping signal detection and parameter estimation method based on deep neural network
CN106098079A (en) Method and device for extracting audio signal
CN109884591A (en) A kind of multi-rotor unmanned aerial vehicle acoustical signal Enhancement Method based on microphone array
CN105304078A (en) Target sound data training device and target sound data training method
Yu et al. Discriminative training for multiple observation likelihood ratio based voice activity detection
CN112394324A (en) Microphone array-based remote sound source positioning method and system
CN102509548B (en) Audio indexing method based on multi-distance sound sensor
Tran et al. GAN-based data augmentation for UWB NLOS identification using machine learning
Malik et al. Automatic threshold optimization in nonlinear energy operator based spike detection
CN111325143A (en) Underwater target identification method under unbalanced data set condition
Wang et al. Binary neural networks for wireless interference identification
CN105743756B (en) Frame detection method based on adaboost algorithm in WiFi system
CN105717490B (en) LFM Signal separators and method for parameter estimation based on time frequency analysis
CN116992263A (en) Method for extracting, identifying and separating self-noise characteristics of underwater vehicle based on multitask learning

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20140611

Termination date: 20210425