CN113158916A - Cough sound automatic identification algorithm, device, medium and equipment - Google Patents

Cough sound automatic identification algorithm, device, medium and equipment Download PDF

Info

Publication number
CN113158916A
CN113158916A CN202110450820.0A CN202110450820A CN113158916A CN 113158916 A CN113158916 A CN 113158916A CN 202110450820 A CN202110450820 A CN 202110450820A CN 113158916 A CN113158916 A CN 113158916A
Authority
CN
China
Prior art keywords
sample
cough sound
cough
svm classifier
signal sequence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110450820.0A
Other languages
Chinese (zh)
Other versions
CN113158916B (en
Inventor
莫鸿强
曾键沣
周樊
章臻
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN202110450820.0A priority Critical patent/CN113158916B/en
Publication of CN113158916A publication Critical patent/CN113158916A/en
Application granted granted Critical
Publication of CN113158916B publication Critical patent/CN113158916B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines

Abstract

The invention provides an automatic cough sound identification algorithm, a device, a medium and equipment; the algorithm comprises the following steps: obtaining a sample to be identified; carrying out end point detection on the sample, intercepting a fixed length sequence with the duration of t from the initial point, setting the fixed length sequence as a signal sequence and carrying out normalization processing; obtaining a prediction signal sequence and a residual signal sequence by utilizing linear prediction coding; determining MFCC parameters MFCC for a predicted signal sequencerAnd the short-time energy en of the residual signal sequencerConstructing a feature vector; and inputting the feature vector into a linear SVM classifier, and judging whether the type of the sample to be recognized is a cough sound sample or a non-cough sound sample according to the output of the linear SVM classifier. The algorithm overcomes the problem that the time domain characteristics of the excitation source are easily influenced by the resonance of the sound channel, and has good identification capability and high identification accuracy.

Description

Cough sound automatic identification algorithm, device, medium and equipment
Technical Field
The invention relates to the technical field of medical instruments and medical signal processing, in particular to an automatic cough sound identification algorithm, device, medium and equipment.
Background
Cough is a natural reflex action of a human body and is a common symptom of respiratory diseases such as asthma, pneumonia, laryngitis, chronic obstructive pulmonary disease and the like. In clinic, doctors often use information such as cough frequency and intensity of patients as important basis for diagnosis, but the information is mainly from subjective description of patients and has poor reliability. The recording and automatic identification of the cough sounds, unlike the subjective description of the patient, facilitates the doctor to obtain the patient cough information more objectively and accurately.
One of the major difficulties in automatic recognition of cough sounds is the small number of samples available for training and testing of recognition models. At present, a large-scale cough sound database is not disclosed, and the number of cough sound samples collected by each research team is limited; in addition, the non-cough sound samples are complex in type, and the number of samples is far larger than that of the cough sound samples, so that the problem of sample imbalance exists. The lack of samples of the cough sounds and the imbalance of the training samples cause great difficulty in establishing an automatic cough sound recognition model.
The selection of targeted features to minimize undetermined parameters of the recognition model is an effective measure to overcome the above difficulties. Many documents currently employ Mel-Frequency Cepstrum Coefficient (MFCC) as a feature of cough sound; however, from the mechanism of cough sound formation, MFCC is not comprehensive enough in characterization of cough sound.
In cough, the contraction of abdominal muscles and diaphragm muscles causes the lungs to exhale a large amount of gas instantly, and the airflow impact of the lungs causes the vocal tract to vibrate to give out cough sound. According to its generation mechanism, the coughing tone can be viewed as a response produced by an excitation source acting on a vocal tract model, where the excitation source reflects the pulmonary airflow and the vocal tract model reflects the vocal tract shape. In order to accurately identify the cough sound, the feature extraction of the cough sound signal needs to fully consider information of both an excitation source and a sound channel, but currently, common automatic cough sound identification algorithms mostly utilize features related to sound channel information, such as MFCC, and excitation source information is not utilized.
For cough sound recognition, the primary information of the excitation source is its intensity over time and its variation in intensity; however, the convolution relation exists between the vocal tract model and the excitation source in the time domain, and the intensity of the excitation source signal cannot be directly determined by the cough sound; if the short-time energy/intensity of the coughing sound is simply used as the intensity of the excitation source, the obtained result will be affected by the resonance of the vocal tract and have a large error. Therefore, the extraction of the excitation source feature has a certain difficulty.
Disclosure of Invention
To overcome the disadvantages and shortcomings of the prior art, the present invention provides an automatic cough sound recognition algorithm, apparatus, medium and device. The algorithm utilizes Linear Predictive Coding (LPC) to realize the separation of a vocal tract model from an excitation source so as to overcome the difficulty in the extraction of the characteristics of the excitation source caused by the influence of vocal tract resonance; the short-time energy of the MFCC and the excitation source is used as the characteristic of the automatic cough sound identification algorithm, and the characteristic simultaneously contains information of a vocal tract model and the excitation source so as to solve the problem that the common algorithm neglects the characteristic of the excitation source; the algorithm utilizes a Support Vector Machine (SVM) as a classifier to solve the problem that the model is difficult to train due to insufficient samples and unbalanced samples; in addition, the algorithm only utilizes the first stage extraction features of the cough sound, the model is simple, the calculation amount is small, and the method is convenient to implement in wearable equipment.
In order to achieve the purpose, the invention is realized by the following technical scheme: an automatic cough sound recognition algorithm, comprising: the method comprises the following steps:
s1, obtaining a sample to be identified, wherein the sampling frequency of the sample to be identified is fs
S2, carrying out end point detection on the sample, intercepting a fixed length sequence with duration of t from a starting point, and setting the fixed length sequence as a signal sequence r (k); wherein k is 1, 2, …, N; n is t × fsAnd N is a positive integer;
s3, normalizing the signal sequence r (k) to obtain
Figure BDA0003038602390000021
S4, using linear predictive coding to encode the signal sequence
Figure BDA0003038602390000022
Decomposition into a predicted signal sequence sr(k) And residual signal sequence er(k) (ii) a Wherein, the order of the linear predictive coding model is p;
s5, obtaining a predicted signal sequence Sr(k) The Q-dimensional MFCC parameter MFCCr=[mr1,mr2,…mrQ]T
Step S6, obtaining a residual signal sequence er(k) Short-time energy en ofr(ii) a Wherein enrIs a scalar;
s7, obtaining the MFCC parameter MFCCrAnd short-term energy enrConstructed as feature vectors cr=[mr1,mr2,…mrQ,enr]T
S8, obtaining the feature vector crInputting the data into a linear SVM classifier, and outputting f (c) according to the linear SVM classifierr) Judging the type of the sample to be identified: if f (c)r) If the value is 1, the sample is a cough sound sample; if f (c)r) If the value is-1, the sample is a non-cough sound sample; the linear SVM classifier is a linear SVM classifier obtained by training and testing an initial linear SVM classifier.
Preferably, the training and testing process of the initial linear SVM classifier includes: the method comprises the following steps:
step Y1, at sampling frequency fsCollecting a plurality of cough sound samples and non-cough sound samples; setting the cough sound sample label to be 1, and setting the non-cough sound sample label to be-1; respectively dividing a training set and a testing set from a cough sound sample and a non-cough sound sample;
y2, processing each sample of the training set according to the steps from S2 to S7 respectively to obtain a feature vector of each sample;
step Y3, establishing a linear SVM classifier with the decision function of f (c) sgn [ wTc+b](ii) a Wherein c is a feature vector of the signal to be decided; f (c) is a decision value of the feature vector c, f (c) ═ 1 indicates that the signal to be decided is a cough sound, and f (c) ═ 1 indicates that the signal to be decided is a non-cough sound; w is a weight vector; b is the hyperplane offset;
y4, characterizing the sampleUsing the quantity as input, using a sample label corresponding to the sample as expected output, and training the established linear SVM classifier to obtain a decision function f (c) ═ sgn [ w [ ]*Tc+b*](ii) a Wherein, w*And b*Respectively obtaining the optimal weight vector and the hyperplane offset obtained by training;
step Y5, processing each sample of the test set according to the step Y2 respectively to obtain a feature vector of each sample;
y6, inputting the feature vector of the sample into a trained linear SVM classifier; comparing the output f (c) of the linear SVM classifier with a sample label corresponding to the sample, and calculating the sensitivity and the specificity according to the comparison result;
judging whether the sensitivity and the specificity meet the preset requirements:
if so, ending the training and testing;
otherwise, jumping to the step Y4 to retrain the decision function f (c) of the linear SVM classifier again*Tc+b*]。
Preferably, in the step Y1, the divided training set includes m cough sound samples and n non-cough sound samples;
the step Y2 is as follows: processing each sample of the training set according to the step S2 to obtain a cough sound signal sequence xi(k) And a sequence of non-cough tone signals yj(k) (ii) a Wherein, i is 1, 2, … m, j is 1, 2, … n;
sequence of cough tone signals xi(k) And a sequence of non-cough tone signals yj(k) Respectively processing according to the steps from S3 to S7 to obtain the characteristic vector of the cough sound sample
Figure BDA0003038602390000041
And feature vectors of non-cough sound samples
Figure BDA0003038602390000042
The feature set of the training sample is
Figure BDA0003038602390000043
Which corresponds to a set of sample tags of
Figure BDA0003038602390000044
Wherein all cough sound samples are labeled as
Figure BDA0003038602390000045
All non-cough sound samples are labeled as
Figure BDA0003038602390000046
The index x of the eigenvector in the characteristic set phi and the sample label in the sample label set TiOr yjOne-to-one correspondence is realized;
and Y4, taking each feature vector of the feature set phi as input, taking a sample label corresponding to the feature vector in the sample label set T as expected output, and training the established linear SVM classifier to obtain a decision function f (c) ═ sgn [ w ]*Tc+b*];
The step Y5, processing each sample of the test set according to the step Y2 respectively to obtain a feature vector of each sample; obtaining a feature set phi 'of the test sample and a corresponding sample label set T';
and Y6, taking each feature vector in the feature set phi 'of the test sample as the input of the linear SVM classifier, and comparing the output f (c) of the linear SVM classifier with the sample label corresponding to the feature vector in the sample label set T'.
Preferably, in step S2, the duration t has a value range of: t is more than or equal to 30ms and less than or equal to 50 ms.
Preferably, in step S3, the normalization formula is:
Figure BDA0003038602390000051
where max (×) represents the maximum of all elements in the set.
Preferably, in the step S4, the order p of the linear predictive coding model is 8 ≦ p ≦ 12;
in step S5, the MFCC dimension Q is 12.
Preferably, in step S6, the calculation formula of the short-time energy is
Figure BDA0003038602390000052
In the step Y6, in order to calculate sensitivity and specificity, the following convention is made:
1) appointing that the cough sound sample is a positive sample, and the non-cough sound sample is a negative sample;
2) if the sample label of a certain sample is 1 and the output of the linear SVM classifier taking the feature vector of the sample as input is 1, the sample is a true positive sample; if the sample label of a certain sample is-1 and the output of the linear SVM classifier taking the feature vector of the sample as input is-1, the sample is a true negative sample; if the sample label of a certain sample is-1 and the output of the linear SVM classifier taking the feature vector of the sample as input is 1, the sample is a false positive sample; if the sample label of a certain sample is 1 and the output of the linear SVM classifier taking the feature vector of the sample as input is-1, the sample is a false negative sample;
based on the convention, the calculation method of the sensitivity and the specificity comprises the following steps: counting the number of samples with true positive, true negative, false positive and false negative in the test set, and respectively recording as TP, TN, FP and FN, then the sensitivity is
Figure BDA0003038602390000053
With a specificity of
Figure BDA0003038602390000054
An automatic cough sound recognition device, characterized in that: the method comprises the following steps:
a sample obtaining module for obtaining a sample to be identified, wherein the sampling frequency of the sample to be identified is fs
The preprocessing module is used for carrying out end point detection on the sample, intercepting a fixed length sequence with the duration of t from a starting point and setting the fixed length sequence as a signal sequence r (k); wherein k is 1, 2, …, N; n is t × fsAnd N is a positive integer; normalizing the signal sequence r (k) to
Figure BDA0003038602390000061
A feature extraction module for encoding the signal sequence using linear predictive coding
Figure BDA0003038602390000063
Decomposition into a predicted signal sequence sr(k) And residual signal sequence er(k) (ii) a Wherein, the order of the linear predictive coding model is p; determining a predicted signal sequence sr(k) The Q-dimensional MFCC parameter MFCCr=[mr1,mr2,…mrQ]T(ii) a Obtaining a residual signal sequence er(k) Short-time energy en ofr(ii) a Wherein enrIs a scalar; the resulting MFCC parameter MFCC is measuredrAnd short-term energy enrConstructed as feature vectors cr=[mr1,mr2,…mrQ,enr]T
An identification module for identifying the obtained feature vector crInputting the data into a linear SVM classifier, and outputting f (c) according to the linear SVM classifierr) Judging the type of the sample to be identified: if f (c)r) If the value is 1, the sample is a cough sound sample; if f (c)r) If the value is-1, the sample is a non-cough sound sample; the linear SVM classifier is a linear SVM classifier obtained by training and testing an initial linear SVM classifier.
A storage medium, wherein the storage medium stores a computer program which, when executed by a processor, causes the processor to execute the above cough sound automatic recognition algorithm.
A computing device comprising a processor and a memory for storing a program executable by the processor, wherein the processor implements the cough sound automatic identification algorithm when executing the program stored in the memory.
The principle of the automatic identification algorithm of the invention is as follows:
linear predictive coding can enable deconvolution of the cough tone signal channel model and the excitation source in the time domain. The cough tone signal can be viewed as a response from an input sequence (characterizing the pulmonary airflow) exciting an all-pole system (characterizing the shape of the vocal tract), which acts as a vocal tract model with a transfer function:
Figure BDA0003038602390000062
wherein R (z) and U (z) are respectively the z-transforms of the signal sequence r (k) and the excitation source u (k). The relationship between r (k) and u (k) can be expressed as a difference equation:
Figure BDA0003038602390000071
therein, the system
Figure BDA0003038602390000072
Called linear predictor, i.e. the current value is predicted by a linear combination of p historical time values, aiReferred to as linear prediction coefficients. Linear prediction coefficients are a description of the vocal tract model, from which the time-domain sequence of the vocal tract model, i.e. the predicted signal sequence s, can be derivedr(k) In that respect While linearly predicted residual
Figure BDA0003038602390000073
The remainder of the original signal sequence minus the predicted signal sequence is described as the excitation source of the cough tone signal.
To find the optimal linear prediction coefficient, the sum of the squares of the errors of the linear prediction is defined as
Figure BDA0003038602390000074
Adjustment of parameter aiWhen E is minimum, the optimal linear prediction coefficient is obtained. To pair
Figure BDA0003038602390000075
Finding E for ajIf the partial derivative of (1) is 0, then
Figure BDA0003038602390000076
A total of p equations is obtained. Order to
Figure BDA0003038602390000077
The system of equations can be simplified to
Figure BDA0003038602390000078
The system of linear equations may be solved using an autocorrelation method. The solving method comprises the following steps: defining a windowed autocorrelation function of r (k) as
Figure BDA0003038602390000081
Thereby the device is provided with
Figure BDA0003038602390000082
Can be expressed as
Figure BDA0003038602390000083
Therefore, the formula (7) can be represented as
Figure BDA0003038602390000084
Equation (9) is called Yule-Walker equation, and the system of equations composed of such equations can be efficiently solved by the Levinson-Durbin algorithm. After the equation set is solved, the predicted signal sequence s can be obtained by linear prediction coefficient recursionr(k) And a residual signal sequence er(k) In that respect Thus, the vocal tract model and the excitation source realize the deconvolution in the time domain.
The initial stage of the cough sound contains the most important information in the cough sound, and if only the initial stage of the cough sound is analyzed, a time-varying system can be simplified into a constant system, and the complexity of a model is reduced. Practice shows that the duration t of the first stage of the cough sound is about 30ms ≦ t ≦ 50ms, so that the target segment intercepted in step S2 of the present invention is preferably about 30ms to 50ms after the original signal from the start point of the end point detection.
During signal acquisition, the distance between a sound source and equipment and the gain of the equipment influence the amplitude of the acquired digital signal, and further unnecessary random errors are brought to the calculation of short-time energy in the step S6. Therefore, step S3 of the present invention normalizes the signal to reduce the error.
Linear predictive coding treats the vocal tract as a resonant cavity, and the formants are the resonant frequencies of this cavity. The vocal tract model is an all-pole model with the number of poles being p, and one pair of poles corresponds to one formant. Practice shows that for most signals, it is sufficient to model the sound channels accurately if the order p of linear predictive coding is 8 ≦ p ≦ 12. Therefore, the value range of the order p of the linear predictive coding in step S4 of the present invention is preferably: p is more than or equal to 8 and less than or equal to 12.
The vocal tract model often selects cepstrum-like parameters that reflect the shape of the vocal tract as features, but generally the cepstrum parameters are particularly sensitive to noise. The human ear is able to hear speech signals from noisy background noise because the human inner ear base membrane modulates incoming signals. For different frequencies, signals within the corresponding critical bandwidth will cause vibrations at different locations on the base membrane. MFCC imitates human ear hearing through band-pass filter bank to reduce the influence of noise to speech, have stronger robustness to noise. Therefore, the present invention, at step S5, selects MFCC as the vocal tract model feature.
The separation of the vocal tract model from the excitation source makes the determination of the time-domain characteristics of the excitation source free from vocal tract resonance, and the determination of the characteristics of the excitation source is thus possible. The short-time energy is defined as:
Figure BDA0003038602390000091
it represents the excitation source intensity of the intercepted segment. The exciting source reflects the airflow impacting the sound channel during coughing, and the cough sound has the characteristics of paroxysmal property and high intensity, so that the signal energy is strong and concentrated, and the time domain characteristics of the cough sound can be effectively reflected by using short-time energy. Therefore, the invention S6 selects the short-time energy of the residual signal as the excitation source characteristic.
Compared with the prior art, the invention has the following advantages and beneficial effects:
1. the invention realizes the separation of the vocal tract model and the excitation source through the linear predictive coding, and overcomes the problem that the time domain characteristic of the excitation source is easily influenced by vocal tract resonance;
2. the invention increases the short-time energy of the excitation source as the characteristic of the cough sound identification, overcomes the problem that the conventional common automatic cough sound identification algorithm only utilizes the characteristic related to the sound channel information but often does not utilize the excitation source information, and has more comprehensive description on the cough sound;
3. the invention overcomes the problems of model difficult training caused by insufficient samples and unbalanced samples;
4. the method has good identification capability on the cough sound and high identification accuracy;
5. the method only identifies the initial stage of the cough sound, simplifies the model, reduces the calculated amount, has low requirement on hardware and is convenient to realize in wearable equipment.
Drawings
FIG. 1 is a flow chart of an embodiment of an automatic cough sound identification algorithm;
fig. 2(a) to 2(c) are respectively a time-domain waveform diagram of a cough sound, a short-time energy diagram directly calculated by a cough sound sequence, and a short-time energy diagram calculated by a cough sound residual signal sequence;
3(a) -3 (i) are time domain waveform diagrams of cough sound, impact sound and fricative sound, time domain waveform diagrams of residual signal sequence and short-time energy diagrams of residual signal sequence, respectively;
FIG. 4 is a classification hyperplane schematic of a linear SVM classifier.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and specific embodiments.
Example one
In the cough sound automatic identification algorithm, sample segments are intercepted and normalized for feature extraction; processing the sample by utilizing linear predictive coding to respectively obtain a predicted signal and a residual signal, wherein for a cough sound signal, the obtained predicted signal reflects the characteristics of a sound channel model, and the residual signal reflects the characteristics of an excitation source; respectively solving MFCC and short-time energy from the prediction signal and the residual signal and constructing a combination characteristic; and inputting the data into a linear SVM classifier to judge whether the sample type is cough sound or non-cough sound.
The working flow is shown in fig. 1, and comprises the following steps:
s1, obtaining a sample to be identified, wherein the sampling frequency of the sample to be identified is fs
S2, carrying out end point detection on the sample, intercepting a fixed length sequence with duration of t from a starting point, and setting the fixed length sequence as a signal sequence r (k); wherein k is 1, 2, …, N; n is t × fsAnd N is a positive integer; the value range of the duration t is preferably: t is more than or equal to 30ms and less than or equal to 50 ms.
S3, normalizing the signal sequence r (k) to obtain
Figure BDA0003038602390000101
The normalized formula is:
Figure BDA0003038602390000102
Figure BDA0003038602390000103
where max (×) represents the maximum of all elements in the set.
S4, using linear predictive coding to encode the signal sequence
Figure BDA0003038602390000104
Decomposition into a predicted signal sequence sr(k)And residual signal sequence er(k) (ii) a Wherein, the order of the linear predictive coding model is p; the order p is preferably 8. ltoreq. p.ltoreq.12.
S5, obtaining a predicted signal sequence Sr(k) The Q-dimensional MFCC parameter MFCCr=[mr1,mr2,…mrQ]T(ii) a The MFCC dimension Q is preferably 12.
Step S6, obtaining a residual signal sequence er(k) Short-time energy en ofr(ii) a Wherein enrIs a scalar; the short-time energy is calculated by the formula
Figure BDA0003038602390000111
S7, obtaining the MFCC parameter MFCCrAnd short-term energy enrConstructed as feature vectors cr=[mr1,mr2,…mrQ,enr]T
S8, obtaining the feature vector crInputting to a linear SVM classifier, classifying according to the linear SVM
Output f (c) of the devicer) Judging the type of the sample to be identified: if f (c)r) If the value is 1, the sample is a cough sound sample; if f (c)r) If the value is-1, the sample is a non-cough sound sample; the linear SVM classifier is a linear SVM classifier obtained by training and testing an initial linear SVM classifier.
The training and testing processing of the initial linear SVM classifier is as follows: the method comprises the following steps:
step Y1, at sampling frequency fsCollecting a plurality of cough sound samples and non-cough sound samples; setting the cough sound sample label to be 1, and setting the non-cough sound sample label to be-1; respectively dividing a training set and a testing set from a cough sound sample and a non-cough sound sample; for example, 2/3 samples accounting for the total number of the samples are respectively selected from the two types of samples to construct a training set; the remaining 1/3 samples, which account for the total number of respective samples, comprise the test set; the divided training set comprises m cough sound samples and n non-cough sound samples; the partitioned test set contains m 'cough sound samples and n' non-cough sound samples.
Y2, processing each sample of the training set according to S2 steps to obtain a cough sound signal sequence xi(k) And a sequence of non-cough tone signals yj(k) (ii) a Wherein, i is 1, 2, … m, j is 1, 2, … n.
Sequence of cough tone signals xi(k) And a sequence of non-cough tone signals yj(k) Respectively processing according to the steps from S3 to S7 to obtain the characteristic vector of the cough sound sample
Figure BDA0003038602390000112
And feature vectors of non-cough sound samples
Figure BDA0003038602390000113
The feature set of the training sample is
Figure BDA0003038602390000121
Which corresponds to a set of sample tags of
Figure BDA0003038602390000122
Wherein all cough sound samples are labeled as
Figure BDA0003038602390000123
All non-cough sound samples are labeled as
Figure BDA0003038602390000124
The index x of the eigenvector in the characteristic set phi and the sample label in the sample label set TiOr yjAnd correspond to each other.
Step Y3, establishing a linear SVM classifier with the decision function of f (c) sgn [ wTc+b](ii) a Wherein c is a feature vector of the signal to be decided; f (c) is a decision value of the feature vector c, f (c) ═ 1 indicates that the signal to be decided is a cough sound, and f (c) ═ 1 indicates that the signal to be decided is a non-cough sound; w is a weight vector; b is the hyperplane offset.
And Y4, taking each feature vector of the feature set phi as input, taking the sample label corresponding to the feature vector in the sample label set T as expected output, and training the established linear SVM classificationDerived decision function f (c) sgn [ w*Tc+b*](ii) a Wherein, w*And b*The optimal weight vector and the hyperplane offset obtained by training are respectively.
Step Y5, processing each sample of the test set according to the step Y2 respectively to obtain a feature vector of each sample; obtaining a feature set phi 'of the test sample and a corresponding sample label set T';
step Y6, taking each feature vector in the feature set phi 'of the test sample as the input of a linear SVM classifier, and comparing the output f (c) of the linear SVM classifier with the sample label corresponding to the feature vector in the sample label set T'; calculating the sensitivity and specificity according to the comparison result;
judging whether the sensitivity and the specificity meet the preset requirements:
if so, ending the training and testing;
otherwise, jumping to the step Y4 to retrain the decision function f (c) of the linear SVM classifier again*Tc+b*]。
To calculate sensitivity and specificity, the following convention is made:
1) appointing that the cough sound sample is a positive sample, and the non-cough sound sample is a negative sample;
2) if the sample label of a certain sample is 1 and the output of the linear SVM classifier taking the feature vector of the sample as input is 1, the sample is a true positive sample; if the sample label of a certain sample is-1 and the output of the linear SVM classifier taking the feature vector of the sample as input is-1, the sample is a true negative sample; if the sample label of a certain sample is-1 and the output of the linear SVM classifier taking the feature vector of the sample as input is 1, the sample is a false positive sample; if the sample label of a certain sample is 1 and the output of the linear SVM classifier taking the feature vector of the sample as input is-1, the sample is a false negative sample;
based on the convention, the calculation method of the sensitivity and the specificity comprises the following steps: counting the number of samples with true positive, true negative, false positive and false negative in the test set, and respectively recording as TP, TN, FP and FN, then the sensitivity is
Figure BDA0003038602390000131
With a specificity of
Figure BDA0003038602390000132
If the sensitivity and specificity of the obtained test result reach the preset requirements, the model can be applied to cough sound recognition.
Fig. 2(a) is a time-domain waveform diagram of a cough sound, fig. 2(b) is a short-time energy diagram obtained by directly calculating the cough sound sequence of fig. 2(a), and fig. 2(c) is a short-time energy diagram obtained by decomposing the cough sound sequence into a prediction signal sequence and a residual signal sequence by linear predictive coding and then calculating the residual signal sequence.
Fig. 3(a) is a time domain waveform diagram of a cough sound, fig. 3(b) is a time domain waveform diagram of a cough sound residual signal sequence obtained by a linear predictive coding process, and fig. 3(c) is a short-term energy diagram corresponding to the residual signal sequence of fig. 3 (b); fig. 3(d) is a time domain waveform diagram of the impact sound, fig. 3(e) is a time domain waveform diagram of the impact sound residual signal sequence obtained by the linear predictive coding process, and fig. 3(f) is a short-time energy diagram corresponding to the residual signal sequence of fig. 3 (e); fig. 3(g) is a time domain waveform diagram of the fricatives, fig. 3(h) is a time domain waveform diagram of the fricatives residual signal sequence obtained by the linear predictive coding process, and fig. 3(i) is a short-time energy diagram corresponding to the residual signal sequence of fig. 3 (h).
The principle of the computational algorithm of the invention is:
as shown in fig. 2(a) to 2(c), during the first stage of the cough, due to the contraction of the abdominal and diaphragm muscles, the lungs exhale a large amount of gas instantaneously, the short-term energy of the residual signal shows a peak as in the front part of fig. 2(c), and then the lung gas is still exhausted continuously and slowly, the short-term energy of the residual signal tends to be flat, and during the third stage, due to the periodic vibration of the vocal cords caused by the closing of the glottis, the small fluctuation appears again, and fig. 2(c) reflects the characteristics of the impulsive airflow during the cough very well. While fig. 2(b) is affected by the vocal tract resonance, a high peak with a large amplitude appears after the first stage, which is obviously inconsistent with the impulsive airflow during the cough, and the excitation source of the cough sound signal cannot be accurately described. Therefore, before calculating the time domain feature of the excitation source, the deconvolution of the acoustic channel and the excitation source should be performed, and if the short-time energy of the original signal is directly used as the feature, a large error exists.
Comparing the cough sound residual signal of fig. 3(b) and the impact sound residual signal of fig. 3(e), the amplitudes of which are close in the initial burst phase, the residual signal of the following impact sound decays rapidly, while the residual signal of the cough sound lasts for a longer time because of the continuous airflow impact during the cough, while the impact sound is excited only at one instant of the impact. Comparing the short-term energy of the cough residual signal of fig. 3(c) with the short-term energy of the impact residual signal of fig. 3(f), the two peaks appear, except that the former has a longer peak duration and is maintained at a certain level after attenuation, while the latter has only one impact with a short duration. Generally, because the duration of an energy burst is short, the short-term energy of a residual signal of a burst sound such as a clap and an impact is generally smaller than that of a cough sound, and therefore, the cough sound and various burst sounds can be effectively distinguished through the short-term energy of the residual signal.
Comparing the cough sound waveform of fig. 3(a) with the frictional sound waveform of fig. 3(g), the two waveforms are relatively close and difficult to distinguish clearly. Comparing the cough sound residual signal of fig. 3(b) with the fricative sound residual signal of fig. 3(h), both have continuous amplitudes, but the former amplitude is significantly higher than the latter due to the fact that the air flow impact intensity at cough is much stronger than the excitation of normal fricative sound. Comparing the short-term energy of the cough residual signal of fig. 3(c) with the short-term energy of the fricative residual signal of fig. 3(i), the latter has no distinct peak and is lower in amplitude overall than the former. Generally, although the friction sound such as the friction of the stool and the fingernail has a continuous excitation effect, the short-time energy of the residual signal is generally smaller than that of the cough sound due to the low burst intensity, so that the cough sound and various friction sounds can be effectively distinguished through the short-time energy of the residual signal. In summary, it is appropriate that the step S6 of the present invention selects the short-term energy of the residual signal as the excitation source characteristic.
The SVM is a data mining method based on the statistical learning theory, which constructs a hyperplane with optimal decision based on the principle of minimizing structural risk, as shown in fig. 4, so that the hyperplane can ensure the classification accuracy and maximize the distance between the convex set (represented by support vector) of two types of samples and the hyperplane. In other words, the SVM classification hyperplane is determined by the support vector, and most training samples except the support vector do not need to be reserved after model training is completed. The characteristic of the SVM enables a large number of samples to participate in model training, and the influence of the missing of part of samples on the final model is small, so that the SVM has inherent advantages in the aspect of processing small sample statistics, can provide good popularization capability, and is insensitive to the problem of sample imbalance. In theory, SVM can achieve optimal classification of linearly separable problems, while for linearly inseparable problems, by introducing an inner product kernel function, the input space can be transformed to a high-dimensional space where it is linearly separable. Therefore, in step S8, a linear SVM classifier is selected as the classifier.
The following describes the training process of the linear SVM classifier in the automatic recognition algorithm of the present invention with specific examples:
step Y1, with fsCollecting a plurality of cough sound samples and non-cough sound samples for 8000Hz, and respectively selecting 2/3 samples accounting for the total number of the samples from the two samples to construct a training set, wherein the obtained training set comprises 104 cough sound samples and 636 non-cough sound samples; the remaining 1/3 samples, which account for the total number of respective samples, constitute a test set, the resulting test set containing 51 cough sound samples and 320 non-cough sound samples; here, the cough sound sample flag is set to 1, and the non-cough sound sample flag is set to-1.
And Y2, respectively carrying out end point detection on each sample of the training set, and intercepting the sample with the length of N being 400 from the starting point and the duration of N being 400
Figure BDA0003038602390000151
Obtaining the sounding signal sequence x from the fixed length sequencei(k) And a sequence of non-cough tone signals yj(k) (ii) a Wherein, i is 1, 2, … 104, j is 1, 2, … 636, k is 1, 2, …, 400;
sequence of cough tone signals xi(k) Normalized to obtain
Figure BDA0003038602390000161
Sequence of non-cough tone signals yj(k) Normalized to obtain
Figure BDA0003038602390000162
Wherein max (×) represents the maximum of all elements in the set;
method for coding a cough signal using linear predictive coding
Figure BDA0003038602390000163
Decomposition into predicted signal sequences
Figure BDA0003038602390000164
And residual signal sequence
Figure BDA0003038602390000165
Sequencing non-cough tone signals
Figure BDA0003038602390000166
Decomposition into predicted signal sequences
Figure BDA0003038602390000167
And residual signal sequence
Figure BDA0003038602390000168
Wherein, the order of the linear predictive coding is p-12;
predicting signal sequences for cough sounds
Figure BDA0003038602390000169
Solving 12-dimensional MFCC parameters
Figure BDA00030386023900001610
Predicting signal sequences for non-cough sounds
Figure BDA00030386023900001611
Solving 12-dimensional MFCC parameters
Figure BDA00030386023900001612
As a cough residual signal sequence
Figure BDA00030386023900001613
Finding short-time energy
Figure BDA00030386023900001614
As non-coughing tone residual signal sequences
Figure BDA00030386023900001615
Finding short-time energy
Figure BDA00030386023900001616
Constructing MFCC parameters and short-time energy resulting from processing of cough sounds as a combined feature
Figure BDA00030386023900001617
Figure BDA00030386023900001618
Constructing MFCC parameters and short-time energy resulting from processing non-cough sounds as a combined signature
Figure BDA00030386023900001619
Thus, the feature set of the training sample is obtained
Figure BDA00030386023900001620
Which corresponds to a set of sample tags of
Figure BDA00030386023900001621
Wherein all cough sound samples are labeled as
Figure BDA00030386023900001622
All non-cough sound samples are labeled as
Figure BDA00030386023900001623
The characteristic vector in phi and the sample label in T are indexed by xiOr yjOne-to-one correspondence。
Step Y3, establishing a linear SVM classifier with the decision function of f (c) sgn [ wTc+b](ii) a Wherein c is a feature vector of the signal to be decided; f (c) is a decision value of the feature vector c, f (c) ═ 1 indicates that the signal to be decided is a cough sound, and f (c) ═ 1 indicates that the signal to be decided is a non-cough sound; w is a weight vector; b is the hyperplane offset.
And step Y4, taking each feature vector of the feature set phi obtained in the step Y2 as input, taking a sample label corresponding to the feature vector in the sample label set T as expected output, and training a decision function f (c) of the linear SVM classifier established in the step Y3 to obtain sgn [ w*Tc+b*](ii) a Wherein the optimal weight vector is w*=[-0.0901,-0.1392,-0.1756,-0.0798,-0.0487,0.0179,-0.0777,-0.2485,0.0359,-0.2898,0.0889,-0.2116,0.1081]TThe hyperplane offset is b*=-6.9896。
And Y5, calculating the feature vectors of the test samples in the test set according to the method of the Y2 step, and constructing a feature set phi 'and a corresponding test sample label set T'.
Y6, taking each feature vector in the test sample feature set phi 'as the input of the linear SVM classifier obtained by the training in the Y4 step, comparing the output f (c) of the linear SVM classifier with the corresponding sample label of the feature vector in the test sample label set T', and calculating the sensitivity and the specificity according to the comparison result; the sensitivity obtained is Sn ═ 0.961, and the specificity is Sp ═ 0.981, meeting the predetermined requirements, and the model can be applied to cough sound recognition.
When the linear SVM classifier obtained by the training and testing processing is applied to the cough sound automatic recognition algorithm, a sample to be recognized is obtained, and the processing is carried out according to steps from S2 to S7 to obtain a feature vector cr=[-10.3715,-14.0365,-19.3825,-6.7138,-0.1099,-4.5139,-1.5521,-8.1192,-1.9553,-1.6719,-0.5847,0.0388,25.6368](ii) a Substituting the linear SVM classifier into a linear SVM classifier obtained by training and testing to obtain classifier output f (c)r) When the value is 1, the sample is judged as a cough sound sample.
In order to implement the cough sound automatic identification algorithm, the embodiment further provides a cough sound automatic identification device, including:
a sample obtaining module for obtaining a sample to be identified, wherein the sampling frequency of the sample to be identified is fs
The preprocessing module is used for carrying out end point detection on the sample, intercepting a fixed length sequence with the duration of t from a starting point and setting the fixed length sequence as a signal sequence r (k); wherein k is 1, 2, …, N; n is t × fsAnd N is a positive integer; normalizing the signal sequence r (k) to
Figure BDA0003038602390000182
A feature extraction module for encoding the signal sequence using linear predictive coding
Figure BDA0003038602390000181
Decomposition into a predicted signal sequence sr(k) And residual signal sequence er(k) (ii) a Wherein, the order of the linear predictive coding model is p; determining a predicted signal sequence sr(k) The Q-dimensional MFCC parameter MFCCr=[mr1,mr2,…mrQ]T(ii) a Obtaining a residual signal sequence er(k) Short-time energy en ofr(ii) a Wherein enrIs a scalar; the resulting MFCC parameter MFCC is measuredrAnd short-term energy enrConstructed as feature vectors cr=[mr1,mr2,…mrQ,enr]T
An identification module for identifying the obtained feature vector crInputting the data into a linear SVM classifier, and outputting f (c) according to the linear SVM classifierr) Judging the type of the sample to be identified: if f (c)r) If the value is 1, the sample is a cough sound sample; if f (c)r) If the value is-1, the sample is a non-cough sound sample; the linear SVM classifier is a linear SVM classifier obtained by training and testing an initial linear SVM classifier.
Example two
A storage medium of the present embodiment stores a computer program, and the computer program, when executed by a processor, causes the processor to execute the cough sound automatic recognition algorithm according to the first embodiment.
EXAMPLE III
The computing device of the embodiment includes a processor and a memory for storing a processor executable program, and when the processor executes the program stored in the memory, the cough sound automatic identification algorithm described in the first embodiment is implemented.
The above embodiments are preferred embodiments of the present invention, but the present invention is not limited to the above embodiments, and any other changes, modifications, substitutions, combinations, and simplifications which do not depart from the spirit and principle of the present invention should be construed as equivalents thereof, and all such changes, modifications, substitutions, combinations, and simplifications are intended to be included in the scope of the present invention.

Claims (10)

1. An automatic cough sound recognition algorithm, comprising: the method comprises the following steps:
s1, obtaining a sample to be identified, wherein the sampling frequency of the sample to be identified is fs
S2, carrying out end point detection on the sample, intercepting a fixed length sequence with duration of t from a starting point, and setting the fixed length sequence as a signal sequence r (k); wherein k is 1, 2, …, N; n is t × fsAnd N is a positive integer;
s3, normalizing the signal sequence r (k) to obtain
Figure FDA0003038602380000011
S4, using linear predictive coding to encode the signal sequence
Figure FDA0003038602380000012
Decomposition into a predicted signal sequence sr(k) And residual signal sequence er(k) (ii) a Wherein, the order of the linear predictive coding model is p;
s5, obtaining a predicted signal sequence Sr(k) The Q-dimensional MFCC parameter MFCCr=[mr1,mr2,…mrQ]T
The step of step S6 is executed,obtaining a residual signal sequence er(k) Short-time energy en ofr(ii) a Wherein enrIs a scalar;
s7, obtaining the MFCC parameter MFCCrAnd short-term energy enrConstructed as feature vectors cr=[mr1,mr2,…mrQ,enr]T
S8, obtaining the feature vector crInputting the data into a linear SVM classifier, and outputting f (c) according to the linear SVM classifierr) Judging the type of the sample to be identified: if f (c)r) If the value is 1, the sample is a cough sound sample; if f (c)r) If the value is-1, the sample is a non-cough sound sample; the linear SVM classifier is a linear SVM classifier obtained by training and testing an initial linear SVM classifier.
2. The cough sound automatic recognition algorithm of claim 1, wherein: the training and testing processing of the initial linear SVM classifier is as follows: the method comprises the following steps:
step Y1, at sampling frequency fsCollecting a plurality of cough sound samples and non-cough sound samples; setting the cough sound sample label to be 1, and setting the non-cough sound sample label to be-1; respectively dividing a training set and a testing set from a cough sound sample and a non-cough sound sample;
y2, processing each sample of the training set according to the steps from S2 to S7 respectively to obtain a feature vector of each sample;
step Y3, establishing a linear SVM classifier with the decision function of f (c) sgn [ wTc+b](ii) a Wherein c is a feature vector of the signal to be decided; f (c) is a decision value of the feature vector c, f (c) ═ 1 indicates that the signal to be decided is a cough sound, and f (c) ═ 1 indicates that the signal to be decided is a non-cough sound; w is a weight vector; b is the hyperplane offset;
step Y4, taking the feature vector of the sample as input, taking the sample label corresponding to the sample as expected output, training the established linear SVM classifier to obtain the decision function f (c) ═ sgn [ w*Tc+b*](ii) a Wherein, w*And b*Respectively obtained by trainingThe optimal weight vector and the hyperplane offset;
step Y5, processing each sample of the test set according to the step Y2 respectively to obtain a feature vector of each sample;
y6, inputting the feature vector of the sample into a trained linear SVM classifier; comparing the output f (c) of the linear SVM classifier with a sample label corresponding to the sample, and calculating the sensitivity and the specificity according to the comparison result;
judging whether the sensitivity and the specificity meet the preset requirements:
if so, ending the training and testing;
otherwise, jumping to the step Y4 to retrain the decision function f (c) of the linear SVM classifier again*Tc+b*]。
3. The cough sound automatic recognition algorithm of claim 2, wherein: the step Y1, namely, the divided training set comprises m cough sound samples and n non-cough sound samples;
the step Y2 is as follows: processing each sample of the training set according to the step S2 to obtain a cough sound signal sequence xi(k) And a sequence of non-cough tone signals yj(k) (ii) a Wherein, i is 1, 2, … m, j is 1, 2, … n;
sequence of cough tone signals xi(k) And a sequence of non-cough tone signals yj(k) Respectively processing according to the steps from S3 to S7 to obtain the characteristic vector of the cough sound sample
Figure FDA0003038602380000021
And feature vectors of non-cough sound samples
Figure FDA0003038602380000031
The feature set of the training sample is
Figure FDA0003038602380000032
Which corresponds to a set of sample tags of
Figure FDA0003038602380000033
Wherein all cough sound samples are labeled as
Figure FDA0003038602380000034
All non-cough sound samples are labeled as
Figure FDA0003038602380000035
The index x of the eigenvector in the characteristic set phi and the sample label in the sample label set TiOr yjOne-to-one correspondence is realized;
and Y4, taking each feature vector of the feature set phi as input, taking a sample label corresponding to the feature vector in the sample label set T as expected output, and training the established linear SVM classifier to obtain a decision function f (c) ═ sgn [ w ]*Tc+b*];
The step Y5, processing each sample of the test set according to the step Y2 respectively to obtain a feature vector of each sample; obtaining a feature set phi 'of the test sample and a corresponding sample label set T';
and Y6, taking each feature vector in the feature set phi 'of the test sample as the input of the linear SVM classifier, and comparing the output f (c) of the linear SVM classifier with the sample label corresponding to the feature vector in the sample label set T'.
4. The cough sound automatic recognition algorithm of claim 1, wherein: in the step S2, the value range of the duration t is: t is more than or equal to 30ms and less than or equal to 50 ms.
5. The cough sound automatic recognition algorithm of claim 1, wherein: in the step S3, the normalization formula is:
Figure FDA0003038602380000036
where max (×) represents the maximum of all elements in the set.
6. The cough sound automatic recognition algorithm of claim 1, wherein: in the step S4, the order p of the linear predictive coding model is more than or equal to 8 and less than or equal to 12;
in step S5, the MFCC dimension Q is 12.
7. The cough sound automatic recognition algorithm of claim 1, wherein: in the step S6, the formula for calculating the short-time energy is
Figure FDA0003038602380000041
8. An automatic cough sound recognition device, characterized in that: the method comprises the following steps:
a sample obtaining module for obtaining a sample to be identified, wherein the sampling frequency of the sample to be identified is fs
The preprocessing module is used for carrying out end point detection on the sample, intercepting a fixed length sequence with the duration of t from a starting point and setting the fixed length sequence as a signal sequence r (k); wherein k is 1, 2, …, N; n is t × fsAnd N is a positive integer; normalizing the signal sequence r (k) to
Figure FDA0003038602380000042
A feature extraction module for encoding the signal sequence using linear predictive coding
Figure FDA0003038602380000043
Decomposition into a predicted signal sequence sr(k) And residual signal sequence er(k) (ii) a Wherein, the order of the linear predictive coding model is p; determining a predicted signal sequence sr(k) The Q-dimensional MFCC parameter MFCCr=[mr1,mr2,…mrQ]T(ii) a Obtaining a residual signal sequence er(k) Short-time energy en ofr(ii) a Wherein enrIs a scalar; the resulting MFCC parameter MFCC is measuredrAnd short-term energy enrConstructed as feature vectors cr=[mr1,mr2,…mrQ,enr]T
An identification module for identifying the obtained feature vector crInputting the data into a linear SVM classifier, and outputting f (c) according to the linear SVM classifierr) Judging the type of the sample to be identified: if f (c)r) If the value is 1, the sample is a cough sound sample; if f (c)r) If the value is-1, the sample is a non-cough sound sample; the linear SVM classifier is a linear SVM classifier obtained by training and testing an initial linear SVM classifier.
9. A storage medium having stored thereon a computer program which, when executed by a processor, causes the processor to execute the cough sound automatic recognition algorithm of any one of claims 1-7.
10. A computing device comprising a processor and a memory for storing processor-executable programs, wherein the processor, when executing a program stored in the memory, implements the cough tone automatic recognition algorithm of any one of claims 1-7.
CN202110450820.0A 2021-04-26 2021-04-26 Cough sound automatic identification algorithm, device, medium and equipment Active CN113158916B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110450820.0A CN113158916B (en) 2021-04-26 2021-04-26 Cough sound automatic identification algorithm, device, medium and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110450820.0A CN113158916B (en) 2021-04-26 2021-04-26 Cough sound automatic identification algorithm, device, medium and equipment

Publications (2)

Publication Number Publication Date
CN113158916A true CN113158916A (en) 2021-07-23
CN113158916B CN113158916B (en) 2022-04-22

Family

ID=76870648

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110450820.0A Active CN113158916B (en) 2021-04-26 2021-04-26 Cough sound automatic identification algorithm, device, medium and equipment

Country Status (1)

Country Link
CN (1) CN113158916B (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101894551A (en) * 2010-07-02 2010-11-24 华南理工大学 Method and device for automatically identifying cough
US20150073306A1 (en) * 2012-03-29 2015-03-12 The University Of Queensland Method and apparatus for processing patient sounds
US20190096389A1 (en) * 2017-09-28 2019-03-28 International Business Machines Corporation Feature extraction of acoustic signals
CN111179967A (en) * 2019-12-17 2020-05-19 华南理工大学 Linear classification algorithm, medium and equipment for true and false cough sounds of cervical spinal cord injury patients

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101894551A (en) * 2010-07-02 2010-11-24 华南理工大学 Method and device for automatically identifying cough
US20150073306A1 (en) * 2012-03-29 2015-03-12 The University Of Queensland Method and apparatus for processing patient sounds
US20190096389A1 (en) * 2017-09-28 2019-03-28 International Business Machines Corporation Feature extraction of acoustic signals
CN111179967A (en) * 2019-12-17 2020-05-19 华南理工大学 Linear classification algorithm, medium and equipment for true and false cough sounds of cervical spinal cord injury patients

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
朱春媚等: ""基于改进反向Mel频率倒谱系数的咳嗽干湿性自动分类"", 《生物医学工程学杂志》 *

Also Published As

Publication number Publication date
CN113158916B (en) 2022-04-22

Similar Documents

Publication Publication Date Title
CN107657964B (en) Depression auxiliary detection method and classifier based on acoustic features and sparse mathematics
Dibazar et al. Feature analysis for automatic detection of pathological speech
JP4959727B2 (en) Speech recognition using speaker adaptation and registration by pitch
Panek et al. Acoustic analysis assessment in speech pathology detection
US11672472B2 (en) Methods and systems for estimation of obstructive sleep apnea severity in wake subjects by multiple speech analyses
CN109192221A (en) It is a kind of that phonetic decision Parkinson severity detection method is used based on cluster
El Emary et al. Towards developing a voice pathologies detection system
CN112190253A (en) Classification method for severity of obstructive sleep apnea
CN113782032B (en) Voiceprint recognition method and related device
US20210090734A1 (en) System, device and method for detection of valvular heart disorders
CN113158916B (en) Cough sound automatic identification algorithm, device, medium and equipment
Perez Carrillo et al. Learning and extraction of violin instrumental controls from audio signal
Porieva et al. Investigation of lung sounds features for detection of bronchitis and COPD using machine learning methods
Costa et al. Pathological voice discrimination using cepstral analysis, vector quantization and hidden Markov models
Liu et al. Classifying respiratory sounds using electronic stethoscope
Neto et al. Feature estimation for vocal fold edema detection using short-term cepstral analysis
JP2020513908A (en) How to characterize sleep-disordered breathing
Paul et al. Speech recognition of throat microphone using MFCC approach
Costa et al. Parametric cepstral analysis for pathological voice assessment
Mostaani et al. On Breathing Pattern Information in Synthetic Speech.
Tucker An ecological approach to the classification of transient underwater acoustic events: Perceptual experiments and auditory models
Chudasama et al. Voice Based Pathology Detection from Respiratory Sounds using Optimized Classifiers
Ghoraani et al. Pathological speech signal analysis using time-frequency approaches
Atmaja et al. Cross-dataset COVID-19 Transfer Learning with Cough Detection, Cough Segmentation, and Data Augmentation
Mamedov et al. Exploring Deep Learning Approaches to Cleft Lip and Palate Speech

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant