CN113537412A - Lung disease auscultation system based on convolutional neural network, signal processing method and equipment - Google Patents

Lung disease auscultation system based on convolutional neural network, signal processing method and equipment Download PDF

Info

Publication number
CN113537412A
CN113537412A CN202111077379.2A CN202111077379A CN113537412A CN 113537412 A CN113537412 A CN 113537412A CN 202111077379 A CN202111077379 A CN 202111077379A CN 113537412 A CN113537412 A CN 113537412A
Authority
CN
China
Prior art keywords
data
neural network
convolutional neural
frequency spectrum
low
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111077379.2A
Other languages
Chinese (zh)
Other versions
CN113537412B (en
Inventor
郭亮
张淼
刘建亚
马悦宁
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Liangrui Artificial Intelligence Jinan Co ltd
Original Assignee
Shandong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University filed Critical Shandong University
Priority to CN202111077379.2A priority Critical patent/CN113537412B/en
Publication of CN113537412A publication Critical patent/CN113537412A/en
Application granted granted Critical
Publication of CN113537412B publication Critical patent/CN113537412B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/243Classification techniques relating to the number of classes
    • G06F18/24323Tree-organised classifiers
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B7/00Instruments for auscultation
    • A61B7/003Detecting lung or respiration noise
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B7/00Instruments for auscultation
    • A61B7/02Stethoscopes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/254Fusion techniques of classification results, e.g. of results related to same input data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/259Fusion by voting

Abstract

The invention provides a pulmonary disease auscultation system based on a convolutional neural network, a signal processing method and equipment, wherein the system comprises: the preprocessing module is configured to sequentially perform normalization processing on the acquired breath sound signals; the data conversion module is configured to extract audio features from the normalized data and generate Mel frequency spectrum data; a data enhancement module configured to perform data amplification on the generated mel-frequency spectrum data by using a data enhancement model; the convolutional neural network module is configured to perform feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model; and the feature classification module is configured to classify the extracted features by using the trained low-difference forest classifier to obtain a classification result. The method can realize accurate feature classification aiming at small samples, has small system operation amount, can realize deployment on embedded computer equipment, and can realize remote diagnosis and treatment.

Description

Lung disease auscultation system based on convolutional neural network, signal processing method and equipment
Technical Field
The invention relates to the technical field related to intelligent medical information technology, in particular to a pulmonary disease auscultation system based on a convolutional neural network, a signal processing method and equipment.
Background
The statements in this section merely provide background information related to the present disclosure and may not necessarily constitute prior art.
Respiratory diseases are a widespread problem worldwide. The mild case manifests as cough, chest pain, and respiratory affection, and the severe case manifests as dyspnea, hypoxia, and even death due to respiratory failure and exhaustion. Smoking is the most common cause of respiratory illness, but is sometimes caused by genetic and environmental factors.
Automatic classification of breath sounds is likely to detect abnormalities in the early stages of respiratory dysfunction, thereby improving the effectiveness of the decision. At present, the mainstream diagnosis method in hospitals is to perform preliminary diagnosis on the lung of a patient through imaging examination or breath sound. However, both diagnostic methods have some problems. The former has accurate data and is easy to judge, but needs related professional equipment; although the latter has low cost, the requirements on the experience level of doctors are high, professional equipment or doctors with abundant experience cannot be timely configured in many areas, and the accurate diagnosis of respiratory diseases is greatly influenced.
Part of research and development personnel begin to use a deep learning model to perform auxiliary diagnosis, and the existing model part has the defects that the algorithm is very complex, the network structure is huge, the model is very difficult to be deployed at a mobile terminal, and the model cannot be applied to a computer with low price, so that the model cannot benefit the vast underdeveloped areas.
Disclosure of Invention
The invention provides a pulmonary disease auscultation system based on a convolutional neural network, a signal processing method and equipment for solving the problems, wherein the acquired respiratory sound signals are subjected to preprocessing, data conversion and data enhancement, so that the accuracy of feature extraction and the accuracy of classification results can be improved for small sample data.
In order to achieve the purpose, the invention adopts the following technical scheme:
in a first aspect, the invention provides a convolutional neural network-based pulmonary disease auscultation system, comprising:
the preprocessing module is configured to preprocess the acquired breath sound signals and normalize the breath sound signals in sequence;
the data conversion module is configured to extract audio features from the normalized data and generate Mel frequency spectrum data;
a data enhancement module configured to perform data amplification on the generated mel-frequency spectrum data by using a data enhancement model;
the convolutional neural network module is configured to perform feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and the feature classification module is configured to classify the extracted features by using the trained low-difference forest classifier to obtain a classification result.
As a further improvement, the method further comprises the following steps:
a training model module configured to train the convolutional neural network module and the low-diversity forest classifier using a public or autonomously acquired breath sound signal data set to find optimal model parameters.
As an alternative implementation, the low-difference forest classifier takes K decision trees as basic classifiers, and a combined classifier is obtained after ensemble learning, and when a sample to be classified is given, the classification result output by the low-difference forest classifier is voted and determined by the classification result of each decision tree.
As an alternative embodiment, an acquisition device is further included for acquiring and storing the breathing sound signal.
As an alternative embodiment, the deep learning model and the low-dissimilarity forest classifier are cascaded;
as an alternative embodiment, the training process of the low-difference forest classifier includes:
(a) generating a low-difference degree sequence based on a low-difference degree sequence sampling method according to the number of samples of the training data set;
(b) according to the principle of ascending or descending, acquiring the sequence numbers of all elements of the low-difference sequence, and generating a sequence number sequence;
(c) taking samples with set sizes of the training data set as training sets of corresponding decision trees;
(d) setting the number of decision trees of the forest with low diversity;
(e) random integers in the sample number range are used as initial sample indexes of a decision tree;
(f) starting from the random integer, taking a set number of continuous elements from a serial number sequence;
(g) taking out a corresponding number of samples from the input rearranged and deleted T-column data according to the taken-out elements as sample serial numbers to form a training sample of a decision tree;
(h) constructing a decision tree, and training the decision tree by using the training sample;
(i) and (e) repeating the steps (e) - (h), and constructing and training the decision trees according with the number of the decision trees.
As an alternative, the data enhancement model is a variational auto-encoder, and when the auto-encoder is trained, the data information is encoded in the latent space and then decoded to reconstruct the original data.
In an alternative embodiment, the deep learning model is a lightweight convolutional neural network, and includes an input layer, two convolutional layers, a pooling layer, a convolutional layer, a flattening layer, and a fully-connected layer, which are connected in sequence.
The invention provides a signal processing method of a pulmonary disease auscultation system based on a convolutional neural network, which comprises the following steps:
preprocessing the acquired breath sound signals, and sequentially normalizing the breath sound signals;
extracting audio features from the normalized data to generate Mel frequency spectrum data;
amplifying the data of the generated Mel frequency spectrum data by using a data enhancement model;
carrying out feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and classifying the extracted features by using the trained low-difference forest classifier to obtain a classification result.
A third aspect of the present invention provides a terminal device comprising a memory and a processor, and computer instructions stored on the memory and executed on the processor, the computer instructions, when executed by the processor, performing the steps of:
preprocessing the acquired breath sound signals, and sequentially normalizing the breath sound signals;
extracting audio features from the normalized data to generate Mel frequency spectrum data;
amplifying the data of the generated Mel frequency spectrum data by using a data enhancement model;
carrying out feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and classifying the extracted features by using the trained low-difference forest classifier to obtain a classification result.
Compared with the prior art, the invention has the beneficial effects that:
the invention uses breath sound as the diagnosis basis, can realize accurate characteristic classification aiming at small samples, has small system operation amount, uses a lightweight class classification model with lightweight class and high accuracy, can realize deployment on low-cost embedded computer equipment, can realize remote diagnosis and treatment, and realizes artificial intelligent medical service in remote or underdeveloped areas.
The invention adopts the Mel frequency spectrum analysis method in the aspect of feature processing, the extracted features are obviously superior to the results of most low-order feature extraction methods, and the further feature extraction of a lightweight network is facilitated.
The invention uses the deep learning method of fusing the low-difference degree series forest, has very high accuracy and stability while having very low requirements on computing resources, and ensures that the whole model can be deployed on cheap embedded equipment.
Advantages of additional aspects of the invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, are included to provide a further understanding of the invention, and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description, serve to explain the invention and not to limit the invention.
FIG. 1(a) is a flow chart of a model training process of an auscultation system according to embodiment 1 of the present invention;
fig. 1(b) is a flowchart of a model deployment process of the auscultation system of embodiment 1 of the present invention;
FIG. 2 is a network structure diagram of a low-diversity forest classifier according to embodiment 1 of the present invention;
FIG. 3 is a schematic diagram of a deep neural network and low-diversity forest classifier cascade model in embodiment 1 of the present invention;
FIG. 4(a) is a model training flowchart of embodiment 1 of the present invention;
FIG. 4(b) is a flowchart of model test in example 1 of the present invention;
FIG. 5 is a schematic diagram of an audio file waveform of an example of embodiment 1 of the present invention;
FIG. 6 is a Mel-spectral diagram of exemplary audio of embodiment 1 of the present invention;
fig. 7 is a system architecture diagram of the auscultation system of embodiment 1 of the invention.
The specific implementation mode is as follows:
the invention is further described with reference to the following figures and examples.
It is to be understood that the following detailed description is exemplary and is intended to provide further explanation of the invention as claimed. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of exemplary embodiments according to the invention. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise. It should be noted that the embodiments and features of the embodiments may be combined with each other without conflict. The embodiments will be described in detail below with reference to the accompanying drawings.
Example 1
In one or more embodiments, as shown in fig. 1-7, a convolutional neural network-based pulmonary disease auscultation system is provided, comprising:
the preprocessing module is configured to preprocess the acquired breath sound signals and normalize the breath sound signals in sequence;
a data conversion module configured to extract audio features from the normalized data and generate Mel Spectrogram (Mel Spectrogram) data;
a data enhancement module configured to perform data amplification on the generated mel-frequency spectrum data by using a data enhancement model;
the convolutional neural network module is configured to perform feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and the feature classification module is configured to classify the extracted features by using the trained low-difference forest classifier to obtain a classification result.
In the embodiment, data conversion and data enhancement are performed in the characteristic extraction stage of the data, so that the processing efficiency of the small sample data can be improved, the signal processing efficiency can be improved, the identification failure rate can be reduced, and the adaptability of system application can be improved.
As a further improvement, the method further comprises the following steps:
a training model module configured to train a convolutional neural network module using public or autonomously acquired breath sound signal data sets to find optimal model parameters; or/and, configured to train a low dissimilarity forest classifier from the acquired breath sound signal dataset.
What is achieved is also a data acquisition module, coupled to the breath sound collection device, configured to acquire the breath sound signal.
The system implementation, as shown in fig. 1(a) and fig. 1(b), mainly includes two parts, model training and model deployment. In this embodiment, the system can be deployed in an embedded computer such as a raspberry pi computer, or an english junson Nano computer, or an RK3399 computer, to make a portable respiratory disease diagnosis device. However, in other embodiments, the system may be deployed on other terminal devices, and is not limited to the above-mentioned devices.
In this embodiment, the training part includes using an open data set, and performing model training and verification in a ten-fold cross-validation manner to obtain an optimal model structure and coefficients.
Specifically, in the training model section, the present embodiment uses an open breath sound data set collected by 2017 ICBHI (International Conference on Biomedical and Health information, International Conference on biomedicine and information, hereinafter abbreviated as ICBHI) to perform detection based on the abnormality level and the pathology level, respectively. The sample statistics for the anomaly level are shown in table 1. The sample statistics for the pathology levels are shown in table 2.
TABLE 1
Figure 100291DEST_PATH_IMAGE001
TABLE 2
Figure 830349DEST_PATH_IMAGE002
In table 1 and table 2, the data sets have unbalanced category distribution and fewer samples, and in order to improve the processing effect, the data sets are expanded by the present embodiment as a further improvement. Specifically, a Variational Auto Encoder (VAE) model may be used for data enhancement.
In training an autoencoder, data information is encoded in the underlying space and then decoded to reconstruct the original data.
The latent space is the lattice space, a mathematical space. An autoencoder consists of two components, an encoder and a decoder. The encoder brings the data from the high dimensional input to the bottleneck layer, which has the smallest number of neurons. The decoder then receives this encoded input and converts it back to the original input shape. The potential space is the space where the data is located at the bottleneck level.
As shown in fig. 2, the architecture of the data enhancement network is shown. The sample conditions for each class after enhancement are shown in tables 3 and 4.
TABLE 3
Figure 30387DEST_PATH_IMAGE003
TABLE 4
Figure 804176DEST_PATH_IMAGE004
And training by adopting a ten-fold cross validation method. I.e. samples of the data set are randomly shuffled and then evenly divided into ten. And selecting 1 part of the test set as a test set every time, providing the rest 9 parts of the test set as a training data set for convolutional neural network learning and low-diversity forest model training, and sending the test set into the test set after the training is finished to obtain the accuracy of the test set. And the circulation is performed for ten times. And finally, taking the average accuracy as the prediction accuracy of the model.
In this embodiment, a convolutional neural network and a low-variance forest model form a cascade model, as shown in fig. 3, a flow of model training and testing is shown in fig. 4(a) and 4(b), and a training process is as follows:
1. a training data set D is obtained and the number of samples of this data set is calculated as N.
2. Carrying out normalization processing on the breath sound audio files;
the first sample (i.e. a breath sound audio file) is normalized such that each audio data point is within the interval-1, 1.
3. For extracting audio features from the normalized audio data, optionally, a mel-frequency spectrum analysis method may be used to extract the audio features.
The audio features are extracted by a mel frequency spectrum analysis method, and the following formula can be adopted:
Figure 89664DEST_PATH_IMAGE005
where f represents the frequency of the sound.
Specifically, generation of the mel frequency spectrum is completed by adopting an Libsora development kit. The parameters are shown in table 5 below:
TABLE 5
Figure 561097DEST_PATH_IMAGE006
4. Repeating the steps 2-3 to complete the Mel-frequency spectrum conversion of all samples of the training data set, and storing in a temporary file D1.
The steps 1 to 4 are a preprocessing process of data.
5. Reading in D1, extracting features by using a convolutional neural network, and outputting N × 64 neuron data, wherein the structure and parameters of a convolutional neural network model are shown in the following table 6:
TABLE 6
Figure 615641DEST_PATH_IMAGE007
6. Storing N x 64 neuron data output by the convolutional neural network feature extraction in the last step and the classification labels of the original training set data into a temporary file D2;
7. performing dimensionality reduction on the feature data D2 output by the convolutional neural network;
specifically, all 64 feature columns are reduced to one column T using a dimension reduction method. T is appended to the last column of D2. The dimension reduction method can be PCA (principal component analysis), FA (factor analysis), kPCA (nonlinear principal component analysis), tSVD (truncated singular value decomposition), and the like.
8. Reordering the characteristic data D2 according to the obtained numerical value after the dimension reduction processing to obtain new characteristic data D3;
optionally, all samples of D2 are rearranged in ascending or descending order, depending on the size of the T column values. The T column is then deleted and stored as a new temporary file D3.
The steps 6-8 are performed on the output data of the convolutional neural network before being input into the low-difference forest classifier, so that the data are more suitable for the low-difference forest algorithm, and the data processing efficiency is improved.
The present embodiment uses a lightweight convolutional neural network, as shown in table 3, which includes an input layer, two convolutional layers, a pooling layer, a convolutional layer, a flattening layer, and a fully-connected layer, which are connected in sequence. And then, outputting the result through a full connection layer, and combining a low-diversity forest algorithm to obtain a final result.
In this embodiment, a low-diversity forest classifier model is used as the final classifier. The Forest with low diversity (Best diversity Sequence Forest) is K decision trees
Figure 576643DEST_PATH_IMAGE008
And the combined classifier is obtained after ensemble learning is carried out on the basic classifier. When a sample to be classified is given, the classification result output by the low-diversity forest is simply voted to decide by the classification result of each decision tree. As shown in fig. 3, here
Figure 100160DEST_PATH_IMAGE009
Is a random variable sequence.
Specifically, the low-diversity forest algorithm flow is as follows:
9.1) generating a low diversity order sequence BDS with N elements according to the number of samples N of the training data set D using the following formula:
Figure 172021DEST_PATH_IMAGE010
that is, each element in the low variance sequence is equal to a natural number and a circumferential ratio (i.e., 3.141592653589793238462) or other transcendental number (e.g., 3.141592653589793238462)
Figure 346650DEST_PATH_IMAGE011
Etc.) and the remaining bits (here, 21 bits) after the decimal point. And N is a continuous natural number starting from 1 and is taken as the sample number N of the training data set D. For example, BDS = {0.142, 0.283, 0.425... } (note: only the 3 bits after the decimal point are reserved here for convenience of illustration).
And 9.2) acquiring the sequence numbers of all the elements of the BDS sequence according to the ascending or descending principle to generate a sequence number sequence R. For example, the first number in the low disparity sequence is 0.142, and is ordered in ascending order among the N elements at r1, the second number is 0.283, the third number is ordered at r2, and the third number is 0.425, the third number is ordered at r3. Therefore, the generated low difference degree number series R is [ R1, R2, r3. ].
9.3) as is customary for the usual integration method, each decision tree uses 65% of the samples of the dataset as the training set for the tree, stored as temporary objects d =65% N.
9.4) setting the number K of the low-difference forests including the decision trees, wherein K =100 in this embodiment, that is, 100 decision trees are used to form one low-difference forest.
9.5) generating a random integer x between 1 and N as an initial sample index of a decision tree.
9.6) starting from x, in R generated in step 9.2, d successive elements are taken.
9.7) taking D samples from the D3 data set generated in the step 8 according to the taken elements as sample serial numbers to form a temporary data set as a training sample of a decision tree.
9.8) to construct a decision tree, any one of the decision tree algorithms of ID3, CART, C4.5, and the like can be used. When splitting each node of the decision tree, a feature column subset is extracted randomly from all the features with equal probability, and is usually taken
Figure 212975DEST_PATH_IMAGE012
And (4) selecting an optimal attribute from the subset to split the nodes, wherein m is the total number of the features. Using the temporary data set of the step 9-7) for training, and storing the parameters after the decision tree training into a temporary container P.
9.9) circulating the steps 9.5-9.8, constructing and training K decision trees together, and storing P and the parameters of the lightweight convolutional neural network in the step 7 into an ONNX pre-training model file.
At this point, one round of training is finished, a reserved test data set is used for classification testing, and then the rest 9 rounds of training and testing are carried out in a ten-fold cross validation mode. The experimental result shows that the indexes of the accurate classification rate of the respiratory disease, the F-Score and the like in the model constructed by the embodiment are all over 99 percent.
System deployment follows. By utilizing the electronic stethoscope, a 3.5mm earphone hole electronic stethoscope which is commonly used in the market can be used; or a sound pick-up is added on the head part of the common stethoscope; continuously collecting respiratory sound signals of 6 parts (such as a left chest, a right chest and a right chest) of a human body in a quiet environment, converting the respiratory sound signals into electric signals and storing the electric signals, wherein the electric signals can be stored as a wav file;
the electronic stethoscope is connected to a raspberry pi 4B card computer (or other similar inexpensive embedded computers) and the wav file is normalized by an algorithm. And extracting features of the preprocessed audio signal by using a Mel frequency spectrum method, sending the feature map into a trained cascade model combining a convolutional neural network and a low-diversity forest, and calculating to obtain a final recognition result to complete preliminary self-diagnosis.
As shown in fig. 1(b), the pulmonary auscultation system based on convolutional neural network performs a preliminary self-diagnosis, and the signal processing procedure of the system is as follows:
the data acquisition module acquires a breathing sound signal; such as the waveforms illustrated in fig. 5.
The preprocessing module preprocesses the acquired breath sound signals and normalizes the breath sound signals in sequence; step 2, the normalization method is the same as that in the model training process;
the data conversion module extracts audio features from the normalized data to generate Mel frequency spectrum data; the mel-frequency spectrum data generation method is the same as step 3 in the model training process, and as shown in fig. 6, mel-frequency spectrum data after the extraction of the exemplary waveform of fig. 5 is obtained.
The data enhancement module utilizes a data enhancement model to perform data amplification on the generated Mel frequency spectrum data; a Variational Auto Encoder (VAE) model may be used for data enhancement.
The convolutional neural network module utilizes the trained deep learning model to extract the features of the amplified Mel frequency spectrum data; receiving the data-enhanced mel-frequency spectrum data, inputting the data to a lightweight convolution neural network module arranged in the embodiment for feature extraction, and outputting N × 64 neuron data;
and the feature classification module classifies the extracted features by using the trained low-difference forest classifier to obtain a classification result. The classification algorithm performs steps 9.1-9.9.
Example 2
Based on embodiment 1, this embodiment provides a terminal device, including a memory, a processor, and computer instructions stored in the memory and executed on the processor, where the computer instructions, when executed by the processor, perform the following steps:
preprocessing the acquired breath sound signals, and sequentially normalizing the breath sound signals;
extracting audio features from the normalized data to generate Mel frequency spectrum data;
amplifying the data of the generated Mel frequency spectrum data by using a data enhancement model;
carrying out feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and classifying the extracted features by using the trained low-difference forest classifier to obtain a classification result.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Although the embodiments of the present invention have been described with reference to the accompanying drawings, it is not intended to limit the scope of the present invention, and it should be understood by those skilled in the art that various modifications and variations can be made without inventive efforts by those skilled in the art based on the technical solution of the present invention.

Claims (10)

1. Pulmonary disease auscultation system based on convolutional neural network, characterized by, including:
the preprocessing module is configured to preprocess the acquired breath sound signals and normalize the breath sound signals in sequence;
the data conversion module is configured to extract audio features from the normalized data and generate Mel frequency spectrum data;
a data enhancement module configured to perform data amplification on the generated mel-frequency spectrum data by using a data enhancement model;
the convolutional neural network module is configured to perform feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and the feature classification module is configured to classify the extracted features by using the trained low-difference forest classifier to obtain a classification result.
2. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: the model training module is configured to train the convolutional neural network module and the low-diversity forest classifier by using a public or self-collected breath sound signal data set to obtain an optimal model parameter.
3. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: and when a sample to be classified is given, the classification result output by the low-difference forest classifier is voted and determined by the classification result of each decision tree.
4. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: the device also comprises acquisition equipment for acquiring and storing the breathing sound signal.
5. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: the deep learning model and the low-difference forest classifier are cascaded.
6. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: the training process of the low-difference forest classifier comprises the following steps:
(a) generating a low-difference degree sequence based on a low-difference degree sequence sampling method according to the number of samples of the training data set;
(b) according to the principle of ascending or descending, acquiring the sequence numbers of all elements of the low-difference sequence, and generating a sequence number sequence;
(c) taking samples with set sizes of the training data set as training sets of corresponding decision trees;
(d) setting the number of decision trees of the forest with low diversity;
(e) random integers in the sample number range are used as initial sample indexes of a decision tree;
(f) starting from the random integer, taking a set number of continuous elements from a serial number sequence;
(g) taking out a corresponding number of samples from the input rearranged and deleted T-column data according to the taken-out elements as sample serial numbers to form a training sample of a decision tree;
(h) constructing a decision tree, and training the decision tree by using the training sample;
(i) and (e) repeating the steps (e) - (h), and constructing and training the decision trees according with the number of the decision trees.
7. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: the deep learning model is a lightweight convolutional neural network and comprises an input layer, two convolutional layers, a pooling layer, a convolutional layer, a flattening layer and a full-connection layer which are sequentially connected.
8. The convolutional neural network-based pulmonary auscultation system of claim 1, wherein: the data enhancement model is a variational auto-encoder, and when the auto-encoder is trained, data information is encoded in a latent space and then decoded to reconstruct original data.
9. The signal processing method of the lung disease auscultation system based on the convolutional neural network is characterized by comprising the following steps of:
preprocessing the acquired breath sound signals, and sequentially normalizing the breath sound signals;
extracting audio features from the normalized data to generate Mel frequency spectrum data;
amplifying the data of the generated Mel frequency spectrum data by using a data enhancement model;
carrying out feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and classifying the extracted features by using the trained low-difference forest classifier to obtain a classification result.
10. A terminal device is characterized in that: comprising a memory and a processor and computer instructions stored on the memory and executed on the processor, which when executed by the processor, perform the steps of:
preprocessing the acquired breath sound signals, and sequentially normalizing the breath sound signals;
extracting audio features from the normalized data to generate Mel frequency spectrum data;
amplifying the data of the generated Mel frequency spectrum data by using a data enhancement model;
carrying out feature extraction on the amplified Mel frequency spectrum data by using the trained deep learning model;
and classifying the extracted features by using the trained low-difference forest classifier to obtain a classification result.
CN202111077379.2A 2021-09-15 2021-09-15 Lung disease auscultation system based on convolutional neural network, signal processing method and equipment Active CN113537412B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111077379.2A CN113537412B (en) 2021-09-15 2021-09-15 Lung disease auscultation system based on convolutional neural network, signal processing method and equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111077379.2A CN113537412B (en) 2021-09-15 2021-09-15 Lung disease auscultation system based on convolutional neural network, signal processing method and equipment

Publications (2)

Publication Number Publication Date
CN113537412A true CN113537412A (en) 2021-10-22
CN113537412B CN113537412B (en) 2021-12-14

Family

ID=78093228

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111077379.2A Active CN113537412B (en) 2021-09-15 2021-09-15 Lung disease auscultation system based on convolutional neural network, signal processing method and equipment

Country Status (1)

Country Link
CN (1) CN113537412B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114496209A (en) * 2022-02-18 2022-05-13 青岛市中心血站 Blood donation intelligent decision method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107315721A (en) * 2017-06-27 2017-11-03 郭亮 A kind of methods of sampling and system of the resident family of community based on low diversity factor ordered series of numbers
CN108417225A (en) * 2018-02-11 2018-08-17 广州市碳码科技有限责任公司 A kind of breath sound monitoring method, device, terminal and computer readable storage medium
US20190008475A1 (en) * 2017-07-04 2019-01-10 Tata Consultancy Services Limited Systems and methods for detecting pulmonary abnormalities using lung sounds
US20200008004A1 (en) * 2016-12-01 2020-01-02 Nokia Technologies Oy Audio processing
CN111640439A (en) * 2020-05-15 2020-09-08 南开大学 Deep learning-based breath sound classification method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200008004A1 (en) * 2016-12-01 2020-01-02 Nokia Technologies Oy Audio processing
CN107315721A (en) * 2017-06-27 2017-11-03 郭亮 A kind of methods of sampling and system of the resident family of community based on low diversity factor ordered series of numbers
US20190008475A1 (en) * 2017-07-04 2019-01-10 Tata Consultancy Services Limited Systems and methods for detecting pulmonary abnormalities using lung sounds
CN108417225A (en) * 2018-02-11 2018-08-17 广州市碳码科技有限责任公司 A kind of breath sound monitoring method, device, terminal and computer readable storage medium
CN111640439A (en) * 2020-05-15 2020-09-08 南开大学 Deep learning-based breath sound classification method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LIANG GUO ET AL.: "Subsampling bias and the best-discrepancy systematic cross validation", 《SCIENCE CHINA MATHEMATICS》 *
陈少琼: "心肺音频信号的分类和应用", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114496209A (en) * 2022-02-18 2022-05-13 青岛市中心血站 Blood donation intelligent decision method and system
CN114496209B (en) * 2022-02-18 2022-09-27 青岛市中心血站 Intelligent decision-making method and system for blood donation

Also Published As

Publication number Publication date
CN113537412B (en) 2021-12-14

Similar Documents

Publication Publication Date Title
CN109934089B (en) Automatic multi-stage epilepsia electroencephalogram signal identification method based on supervised gradient raiser
JP2021506486A (en) A method for analyzing cough sounds using disease signatures to diagnose respiratory disease
CN108090509B (en) Data length self-adaptive electrocardiogram classification method
CN110298264B (en) Human body daily behavior activity recognition optimization method based on stacked noise reduction self-encoder
Göğüş et al. Classification of asthmatic breath sounds by using wavelet transforms and neural networks
Dentamaro et al. AUCO ResNet: an end-to-end network for Covid-19 pre-screening from cough and breath
CN111759345B (en) Heart valve abnormality analysis method, system and device based on convolutional neural network
Liu et al. Research on medical data feature extraction and intelligent recognition technology based on convolutional neural network
Wang et al. Application of fuzzy cluster analysis for medical image data mining
Singh et al. Short unsegmented PCG classification based on ensemble classifier
CN111354338A (en) Parkinson speech recognition system based on PSO convolution kernel optimization sparse transfer learning
CN113537412B (en) Lung disease auscultation system based on convolutional neural network, signal processing method and equipment
Salhi et al. Voice disorders identification using hybrid approach: Wavelet analysis and multilayer neural networks
CN113349801A (en) Imaginary speech electroencephalogram signal decoding method based on convolutional neural network
Sengupta et al. Lung sound classification using local binary pattern
CN113413163A (en) Heart sound diagnosis system for mixed deep learning and low-difference forest
Jayasekara et al. Timecaps: Capturing time series data with capsule networks
CN111938691B (en) Basic heart sound identification method and equipment
CN113116300A (en) Physiological signal classification method based on model fusion
Liu et al. Respiratory sounds feature learning with deep convolutional neural networks
Bashar et al. Heart abnormality classification using phonocardiogram (PCG) signals
CN114999628B (en) Method for searching for obvious characteristic of degenerative knee osteoarthritis by using machine learning
Tiwari et al. Deep lung auscultation using acoustic biomarkers for abnormal respiratory sound event detection
CN115376560A (en) Voice feature coding model for early screening of mild cognitive impairment and training method thereof
Choi et al. Performance improvement of deep learning based multi-class ECG classification model using limited medical dataset

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20230615

Address after: No. 6, 13th Floor, Building 8, No. 399, West Section of Fucheng Avenue, Chengdu High tech Zone, China (Sichuan) Pilot Free Trade Zone, Chengdu City, Sichuan Province, 610000

Patentee after: Sichuan Hongxue Jinlian Technology Group Co.,Ltd.

Address before: 250014, No. 27, Da Nan Road, Licheng District, Shandong, Ji'nan

Patentee before: SHANDONG University

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20230802

Address after: Room 303, Building 3, National University Science and Technology Park, Shizhong District, Jinan City, Shandong Province, 250000

Patentee after: Liangrui Artificial Intelligence (Jinan) Co.,Ltd.

Address before: No. 6, 13th Floor, Building 8, No. 399, West Section of Fucheng Avenue, Chengdu High tech Zone, China (Sichuan) Pilot Free Trade Zone, Chengdu City, Sichuan Province, 610000

Patentee before: Sichuan Hongxue Jinlian Technology Group Co.,Ltd.