CN109360584A - Cough monitoring method and device based on deep learning - Google Patents

Cough monitoring method and device based on deep learning Download PDF

Info

Publication number
CN109360584A
CN109360584A CN201811259953.4A CN201811259953A CN109360584A CN 109360584 A CN109360584 A CN 109360584A CN 201811259953 A CN201811259953 A CN 201811259953A CN 109360584 A CN109360584 A CN 109360584A
Authority
CN
China
Prior art keywords
cough
audio
audio data
data
sounder
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811259953.4A
Other languages
Chinese (zh)
Inventor
彭话易
蔡元哲
黄章成
王健宗
肖京
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Priority to CN201811259953.4A priority Critical patent/CN109360584A/en
Publication of CN109360584A publication Critical patent/CN109360584A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/66Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/02Preprocessing operations, e.g. segment selection; Pattern representation or modelling, e.g. based on linear discriminant analysis [LDA] or principal components; Feature selection or extraction
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/18Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Multimedia (AREA)
  • Acoustics & Sound (AREA)
  • Human Computer Interaction (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Public Health (AREA)
  • General Health & Medical Sciences (AREA)
  • Epidemiology (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)

Abstract

This disclosure relates to which field of artificial intelligence, discloses a kind of cough monitoring method and device based on deep learning, comprising: pre-processed to obtain several frame tonic trains to the audio data of acquisition;Cough identification is carried out to several frame tonic trains, whether is cough audio with the determination audio data;And Application on Voiceprint Recognition is carried out to several frame tonic trains, with the sounder of the determination audio data;If the audio data is cough audio, processing is updated according to cough monitoring data of the audio data to the sounder.Cough identification and Application on Voiceprint Recognition are carried out to audio data using the method for deep learning, to obtain the cough monitoring data that audio data corresponds to sounder, realize the automatic monitoring to cough, it is convenient and efficient, it does not need special messenger and carries out cough monitoring, the efficiency for improving cough monitoring ensure that the real-time of cough monitoring data.

Description

Cough monitoring method and device based on deep learning
Technical field
This disclosure relates to field of artificial intelligence, in particular to a kind of cough monitoring method and dress based on deep learning It sets.
Background technique
Cough is the common sympton of respiratory disease, by the status of cough of patient can reflect patient physical condition or The recovery situation of patient's person state of an illness.The status of cough of general patient is entourage's (such as families of patients) by patient in day Often accompany and attend to what Cough length recorded in process, cough number etc. were got.But entourage is not to accompany at any time At one's side, thus by entourage, that status of cough obtained is recorded during accompanying and attending to is not comprehensive, and by special by patient Entourage it is also big come the status of cough workload for recording patient.
Therefore cough monitoring effect and cough monitoring efficiency need to be improved.
Summary of the invention
In order to solve the problems, such as present in the relevant technologies, present disclose provides a kind of cough monitoring side based on deep learning Method and device.
A kind of cough monitoring method based on deep learning, comprising:
The audio data of acquisition is pre-processed to obtain several frame tonic trains;
Cough identification is carried out to several frame tonic trains, whether is cough audio with the determination audio data;With And
Application on Voiceprint Recognition is carried out to several frame tonic trains, with the sounder of the determination audio data;
If the audio data is cough audio, number is monitored to the cough of the sounder according to the audio data It is handled according to being updated.
A kind of cough monitoring device based on deep learning, comprising:
Preprocessing module is configured as executing: being pre-processed to obtain several frame tonic trains to the audio data of acquisition;
Cough identification module, is configured as executing: cough identification is carried out to several frame tonic trains, described in determination Whether audio data is cough audio;And
Voiceprint identification module is configured as executing: Application on Voiceprint Recognition is carried out to several frame tonic trains, described in determination The sounder of audio data;
Monitoring data update module is configured as executing: if the audio data is cough audio, according to the sound Frequency evidence is updated processing to the cough monitoring data of the sounder.
In one embodiment, the preprocessing module includes:
Short Time Fourier Transform unit is configured as executing: carrying out Short Time Fourier Transform to the audio data of acquisition and obtains To the corresponding sonograph of the audio data;
Segmenting unit is configured as executing: carrying out the segmentation of the sonograph according to specified duration, obtains several frames Tonic train.
In one embodiment, the cough identification module includes:
Local feature vectors extraction unit is configured as executing: is extracted described in obtaining from several frame tonic trains Several local feature vectors of audio data;
Full connection unit is configured as executing: carrying out the full connection of several local feature vectors, obtains the audio The global characteristics vector of data;
Cough label prediction unit, is configured as executing: carrying out cough Tag Estimation to the global characteristics vector, obtain The cough label of the audio data;
Interpretation unit is configured as executing: judging whether the audio data is cough audio according to the cough label.
In one embodiment, the voiceprint identification module includes:
Vocal print feature extraction unit is configured as executing: extracting from several frame tonic trains and obtains the audio The vocal print feature of data;
Sounder Tag Estimation module, is configured as executing: carrying out sounder Tag Estimation according to the vocal print feature, obtains To the sounder label of the audio data;
Sounder determining module is configured as executing: the sounding of the audio data is determined according to the sounder label Person.
In one embodiment, monitoring data update module includes:
Timestamp information extraction unit is configured as executing: if the audio data is cough audio, from the sound Frequency is extracted in and obtains timestamp information;
Cough length information and cough frequency information generating unit, are configured as executing: according to the cough audio and The corresponding timestamp information generates Cough length information and cough frequency information;
Cough monitoring data acquiring unit in short-term, is configured as executing: by the Cough length information and cough frequency letter Cease the cough monitoring data in short-term as the sounder in the audio data;
Monitoring data updating unit is configured as executing: the monitoring data of cough in short-term are uploaded to monitor database In, to update the cough monitoring data of sounder described in the monitor database.
In one embodiment, Cough length information and cough frequency information generating unit include:
End-point detection unit is configured as executing: carrying out end-point detection to the cough audio with the determination cough sound The beginning and end coughed every time in frequency;
Acquiring unit is configured as executing: being obtained according to timestamp information corresponding to the beginning and end coughed every time The points or terminal number coughed into the Cough length information, and the statistics cough audio obtain the cough frequency Information.
A kind of cough monitoring device based on deep learning, comprising:
Processor;And
Memory for storage processor executable instruction;
Wherein, the above-described cough based on deep learning is realized when the executable instruction is executed by the processor Monitoring method.
A kind of computer readable storage medium, is stored thereon with computer program, and the computer program is held by processor The above-described cough monitoring method based on deep learning is realized when row.
The technical scheme provided by this disclosed embodiment can include the following benefits: by carrying out cough to audio data It coughs identification and Application on Voiceprint Recognition, to determine whether audio data is cough audio and the corresponding sounder of audio data, in turn The cough monitoring data of sounder are obtained to cough audio analysis, realize the automatic monitoring to cough, it is convenient and efficient, it does not need Special messenger carries out cough monitoring, improves the efficiency of cough monitoring, ensure that the real-time of cough monitoring data.
It should be understood that the above general description and the following detailed description are merely exemplary, this can not be limited It is open.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows and meets implementation of the invention Example, and in specification together principle for explaining the present invention.
Fig. 1 is the schematic diagram of the implementation environment according to involved in the disclosure;
Fig. 2 is a kind of block diagram of monitoring server of coughing shown according to an exemplary embodiment;
Fig. 3 is a kind of flow chart of cough monitoring method based on deep learning shown according to an exemplary embodiment;
Fig. 4 is the flow chart of the step S110 of embodiment illustrated in fig. 3;
Fig. 5 is the flow chart of the step S130 of embodiment illustrated in fig. 3;
Fig. 6 is the flow chart of the step S140 of embodiment illustrated in fig. 3;
Fig. 7 is the flow chart of the step S150 of embodiment illustrated in fig. 3;
Fig. 8 is the flow chart of the step S152 of embodiment illustrated in fig. 7;
Fig. 9 is the flow chart after the step S150 of embodiment illustrated in fig. 3;
Figure 10 is the schematic diagram for the neural network model for identification of coughing;
Figure 11 is the block diagram of the cough monitoring device shown according to an exemplary embodiment based on deep learning;
Figure 12 is the block diagram of the cough monitoring device based on deep learning shown according to another exemplary embodiment.
Specific embodiment
Here will the description is performed on the exemplary embodiment in detail, the example is illustrated in the accompanying drawings.Following description is related to When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment Described in embodiment do not represent all embodiments consistented with the present invention.On the contrary, they be only with it is such as appended The example of device and method being described in detail in claims, some aspects of the invention are consistent.
Fig. 1 is the schematic diagram of the implementation environment according to involved in the disclosure.The implementation environment includes: cough monitoring server 200.Cough monitoring server 200 can carry out cough identification and Application on Voiceprint Recognition to audio data using the method that the disclosure provides To obtain cough monitoring data.
As needed, in the implementation environment will also include audio collecting device 100, for provide cough monitoring server into The audio data of row cough identification and Application on Voiceprint Recognition.Audio collecting device 100 can be arranged in the living environment of related sounder In, such as family, for the application scenarios of hospital, audio collecting device 100 can be arranged in the ward of patient (for this kind Situation, patient are sounder involved in the disclosure.) the implementation progress audio data collecting of audio collecting device 100, then will Collected audio data is transferred to cough monitoring server 200 and is handled, and then obtains the cough audio data of sounder.
Wherein audio collecting device 100 can be independent audio collecting device, such as microphone, recorder, can be with Electronic equipment of microphone, such as video recorder, smart phone, computer etc. are carried, is not defined herein.
As needed, in the implementation environment will also include cough monitoring terminal 300, the cough monitor terminal 300 be used for into The display etc. of row cough monitoring data.The cough monitoring data for the sounder that cough monitoring server 200 obtains processing return Into cough monitoring terminal 300, consequently facilitating user knows the cough monitoring data of implementation.
Wireless or wired network is pre-established between audio collecting device 100 and cough monitoring server 200 to connect It connects, and then is transmitted by the data that network connection realizes audio collecting device 100 between monitoring server 200 of coughing, for example, The audio data that audio collecting device 100 acquires.Correspondingly, pre- between cough monitoring terminal 300 and cough monitoring server 200 Wired or wireless network connection is first established, and then cough monitoring server 200 and cough can be realized by network connection The data monitored between terminal 300 of coughing are transmitted.
It should be noted that cough monitoring method of the disclosure based on deep learning, be not limited in cough monitoring server Corresponding processing logic is disposed in 200, is also possible to the processing logic being deployed in other machines.For example, being calculated having Deployment carries out the processing logic of cough identification and Application on Voiceprint Recognition to the audio data of acquisition in the cough monitoring terminal 300 of ability Deng.
Fig. 2 is a kind of block diagram of monitoring server of coughing shown according to an exemplary embodiment.With this hardware configuration Server can be used for carrying out cough monitoring and be deployed in implementation environment shown in FIG. 1.
It should be noted that the cough monitoring server is the example for adapting to the disclosure, it is not construed as mentioning Any restrictions to disclosure use scope are supplied.The cough monitoring server can not be construed to need to rely on or necessary With one or more component in illustrative cough monitoring server 200 shown in Figure 2.
The hardware configuration of the cough monitoring server can generate biggish difference due to the difference of configuration or performance, such as scheme Shown in 2, cough monitoring server 200 includes: power supply 210, interface 230, at least a memory 250 and an at least centre Manage device (CPU, Central Processing Units) 270.
Wherein, power supply 210 is used to provide operating voltage for each hardware device in cough monitoring server 200.
Interface 230 includes an at least wired or wireless network interface 231, at least a string and translation interface 233, at least one defeated Enter output interface 235 and at least USB interface 237 etc., is used for and external device communication.
The carrier that memory 250 is stored as resource, can be read-only memory, random access memory, disk or CD Deng the resource stored thereon includes operating system 251, application program 253 and data 255 etc., and storage mode can be of short duration It stores or permanently stores.Wherein, each hardware that operating system 251 is used to manage and control in cough monitoring server 200 is set Standby and application program 253 can be with realizing calculating and processing of the central processing unit 270 to mass data 255 Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM etc..Application program 253 is based on operation The computer program that at least one particular job is completed on system 251, may include that an at least module (is not shown in Fig. 2 Out), each module can separately include the series of computation machine readable instruction to cough monitoring server 200.Data 255 Photo, picture, the audio etc. that can be stored in disk.
Central processing unit 270 may include the processor of one or more or more, and be set as through bus and memory 250 communications, for the mass data 255 in operation and processing memory 250.
As described in detail above, the cough monitoring server 200 for being applicable in the disclosure will be read by central processing unit 270 The form of the series of computation machine readable instruction stored in access to memory 250 come complete cough monitoring method.
In the exemplary embodiment, cough monitoring server 200 can be by one or more application specific integrated circuit At (Application Specific Integrated Circuit, abbreviation ASIC), digital signal processor, digital signal Manage equipment, programmable logic device, field programmable gate array, controller, microcontroller, microprocessor or other electronic components It realizes, for executing following methods.Therefore, realize that the present invention is not limited to any specific hardware circuit, software and the two Combination.
Fig. 3 is a kind of flow chart of cough monitoring method based on deep learning shown according to an exemplary embodiment. This method can be used for the cough monitoring server 200 in implementation environment shown in Fig. 1, as shown in figure 3, this method, it can be by cough Monitoring server of coughing 200 executes, and may comprise steps of.
Step S110 pre-processes the audio data of acquisition to obtain several frame tonic trains.
Wherein, audio data can the audio collecting device as shown in Figure 1 in implementation environment collect.It then will acquisition To audio data upload to cough monitoring server 200 in handled.
In one exemplary embodiment, as shown in Fig. 2, step S110 may further include:
Step S111 carries out Short Time Fourier Transform to the audio data of acquisition and obtains the corresponding sonograph of audio data.
Short Time Fourier Transform (STFT, short-time Fourier transform) is by a Time-Frequency Localization Window function, it is assumed that audio of coughing in the short time interval that the window function is included is smoothly then Moving Window letter Number, making audio data is stationary signal in different finite time width, to calculate audio data when each different The power spectrum at quarter.The power spectrum at each moment is stacked up along time dimension finally, obtains the two dimension letter similar to a width figure Number form is to get arriving the corresponding sonograph of audio data.
Step S112 carries out the segmentation of sonograph according to specified duration, obtains several frame tonic trains.
Duration is wherein specified to refer to time span corresponding to each frame cough audio frame sequence set by user.It is specified Duration can set the processing requirement of audio data according to cough monitoring server 200.By specified duration to sound spectrum Figure is segmented, and if then obtaining dry cough audio frame sequence.
In one embodiment, with audio of coughing for 35s, according to it is specified when it is a length of 4 milliseconds progress sonograph segmentation, Available 8750 frame cough audio frame sequence, wherein each frame contains 64 frequency components.
Step S130 carries out cough identification to several frame tonic trains, to determine whether audio data is cough audio.
Cough wherein carried out to several audio frame sequences be identified by the mode of deep learning to carry out, i.e., by training after Neural network model carry out cough identification, thus audio data whether be cough audio.
In one exemplary embodiment, as shown in figure 5, step S130 includes:
Step S131 is extracted from several frame tonic trains and is obtained several local feature vectors of audio data.
Local feature vectors are the features for characterizing tonic train, and audio corresponding to different tonic trains is in time domain It is different with the feature that is shown on frequency domain, such as tone, loudness, mel-frequency, mel-frequency, mel-frequency cepstrum system Number (MFCC), linear predictor coefficient (LPC), linear prediction residue error (LPCC) etc..Wherein constructed local feature vectors It can be the combination based on feature a certain in feature enumerated above or certain several feature to construct.Certainly, the above institute The feature enumerated is only illustrative examples, can not be considered the limitation to disclosure use scope.
For each frame tonic train one local feature vectors of corresponding building, for characterizing the spy of this frame tonic train Sign.
Step S132 carries out the full connection of several local feature vectors, obtains the global characteristics vector of audio data.
Step S133 carries out cough Tag Estimation to global characteristics vector, obtains the cough label of audio data.
Whether step S134 is cough audio according to cough label audio data.
Wherein cough label includes the label " cough " for being directed to cough audio and label " the non-cough for non-cough audio It coughs ".In a specific embodiment, cough label can be indicated by encoding, such as with digital " 1 " expression " cough " label, " non-cough label " is indicated with digital " 0 ".To be sentenced by coding rule after obtaining the cough label of audio data Break meaning represented by the cough label, such as cough or non-cough, so according to cough label to judge the audio number According to whether be cough audio.
When carrying out cough Tag Estimation, prediction obtains the probability P 1 and audio number that audio data is " cough " label respectively According to the probability P 2 for " non-cough " label, then P1 and P2 are compared, if P1 is greater than P2, export the cough of audio data Label of coughing is " cough " label;If P1 is less than P2, the cough label for exporting audio data is " non-cough " audio.
In one exemplary embodiment, before carrying out cough identification to audio data, cough is constructed based on neural network Cough of the identification model for audio data identifies.Figure 10 shows the illustrative diagram of cough identification model, wherein the cough Identification model of coughing is constructed using Recognition with Recurrent Neural Network, so as to improve cough using the time related information in cough audio The precision of identification.As shown in Figure 10, which includes: encoder layer, decoder layer, full articulamentum and classification Layer.
Wherein, encoder layer and decoder layer are used to construct each frame sound by operations such as coding, compression, dimensionality reduction, decodings The local feature vectors of frequency sequence.Specifically, what encoder layer was formed by 3 layers, including two layers of bidirectional circulating layer and a unidirectional ply, One of bidirectional circulating layer includes 128 circulation neurons, another bidirectional circulating layer includes 64 circulation neurons, unidirectionally Layer has 32 circulation neurons.Several frame tonic trains of input pass sequentially through 128 neuronal layers, 64 neuronal layers and 32 minds Through first layer, subsequently into decoder layer.Several frame tonic trains of input can obtain a tool by 128 circulation neuronal layers There is the output of 128 dimensions, an output with 64 dimensions can be obtained by recycling neuronal layers by 64, then pass through 32 The layer of a circulation neuron can obtain the output with 32 dimensions.The output for 32 dimensions that encoder layer is exported is made For the input of decoder layer.
Wherein two-way circulation neuron is to allow neuron when calculating not merely with going through in tonic train History information can also utilize its following information.The building and transformation for being carried out vector using circulation neuron in encoder layer, are followed The calculating of ring neuron is and time correlation.
Decoder layer is made of an individual circulation layer, it have 64 long short-term memory (LSTM) neuron, 64 Long memory unit in short-term carries out linear transformation and activation to the output of 32 dimensions of input, and it is right to obtain each frame tonic train institute The local feature vectors for only one dimension answered.
Further, the LSTM neuron of decoder layer combines attention mechanism, and attention mechanism refers to that decoder exists When exporting decoding result, an attention range can be also exported, identifying next decoding will pay close attention in sequence Then which part is exported according to these parts.Because state sometime may be by certain in a time series The state influence at a little moment is bigger, therefore noted that power mechanism is that the neuron of decoder layer is allowed to calculate shape sometime in decoding It is average to the implicit variable weighting of different moments when state.It is defeated by being that the LSTM neuron mainly pours into conjunction with attention mechanism Enter the signal portion of information, and then improves the accuracy of subsequent cough identification.
Full articulamentum has 256 neurons using ReLU activation primitive, the local feature that decoder layer is exported to Amount is integrated and is converted, that is, carries out the full connection of local feature vectors, obtains the global characteristics vector for the audio data.
Last classification layer uses softmax classifier, carries out classification prediction to the global characteristics vector of input, obtains cough It coughs label, i.e., " cough " label or " non-cough " label.
Different types of neuron is targetedly set by each layer in neural network, audio sequence can be made full use of Information in column carries out cough identification, ensure that the precision of prediction of cough disease identification model.
Certainly, cough identification model shown in Fig. 10 is only an illustrative examples, is a preferred cough identification model, Other neural network models that cough identification may be implemented are equally applicable to the present invention, can not think schematic diagram shown in Fig. 10 It is the limitation to disclosure use scope.
It in one embodiment, further include using sample data to cough before carrying out cough identification using cough identification model Identification model of coughing is trained.Several sample audio datas are acquired, manually each sample audio data is labeled, for Sample audio data for audio of coughing is labeled as " coughing ", is labeled as " non-cough for the sample audio data of non-cough audio It coughs ".Then sample audio data and its corresponding mark are input to the instruction that cough identification model is carried out in cough identification model Practice.In the training process, for the sample audio data of each input, the classification layer for identification model of coughing one mark of corresponding output Label, the label of output is compared with the annotation results to the sample audio data, if the label of output with mark not phase Together, the parameter of cough identification model is adjusted until the label of output is identical as mark;If identical, continue with other samples Audio data is trained.Finally until the accuracy of identification of cough identification model reaches permissible accuracy, then completing to cough identifies The training of model.The cough identification model that training is completed is used for the cough identification of audio data.
And step S140, Application on Voiceprint Recognition is carried out to several frame tonic trains, to determine the sounder of audio data.
Application on Voiceprint Recognition is that the identification of sounder identity is carried out by the characteristic voice of sounder, i.e., according to an audio data Audio features carry out sounder identification.
In one embodiment, in one exemplary embodiment, as shown in fig. 6, step S140 includes:
Step S141 is extracted from several frame tonic trains and is obtained the vocal print feature of audio data.
Step S142 carries out sounder Tag Estimation according to vocal print feature, obtains the sounder label of audio data.
Step S143 determines the sounder of audio data according to sounder label.
Wherein Application on Voiceprint Recognition can be carried out by Application on Voiceprint Recognition model, Application on Voiceprint Recognition model is also based on neural network and is built It is vertical.Application on Voiceprint Recognition model carries out the process of Application on Voiceprint Recognition as shown in step S141, S142, S143.
Extracted vocal print feature can be the tamber characteristic of audio data, it is, of course, also possible to be that other can characterize sound Frequency is according to the feature for corresponding to sounder feature, such as tamber characteristic combination mel-frequency, mel-frequency cepstrum coefficient, linear prediction Some or certain several features in cepstrum coefficient.Then sounder label is carried out according to extracted vocal print feature.Finally The sounder of audio data is determined according to sounder label.
Before carrying out Application on Voiceprint Recognition using Application on Voiceprint Recognition model, need to be trained Application on Voiceprint Recognition model, i.e., with several The audio data of people is trained, and marks corresponding sounder to each audio data, passes through the hair of audio data and mark The training of sound person progress Application on Voiceprint Recognition model, so that the Application on Voiceprint Recognition model after training can be determined for the audio of an input and is somebody's turn to do The sounder of audio.
Step S150, if audio data is cough audio, according to audio data to the cough monitoring data of sounder It is updated processing.
If audio data is cough audio, Cough length, cough number etc. in the audio data can be calculated Then cough information will be added in the cough monitoring data monitored from cough information obtained in the audio data, real The update for monitoring data of now coughing.
In one exemplary embodiment, as shown in fig. 7, step S150 includes:
Step S151 is extracted from audio data if audio data is cough audio and is obtained timestamp information.
Audio generation time corresponding to any node audio in timestamp information, that is, audio data.Wherein timestamp information It can be the timestamp information that audio data is automatically recorded when audio collecting device 100 acquires audio data and by the timestamp Information is added in audio data, thus in step S151 can from audio data extraction time stab information.
Step S152 generates Cough length information and the cough frequency according to cough audio and corresponding timestamp information Information.
The multiple cough that may include sounder in audio of wherein coughing passes through analysis cough audio and cough audio pair The timestamp information answered, the available Cough length information for the cough audio and cough frequency information.
Wherein Cough length information includes but is not limited to: every time at the beginning of cough, end time for coughing every time, every The duration of secondary cough.Frequency information of coughing includes but is not limited to the cough in the corresponding time range of this section cough audio Number, cough frequency etc..
In one exemplary embodiment, as shown in figure 8, step S152 includes:
Step S210 carries out end-point detection to cough audio to determine the beginning and end coughed every time in cough audio.
End-point detection is exactly that effective audio section is detected from continuous audio stream.In the scheme of the disclosure, cough Corresponding audio is effective audio end.Breaking point detection includes two aspects, detects the starting point of effective audio, detects The terminal of effective audio.Breaking point detection determines the beginning and end of effective audio by the energy value of audio.It coughed from non- It crosses to cough, the reflection on audio power is: audio power sharply increases.And it is transitioned into non-cough from cough, in audio energy Reflection in amount is that audio power drastically reduces.
So can by set energy threshold, if in a certain node, the energy of continuous N frame tonic train before the node Value is greater than the energy of setting in the energy value of the next continuous N frame tonic train of the node lower than the energy threshold set Threshold value, then the node can be determined as the starting point once coughed.Likewise, if in a certain node, continuous N frame before the node The energy value of tonic train is greater than the energy threshold of setting, and the energy value of the next continuous N frame tonic train of the node is small In the energy threshold of setting, then the node can be determined as the terminal once coughed.
In the primary cough of corresponding sounder, this time audio caused by cough is continuous, can by end-point detection To determine the beginning and end for corresponding audio of this time coughing.Thus in audio data collected, it may be possible to corresponding one The audio of secondary cough, it is also possible to the audio repeatedly coughed.When being the audio repeatedly coughed for audio data collected, often Audio corresponding to secondary cough is effective audio of end-point detection, and the audio between cough in the period may be background twice Noise is mute, and the audio between cough in the period is invalid audio twice.
Step S230, the timestamp information according to corresponding to the beginning and end coughed every time obtain Cough length information, with And the points or terminal number coughed in statistics cough audio obtain cough frequency information.
Cough time started, cough end time and the cough that wherein Cough length information is including but not limited to coughed every time Duration.Cough frequency information includes but is not limited to correspond to the number coughed in the period, cough frequency in the cough audio.
When the cough that wherein the secondary cough can be obtained in the timestamp information according to corresponding to the starting point coughed every time starts Between, the cough end time of the secondary cough can be obtained according to timestamp information corresponding to the terminal coughed every time.Then lead to It spends the cough time started and the cough duration of the secondary cough is calculated in the cough end time.
Certainly in cough audio has been determined after the beginning and end coughed every time, can by count starting point number or The number of person's terminal obtains number of coughing in this section cough audio, and the frequency of cough can be obtained by obtained cough number Rate.
Step S153, the cough in short-term using Cough length information and cough frequency information as sounder in audio data Monitoring data.
Step S154, the monitoring data that will cough in short-term upload in monitor database, to update sounding in monitor database The cough monitoring data of person.
Monitor database be used for store sounder audio data and corresponding cough monitoring data, wherein each send out Sound person corresponds to a subdata base, and the cough monitoring data of the sounder can be obtained from the subdata base.For example, in hospital Application scenarios in, monitor database can be the medical data base for patient, so as to from the subdata of the patient Transfer the cough monitoring data of the patient.In a particular embodiment, in the subdata base of patient can also comprising patient case, The information such as treatment record.
Obtained cough monitoring data be based on before the audio data collected data analyze, so The monitoring data of cough in short-term obtained based on the audio data analysis are added in obtained cough monitoring data, are carried out The update for monitoring data of coughing, the including but not limited to update of Cough length information, the update for frequency information of coughing.
For monitoring data of coughing after monitor database update, related personnel can view cough by the monitor database Monitoring data are realized to the remote real-time monitoring of patient's cough monitoring situation, are not needed special messenger and supervise near sounder Control.In a particular embodiment, such as in hospital, sounder is patient, then the attending physician of the patient or nurse can be at any time Cough monitoring data are viewed in monitor database, further, doctor can also be supervised by the cough within a period of time Measured data carries out the prediction of patient's state of an illness and recovery situation.
After monitoring data of coughing update, updated cough monitoring data can be carried out by cough monitoring terminal 300 It has been shown that, cough monitoring terminal 300 can be the ward positioned at the house where sounder, such as patient, thus either family Category, nurse, doctor can monitor the cough monitoring situation that terminal 300 quickly understands patient by the cough.
Shown by a particular embodiment, being adjusted by setting time range in cough monitoring terminal 300 Cough monitoring data.For example, if the cough frequency of statistics is the cough frequency in 1 hour, it is only whole in cough monitoring The cough frequency in 1 hour is shown on end 300, and carries out real-time update.Cough monitoring data in certain other times can To be stored in specified data library, it can transfer and check in real time for user.
In one exemplary embodiment, the method for the embodiment further include:
Step S170 judges whether the cough frequency of sounder exceeds the safe frequency according to cough frequency information.
Step S190 generates information warning if exceeding the safe frequency.
The cough frequency may cause expiratory dyspnea beyond the safe frequency, thus by setting the safe frequency, if beyond peace Full range, then generate information warning.In a particular embodiment, external warning device (example can be triggered by the information warning Such as buzzer, alarm lamp) movement, so as to allow the personnel of surrounding to learn and handle in time.Certainly, warning generated Information can also be broadcasted automatically by voice, and related personnel is allowed to recognize the situation in time.
It is worth noting that the technical solution of the disclosure cannot only be applied in the scene of hospital, for needing to monitor The patient of status of cough can be used for the scenes such as home for destitute, sanatorium, family and carry out cough monitoring, herein without specific It limits.
Following is embodiment of the present disclosure, can be used for executing that the above-mentioned mobile terminal 110 of the disclosure executes based on depth Spend the cough monitoring method embodiment of study.For those undisclosed details in the apparatus embodiments, disclosure base is please referred to In the cough monitoring method embodiment of deep learning.
Figure 11 is a kind of block diagram of cough monitoring device based on deep learning shown according to an exemplary embodiment, should Device can be used in the cough monitoring server 200 of implementation environment shown in Fig. 1, execute shown in any of the above embodiment of the method The all or part of step of cough monitoring method based on deep learning.As shown in figure 11, which includes but unlimited In: preprocessing module 110, cough identification module 130, voiceprint identification module 140 and monitoring data update module 150.
Preprocessing module 110 is configured as executing: being pre-processed to obtain several frame audio sequences to the audio data of acquisition Column.
Cough identification module 130, the module connect with preprocessing module 110, are configured as executing: to several frame audio sequences Column carry out cough identification, to determine whether audio data is cough audio.And
Voiceprint identification module 140, the module are connect with preprocessing module 110, are configured as executing: to several frame audio sequences Column carry out Application on Voiceprint Recognition, to determine the sounder of audio data.
Monitoring data update module 150, the module connect with cough identification module 130, voiceprint identification module 140, are matched It is set to execution: if audio data is cough audio, being updated according to cough monitoring data of the audio data to sounder Processing.
In one embodiment, preprocessing module 110 includes:
Short Time Fourier Transform unit is configured as executing: carrying out Short Time Fourier Transform to the audio data of acquisition and obtains To the corresponding sonograph of audio data.
Segmenting unit is configured as executing: carrying out the segmentation of sonograph according to specified duration, obtains several frame audio sequences Column.
In one embodiment, cough identification module includes:
Local feature vectors extraction unit is configured as executing: extracting from several frame tonic trains and obtains audio data Several local feature vectors.
Full connection unit is configured as executing: carrying out the full connection of several local feature vectors, obtains the complete of audio data Office's feature vector.
Cough label prediction unit, is configured as executing: carrying out cough Tag Estimation to global characteristics vector, obtain audio The cough label of data.
Whether interpretation unit is configured as executing: being cough audio according to cough label audio data.
In one embodiment, voiceprint identification module includes:
Vocal print feature extraction unit is configured as executing: extracting from several frame tonic trains and obtains the sound of audio data Line feature.
Sounder Tag Estimation module, is configured as executing: carrying out sounder Tag Estimation according to vocal print feature, obtains sound The sounder label of frequency evidence.
Sounder determining module is configured as executing: the sounder of audio data is determined according to sounder label.
In one embodiment, monitoring data update module includes:
Timestamp information extraction unit is configured as executing: if audio data is cough audio, from audio data Extraction obtains timestamp information.
Cough length information and cough frequency information generating unit, are configured as executing: right according to cough audio and institute The timestamp information answered generates Cough length information and cough frequency information.
Cough monitoring data acquiring unit in short-term, is configured as executing: Cough length information and cough frequency information are made The cough monitoring data in short-term for being sounder in audio data.
Monitoring data updating unit is configured as executing: the monitoring data that will cough in short-term upload in monitor database, with Update the cough monitoring data of sounder in monitor database.
In one embodiment, Cough length information and cough frequency information generating unit include:
End-point detection unit is configured as executing: carrying out end-point detection to cough audio to determine in cough audio every time The beginning and end of cough.
Acquiring unit, be configured as executing: the timestamp information according to corresponding to the beginning and end coughed every time obtains cough The points or terminal number coughed in the temporal information, and statistics cough audio of coughing obtain cough frequency information.
Modules/unit function and the realization process of effect are specifically detailed in above-mentioned based on deep learning in above-mentioned apparatus Cough monitoring method in correspond to the realization process of step, details are not described herein.
It is appreciated that these module/units can by hardware, software, or a combination of both realize.When in hardware When realization, these modules may be embodied as one or more hardware modules, such as one or more specific integrated circuits.When with soft When part mode is realized, these modules may be embodied as the one or more computer journeys executed on the one or more processors Sequence, such as the program being stored in performed by the central processing unit 270 of Fig. 2 in memory 250.
Optionally, the disclosure also provides a kind of cough monitoring device 1000 based on deep learning, as shown in figure 12, packet It includes:
Processor 1001;And
Memory 1002 for 1001 executable instruction of storage processor;
Wherein, executable instruction by processor 1001 execute when realize above method embodiment in it is any shown in all or Person's part steps, executable instruction can be computer-readable instruction, in processor work, can pass through data line/communication Line 1003 reads the computer-readable instruction in memory 1002 and execution.
The concrete mode that the processor of device in the embodiment executes operation is somebody's turn to do related based on deep learning Detailed description is performed in the embodiment for monitoring method of coughing, no detailed explanation will be given here.
Optionally, the disclosure also provides a kind of computer readable storage medium, is stored thereon with computer program, computer The cough monitoring method based on deep learning shown in any of the above embodiment is realized when program is executed by processor.Wherein, should Computer readable storage medium can be the provisional and non-transitorycomputer readable storage medium for including computer program, should Memory 250 of the computer readable storage medium for example including computer program, above-mentioned computer program can be by cough monitoring clothes The central processing unit 270 of business device 200 is executed to complete the above-mentioned cough monitoring method based on deep learning.
It should be understood that the present invention is not limited to the precise structure already described above and shown in the accompanying drawings, and And various modifications and change can executed without departing from the scope.The scope of the present invention is limited only by the attached claims.

Claims (10)

1. a kind of cough monitoring method based on deep learning characterized by comprising
The audio data of acquisition is pre-processed to obtain several frame tonic trains;
Cough identification is carried out to several frame tonic trains, whether is cough audio with the determination audio data;And
Application on Voiceprint Recognition is carried out to several frame tonic trains, with the sounder of the determination audio data;
If the audio data is cough audio, according to the audio data to the cough monitoring data of the sounder into Row update processing.
2. if the method according to claim 1, wherein the audio data of described pair of acquisition is pre-processed to obtain Dry frame tonic train, comprising:
Short Time Fourier Transform is carried out to the audio data of acquisition and obtains the corresponding sonograph of the audio data;
The segmentation that the sonograph is carried out according to specified duration obtains several frame tonic trains.
3. the method according to claim 1, wherein described carry out cough identification to several frame tonic trains It whether is cough audio with the determination audio data, comprising:
It is extracted from several frame tonic trains and obtains several local feature vectors of the audio data;
The full connection for carrying out several local feature vectors obtains the global characteristics vector of the audio data;
Cough Tag Estimation is carried out to the global characteristics vector, obtains the cough label of the audio data;
Judge whether the audio data is cough audio according to the cough label.
4. the method according to claim 1, wherein described carry out vocal print knowledge to several frame tonic trains Not, with the sounder of the determination audio data, comprising:
It is extracted from several frame tonic trains and obtains the vocal print feature of the audio data;
Sounder Tag Estimation is carried out according to the vocal print feature, obtains the sounder label of the audio data;
The sounder of the audio data is determined according to the sounder label.
5. method according to any one of claims 1 to 4, which is characterized in that if the audio data is cough Audio is then updated processing according to cough monitoring data of the audio data to the sounder, comprising:
If the audio data is cough audio, is extracted from the audio data and obtain timestamp information;
Cough length information and cough frequency information are generated according to the cough audio and the corresponding timestamp information;
Cough in short-term using the Cough length information and cough frequency information as the sounder in the audio data Monitoring data;
The monitoring data of cough in short-term are uploaded in monitor database, to update sounder described in the monitor database Cough monitoring data.
6. according to the method described in claim 5, it is characterized in that, it is described according to the cough audio and it is corresponding described in Timestamp information generates Cough length information and cough frequency information, comprising:
The beginning and end that end-point detection is carried out to cough every time in the determination cough audio to the cough audio;
The Cough length information, and statistics are obtained according to timestamp information corresponding to the beginning and end coughed every time The points or terminal number coughed in the cough audio obtain the cough frequency information.
7. according to the method described in claim 5, it is characterized in that, the method also includes:
Judge whether the cough frequency of the sounder exceeds the safe frequency according to the cough frequency information;
If exceeding the safe frequency, information warning is generated.
8. a kind of cough monitoring device based on deep learning characterized by comprising
Preprocessing module is configured as executing: being pre-processed to obtain several frame tonic trains to the audio data of acquisition;
Cough identification module, is configured as executing: cough identification is carried out to several frame tonic trains, with the determination audio Whether data are cough audio;And
Voiceprint identification module is configured as executing: Application on Voiceprint Recognition is carried out to several frame tonic trains, with the determination audio The sounder of data;
Monitoring data update module is configured as executing: if the audio data is cough audio, according to the audio number Processing is updated according to the cough monitoring data to the sounder.
9. a kind of cough monitoring device based on deep learning characterized by comprising
Processor;And
Memory for storage processor executable instruction;
Wherein, method as described in any one of claim 1 to 7 is realized when the executable instruction is executed by the processor.
10. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the computer program The method as described in any one of claims 1 to 7 is realized when being executed by processor.
CN201811259953.4A 2018-10-26 2018-10-26 Cough monitoring method and device based on deep learning Pending CN109360584A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811259953.4A CN109360584A (en) 2018-10-26 2018-10-26 Cough monitoring method and device based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811259953.4A CN109360584A (en) 2018-10-26 2018-10-26 Cough monitoring method and device based on deep learning

Publications (1)

Publication Number Publication Date
CN109360584A true CN109360584A (en) 2019-02-19

Family

ID=65346788

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811259953.4A Pending CN109360584A (en) 2018-10-26 2018-10-26 Cough monitoring method and device based on deep learning

Country Status (1)

Country Link
CN (1) CN109360584A (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110111815A (en) * 2019-04-16 2019-08-09 平安科技(深圳)有限公司 Animal anomaly sound monitoring method and device, storage medium, electronic equipment
CN110890102A (en) * 2019-09-07 2020-03-17 创新奇智(重庆)科技有限公司 Engine defect detection algorithm based on RNN voiceprint recognition
CN111028862A (en) * 2019-12-17 2020-04-17 北京海益同展信息科技有限公司 Method, apparatus, computer device and storage medium for processing voice data
CN111091849A (en) * 2020-03-03 2020-05-01 龙马智芯(珠海横琴)科技有限公司 Snore identification method and device, storage medium snore stopping equipment and processor
CN111524537A (en) * 2020-03-24 2020-08-11 苏州数言信息技术有限公司 Method for identifying coughing and sneezing aiming at real-time voice flow
CN111860415A (en) * 2020-07-29 2020-10-30 浙江大华技术股份有限公司 Cough detection method, device, storage medium and electronic device
CN112687290A (en) * 2020-12-30 2021-04-20 同济大学 Compressed cough automatic detection method and embedded device
CN113327592A (en) * 2021-04-20 2021-08-31 青岛海尔空调器有限总公司 Air conditioning system control method and device, electronic equipment and storage medium
WO2021169742A1 (en) * 2020-02-26 2021-09-02 Oppo广东移动通信有限公司 Method and device for predicting operating state of transportation means, and terminal and storage medium
CN113488071A (en) * 2021-07-16 2021-10-08 河南牧原智能科技有限公司 Pig cough recognition method, device, equipment and readable storage medium
CN113948109A (en) * 2021-10-14 2022-01-18 广州蓝仕威克软件开发有限公司 System for recognizing physiological phenomenon based on voice

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN202235373U (en) * 2011-09-05 2012-05-30 苏州大学 Cough monitor
CN202723829U (en) * 2012-04-06 2013-02-13 肖遥 Intelligent cough monitoring and evaluation system
CN107242857A (en) * 2017-06-12 2017-10-13 南开大学 The intelligent traditional Chinese medical science based on deep learning integrates diagnosis and therapy system
CN108403119A (en) * 2018-02-26 2018-08-17 深圳乐测物联网科技有限公司 A kind of cough detection method
CN108701469A (en) * 2017-07-31 2018-10-23 深圳和而泰智能家居科技有限公司 Cough sound recognition methods, equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN202235373U (en) * 2011-09-05 2012-05-30 苏州大学 Cough monitor
CN202723829U (en) * 2012-04-06 2013-02-13 肖遥 Intelligent cough monitoring and evaluation system
CN107242857A (en) * 2017-06-12 2017-10-13 南开大学 The intelligent traditional Chinese medical science based on deep learning integrates diagnosis and therapy system
CN108701469A (en) * 2017-07-31 2018-10-23 深圳和而泰智能家居科技有限公司 Cough sound recognition methods, equipment and storage medium
CN108403119A (en) * 2018-02-26 2018-08-17 深圳乐测物联网科技有限公司 A kind of cough detection method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JUSTICE AMOH 等: "Deep Neural Networks for Identifying Cough Sounds", 《IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110111815A (en) * 2019-04-16 2019-08-09 平安科技(深圳)有限公司 Animal anomaly sound monitoring method and device, storage medium, electronic equipment
CN110890102A (en) * 2019-09-07 2020-03-17 创新奇智(重庆)科技有限公司 Engine defect detection algorithm based on RNN voiceprint recognition
CN111028862B (en) * 2019-12-17 2022-09-06 京东科技信息技术有限公司 Method, apparatus, computer device and storage medium for processing voice data
CN111028862A (en) * 2019-12-17 2020-04-17 北京海益同展信息科技有限公司 Method, apparatus, computer device and storage medium for processing voice data
WO2021169742A1 (en) * 2020-02-26 2021-09-02 Oppo广东移动通信有限公司 Method and device for predicting operating state of transportation means, and terminal and storage medium
CN111091849A (en) * 2020-03-03 2020-05-01 龙马智芯(珠海横琴)科技有限公司 Snore identification method and device, storage medium snore stopping equipment and processor
CN111524537A (en) * 2020-03-24 2020-08-11 苏州数言信息技术有限公司 Method for identifying coughing and sneezing aiming at real-time voice flow
CN111524537B (en) * 2020-03-24 2023-04-14 苏州数言信息技术有限公司 Cough and sneeze identification method aiming at real-time voice flow
CN111860415A (en) * 2020-07-29 2020-10-30 浙江大华技术股份有限公司 Cough detection method, device, storage medium and electronic device
CN112687290A (en) * 2020-12-30 2021-04-20 同济大学 Compressed cough automatic detection method and embedded device
CN112687290B (en) * 2020-12-30 2022-09-20 同济大学 Compressed cough automatic detection method and embedded device
CN113327592B (en) * 2021-04-20 2022-08-19 青岛海尔空调器有限总公司 Air conditioning system control method and device, electronic equipment and storage medium
CN113327592A (en) * 2021-04-20 2021-08-31 青岛海尔空调器有限总公司 Air conditioning system control method and device, electronic equipment and storage medium
CN113488071A (en) * 2021-07-16 2021-10-08 河南牧原智能科技有限公司 Pig cough recognition method, device, equipment and readable storage medium
CN113948109A (en) * 2021-10-14 2022-01-18 广州蓝仕威克软件开发有限公司 System for recognizing physiological phenomenon based on voice

Similar Documents

Publication Publication Date Title
CN109360584A (en) Cough monitoring method and device based on deep learning
CN109431507A (en) Cough disease identification method and device based on deep learning
AU2013239327B2 (en) A method and apparatus for processing patient sounds
Istrate et al. Information extraction from sound for medical telemonitoring
Vacher et al. The sweet-home project: Audio technology in smart homes to improve well-being and reliance
Aminikhanghahi et al. Using change point detection to automate daily activity segmentation
US20240127078A1 (en) Transferring learning in classifier-based sensing systems
CN109044396B (en) Intelligent heart sound identification method based on bidirectional long-time and short-time memory neural network
US20210267488A1 (en) Sensing system and method for monitoring time-dependent processes
CN111161715A (en) Specific sound event retrieval and positioning method based on sequence classification
JP2015207179A (en) behavior understanding system and program
Ghods et al. Activity2vec: Learning adl embeddings from sensor data with a sequence-to-sequence model
CN110111815A (en) Animal anomaly sound monitoring method and device, storage medium, electronic equipment
CN111429943B (en) Joint detection method for music and relative loudness of music in audio
CN113080907B (en) Pulse wave signal processing method and device
Min et al. Semi-supervised event detection using higher order statistics for multidimensional time series accelerometer data
CN114830257A (en) Cough detection system and method
CN117198468A (en) Intervention scheme intelligent management system based on behavior recognition and data analysis
CN114581749B (en) Audio-visual feature fusion target behavior identification method and device and application
CN115205972A (en) Six-step hand washing evaluation method and device based on deep learning
Pettinati et al. Automatic and robust identification of spontaneous coughs from covid-19 patients
Hervás et al. homeSound: A High Performance Platform for Massive Data Acquisition and Processing in Ambient Assisted Living Environments.
Wang et al. Epileptic Seizures Prediction Based on Unsupervised Learning for Feature Extraction
Kapoor et al. IoT based real-time face detection and recognition system
CN116052725B (en) Fine granularity borborygmus recognition method and device based on deep neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190219

RJ01 Rejection of invention patent application after publication