CN110890102A - Engine defect detection algorithm based on RNN voiceprint recognition - Google Patents

Engine defect detection algorithm based on RNN voiceprint recognition Download PDF

Info

Publication number
CN110890102A
CN110890102A CN201910844907.9A CN201910844907A CN110890102A CN 110890102 A CN110890102 A CN 110890102A CN 201910844907 A CN201910844907 A CN 201910844907A CN 110890102 A CN110890102 A CN 110890102A
Authority
CN
China
Prior art keywords
segments
detection algorithm
rnn
engine
defect detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910844907.9A
Other languages
Chinese (zh)
Inventor
张发恩
王忠强
唐永亮
黄家水
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Innovation Qizhi (chongqing) Technology Co Ltd
Original Assignee
Innovation Qizhi (chongqing) Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Innovation Qizhi (chongqing) Technology Co Ltd filed Critical Innovation Qizhi (chongqing) Technology Co Ltd
Priority to CN201910844907.9A priority Critical patent/CN110890102A/en
Publication of CN110890102A publication Critical patent/CN110890102A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • G10L25/30Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/18Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/21Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being power information
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Acoustics & Sound (AREA)
  • Human Computer Interaction (AREA)
  • Evolutionary Computation (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Signal Processing (AREA)
  • Biophysics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)

Abstract

The invention discloses an engine defect detection algorithm based on RNN voiceprint recognition, which belongs to the field of intelligent quality inspection and comprises the following specific steps: s1: segmenting all recorded audios, and extracting key segments of the generator during oiling and high-speed running; s2: marking the extracted key segments to construct a training set; s3: building a deep learning network model and training; s4: the trained network model is used for detecting the sound of the unknown generator, and the data amount required to be processed is greatly reduced by combining the engine sound key fragment extraction algorithm, so that the accuracy of the whole method is improved; the abnormal sound detection algorithm for deep learning can automatically learn the characteristics of normal sounds and various abnormal sounds through a large number of samples of known labels without manually analyzing the characteristics of the sounds, and can process the cyclic neural network model with variable length sequence input, so that the problem of processing audio by using a convolutional neural network in the prior art is solved.

Description

Engine defect detection algorithm based on RNN voiceprint recognition
Technical Field
The invention relates to the technical field of intelligent quality inspection, in particular to an engine defect detection algorithm based on RNN voiceprint recognition.
Background
At present, an algorithm based on a deep neural network is increasingly applied to the field of intelligent quality inspection. However, most of the existing algorithms detect the appearance of the part, the processing object is an image, and the network structure used is a convolutional neural network. However, for products that cannot visually recognize defects, such as engines, the image-based method described above cannot be used. In an actual production process, a worker can perform defect detection by recognizing abnormal sounds when an engine is operated, but in the field of computers, a technology for detecting defects of products by audio is still blank.
Most of the existing abnormal sound detection algorithms recognize the characteristics of abnormal sounds manually and summarize a set of algorithm flows to judge unknown sounds. The algorithm cannot automatically learn the characteristics of abnormal sounds, so the application range is small, and the algorithm cannot be repeatedly used for different types of sound detection. There is also a sound classification algorithm based on a convolutional neural network. Firstly, each audio frequency is regulated into a window with the same length of 20 frames; then extracting 12-dimensional Mel cepstrum coefficients and first-order and second-order differences thereof from each window, and summing up 36-dimensional characteristic vectors; then, taking the feature vectors of all windows of each video as 720 × 1 or 36 × 20 images, and respectively training the sounds of known classes by using one-dimensional and two-dimensional convolutional neural networks; and finally, predicting the sound of unknown class by using the trained network, thus obtaining the class of the sound. However, the convolutional neural network is not completely applicable to audio, and since the audio length is usually not fixed, preprocessing operations such as cropping need to be performed on the audio.
In summary, the prior art has the following disadvantages:
firstly, most of the sound features used by the existing algorithm are mel cepstrum coefficients, and the features process actual sound according to the perception features of human ears, so the algorithm is very suitable for processing the sound, but for non-speech sounds such as engines, the features cannot fully reflect the characteristics of the sound;
secondly, most abnormal sound detection algorithms usually find out the characteristics of abnormal sounds manually, and design a series of fixed steps for judgment, rather than automatically learning out the characteristics of abnormal sounds, so that the algorithms have no universality, and different algorithms need to be designed for different practical problems;
thirdly, some algorithms process the audio by using a convolutional neural network, so that although the characteristics of abnormal sounds can be learned and detected by self, the convolutional neural network requires that the input has the same size, and the audio generally cannot meet the requirement; if the audio is pre-processed by cropping or the like, part of the information may be lost.
Based on the method, the engine defect detection algorithm based on RNN voiceprint recognition is designed to solve the problem that some faults in the assembled engine cannot be recognized through visual observation.
Disclosure of Invention
The invention aims to provide an engine defect detection algorithm based on RNN voiceprint recognition, which is combined with an engine sound key segment extraction algorithm, a deep learning abnormal sound detection algorithm and a recurrent neural network model capable of processing variable length sequence input so as to solve the problems in the background art.
In order to achieve the purpose, the invention provides the following technical scheme: an engine defect detection algorithm based on RNN voiceprint recognition comprises the following specific steps:
s1: segmenting all recorded audios, and extracting key segments of the generator during oiling and high-speed running;
s2: marking the extracted key segments to construct a training set;
s3: building a deep learning network model and training;
s4: and detecting the sound of the unknown generator by using the trained network model.
Preferably, the specific steps of step S1 are as follows:
s1.1: generating a two-dimensional spectrogram from the recorded original audio;
s1.2: finding out a high-frequency part from a two-dimensional spectrogram of sound by observing, and if the content of the high-frequency sound at a certain time is lower than a threshold value, segmenting and removing the high-frequency part; the remaining segments are segments containing high frequency sounds, i.e., key segments of sounds when the engine is running at high speed.
Preferably, the specific steps of step S1.1 are: by acquiring one-dimensional amplitude information in the recorded original audio, the window function with length N of 2048 and sliding distance of 512 is used as
Figure BDA0002194914500000031
Dividing the frame into a plurality of frames; and performs a discrete fourier transform on each 2048-length frame:
Figure BDA0002194914500000032
calculating characteristic vectors with the length of 2048 dimensions for each frame, and stacking the characteristic vectors of all the frames to obtain a two-dimensional spectrogram with the size of 2048 multiplied by n, wherein n is the number of the frames.
Preferably, the specific steps of step S1.2 are: taking the first 800 dimensions of the 2048-dimensional frequency feature vector as a low-frequency signal and the rest as a high-frequency signal, the energy of the high-frequency signal of each frame is:
Figure BDA0002194914500000033
and combining the continuous frames with high frequency energy higher than the threshold value to obtain a plurality of audio segments, removing segments with shorter duration, and using the rest segments for the next operation.
Preferably, in step S2, the labeling manner is performed one by one.
Preferably, in step S3, the deep learning network model adopts a long-short term memory network LSTM, and the specific implementation steps are as follows:
s3.1: taking a spectrogram with indefinite length and height of 2048 as input, and searching and learning the relation of the sequence in time;
s3.2: extracting the characteristics of an input sequence by using a two-layer LSTM network, and classifying by using two full connection layers DENSE;
s3.3: the classification of the audio is output as a vector with a length of M +1, where M is the type of the anomaly, and the value in the vector is between 0 and 1, which respectively indicates the probability that the audio is normal or has some anomaly, i.e. whether the anomaly exists and the type of the anomaly.
Preferably, in step S4, if all the segments are identified as normal after the detection, the engine is considered to have no defect; if any of the segments are identified as abnormal, the engine is considered to be defective.
Compared with the prior art, the invention has the beneficial effects that:
1. an engine sound key segment extraction algorithm is adopted: by utilizing the characteristics of the engine sound, the key segments in high-speed operation are segmented, so that the data volume to be processed is greatly reduced, the interference of a large number of useless segments is avoided, and the accuracy of the whole method is improved;
2. adopting an abnormal sound detection algorithm based on deep learning: the deep learning algorithm is adopted, the characteristics of normal and various abnormal sounds can be automatically learned through a large number of samples of known labels without manually analyzing the characteristics of the sounds, and meanwhile, the method can be conveniently used for processing other similar problems without largely modifying;
3. adopting a recurrent neural network model capable of processing variable-length sequence input: a cyclic neural network which can accept variable-length sequences as input is used as an input layer, and a deep neural network model is constructed on the basis of the cyclic neural network, so that the problem that all the input is required to have the same scale and the length of the audio is not fixed in the traditional method for processing the audio by using a convolutional neural network is successfully solved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic block diagram of the present invention;
FIG. 2 is a schematic diagram of a key segment extraction algorithm of the present invention;
FIG. 3 is a diagram of a recurrent neural network of the present invention;
FIG. 4 is a schematic diagram of a detection algorithm according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1-3, the present invention provides a technical solution: an engine defect detection algorithm based on RNN voiceprint recognition comprises the following specific steps:
s1: segmenting all recorded audios, and extracting key segments of the generator during oiling and high-speed running;
the method comprises the following specific steps:
s1.1: generating a two-dimensional spectrogram from the recorded original audio, acquiring one-dimensional amplitude information in the recorded original audio, and taking a window function with a length N of 2048 and a sliding distance of 512 as
Figure BDA0002194914500000051
Divide it into severalA frame; and performs a discrete fourier transform on each 2048-length frame:
Figure BDA0002194914500000052
calculating characteristic vectors with length of 2048 dimensions for each frame, and stacking the characteristic vectors of all the frames to obtain a two-dimensional spectrogram with the size of 2048 multiplied by n, wherein n is the number of the frames;
s1.2: finding out a high-frequency part from a two-dimensional spectrogram of sound by observing, and if the content of the high-frequency sound at a certain time is lower than a threshold value, segmenting and removing the high-frequency part; the remaining segment is a segment containing high-frequency sounds, i.e., a key segment of sounds during high-speed operation of the engine, and in the spectrogram of fig. 2, the horizontal axis represents time, the vertical axis represents frequency, the value of the corresponding point represents the intensity of the frequency at that time,
the method comprises the following specific steps: taking the first 800 dimensions of the 2048-dimensional frequency feature vector as a low-frequency signal and the rest as a high-frequency signal, the energy of the high-frequency signal of each frame is:
Figure BDA0002194914500000053
combining successive frames with high frequency energy above the threshold of 0.05, resulting in several audio segments, removing segments of less than 20 frames in duration, the remaining segments being used for the next operation.
S2: and marking the extracted key segments one by one in a marking mode, and ensuring enough normal and abnormal sound segments in a training set. If an audio is normal, all its segments can be marked as normal; if one audio is abnormal, all the segments are not abnormal, so that the segments need to be labeled one by one to construct a training set;
s3: a deep learning network model is built and trained, a long and short term memory network LSTM is adopted for the deep learning network model, and the method is one of a recurrent neural network RNN and specifically comprises the following steps:
s3.1: receiving a spectrogram which is obtained before and has an indefinite length and a height of 2048 as input, and searching and learning the relation of the sequence on time;
s3.2: extracting the characteristics of an input sequence by using a two-layer LSTM network, and classifying by using two full connection layers DENSE;
s3.3: outputting the classification of the audio, wherein the classification of the audio is a vector with the length of M +1, M is the type of the abnormity, and the value in the vector is between 0 and 1, which respectively represents the probability that the audio is normal or has some abnormity, namely whether the abnormity exists and the type of the abnormity exists;
s4: detecting the sound of an unknown generator by using a trained network model, similarly extracting key segments from a section of audio to be detected, then performing anomaly detection on the segments by using the trained model, and determining that the engine has no defects if all the segments are identified as normal after detection; if any of the segments are identified as abnormal, the engine is considered to be defective.
The specific working principle is as follows:
extracting all audio frequency segments of the engine in high-speed operation by a key segment extraction algorithm; then, in the training process, the training sample is sent to a neural network, and the estimation result and the real label of the neural network are jointly used for optimizing the network; after training is completed, the network can be used for predicting whether an audio frequency is abnormal, and the prediction label can be obtained only by inputting the segments into the trained network.
1. Advantage of engine sound key segment extraction algorithm
The unusual sound is usually noticeable only when the engine is running at high speed, so the unusual sound detection is generally performed only for these segments, while the other segments are ignored. The algorithm utilizes the characteristics of the engine sound to segment the key segments during high-speed operation, thereby greatly reducing the data volume to be processed, avoiding the interference of a large number of useless segments and improving the accuracy of the whole method.
2. Abnormal sound detection algorithm based on deep learning
The method adopts a deep learning algorithm, and can automatically learn the characteristics of normal and various abnormal sounds through a large number of samples of known labels without manually analyzing the characteristics of the sounds. At the same time, the method can be conveniently used for processing other similar problems without a great deal of modification.
3. Cyclic neural network model capable of processing variable-length sequence input
Although audio can also be processed using convolutional neural networks, they require all inputs to be of the same scale, but the length of the audio is usually not fixed, and thus it is inconvenient to use convolutional neural networks. In the method, a cyclic neural network which can accept variable-length sequences as input is used as an input layer, and a deep neural network model is constructed on the basis of the input layer, so that the problems are successfully avoided.
Example (b):
as in the section of engine audio in fig. 4, 11 key segments can be extracted from the section of engine audio by the key segment extracting step; the segments are predicted through a trained neural network, wherein one segment is judged to be abnormal in howling and the other segments are judged to be normal; and therefore it is finally judged that the engine has the time-howling abnormality.
In the description herein, references to the description of "one embodiment," "an example," "a specific example" or the like are intended to mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
The preferred embodiments of the invention disclosed above are intended to be illustrative only. The preferred embodiments are not intended to be exhaustive or to limit the invention to the precise embodiments disclosed. Obviously, many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the invention and the practical application, to thereby enable others skilled in the art to best utilize the invention. The invention is limited only by the claims and their full scope and equivalents.

Claims (7)

1. An engine defect detection algorithm based on RNN voiceprint recognition is characterized in that: the method comprises the following specific steps:
s1: segmenting all recorded audios, and extracting key segments of the generator during oiling and high-speed running;
s2: marking the extracted key segments to construct a training set;
s3: building a deep learning network model and training;
s4: and detecting the sound of the unknown generator by using the trained network model.
2. The RNN voiceprint recognition based engine defect detection algorithm of claim 1, wherein: the specific steps of step S1 are as follows:
s1.1: generating a two-dimensional spectrogram from the recorded original audio;
s1.2: finding out a high-frequency part from a two-dimensional spectrogram of sound by observing, and if the content of the high-frequency sound at a certain time is lower than a threshold value, segmenting and removing the high-frequency part; the remaining segments are segments containing high frequency sounds, i.e., key segments of sounds when the engine is running at high speed.
3. The RNN voiceprint recognition based engine defect detection algorithm of claim 2, wherein: the specific steps of step S1.1 are: by acquiring one-dimensional amplitude information in the recorded original audio, the window function with length N of 2048 and sliding distance of 512 is used as
Figure FDA0002194914490000011
Dividing the frame into a plurality of frames; and performs a discrete fourier transform on each 2048-length frame:
Figure FDA0002194914490000012
calculating characteristic vectors with the length of 2048 dimensions for each frame, and stacking the characteristic vectors of all the frames to obtain a two-dimensional spectrogram with the size of 2048 multiplied by n, wherein n is the number of the frames.
4. The RNN voiceprint recognition based engine defect detection algorithm of claim 3, wherein: the specific steps of step S1.2 are: taking the first 800 dimensions of the 2048-dimensional frequency feature vector as a low-frequency signal and the rest as a high-frequency signal, the energy of the high-frequency signal of each frame is:
Figure FDA0002194914490000021
and combining the continuous frames with high frequency energy higher than the threshold value to obtain a plurality of audio segments, removing segments with shorter duration, and using the rest segments for the next operation.
5. The RNN voiceprint recognition based engine defect detection algorithm of claim 1, wherein: in the step S2, the labeling manner is performed one by one.
6. The RNN voiceprint recognition based engine defect detection algorithm of claim 4, wherein: in step S3, the deep learning network model adopts a long-short term memory network LSTM, and the specific implementation steps are as follows:
s3.1: taking a spectrogram with indefinite length and height of 2048 as input, and searching and learning the relation of the sequence in time;
s3.2: extracting the characteristics of an input sequence by using a two-layer LSTM network, and classifying by using two full connection layers DENSE;
s3.3: the classification of the audio is output as a vector with a length of M +1, where M is the type of the anomaly, and the value in the vector is between 0 and 1, which respectively indicates the probability that the audio is normal or has some anomaly, i.e. whether the anomaly exists and the type of the anomaly.
7. The RNN voiceprint recognition based engine defect detection algorithm of claim 6, wherein: in step S4, if all the detected segments are identified as normal, the engine is considered to have no defect; if any of the segments are identified as abnormal, the engine is considered to be defective.
CN201910844907.9A 2019-09-07 2019-09-07 Engine defect detection algorithm based on RNN voiceprint recognition Pending CN110890102A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910844907.9A CN110890102A (en) 2019-09-07 2019-09-07 Engine defect detection algorithm based on RNN voiceprint recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910844907.9A CN110890102A (en) 2019-09-07 2019-09-07 Engine defect detection algorithm based on RNN voiceprint recognition

Publications (1)

Publication Number Publication Date
CN110890102A true CN110890102A (en) 2020-03-17

Family

ID=69745924

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910844907.9A Pending CN110890102A (en) 2019-09-07 2019-09-07 Engine defect detection algorithm based on RNN voiceprint recognition

Country Status (1)

Country Link
CN (1) CN110890102A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111640437A (en) * 2020-05-25 2020-09-08 中国科学院空间应用工程与技术中心 Voiceprint recognition method and system based on deep learning
CN112259078A (en) * 2020-10-15 2021-01-22 上海依图网络科技有限公司 Method and device for training audio recognition model and recognizing abnormal audio
CN112509599A (en) * 2020-10-21 2021-03-16 中国人民解放军陆军炮兵防空兵学院 Acoustic spectrum fault analysis and diagnosis method based on BP neural network and Mel cepstrum
CN112836315A (en) * 2021-02-24 2021-05-25 上海交通大学 Neural network-based limit switch production line abnormity monitoring method
CN112992179A (en) * 2021-02-05 2021-06-18 安徽绿舟科技有限公司 Recognition method for detecting faults of gas turbine based on voiceprint signals
CN113571092A (en) * 2021-07-14 2021-10-29 东软集团股份有限公司 Method for identifying abnormal sound of engine and related equipment thereof
CN114088405A (en) * 2021-11-10 2022-02-25 中国人民解放军陆军炮兵防空兵学院 Engine fault diagnosis method of CNN fault diagnosis model based on spectrogram
CN114112401A (en) * 2021-11-10 2022-03-01 中国人民解放军陆军炮兵防空兵学院 Engine fault diagnosis method of LSTM fault diagnosis model based on spectrogram
CN115083395A (en) * 2022-08-23 2022-09-20 聊城大学 Engine sound detection system based on convolutional neural network and support vector machine
CN115134485A (en) * 2021-03-10 2022-09-30 霍尼韦尔国际公司 Video surveillance system with audio analysis adapted to specific environments to facilitate identification of anomalous events in specific environments
CN115602196A (en) * 2022-12-12 2023-01-13 杭州兆华电子股份有限公司(Cn) Abnormal sound recognition system and method for fixed-speed motor
CN117116290A (en) * 2023-08-03 2023-11-24 中科航迈数控软件(深圳)有限公司 Method and related equipment for positioning defects of numerical control machine tool parts based on multidimensional characteristics
CN117116290B (en) * 2023-08-03 2024-05-24 中科航迈数控软件(深圳)有限公司 Method and related equipment for positioning defects of numerical control machine tool parts based on multidimensional characteristics

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140019390A1 (en) * 2012-07-13 2014-01-16 Umami, Co. Apparatus and method for audio fingerprinting
CN104167207A (en) * 2014-06-20 2014-11-26 国家电网公司 Equipment sound identification method based on transformer substation patrol inspection robot
CN104819846A (en) * 2015-04-10 2015-08-05 北京航空航天大学 Rolling bearing sound signal fault diagnosis method based on short-time Fourier transform and sparse laminated automatic encoder
CN105275833A (en) * 2015-10-30 2016-01-27 北京航空航天大学 CEEMD (Complementary Empirical Mode Decomposition)-STFT (Short-Time Fourier Transform) time-frequency information entropy and multi-SVM (Support Vector Machine) based fault diagnosis method for centrifugal pump
CN106873440A (en) * 2015-12-14 2017-06-20 富士施乐株式会社 Diagnostic device, diagnostic system, equipment and diagnostic method
CN109087655A (en) * 2018-07-30 2018-12-25 桂林电子科技大学 A kind of monitoring of traffic route sound and exceptional sound recognition system
CN109192222A (en) * 2018-07-23 2019-01-11 浙江大学 A kind of sound abnormality detecting system based on deep learning
CN109360584A (en) * 2018-10-26 2019-02-19 平安科技(深圳)有限公司 Cough monitoring method and device based on deep learning
KR20190018798A (en) * 2017-08-16 2019-02-26 강병수 car noise sound with Convolution Nueral Network classification method
CN109431507A (en) * 2018-10-26 2019-03-08 平安科技(深圳)有限公司 Cough disease identification method and device based on deep learning
CN109616140A (en) * 2018-12-12 2019-04-12 浩云科技股份有限公司 A kind of abnormal sound analysis system
CN109767785A (en) * 2019-03-06 2019-05-17 河北工业大学 Ambient noise method for identifying and classifying based on convolutional neural networks
CN110189769A (en) * 2019-05-23 2019-08-30 复钧智能科技(苏州)有限公司 Abnormal sound detection method based on multiple convolutional neural networks models couplings

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140019390A1 (en) * 2012-07-13 2014-01-16 Umami, Co. Apparatus and method for audio fingerprinting
CN104167207A (en) * 2014-06-20 2014-11-26 国家电网公司 Equipment sound identification method based on transformer substation patrol inspection robot
CN104819846A (en) * 2015-04-10 2015-08-05 北京航空航天大学 Rolling bearing sound signal fault diagnosis method based on short-time Fourier transform and sparse laminated automatic encoder
CN105275833A (en) * 2015-10-30 2016-01-27 北京航空航天大学 CEEMD (Complementary Empirical Mode Decomposition)-STFT (Short-Time Fourier Transform) time-frequency information entropy and multi-SVM (Support Vector Machine) based fault diagnosis method for centrifugal pump
CN106873440A (en) * 2015-12-14 2017-06-20 富士施乐株式会社 Diagnostic device, diagnostic system, equipment and diagnostic method
KR20190018798A (en) * 2017-08-16 2019-02-26 강병수 car noise sound with Convolution Nueral Network classification method
CN109192222A (en) * 2018-07-23 2019-01-11 浙江大学 A kind of sound abnormality detecting system based on deep learning
CN109087655A (en) * 2018-07-30 2018-12-25 桂林电子科技大学 A kind of monitoring of traffic route sound and exceptional sound recognition system
CN109360584A (en) * 2018-10-26 2019-02-19 平安科技(深圳)有限公司 Cough monitoring method and device based on deep learning
CN109431507A (en) * 2018-10-26 2019-03-08 平安科技(深圳)有限公司 Cough disease identification method and device based on deep learning
CN109616140A (en) * 2018-12-12 2019-04-12 浩云科技股份有限公司 A kind of abnormal sound analysis system
CN109767785A (en) * 2019-03-06 2019-05-17 河北工业大学 Ambient noise method for identifying and classifying based on convolutional neural networks
CN110189769A (en) * 2019-05-23 2019-08-30 复钧智能科技(苏州)有限公司 Abnormal sound detection method based on multiple convolutional neural networks models couplings

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
杨毫鸽: "《飞机发动机异常声音识别方法研究》", 《中国优秀硕士学位论文全文数据库(电子期刊)》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111640437A (en) * 2020-05-25 2020-09-08 中国科学院空间应用工程与技术中心 Voiceprint recognition method and system based on deep learning
CN112259078A (en) * 2020-10-15 2021-01-22 上海依图网络科技有限公司 Method and device for training audio recognition model and recognizing abnormal audio
CN112509599A (en) * 2020-10-21 2021-03-16 中国人民解放军陆军炮兵防空兵学院 Acoustic spectrum fault analysis and diagnosis method based on BP neural network and Mel cepstrum
CN112992179A (en) * 2021-02-05 2021-06-18 安徽绿舟科技有限公司 Recognition method for detecting faults of gas turbine based on voiceprint signals
CN112836315A (en) * 2021-02-24 2021-05-25 上海交通大学 Neural network-based limit switch production line abnormity monitoring method
CN115134485A (en) * 2021-03-10 2022-09-30 霍尼韦尔国际公司 Video surveillance system with audio analysis adapted to specific environments to facilitate identification of anomalous events in specific environments
CN113571092A (en) * 2021-07-14 2021-10-29 东软集团股份有限公司 Method for identifying abnormal sound of engine and related equipment thereof
CN113571092B (en) * 2021-07-14 2024-05-17 东软集团股份有限公司 Engine abnormal sound identification method and related equipment thereof
CN114088405A (en) * 2021-11-10 2022-02-25 中国人民解放军陆军炮兵防空兵学院 Engine fault diagnosis method of CNN fault diagnosis model based on spectrogram
CN114112401A (en) * 2021-11-10 2022-03-01 中国人民解放军陆军炮兵防空兵学院 Engine fault diagnosis method of LSTM fault diagnosis model based on spectrogram
CN115083395A (en) * 2022-08-23 2022-09-20 聊城大学 Engine sound detection system based on convolutional neural network and support vector machine
CN115602196A (en) * 2022-12-12 2023-01-13 杭州兆华电子股份有限公司(Cn) Abnormal sound recognition system and method for fixed-speed motor
CN117116290A (en) * 2023-08-03 2023-11-24 中科航迈数控软件(深圳)有限公司 Method and related equipment for positioning defects of numerical control machine tool parts based on multidimensional characteristics
CN117116290B (en) * 2023-08-03 2024-05-24 中科航迈数控软件(深圳)有限公司 Method and related equipment for positioning defects of numerical control machine tool parts based on multidimensional characteristics

Similar Documents

Publication Publication Date Title
CN110890102A (en) Engine defect detection algorithm based on RNN voiceprint recognition
CN111784633B (en) Insulator defect automatic detection algorithm for electric power inspection video
CN112381788B (en) Part surface defect increment detection method based on double-branch matching network
CN111368764B (en) False video detection method based on computer vision and deep learning algorithm
KR102132407B1 (en) Method and apparatus for estimating human emotion based on adaptive image recognition using incremental deep learning
CN112329438B (en) Automatic lie detection method and system based on domain countermeasure training
CN112766218B (en) Cross-domain pedestrian re-recognition method and device based on asymmetric combined teaching network
CN111738054B (en) Behavior anomaly detection method based on space-time self-encoder network and space-time CNN
US20210096530A1 (en) System and method for identifying manufacturing defects
CN111901627B (en) Video processing method and device, storage medium and electronic equipment
CN110909131A (en) Model generation method, emotion recognition method, system, device and storage medium
CN114898466A (en) Video motion recognition method and system for smart factory
CN110610123A (en) Multi-target vehicle detection method and device, electronic equipment and storage medium
CN114495983A (en) Equipment failure voiceprint monitoring system based on cloud edge collaboration
US11238289B1 (en) Automatic lie detection method and apparatus for interactive scenarios, device and medium
CN114782410A (en) Insulator defect detection method and system based on lightweight model
CN110618129A (en) Automatic power grid wire clamp detection and defect identification method and device
CN112562727B (en) Audio scene classification method, device and equipment applied to audio monitoring
Jayanthi et al. Fruit quality inspection system using image processing
CN113450827A (en) Equipment abnormal condition voiceprint analysis algorithm based on compressed neural network
EP3847646B1 (en) An audio processing apparatus and method for audio scene classification
CN112419304B (en) Multi-stage target detection method and device for one-dimensional data
CN115457620A (en) User expression recognition method and device, computer equipment and storage medium
CN111898452A (en) Video monitoring networking system
CN111539277A (en) Detection method and system for construction machinery in power transmission line area

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200317