CN113671031A - Wall hollowing detection method and device - Google Patents

Wall hollowing detection method and device Download PDF

Info

Publication number
CN113671031A
CN113671031A CN202110958298.7A CN202110958298A CN113671031A CN 113671031 A CN113671031 A CN 113671031A CN 202110958298 A CN202110958298 A CN 202110958298A CN 113671031 A CN113671031 A CN 113671031A
Authority
CN
China
Prior art keywords
network
neural network
feature
output
convolutional neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110958298.7A
Other languages
Chinese (zh)
Other versions
CN113671031B (en
Inventor
解传栋
李先刚
邹伟
汤志远
赵帅江
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Seashell Housing Beijing Technology Co Ltd
Original Assignee
Beijing Fangjianghu Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Fangjianghu Technology Co Ltd filed Critical Beijing Fangjianghu Technology Co Ltd
Priority to CN202110958298.7A priority Critical patent/CN113671031B/en
Publication of CN113671031A publication Critical patent/CN113671031A/en
Application granted granted Critical
Publication of CN113671031B publication Critical patent/CN113671031B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/04Analysing solids
    • G01N29/045Analysing solids by imparting shocks to the workpiece and detecting the vibrations or the acoustic waves caused by the shocks
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/44Processing the detected response signal, e.g. electronic circuits specially adapted therefor
    • G01N29/4472Mathematical theories or simulation
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N29/00Investigating or analysing materials by the use of ultrasonic, sonic or infrasonic waves; Visualisation of the interior of objects by transmitting ultrasonic or sonic waves through the object
    • G01N29/44Processing the detected response signal, e.g. electronic circuits specially adapted therefor
    • G01N29/4481Neural networks
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2291/00Indexing codes associated with group G01N29/00
    • G01N2291/02Indexing codes associated with the analysed material
    • G01N2291/028Material parameters
    • G01N2291/0289Internal structure, e.g. defects, grain size, texture

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Chemical & Material Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Biochemistry (AREA)
  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Algebra (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Mathematical Physics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Machine Translation (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

The application provides a wall hollowing detection method and device, wherein the method comprises the following steps: acquiring an audio signal generated by knocking a wall body; acquiring a first feature and a second feature of the audio signal; splicing the first feature and the second feature to obtain a spliced feature; acquiring an identification result corresponding to the splicing characteristic based on a preset audio detection network model; and determining whether the wall is in a hollow state or not according to the identification result. The method can improve the accuracy of wall hollowing detection on the premise of reducing labor cost.

Description

Wall hollowing detection method and device
Technical Field
The embodiment of the disclosure relates to a wall hollowing detection method and device.
Background
The facing layer of the building structure system is subjected to natural force action and thermal force action for a long time, so that the problems of aging, damage, looseness and the like of the facing layer material and a connecting material and a supporting structure of the facing layer material are solved, the integral or local debonding, looseness, hollowing and even shedding of the facing layer are caused, and personal and property injuries are easily caused.
At present, the hollowing detection of the wall body is mainly realized by methods such as manual knocking, ear identification or infrared imaging.
In the process of realizing the application, the method for identifying through manual knocking and ear identification has higher labor cost, and inaccurate detection is caused due to the influence of human factors; the infrared imaging detection is easily interfered by the outside world, such as noise, heat source and the like, which causes inaccurate detection.
Disclosure of Invention
In view of this, the application provides a wall hollowing detection method and device, which can improve the accuracy of wall hollowing detection on the premise of reducing labor cost.
In order to solve the technical problem, the technical scheme of the application is realized as follows:
in one embodiment, a wall hollowing detection method is provided, the method comprising:
acquiring an audio signal generated by knocking a wall body;
acquiring a first feature and a second feature of the audio signal;
splicing the first feature and the second feature to obtain a spliced feature;
acquiring an identification result corresponding to the splicing characteristic based on a preset audio detection network model;
and determining whether the wall is in a hollow state or not according to the identification result.
In another embodiment, there is provided a wall hollowing detection apparatus, the apparatus comprising: the device comprises a storage unit, a first acquisition unit, a second acquisition unit, a third acquisition unit and a determination unit;
the storage unit is used for storing a preset audio detection network model;
the first acquisition unit is used for acquiring an audio signal generated by knocking a wall body;
the second obtaining unit is used for obtaining the first characteristic and the second characteristic of the audio signal obtained by the first obtaining unit; splicing the first characteristic and the second characteristic to obtain a spliced characteristic;
the third obtaining unit is configured to obtain, based on the preset audio detection network model stored in the storage unit, an identification result corresponding to the splicing feature obtained by the second obtaining unit;
the determining unit is configured to determine whether the wall is in an empty drum state according to the identification result obtained by the third obtaining unit.
In another embodiment, an electronic device is provided, comprising a memory, a processor and a computer program stored on the memory and executable on the processor, the processor implementing the steps of the wall hollowing detection method when executing the program.
In another embodiment, a computer-readable storage medium is provided, on which a computer program is stored which, when being executed by a processor, carries out the steps of the wall hollowing detection method.
In another embodiment, a computer program product is provided, comprising a computer program which, when being executed by a processor, carries out the wall hollowing detection method.
According to the technical scheme, the characteristics of the collected audio signals are obtained, the obtained characteristics are spliced to be used as the overall characteristics of the audio signals, and the recognition results corresponding to the spliced characteristics are obtained based on the preset audio detection network model; and determining whether the wall is in the empty drum state or not according to the identification result. According to the scheme, the accuracy of wall hollowing detection can be improved on the premise of reducing labor cost.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without inventive labor.
Fig. 1 is a schematic flow chart illustrating a process of acquiring a preset audio detection network model in an embodiment of the present application;
FIG. 2 is a schematic diagram of a residual CRNN-GLU structure in an embodiment of the present application;
FIG. 3 is a schematic view illustrating a process of detecting hollowing of a wall according to an embodiment of the present disclosure;
FIG. 4 is a schematic diagram of an audio signal acquisition device;
FIG. 5 is a schematic view illustrating a process of detecting hollowing of a wall according to a second embodiment of the present application;
FIG. 6 is a schematic structural diagram of a wall hollowing detection device in an embodiment of the present application;
fig. 7 is a schematic physical structure diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms "first," "second," "third," "fourth," and the like in the description and in the claims, as well as in the drawings, if any, are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the invention described herein are, for example, capable of operation in sequences other than those illustrated or otherwise described herein. Furthermore, the terms "comprising" and "having," as well as any variations thereof, are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements explicitly listed, but may include other steps or elements not explicitly listed or inherent to such process, method, article, or apparatus.
The technical solution of the present invention will be described in detail with specific examples. Several of the following embodiments may be combined with each other and some details of the same or similar concepts or processes may not be repeated in some embodiments.
The embodiment of the application provides a wall hollowing detection method, which comprises the steps of acquiring characteristics of collected audio signals, splicing the acquired characteristics to serve as integral characteristics of the audio signals, and acquiring identification results corresponding to the spliced characteristics based on a preset audio detection network model; and determining whether the wall is in the empty drum state or not according to the identification result. According to the scheme, the accuracy of wall hollowing detection can be improved on the premise of reducing labor cost.
The preset audio detection network model in the embodiment of the application can be pre-established for storage, can be directly used when needed, and can also be established and used when needed.
The method for establishing the preset audio detection network model specifically comprises the following steps:
referring to fig. 1, fig. 1 is a schematic flowchart of acquiring a preset audio detection network model in the embodiment of the present application. The method comprises the following specific steps:
step 101, obtaining a training sample.
The obtained training samples can be unlabeled, partially labeled or fully labeled samples, and the unlabeled samples can be labeled by using the initial audio detection network model as the training samples.
And 102, acquiring a first feature and a second feature of the training sample, and splicing the first feature and the second feature to obtain a spliced feature.
Wherein the first feature is a fundamental frequency feature;
the fundamental frequency features are used for representing tone characteristics, and the multi-dimensional fundamental frequency features can be selected and extracted according to practical application, for example, the following five fundamental frequency features can be selected:
1) the base frequency value of the voice frame;
2) the fundamental frequency first-order difference value of the adjacent frames;
3) the fundamental frequency second-order difference value of the adjacent frames;
4) the length of the current continuous base frequency band;
5) the difference between the current frame fundamental frequency value and the average value of the next N frame fundamental frequencies of the previous continuous fundamental frequency band, where N is usually selected to be 10.
The five-dimensional fundamental frequency features can achieve better feature combinations, but the method involved in the concrete implementation of the application is not limited to the five-dimensional feature combinations, and the increase or decrease of the feature number does not affect the application of the application.
The above method for extracting the fundamental frequency feature is an exemplary implementation manner, and the embodiment of the present application is not limited to the implementation manner.
The second characteristic is any one of the following characteristics or a characteristic obtained by splicing at least two of the following characteristics:
speech features, decoding features, neural network features.
That is, the second feature may be a speech feature, a decoding feature, or a neural network feature;
or the like, or, alternatively,
features after the speech features and the decoding features are spliced;
or the like, or, alternatively,
features after the phonetic features and neural network features are spliced;
or the like, or, alternatively,
decoding the characteristics after the characteristics and the neural network characteristics are spliced;
or the like, or, alternatively,
the characteristics after the speech characteristics, the decoding characteristics and the neural network characteristics are spliced.
And when the second feature is the spliced feature, respectively acquiring the features before splicing, splicing the acquired features, and acquiring the spliced feature, wherein if the second feature is the feature after splicing the decoding feature and the neural network feature, respectively acquiring the decoding feature and the neural network feature of the sample, and splicing the decoding feature and the neural network feature to acquire the spliced feature.
Wherein the content of the first and second substances,
the speech features are as follows: filter Bank (Fbank) characteristics, Mel-Frequency Cepstral Coefficients (MFCC), etc.;
neural network features are as follows: a bottle neck feature, a DNN feature, a CNN feature, an LSTM feature, etc.;
the decoding characteristics are as follows: threshold, maximum, minimum, confidence, grid distribution statistics, etc. for the sentence.
If any one of the acquired voice features, decoding features and neural network features has multiple features, the same type of features can be spliced.
In the embodiment of the present application, the manner of obtaining the speech feature, the decoding feature, and the neural network feature is not limited, and here, the manner of obtaining is not illustrated.
In the embodiment of the present application, when the first feature and the second feature are combined, the order of the two features is not limited, for example, the first feature is [ a ]1,A2,……,An]The second characteristic is [ B1,B2,……,Bm]Then the spliced feature may be [ A ]1,A2,……,An,B1,B2,……,Bm]May also be [ B ]1,B2,……,Bm,A1,A2,……,An]。
Step 103, establishing an initial audio detection network model.
The initial audio detection network model is established based on any one of the following neural networks: deep Neural Networks (DNNs), Convolutional Recurrent Neural Networks (CRNN), Convolutional Neural Networks (CNNs), Long-Short Term Memory Networks (LSTM), i.e., time-cycled Neural Networks;
when a Neural Network for establishing an initial audio detection Network model is a CRNN, the CRNN adopts a residual convolution recursive Network (CRNN) -Gated Linear Unit (GLU) structure;
when residual connection is carried out in the residual convolution recursive network-gate control linear unit structure, the network input before the gate control linear unit is added with the results of two convolution networks through gate control linear unit operation, and the results after residual connection are used as the input of the next CNN network.
In the embodiment of the present application, taking an example that the convolutional recurrent neural network includes two residual connections, the residual convolutional recurrent network-gated linear unit structure includes: a first residual connecting network, a second residual connecting network, a first convolutional neural network, a second convolutional neural network, a Long Short Term Memory (LSTM) network, a full connecting layer and an output layer;
wherein the first residual connecting network comprises a third convolutional neural network, a fourth convolutional neural network and a first gated linear unit, the third convolutional neural network and the fourth convolutional neural network are used for receiving network input, the output of the third convolutional neural network and the output of the fourth convolutional neural network are input to the first gated linear unit, and the output of the first gated linear unit is added with the network input to be used as the output of the first residual connecting network;
the first convolution neural network is in signal connection with the first residual error connecting network and is used for receiving the output of the first residual error connecting network;
the second residual connecting network is in signal connection with the first convolution neural network and used for receiving the output of the first convolution neural network, the second residual connecting network comprises a fifth convolution neural network, a sixth convolution neural network and a second gating linear unit, the fifth convolution neural network and the sixth convolution neural network are used for receiving the output of the first convolution neural network, the output of the fifth convolution neural network and the output of the sixth convolution neural network are input to the second gating linear unit, and the output of the second gating linear unit is added with the output of the first convolution neural network to serve as the output of the second residual connecting network;
the second convolutional neural network is in signal connection with the second residual error connecting network and receives the output of the second residual error connecting network, and the output of the second convolutional neural network is used as the network input of the long-term and short-term memory network;
the network output of the long-short term memory network is used as the input of the full connection layer;
the output of the full connection layer is used as the input of the output layer;
the output of the output layer is used as the output of the residual convolution recursive network-gated linear unit structure.
The first convolutional neural network to the sixth convolutional neural network may be the same network or different networks, which is not limited in this embodiment of the present application.
The network structure of the convolutional recurrent neural network is illustrated below with reference to the drawings.
Referring to fig. 2, fig. 2 is a schematic structural diagram of residual CRNN-GLU in the embodiment of the present application. In fig. 2, from top to bottom, a first gated linear unit (a first gated linear unit) corresponds to a first residual connecting network, a second gated linear unit (a second gated linear unit) corresponds to a second residual connecting network, and two CNNs in the first residual connecting network are a third convolutional neural network and a fourth convolutional neural network; two CNNs in the second residual error connection network are a fifth convolutional neural network and a sixth convolutional neural network, the CNN between the first residual error connection network and the second residual error connection network is a first convolutional neural network, and the CNN between the second residual error connection network and a Long Short Term Memory (LSTM) network is a second convolutional neural network.
The following four implementation processes are given for the structure:
a first part:
loss function: since there may be overlap between events, a LOSS function (LOSS) may be used in embodiments of the present application: BCE (binary cross-entry) loss function, defined as follows:
Figure BDA0003221151080000071
wherein, PnRepresents a prediction signal obtained by a residual error CRNN-GLU structure on the nth frame, OnIndicating that, at the nth frame, the audio corresponds to the correct signal,n denotes the total number of frames of one sample.
A second part:
GLU operation: in the network structure, the network is provided with a plurality of network interfaces,
Figure BDA0003221151080000072
represents a gating operation, the GLU (gated Linear Unit), is defined as follows:
Onext=(W*X+b)⊙σ(V*X+。)
wherein, PnextThe output of gating is represented, W and V represent convolution kernels of CNN, b and c represent bias, and the convolution kernels of the two CNN networks can be the same or different; x represents the input characteristic before GLU, sigma is a nonlinear sigmoid function, and an indicator is a matrix operator and multiplies the value at the corresponding position of the matrix, wherein the indicator represents convolution operation.
And a third part: as shown in fig. 2, when Residual connection is performed, the network input before GLU is added to the result of GLU operation by two convolutional neural networks, specifically as follows:
Figure BDA0003221151080000081
wherein, IiRepresenting features input prior to the ith GLU,
Figure BDA0003221151080000082
representing the output of the ith gate.
Connecting the residual error results RESiAs input to the next CNN network.
As shown in fig. 2, in the embodiment of the present application, two residual convolution recursive network-gated linear unit structures are used to perform two residual connections, but the implementation is not limited to the implementation of two residual connections;
if the requirement on the real-time performance is not high, the residual error part can be increased, the network can be designed to be deep, and the accuracy can be increased.
The fourth part:
after the CNN network, Pooling dimensionality reduction operation can be carried OUT, then the CNN network is sent to an LSTM (a Bi-LSTM can be adopted in a non-real-time scene), then a full connection (FNN) layer is connected, finally an output layer (OUT-PUT) is reached, then a loss function is calculated, and parameter iteration updating and optimization are carried OUT through an SGD (random gradient descent) method.
In the embodiment of the present application, implementation manners of the first part, the second part, and the fourth part are not limited, and when residual connection is implemented for the third part, the network input before the GLU is added to the result of two convolutions through the GLU operation, and the result RES after residual connection is addediAs input to the next CNN network. The realization of the scheme can improve the accuracy of network identification.
The output of the model is either an empty state or a non-empty state.
And 104, training the initial audio detection network model by using the splicing characteristics and the labels corresponding to the training samples to obtain a preset audio detection network model.
And at this point, the establishment of the preset audio detection network model is completed.
The process for detecting the wall hollowing in the embodiment of the application is described in detail below with reference to the accompanying drawings.
Example one
Referring to fig. 3, fig. 3 is a schematic view illustrating a process of detecting hollowing of a wall body according to an embodiment of the present application. The method comprises the following specific steps:
step 301, acquiring an audio signal generated by knocking a wall.
The audio signal acquisition can be carried out on the current wall body knocking, or the stored audio signal can be acquired in advance.
The capturing device for capturing the audio signal is not limited, such as but not limited to the capturing device shown in fig. 4, and fig. 4 is a schematic diagram of an audio signal capturing device.
The small circle part in fig. 4 is a microphone pickup device, the number of which is 1, 2, 4, 6 or 8 can be selected according to the cost, and the acquisition device in fig. 4 takes 8 microphones as an example.
The collecting device is provided with a liquid crystal display screen which can display the judgment result and other items which can assist in judgment in real time.
The computing function of the acquisition device can be selected to be nested in a device chip or partially adopt cloud computing and return a result.
The acquisition equipment is provided with a handle, so that the audio signal can be conveniently acquired for use.
Step 302, extracting a first feature and a second feature of the audio signal.
The first feature is a fundamental frequency feature;
the second characteristic is any one of the following characteristics or a characteristic obtained by splicing at least two of the following characteristics:
speech features, decoding features, neural network features.
And 303, splicing the first characteristic and the second characteristic to obtain a spliced characteristic.
In the embodiment of the present application, when the first feature and the second feature are combined, the order of the two features is not limited, for example, the first feature is [ a ]1,A2,……,An]The second characteristic is [ B1,B2,……,Bm]Then the spliced feature may be [ A ]1,A2,……,An,B1,B2,……,Bm]May also be [ B ]1,B2,……,Bm,A1,A2,……,An](ii) a But needs to be consistent with the sequence of feature splicing when the preset audio detection network model is established.
The first characteristic and the second characteristic are spliced to serve as the characteristic information of the audio signal, the multi-dimensional characteristic of the audio signal is considered, and the detection result can be obtained more accurately.
And 304, acquiring an identification result corresponding to the splicing characteristic based on a preset audio detection network model.
Wherein the identification result is: in an empty state or not.
And 305, determining whether the wall is in a hollow state according to the identification result.
When the identification result is in the empty drum state or not, determining whether the wall is in the empty drum state according to the identification result, including:
when the identification result is in the empty drum state, determining that the wall body is in the empty drum state;
and when the identification result is that the wall body is not in the empty drum state, determining that the wall body is not in the empty drum state.
According to the embodiment of the application, the first feature and the second feature are extracted from the collected audio signal, the first feature and the second feature are spliced to be used as the overall feature of the audio signal, and the identification result corresponding to the spliced feature is obtained based on a preset audio detection network model; and determining whether the wall is in the empty drum state or not according to the identification result. According to the scheme, the accuracy of wall hollowing detection can be improved on the premise of reducing labor cost
Example two
Referring to fig. 5, fig. 5 is a schematic view illustrating a process of detecting hollowing of a wall body according to a second embodiment of the present application. The method comprises the following specific steps:
step 501, obtaining an audio signal generated by knocking a wall.
The audio signal acquisition can be carried out on the current wall body knocking, or the stored audio signal can be acquired in advance.
The embodiment of the present application does not limit the acquisition device for acquiring audio signals.
Step 502, performing endpoint detection on the audio signal, and performing noise reduction processing.
Endpoint Detection, also called Voice Activity Detection, VAD, aims to distinguish between speech and non-speech areas. The end point detection is to accurately locate the starting point and the ending point of the voice from the voice with noise, remove the mute part and the noise part, and find a piece of content really effective in the voice.
VAD algorithms can be roughly classified into three categories: threshold-based VAD, VAD as classifier, model VAD.
Threshold-based VAD: the aim of distinguishing the voice from the non-voice is achieved by extracting characteristics of a time domain (short-time energy, short-time zero crossing rate and the like) or a frequency domain (MFCC, spectral entropy and the like) and reasonably setting a threshold. This is the conventional VAD method.
VAD as classifier: the voice detection can be regarded as a voice/non-voice two-classification problem, and a classifier is trained by a machine learning method to achieve the purpose of voice detection.
Model VAD: the speech segments and the non-speech segments can be distinguished by global information on the basis of decoding by using a complete acoustic model (the granularity of the modeling unit can be very coarse).
The embodiment of the present application does not limit the specific implementation manner of endpoint detection.
The specific implementation may be noise reduction or dereverberation.
In the specific implementation of the method, after the endpoint detection, the noise reduction or dereverberation processing is carried out, so that the external interference in the audio signal can be removed to a greater extent.
The method further comprises the following steps:
and if the signal intensity of the audio signal is smaller than a preset value, performing enhancement processing on the audio signal.
I.e. when the signal strength of the audio signal is relatively small, enhancement processing is performed.
The audio signal after preprocessing can easily extract meaningful Fbank characteristics and fundamental frequency characteristics, and the identification of the final audio signal is convenient.
Step 503, extracting a first feature and a second feature of the audio signal.
The first feature is a fundamental frequency feature;
the second characteristic is any one of the following characteristics or a characteristic obtained by splicing at least two of the following characteristics:
speech features, decoding features, neural network features.
And step 504, splicing the first feature and the second feature to obtain a spliced feature.
In the embodiment of the present application, when the first feature and the second feature are combined, the order of the two features is not limited, for example, the first feature is [ a ]1,A2,……,An]The second characteristic is [ B1,B2,……,Bm]Then the spliced feature may be [ A ]1,A2,……,An,B1,B2,……,Bm]May also be [ B ]1,B2,……,Bm,A1,A2,……,An](ii) a But needs to be consistent with the sequence of feature splicing when the preset audio detection network model is established.
The first characteristic and the second characteristic are spliced to serve as the characteristic information of the audio signal, the multi-dimensional characteristic of the audio signal is considered, and the detection result can be obtained more accurately.
And 505, acquiring an identification result corresponding to the splicing characteristic based on a preset audio detection network model.
Wherein the identification result is: in an empty state or not.
Step 506, determining whether the wall is in a hollow state according to the identification result.
When the identification result is in the empty drum state or not, determining whether the wall is in the empty drum state according to the identification result, including:
when the identification result is in the empty drum state, determining that the wall body is in the empty drum state;
and when the identification result is that the wall body is not in the empty drum state, determining that the wall body is not in the empty drum state.
According to the embodiment of the application, the collected audio signals are preprocessed, first features and second features of the preprocessed audio signals are extracted, the first features and the second features are spliced to be used as overall features of the audio signals, and recognition results corresponding to the spliced features are obtained based on a preset audio detection network model; and determining whether the wall is in the empty drum state or not according to the identification result. According to the scheme, the accuracy of wall hollowing detection can be improved on the premise of reducing labor cost.
Based on the same inventive concept, the embodiment of the application also provides a wall hollowing detection device. Referring to fig. 6, fig. 6 is a schematic structural diagram of a wall hollowing detection device in the embodiment of the present application. The device comprises: a storage unit 601, a first acquisition unit 602, a second acquisition unit 603, a third acquisition unit 604, and a determination unit 605;
the storage unit 601 is used for storing a preset audio detection network model;
a first obtaining unit 602, configured to obtain an audio signal generated by tapping a wall;
a second obtaining unit 603 configured to extract a first feature and a second feature of the audio signal obtained by the first obtaining unit 602; splicing the first characteristic and the second characteristic to obtain a spliced characteristic;
a third obtaining unit 604, configured to obtain, based on the preset audio detection network model stored in the storage unit 601, an identification result corresponding to the splicing feature obtained by the second obtaining unit 603;
a determining unit 605, configured to determine whether the wall is in an empty drum state according to the identification result obtained by the third obtaining unit 604.
In another embodiment of the present invention, the substrate is,
the first feature is a fundamental frequency feature;
the second characteristic is any one of the following characteristics or a characteristic obtained by splicing at least two of the following characteristics:
speech features, decoding features, neural network features.
In another embodiment, the apparatus comprises: a processing unit 606;
the processing unit 606 is configured to, after the first obtaining unit 602 obtains the audio signal generated by tapping the wall, perform endpoint detection on the audio signal before the second obtaining unit 603 obtains the first feature and the second feature of the audio signal, and perform noise reduction processing on the audio signal.
In another embodiment of the present invention, the substrate is,
the storage unit is specifically used for acquiring training samples; acquiring a first feature and a second feature of the audio signal in the training sample, and splicing the first feature and the second feature to obtain a spliced feature; establishing an initial audio detection network model; wherein the initial audio detection network model is established based on any one of the following neural networks: a deep neural network DNN, a convolution recurrent neural network CRNN, a convolution neural network CNN and a long-short term memory network LSTM; and training the initial audio detection network model by using the splicing characteristics and the labels corresponding to the training samples to obtain the preset audio detection network model.
In another embodiment of the present invention, the substrate is,
when a neural network for establishing an initial audio detection network model is a CRNN, the CRNN adopts a residual convolution recursive network-gate control linear unit structure;
when residual connection is carried out in the residual convolution recursive network-gate control linear unit structure, the network input before the gate control linear unit is added with the results of two convolution networks through gate control linear unit operation, and the results after residual connection are used as the input of the next CNN network.
In another embodiment, a residual convolutional recursive network-gated linear cell structure comprises: a first residual connecting network, a second residual connecting network, a first convolutional neural network, a second convolutional neural network, a Long Short Term Memory (LSTM) network, a full connecting layer and an output layer;
wherein the first residual connecting network comprises a third convolutional neural network, a fourth convolutional neural network and a first gated linear unit, the third convolutional neural network and the fourth convolutional neural network are used for receiving network input, the output of the third convolutional neural network and the output of the fourth convolutional neural network are input to the first gated linear unit, and the output of the first gated linear unit is added with the network input to be used as the output of the first residual connecting network;
the first convolution neural network is in signal connection with the first residual error connecting network and is used for receiving the output of the first residual error connecting network;
the second residual connecting network is in signal connection with the first convolution neural network and used for receiving the output of the first convolution neural network, the second residual connecting network comprises a fifth convolution neural network, a sixth convolution neural network and a second gating linear unit, the fifth convolution neural network and the sixth convolution neural network are used for receiving the output of the first convolution neural network, the output of the fifth convolution neural network and the output of the sixth convolution neural network are input to the second gating linear unit, and the output of the second gating linear unit is added with the output of the first convolution neural network to serve as the output of the second residual connecting network;
the second convolutional neural network is in signal connection with the second residual error connecting network and receives the output of the second residual error connecting network, and the output of the second convolutional neural network is used as the network input of the long-term and short-term memory network;
the network output of the long-short term memory network is used as the input of the full connection layer;
the output of the full connection layer is used as the input of the output layer;
the output of the output layer is used as the output of the residual convolution recursive network-gated linear unit structure.
In another embodiment of the present invention, the substrate is,
the determining unit 605 is specifically configured to, when the identification result is in an empty drum state or not, determine whether the wall is in the empty drum state according to the identification result, including: when the identification result is in the empty drum state, determining that the wall body is in the empty drum state; and when the identification result is that the wall body is not in the empty drum state, determining that the wall body is not in the empty drum state.
The units of the above embodiments may be integrated into one body, or may be separately deployed; may be combined into one unit or further divided into a plurality of sub-units.
In another embodiment, an electronic device is also provided, which includes a memory, a processor, and a computer program stored on the memory and executable on the processor, wherein the processor implements the steps of the wall hollowing detection method when executing the program.
In another embodiment, a computer readable storage medium is also provided, having stored thereon computer instructions, which when executed by a processor, may implement the steps in the wall hollowing detection method.
In another embodiment, a computer program product is provided, comprising a computer program which, when being executed by a processor, carries out the wall hollowing detection method.
Fig. 7 is a schematic physical structure diagram of an electronic device according to an embodiment of the present invention. As shown in fig. 7, the electronic device may include: a Processor (Processor)710, a communication Interface 720, a Memory (Memory)730 and a communication bus 740, wherein the Processor 710, the communication Interface 720 and the Memory 730 communicate with each other via the communication bus 740. Processor 710 may call logic instructions in memory 730 to perform the following method:
acquiring an audio signal generated by knocking a wall body;
acquiring a first feature and a second feature of the audio signal;
splicing the first feature and the second feature to obtain a spliced feature;
acquiring an identification result corresponding to the splicing characteristic based on a preset audio detection network model;
and determining whether the wall is in a hollow state or not according to the identification result.
In addition, the logic instructions in the memory 730 can be implemented in the form of software functional units and stored in a computer readable storage medium when the software functional units are sold or used as independent products. Based on such understanding, the technical solution of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment. One of ordinary skill in the art can understand and implement it without inventive effort.
Through the above description of the embodiments, those skilled in the art will clearly understand that each embodiment can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware. With this understanding in mind, the above-described technical solutions may be embodied in the form of a software product, which can be stored in a computer-readable storage medium such as ROM/RAM, magnetic disk, optical disk, etc., and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the methods described in the embodiments or some parts of the embodiments.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like made within the spirit and principle of the present invention should be included in the scope of the present invention.

Claims (10)

1. A wall hollowing detection method is characterized by comprising the following steps:
acquiring an audio signal generated by knocking a wall body;
acquiring a first feature and a second feature of the audio signal;
splicing the first feature and the second feature to obtain a spliced feature;
acquiring an identification result corresponding to the splicing characteristic based on a preset audio detection network model;
and determining whether the wall is in a hollow state or not according to the identification result.
2. The method of claim 1,
the first feature is a fundamental frequency feature;
the second characteristic is any one of the following characteristics or a characteristic obtained by splicing at least two of the following characteristics:
speech features, decoding features, neural network features.
3. The method of claim 1, wherein after obtaining the audio signal generated by tapping the wall and before obtaining the first and second features of the audio signal, the method further comprises:
and carrying out endpoint detection on the audio signal and carrying out noise reduction processing.
4. The method of claim 1, further comprising:
obtaining a training sample;
acquiring a first feature and a second feature of the audio signal in the training sample, and splicing the first feature and the second feature to obtain a spliced feature;
establishing an initial audio detection network model; wherein the initial audio detection network model is established based on any one of the following neural networks: a deep neural network, a convolution recurrent neural network, a convolution neural network and a long-short term memory network;
and training the initial audio detection network model by using the splicing characteristics and the labels corresponding to the training samples to obtain the preset audio detection network model.
5. The method of claim 4,
when the neural network for establishing the initial audio detection network model is a convolution recurrent neural network, the convolution recurrent neural network adopts a residual convolution recurrent network-gate control linear unit structure;
when residual connection is carried out in the residual convolution recursive network-gate control linear unit structure, the network input before the gate control linear unit is added with the results of two convolution networks through gate control linear unit operation, and the results after residual connection are used as the input of the next convolution neural network.
6. The method of claim 5,
the residual convolution recursive network-gated linear unit structure comprises: the system comprises a first residual error connecting network, a second residual error connecting network, a first convolutional neural network, a second convolutional neural network, a long-term and short-term memory network, a full connecting layer and an output layer;
wherein the first residual connecting network comprises a third convolutional neural network, a fourth convolutional neural network and a first gated linear unit, the third convolutional neural network and the fourth convolutional neural network are used for receiving network input and the output of the third convolutional neural network and the fourth convolutional neural network is input to the first gated linear unit, the output of the first gated linear unit is added with the network input as the output of the first residual connecting network;
the first convolutional neural network is in signal connection with the first residual connecting network and is used for receiving the output of the first residual connecting network;
the second residual connecting network is in signal connection with the first convolutional neural network and is used for receiving the output of the first convolutional neural network, the second residual connecting network comprises a fifth convolutional neural network, a sixth convolutional neural network and a second gated linear unit, the fifth convolutional neural network and the sixth convolutional neural network are used for receiving the output of the first convolutional neural network, the output of the fifth convolutional neural network and the output of the sixth convolutional neural network are input to the second gated linear unit, and the output of the second gated linear unit is added with the output of the first convolutional neural network to be used as the output of the second residual connecting network;
the second convolutional neural network is in signal connection with the second residual error connection network and receives the output of the second residual error connection network, and the output of the second convolutional neural network is used as the network input of the long-short term memory network;
the network output of the long-short term memory network is used as the input of the full connection layer;
the output of the full connection layer is used as the input of the output layer;
and the output of the output layer is used as the output of the residual convolution recursive network-gated linear unit structure.
7. The method according to any one of claims 1 to 6,
when the identification result is in the empty drum state or not, determining whether the wall is in the empty drum state according to the identification result, including:
when the identification result is in the empty drum state, determining that the wall body is in the empty drum state;
and when the identification result is that the wall body is not in the empty drum state, determining that the wall body is not in the empty drum state.
8. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the processor implements the method according to any of claims 1-7 when executing the program.
9. A computer-readable storage medium, on which a computer program is stored, which program, when being executed by a processor, is adapted to carry out the method of any one of claims 1 to 7.
10. A computer program product comprising a computer program, characterized in that the computer program realizes the method of any of claims 1-7 when executed by a processor.
CN202110958298.7A 2021-08-20 2021-08-20 Wall hollowing detection method and device Active CN113671031B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110958298.7A CN113671031B (en) 2021-08-20 2021-08-20 Wall hollowing detection method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110958298.7A CN113671031B (en) 2021-08-20 2021-08-20 Wall hollowing detection method and device

Publications (2)

Publication Number Publication Date
CN113671031A true CN113671031A (en) 2021-11-19
CN113671031B CN113671031B (en) 2024-06-21

Family

ID=78544181

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110958298.7A Active CN113671031B (en) 2021-08-20 2021-08-20 Wall hollowing detection method and device

Country Status (1)

Country Link
CN (1) CN113671031B (en)

Citations (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002318155A (en) * 2001-04-24 2002-10-31 Fuji Xerox Co Ltd Signal determination device
US20170140240A1 (en) * 2015-07-27 2017-05-18 Salesforce.Com, Inc. Neural network combined image and text evaluator and classifier
CN106841403A (en) * 2017-01-23 2017-06-13 天津大学 A kind of acoustics glass defect detection method based on neutral net
CN107545902A (en) * 2017-07-14 2018-01-05 清华大学 A kind of article Material Identification method and device based on sound characteristic
CN107704924A (en) * 2016-07-27 2018-02-16 中国科学院自动化研究所 Synchronous self-adapting space-time characteristic expresses the construction method and correlation technique of learning model
CN108241024A (en) * 2018-01-25 2018-07-03 上海众材工程检测有限公司 A kind of hollowing detection method and system based on wall
US20180336889A1 (en) * 2017-05-19 2018-11-22 Baidu Online Network Technology (Beijing) Co., Ltd . Method and Apparatus of Building Acoustic Feature Extracting Model, and Acoustic Feature Extracting Method and Apparatus
CN109086892A (en) * 2018-06-15 2018-12-25 中山大学 It is a kind of based on the visual problem inference pattern and system that typically rely on tree
CN109142523A (en) * 2018-08-14 2019-01-04 浙江核力建筑特种技术有限公司 A kind of metope hollowing recognition quantitative analytic approach based on acoustic imaging
CN208505984U (en) * 2018-08-10 2019-02-15 平湖市佳诚房地产评估事务所(普通合伙) A kind of hollowing detector for Real Estate Appraisal
US20190130257A1 (en) * 2017-10-27 2019-05-02 Sentient Technologies (Barbados) Limited Beyond Shared Hierarchies: Deep Multitask Learning Through Soft Layer Ordering
JP2019148445A (en) * 2018-02-26 2019-09-05 日立オートモティブシステムズ株式会社 Hammering sound inspection device
US20190341052A1 (en) * 2018-05-02 2019-11-07 Simon Says, Inc. Machine Learning-Based Speech-To-Text Transcription Cloud Intermediary
CN110491410A (en) * 2019-04-12 2019-11-22 腾讯科技(深圳)有限公司 Speech separating method, audio recognition method and relevant device
CN110992959A (en) * 2019-12-06 2020-04-10 北京市科学技术情报研究所 Voice recognition method and system
WO2020093042A1 (en) * 2018-11-02 2020-05-07 Deep Lens, Inc. Neural networks for biomedical image analysis
AU2020101229A4 (en) * 2020-07-02 2020-08-06 South China University Of Technology A Text Line Recognition Method in Chinese Scenes Based on Residual Convolutional and Recurrent Neural Networks
US20200256834A1 (en) * 2017-10-11 2020-08-13 Bp Exploration Operating Company Limited Detecting Events Using Acoustic Frequency Domain Features
CN111798840A (en) * 2020-07-16 2020-10-20 中移在线服务有限公司 Voice keyword recognition method and device
CN111916101A (en) * 2020-08-06 2020-11-10 大象声科(深圳)科技有限公司 Deep learning noise reduction method and system fusing bone vibration sensor and double-microphone signals
US20200355649A1 (en) * 2017-08-02 2020-11-12 United States Of America As Represented By The Secretary Of The Navy System and Method for Detecting Failed Electronics Using Acoustics
US20200375512A1 (en) * 2017-07-27 2020-12-03 Vita-Course Technologies (Hainan) Co., Ltd. Method and system for detecting the oxygen saturation within the blood
CN112034044A (en) * 2020-09-01 2020-12-04 南京邮电大学 Intelligent recognition and detection device for hollowing of ceramic tile based on neural network
US20200408718A1 (en) * 2018-03-06 2020-12-31 Lg Chem, Ltd. Apparatus for Diagnosing Crack in Battery Pack and Battery Pack and Vehicle Including the Same
US20210003543A1 (en) * 2018-03-29 2021-01-07 Taiyo Yuden Co., Ltd. Sensing system and information processing apparatus
KR102221617B1 (en) * 2019-10-01 2021-03-03 주식회사 에스아이웨어 Handheld ultrasound scanner for defect detection of weldments
CN112684012A (en) * 2020-12-02 2021-04-20 青岛科技大学 Equipment key force-bearing structural part fault diagnosis method based on multi-parameter information fusion
CN112927696A (en) * 2019-12-05 2021-06-08 中国科学院深圳先进技术研究院 System and method for automatically evaluating dysarthria based on voice recognition
CN113075296A (en) * 2021-04-01 2021-07-06 湖南翰坤实业有限公司 Method and device for detecting safety of outer wall structure based on sound wave detection and BIM model
CN113111804A (en) * 2021-04-16 2021-07-13 北京房江湖科技有限公司 Face detection method and device, electronic equipment and storage medium
US20210216874A1 (en) * 2020-01-10 2021-07-15 Facebook Technologies, Llc Radioactive data generation
CN113125556A (en) * 2021-03-05 2021-07-16 南京智慧基础设施技术研究院有限公司 Structural damage detection system and method based on voiceprint recognition
CN113220933A (en) * 2021-05-12 2021-08-06 北京百度网讯科技有限公司 Method and device for classifying audio segments and electronic equipment

Patent Citations (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002318155A (en) * 2001-04-24 2002-10-31 Fuji Xerox Co Ltd Signal determination device
US20170140240A1 (en) * 2015-07-27 2017-05-18 Salesforce.Com, Inc. Neural network combined image and text evaluator and classifier
CN107704924A (en) * 2016-07-27 2018-02-16 中国科学院自动化研究所 Synchronous self-adapting space-time characteristic expresses the construction method and correlation technique of learning model
CN106841403A (en) * 2017-01-23 2017-06-13 天津大学 A kind of acoustics glass defect detection method based on neutral net
US20180336889A1 (en) * 2017-05-19 2018-11-22 Baidu Online Network Technology (Beijing) Co., Ltd . Method and Apparatus of Building Acoustic Feature Extracting Model, and Acoustic Feature Extracting Method and Apparatus
CN107545902A (en) * 2017-07-14 2018-01-05 清华大学 A kind of article Material Identification method and device based on sound characteristic
US20200375512A1 (en) * 2017-07-27 2020-12-03 Vita-Course Technologies (Hainan) Co., Ltd. Method and system for detecting the oxygen saturation within the blood
US20200355649A1 (en) * 2017-08-02 2020-11-12 United States Of America As Represented By The Secretary Of The Navy System and Method for Detecting Failed Electronics Using Acoustics
US20200256834A1 (en) * 2017-10-11 2020-08-13 Bp Exploration Operating Company Limited Detecting Events Using Acoustic Frequency Domain Features
US20190130257A1 (en) * 2017-10-27 2019-05-02 Sentient Technologies (Barbados) Limited Beyond Shared Hierarchies: Deep Multitask Learning Through Soft Layer Ordering
CN108241024A (en) * 2018-01-25 2018-07-03 上海众材工程检测有限公司 A kind of hollowing detection method and system based on wall
JP2019148445A (en) * 2018-02-26 2019-09-05 日立オートモティブシステムズ株式会社 Hammering sound inspection device
US20200408718A1 (en) * 2018-03-06 2020-12-31 Lg Chem, Ltd. Apparatus for Diagnosing Crack in Battery Pack and Battery Pack and Vehicle Including the Same
US20210003543A1 (en) * 2018-03-29 2021-01-07 Taiyo Yuden Co., Ltd. Sensing system and information processing apparatus
US20190341052A1 (en) * 2018-05-02 2019-11-07 Simon Says, Inc. Machine Learning-Based Speech-To-Text Transcription Cloud Intermediary
CN109086892A (en) * 2018-06-15 2018-12-25 中山大学 It is a kind of based on the visual problem inference pattern and system that typically rely on tree
CN208505984U (en) * 2018-08-10 2019-02-15 平湖市佳诚房地产评估事务所(普通合伙) A kind of hollowing detector for Real Estate Appraisal
CN109142523A (en) * 2018-08-14 2019-01-04 浙江核力建筑特种技术有限公司 A kind of metope hollowing recognition quantitative analytic approach based on acoustic imaging
WO2020093042A1 (en) * 2018-11-02 2020-05-07 Deep Lens, Inc. Neural networks for biomedical image analysis
CN110491410A (en) * 2019-04-12 2019-11-22 腾讯科技(深圳)有限公司 Speech separating method, audio recognition method and relevant device
KR102221617B1 (en) * 2019-10-01 2021-03-03 주식회사 에스아이웨어 Handheld ultrasound scanner for defect detection of weldments
CN112927696A (en) * 2019-12-05 2021-06-08 中国科学院深圳先进技术研究院 System and method for automatically evaluating dysarthria based on voice recognition
CN110992959A (en) * 2019-12-06 2020-04-10 北京市科学技术情报研究所 Voice recognition method and system
US20210216874A1 (en) * 2020-01-10 2021-07-15 Facebook Technologies, Llc Radioactive data generation
AU2020101229A4 (en) * 2020-07-02 2020-08-06 South China University Of Technology A Text Line Recognition Method in Chinese Scenes Based on Residual Convolutional and Recurrent Neural Networks
CN111798840A (en) * 2020-07-16 2020-10-20 中移在线服务有限公司 Voice keyword recognition method and device
CN111916101A (en) * 2020-08-06 2020-11-10 大象声科(深圳)科技有限公司 Deep learning noise reduction method and system fusing bone vibration sensor and double-microphone signals
CN112034044A (en) * 2020-09-01 2020-12-04 南京邮电大学 Intelligent recognition and detection device for hollowing of ceramic tile based on neural network
CN112684012A (en) * 2020-12-02 2021-04-20 青岛科技大学 Equipment key force-bearing structural part fault diagnosis method based on multi-parameter information fusion
CN113125556A (en) * 2021-03-05 2021-07-16 南京智慧基础设施技术研究院有限公司 Structural damage detection system and method based on voiceprint recognition
CN113075296A (en) * 2021-04-01 2021-07-06 湖南翰坤实业有限公司 Method and device for detecting safety of outer wall structure based on sound wave detection and BIM model
CN113111804A (en) * 2021-04-16 2021-07-13 北京房江湖科技有限公司 Face detection method and device, electronic equipment and storage medium
CN113220933A (en) * 2021-05-12 2021-08-06 北京百度网讯科技有限公司 Method and device for classifying audio segments and electronic equipment

Also Published As

Publication number Publication date
CN113671031B (en) 2024-06-21

Similar Documents

Publication Publication Date Title
CN111179975B (en) Voice endpoint detection method for emotion recognition, electronic device and storage medium
CN107369439B (en) Voice awakening method and device
CN102543073B (en) Shanghai dialect phonetic recognition information processing method
CN106601230B (en) Logistics sorting place name voice recognition method and system based on continuous Gaussian mixture HMM model and logistics sorting system
US9799333B2 (en) System and method for processing speech to identify keywords or other information
CN110444202B (en) Composite voice recognition method, device, equipment and computer readable storage medium
JPH06222791A (en) Method and apparatus for measurement of similar point between sound samples
Huang et al. Intelligent feature extraction and classification of anuran vocalizations
CN111048071A (en) Voice data processing method and device, computer equipment and storage medium
CN113327626A (en) Voice noise reduction method, device, equipment and storage medium
CN110428853A (en) Voice activity detection method, Voice activity detection device and electronic equipment
WO2019232833A1 (en) Speech differentiating method and device, computer device and storage medium
CN114333865A (en) Model training and tone conversion method, device, equipment and medium
CN111933148A (en) Age identification method and device based on convolutional neural network and terminal
CN114996489A (en) Method, device and equipment for detecting violation of news data and storage medium
CN111540368A (en) Stable bird sound extraction method and device and computer readable storage medium
CN112309398B (en) Method and device for monitoring working time, electronic equipment and storage medium
CN106340310B (en) Speech detection method and device
CN116884435A (en) Voice event detection method and device based on audio prompt learning
CN113671031B (en) Wall hollowing detection method and device
Dua et al. Gujarati language automatic speech recognition using integrated feature extraction and hybrid acoustic model
CN113327596B (en) Training method of voice recognition model, voice recognition method and device
CN107437414A (en) Parallelization visitor's recognition methods based on embedded gpu system
CN111326161B (en) Voiceprint determining method and device
CN113903328A (en) Speaker counting method, device, equipment and storage medium based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
TA01 Transfer of patent application right

Effective date of registration: 20240521

Address after: Room 102, floor 1, building 1, No. 2, Chuangye Road, Haidian District, Beijing 100085

Applicant after: Seashell Housing (Beijing) Technology Co.,Ltd.

Country or region after: China

Address before: 101399 room 24, 62 Farm Road, Erjie village, Yangzhen, Shunyi District, Beijing

Applicant before: Beijing fangjianghu Technology Co.,Ltd.

Country or region before: China

GR01 Patent grant
GR01 Patent grant