CN111105812A - Audio feature extraction method and device, training method and electronic equipment - Google Patents

Audio feature extraction method and device, training method and electronic equipment Download PDF

Info

Publication number
CN111105812A
CN111105812A CN201911409010.XA CN201911409010A CN111105812A CN 111105812 A CN111105812 A CN 111105812A CN 201911409010 A CN201911409010 A CN 201911409010A CN 111105812 A CN111105812 A CN 111105812A
Authority
CN
China
Prior art keywords
audio
spectrum
fitting
frequency
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911409010.XA
Other languages
Chinese (zh)
Inventor
何维祯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Pulian International Co Ltd
Original Assignee
Pulian International Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Pulian International Co Ltd filed Critical Pulian International Co Ltd
Priority to CN201911409010.XA priority Critical patent/CN111105812A/en
Publication of CN111105812A publication Critical patent/CN111105812A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/18Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being spectral information of each sub-band

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

The invention discloses an audio feature extraction method, an audio feature extraction device, a training method and electronic equipment, wherein the method comprises the following steps: acquiring audio to be extracted according to a preset window length, and dividing the audio to be extracted into M audio frames according to a preset frame length; calculating a frequency spectrum corresponding to each audio frame; obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum; calculating to obtain the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula; dividing each frequency spectrum into m sections of frequency spectrum bands, calculating to obtain a logarithmic frequency spectrum corresponding to each section of frequency spectrum band, and further calculating to obtain the frequency spectrum contrast of each frequency spectrum; and obtaining the characteristic quantity of the audio frame according to the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of the audio frame, and further extracting the audio characteristic of the audio to be extracted. When the audio features extracted by the invention are used in detection scenes such as baby crying and the like, the accuracy of audio detection is favorably improved.

Description

Audio feature extraction method and device, training method and electronic equipment
Technical Field
The invention relates to the technical field of audio processing, in particular to an audio feature extraction method, an audio feature extraction device, an audio feature training method and electronic equipment.
Background
With the development of society, problems such as high labor cost and the like gradually emerge, consumption cost or time cost in the aspect of infant nursing is higher nowadays, and nursing products or household security products such as infant nursing devices capable of identifying infant crying are more and more popular with parents. When the nursing product or the household security product detects the crying of the baby, the warning is automatically sent to a nursing person or a parent, so that the baby can be watched in time.
In the prior art, when detecting a baby crying, it is usually determined whether the baby crying exists according to the energy characteristics of the detected audio, and when the energy characteristics of the detected audio are consistent with the energy characteristics of the baby crying, it is determined that the baby crying exists in the audio, and a warning is given to a caregiver or a parent.
However, the detection environment is often complicated, other sounds of the surrounding environment may exist, the audio may have an environmental noise that is not the baby cry, and when the energy characteristics of the other sounds are similar to the energy characteristics of the baby cry, if the baby cry is detected only by the energy of the audio, the environmental noise may be detected as the baby cry, and the detection accuracy is low.
Disclosure of Invention
The technical problem to be solved by the embodiments of the present invention is to provide an audio feature extraction method, an audio feature extraction device, an audio training method, and an electronic device, which can extract audio features according to a fitting slope, a fitting intercept, a spectral contrast, and a spectral flatness of audio, and are beneficial to multi-dimensional fine audio detection and improvement of audio detection accuracy when used in audio detection scenes such as baby cry.
In order to solve the above technical problem, in a first aspect, the present invention provides an audio feature extraction method, including:
acquiring audio to be extracted according to a preset window length, and dividing the audio to be extracted into M audio frames according to a preset frame length, wherein M is greater than 1;
calculating a frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
calculating to obtain the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula;
dividing each frequency spectrum into m sections of frequency spectrum bands, and calculating to obtain a logarithmic frequency spectrum corresponding to each section of frequency spectrum band; m is greater than 1;
obtaining the spectrum contrast of each spectrum according to the m sections of log spectrums corresponding to each spectrum;
obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of each audio frame;
and extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of the audio frames.
Further, the linear fitting algorithm is a linear least square algorithm, and the obtaining of the fitting slope and the fitting intercept corresponding to each frequency spectrum based on the linear fitting algorithm according to the N frequency domain points of each frequency spectrum specifically includes:
selecting frequency domain points with corresponding frequencies within a preset frequency range from the N frequency domain points of each frequency spectrum;
and performing linear fitting on the selected frequency domain points with the corresponding frequencies within a preset frequency range based on a linear least square algorithm to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum.
Further, the dividing each spectrum into m sections of spectrum bands, and calculating to obtain a logarithmic spectrum corresponding to each section of spectrum band specifically includes:
dividing each frequency spectrum into m sections of frequency spectrum bands, and respectively carrying out K-L conversion processing on each section of frequency spectrum band;
obtaining a logarithmic spectrum corresponding to the spectrum band after each section of the spectrum band is subjected to the K-L conversion processing according to the following formula:
si(f″)=10×log10si(f′);
wherein s isi(f') is the ith spectral band, s, after K-L transformationi(f') is si(f') i is more than or equal to 1 and less than or equal to m.
Further, the obtaining the spectrum contrast of each spectrum according to the m segments of log spectrums corresponding to each spectrum specifically includes:
for each section of the log spectrum, acquiring a spectrum peak value and a spectrum valley value of the log spectrum, and calculating a peak-valley difference value between the spectrum peak value and the spectrum valley value;
and for each spectrum, calculating the average value of m peak-valley difference values of the corresponding m sections of log spectrum to obtain the spectrum contrast of the spectrum.
Further, the calculation formula is as follows:
Figure BDA0002349466460000031
wherein s (f) is the frequency spectrum; (s (f)) is the spectral Flatness corresponding to spectrum s (f); n is the number of frequency domain points included in the spectrum, and x (N) is the amplitude of the nth frequency domain point of the spectrum s (f).
Further, the method further comprises:
calculating to obtain a Mel cepstrum coefficient of each audio frame;
then, the obtaining the feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame specifically includes:
and obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, the spectral contrast and the Mel cepstrum coefficient of each audio frame.
In order to solve the above technical problem, in a second aspect, the present invention further provides an audio feature extraction apparatus, including:
the audio frame acquisition module is used for acquiring audio to be extracted according to a preset window length and dividing the audio to be extracted into M audio frames according to a preset frame length, wherein M is greater than 1;
the first calculating module is used for calculating the frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
the fitting module is used for obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
the spectrum flatness calculation module is used for calculating and obtaining the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula;
the second calculation module is used for dividing each spectrum into m sections of spectrum bands and calculating to obtain a logarithmic spectrum corresponding to each section of spectrum band; m is greater than 1;
the spectrum contrast calculation module is used for obtaining the spectrum contrast of each spectrum according to the m sections of log spectrums corresponding to each spectrum;
a feature quantity obtaining module, configured to obtain a feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame;
and the extraction module is used for extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of the audio frames.
In order to solve the above technical problem, in a third aspect, the present invention further provides a method for training an audio classification model, where the method includes:
constructing an audio classification initial model; the audio classification initial model is used for classifying the audio;
obtaining a plurality of training audios corresponding to each classification result; each training audio is pre-allocated with a classification identifier matched with the corresponding classification result;
taking the training audio as the audio to be extracted, and extracting the audio feature corresponding to each training audio according to the audio feature extraction method of any one of claims 1 to 6;
carrying out standardization processing on the audio features corresponding to each training audio, and constructing a training sample set according to each audio feature after standardization processing and the matched classification identification;
and training the audio classification initial model according to the training sample set to obtain an audio classification model.
Further, the audio features corresponding to each of the training audios are:
Figure BDA0002349466460000051
Ai=[ai1,ai2,…,aiq]
wherein X is an audio feature corresponding to the training audio; a. theiI is more than or equal to 1 and less than or equal to M for the characteristic quantity of the audio frame of the ith frame in the training audio; q is the number of elements of the characteristic quantity, q>1;
Then, the normalizing the audio features corresponding to each of the training audios specifically includes:
normalizing the audio features corresponding to each training audio according to the following formula:
Figure BDA0002349466460000052
Figure BDA0002349466460000053
wherein, X' is the audio frequency characteristic after the standardization treatment; a'iThe feature quantity of the audio frame of the ith frame in the training audio after the normalization processing is carried out; a isk-meanK is more than or equal to 1 and less than or equal to q and is the average value of the kth element in the characteristic quantity of the M frames of audio frames of the training audio; std (a)k) Is the variance of the kth element in the feature quantities of the M frames of audio of the training audio.
In order to solve the above technical problem, in a fourth aspect, the present invention further provides an electronic device, including a processor, a memory, and a computer program stored in the memory and configured to be executed by the processor, wherein the processor, when executing the computer program, implements the audio feature extraction method according to any one of the first aspects.
The audio feature extraction method, the device, the training method and the electronic equipment can extract the audio features including the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of the audio, and compared with the audio features only with the energy of the audio, the audio features extracted by the method can distinguish different audios through more-dimensional information. When the method is used for audio detection, noise and target detection sound can be better identified and classified, and the accuracy of audio detection is favorably improved. For example, the method is used for detecting the baby crying, can better identify and classify the environmental noise and the baby crying, and is favorable for improving the accuracy of the baby crying detection.
Drawings
FIG. 1 is a flow chart of an audio feature extraction method according to a preferred embodiment of the present invention;
fig. 2 is a schematic structural diagram of a preferred embodiment of an audio feature extraction apparatus provided in the present invention;
FIG. 3 is a flowchart illustrating a method for training an audio classification model according to a preferred embodiment of the present invention;
fig. 4 is a schematic structural diagram of a preferred embodiment of an electronic device provided in the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example one
Referring to fig. 1, fig. 1 is a schematic flow chart of a preferred embodiment of an audio feature extraction method according to the present invention; specifically, the method comprises the following steps:
s1, acquiring the audio to be extracted according to the preset window length, and dividing the audio to be extracted into M audio frames according to the preset frame length, wherein M is greater than 1;
s2, calculating the frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
s3, obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
s4, calculating and obtaining the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula;
s5, dividing each frequency spectrum into m sections of frequency spectrum bands, and calculating to obtain a logarithmic frequency spectrum corresponding to each section of frequency spectrum band; m is greater than 1;
s6, obtaining the spectrum contrast of each spectrum according to the m sections of log spectrums corresponding to each spectrum;
s7, obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of each audio frame;
and S8, extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of the audio frames.
In specific implementation, for detecting an input audio, acquiring an audio to be extracted according to a preset window length, for example, intercepting each section of the audio to be extracted according to a 5s window length, and dividing the audio to be extracted into M audio frames according to a preset frame length, where M >1, for example, dividing the 5s audio to be extracted into at least 200 audio frames according to a 25ms frame length; preferably, there is an overlap between adjacent audio frames, for example, there is an overlap of 1/4 frame lengths between adjacent audio frames, and it is anticipated that the number of audio frames will be greater than 200.
The audio frame of each frame has a plurality of time domain points, the time domain points are mapped to a frequency domain through Fourier transform calculation, and the frequency spectrum corresponding to each frame of audio frame is obtained through calculation; wherein the frequency spectrum comprises N frequency domain points. It should be noted that, it is preferable to obtain the frequency spectrum by using the fast fourier transform, and since the obtained frequency spectrum by using the fast fourier transform has symmetry, the N frequency domain points may only select the frequency domain points of the first half of the frequency spectrum, that is, if there are C time domain points of each frame of audio frame, N is equal to C/2.
Performing linear fitting on the N frequency domain points of each frequency spectrum based on a linear fitting algorithm to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum; and calculating to obtain the spectral flatness of each frequency spectrum according to the frequency spectrums and a preset calculation formula.
Further, dividing each frequency spectrum into m sections of frequency spectrum bands, and calculating to obtain a logarithmic frequency spectrum corresponding to each section of frequency spectrum band; m is greater than 1. And the spectrum contrast of each spectrum is obtained by calculating according to the corresponding m sections of log spectrums.
And traversing the frequency spectrums of all the audio frames to obtain all the fitting slopes, the fitting intercepts, the frequency spectrum flatness and the frequency spectrum contrast of the M frames of audio frames of the audio to be extracted. Obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of the audio frame; and extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of audio frames.
The audio feature extraction method provided by the invention can extract the audio features including the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of the audio, and compared with the audio features only with the audio energy, the audio features extracted by the invention can distinguish different audios through more-dimensional information. When the method is used for audio detection, noise and target detection sound can be better identified and classified, and the accuracy of audio detection is favorably improved. For example, the method is used for detecting the baby crying, can better identify and classify the environmental noise and the baby crying, and is favorable for improving the accuracy of the baby crying detection.
Preferably, the linear fitting algorithm is a linear least square algorithm, and the obtaining of the fitting slope and the fitting intercept corresponding to each frequency spectrum based on the linear fitting algorithm according to the N frequency domain points of each frequency spectrum specifically includes:
selecting frequency domain points with corresponding frequencies within a preset frequency range from the N frequency domain points of each frequency spectrum;
and performing linear fitting on the selected frequency domain points with the corresponding frequencies within a preset frequency range based on a linear least square algorithm to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum.
In this embodiment, all frequency domain points of the frequency spectrum are not fitted, but frequency domain points located in a preset frequency range are selected for linear fitting, the preset frequency range is set according to a target detection audio, for example, when an audio characteristic is used for detecting baby cry, since the frequency of baby cry has a certain range, the fitting is performed only on signals in the preset frequency range in the audio, for example, the preset frequency range is set to 250HZ to 600HZ, so as to extract the audio characteristic for baby cry detection. And performing linear simulation by using a linear down-least square algorithm to obtain a fitting slope and a fitting intercept corresponding to each frame of audio frame.
Preferably, the dividing each spectrum into m segments of spectrum bands and obtaining a logarithmic spectrum corresponding to each segment of spectrum band by calculation specifically includes:
dividing each frequency spectrum into m sections of frequency spectrum bands, and respectively carrying out K-L conversion processing on each section of frequency spectrum band;
obtaining a logarithmic spectrum corresponding to the spectrum band after each section of the spectrum band is subjected to the K-L conversion processing according to the following formula:
si(f″)=10×log10si(f′) (1);
wherein s isi(f') is the ith spectral band, s, after K-L transformationi(f') is si(f') i is more than or equal to 1 and less than or equal to m.
In this embodiment, the correlations between different spectral bands are eliminated through K-L transformation, so that the spectra in different spectral band ranges are uncorrelated with each other, and the peak-to-valley difference corresponding to each spectral band can be separately reflected in the spectral contrast of the spectral band, and the spectral contrast of the spectrum is obtained after averaging.
Specifically, assume that the obtained spectrum is s (f), si(f) For the ith frequency spectrum band in s (f), i is more than or equal to 1 and less than or equal to m, and the mean value of the solved frequency spectrum is u, the covariance matrix obtained by K-L transformation is as follows:
C=E[(s(f)-u)(s(f)-u)T](2)
solving eigenvalues λ of covariance matrixiAnd the feature vector phii
i=λiφi
The K-L transform processing is performed on the spectrum s (F) to become F:
F=φT(s(f)-u)
wherein phi is [ phi ]12,…,φm]。
Each spectral band s is obtained from Fi(f) Spectrum after K-L conversion. Then, the logarithmic spectrum corresponding to each spectral band is obtained through the formula (2).
Preferably, the obtaining the spectrum contrast of each spectrum according to the m segments of log spectrums corresponding to each spectrum specifically includes:
for each section of the log spectrum, acquiring a spectrum peak value and a spectrum valley value of the log spectrum, and calculating a peak-valley difference value between the spectrum peak value and the spectrum valley value;
and for each spectrum, calculating the average value of m peak-valley difference values of the corresponding m sections of log spectrum to obtain the spectrum contrast of the spectrum.
In the embodiment, the average value of m peak-to-valley difference values of m segments of log spectrum is used as the spectrum contrast of the spectrum.
Preferably, the calculation formula is:
Figure BDA0002349466460000101
wherein s (f) is the frequency spectrum; (s (f)) is the spectral Flatness corresponding to spectrum s (f); n is the number of frequency domain points included in the spectrum, and x (N) is the amplitude of the nth frequency domain point of the spectrum s (f).
The present embodiment calculates the spectral flatness of the obtained spectrum by the above calculation formula.
Preferably, the method further comprises:
calculating to obtain a Mel cepstrum coefficient of each audio frame;
then, the obtaining the feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame specifically includes:
and obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, the spectral contrast and the Mel cepstrum coefficient of each audio frame.
In this embodiment, the audio features of the audio further include mel-frequency cepstrum coefficients, which further increases the information dimensionality of the audio features and is further beneficial to improving the accuracy of audio identification and classification.
The invention provides an audio characteristic extraction method, which is implemented specifically, the audio to be extracted is obtained according to the preset window length, and the audio to be extracted is divided into M audio frames according to the preset frame length, wherein M is greater than 1; calculating a frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points; obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum; calculating to obtain the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula; dividing each frequency spectrum into m sections of frequency spectrum bands, and calculating to obtain a logarithmic frequency spectrum corresponding to each section of frequency spectrum band; m is greater than 1; obtaining the spectrum contrast of each spectrum according to the m sections of logarithmic spectra corresponding to each spectrum; obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of each audio frame; and extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of audio frames.
The audio feature extraction method provided by the invention can extract the audio features including the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of the audio, and compared with the audio features only with the audio energy, the audio features extracted by the invention can distinguish different audios through more-dimensional information. When the method is used for audio detection, noise and target detection sound can be better identified and classified, and the accuracy of audio detection is favorably improved. For example, the method is used for detecting the baby crying, can better identify and classify the environmental noise and the baby crying, and is favorable for improving the accuracy of the baby crying detection.
Example two
Fig. 2 shows a schematic structural diagram of an audio feature extraction apparatus according to a preferred embodiment of the present invention; specifically, the apparatus comprises:
the audio frame obtaining module 11 is configured to obtain an audio to be extracted according to a preset window length, and divide the audio to be extracted into M audio frames according to a preset frame length, where M is greater than 1;
a first calculating module 12, configured to calculate a frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
the fitting module 13 is configured to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
the spectrum flatness calculation module 14 is configured to calculate and obtain a spectrum flatness of each spectrum according to the spectrum and a preset calculation formula;
the second calculating module 15 is configured to divide each spectrum into m segments of spectrum bands, and calculate to obtain a logarithmic spectrum corresponding to each segment of spectrum band; m is greater than 1;
the spectrum contrast calculation module 16 is configured to obtain a spectrum contrast of each spectrum according to the m segments of log spectrums corresponding to each spectrum;
a feature quantity obtaining module 17, configured to obtain a feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame;
and the extracting module 18 is configured to extract the audio features of the audio to be extracted according to the feature quantities of the M frames of the audio frames.
Preferably, the fitting module 13 is specifically configured to:
selecting frequency domain points with corresponding frequencies within a preset frequency range from the N frequency domain points of each frequency spectrum;
and performing linear fitting on the selected frequency domain points with the corresponding frequencies within a preset frequency range based on a linear least square algorithm to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum.
Preferably, the second calculating module 15 is specifically configured to:
dividing each frequency spectrum into m sections of frequency spectrum bands, and respectively carrying out K-L conversion processing on each section of frequency spectrum band;
obtaining a logarithmic spectrum corresponding to the spectrum band after each section of the spectrum band is subjected to the K-L conversion processing according to the following formula:
si(f″)=10×log10si(f′) (1);
wherein s isi(f') is the ith spectral band, s, after K-L transformationi(f') is si(f') i is more than or equal to 1 and less than or equal to m.
Preferably, the spectral contrast calculation module 16 is specifically configured to:
for each section of the log spectrum, acquiring a spectrum peak value and a spectrum valley value of the log spectrum, and calculating a peak-valley difference value between the spectrum peak value and the spectrum valley value;
and for each spectrum, calculating the average value of m peak-valley difference values of the corresponding m sections of log spectrum to obtain the spectrum contrast of the spectrum.
Preferably, the calculation formula is:
Figure BDA0002349466460000131
wherein s (f) is the frequency spectrum; (s (f)) is the spectral Flatness corresponding to spectrum s (f); n is the number of frequency domain points included in the spectrum, and x (N) is the amplitude of the nth frequency domain point of the spectrum s (f).
Preferably, the apparatus further comprises a third calculation module, and the third calculation module is specifically configured to:
calculating to obtain a Mel cepstrum coefficient of each audio frame;
then, the feature quantity obtaining module 17 is specifically configured to:
and obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, the spectral contrast and the Mel cepstrum coefficient of each audio frame.
The audio feature extraction device provided by the invention can extract audio features including fitting slope, fitting intercept, spectral flatness and spectral contrast of audio, and compared with the audio features only having audio energy, the audio features extracted by the invention can distinguish different audios through information with more dimensionalities. When the method is used for audio detection, noise and target detection sound can be better identified and classified, and the accuracy of audio detection is favorably improved. For example, the method is used for detecting the baby crying, can better identify and classify the environmental noise and the baby crying, and is favorable for improving the accuracy of the baby crying detection.
It should be noted that, the audio feature extraction device provided in the embodiment of the present invention is configured to execute the steps of the audio feature extraction method described in the above embodiment, and working principles and beneficial effects of the two are in one-to-one correspondence, so that details are not repeated.
It will be understood by those skilled in the art that the schematic diagram of the audio feature extraction apparatus is merely an example of the audio feature extraction apparatus, and does not constitute a limitation of the audio feature extraction apparatus, and may include more or less components than those shown in the drawings, or combine some components, or different components, for example, the audio feature extraction apparatus may further include an input and output device, a network access device, a bus, and the like.
EXAMPLE III
Referring to fig. 3, fig. 3 is a schematic flowchart of a preferred embodiment of a method for training an audio classification model according to the present invention; specifically, the method comprises the following steps:
s9, constructing an audio classification initial model; the audio classification initial model is used for classifying the audio;
s10, obtaining a plurality of training audios corresponding to each classification result; each training audio is pre-allocated with a classification identifier matched with the corresponding classification result;
s11, taking the training audio as the audio to be extracted, and extracting the audio feature corresponding to each training audio according to the audio feature extraction method provided in any one of the above embodiments;
s12, carrying out standardization processing on the audio features corresponding to each training audio, and constructing a training sample set according to each audio feature after standardization processing and the matched classification identification;
and S13, training the audio classification initial model according to the training sample set to obtain an audio classification model.
According to the training method of the audio classification model provided by the embodiment of the invention, the audio feature corresponding to each training audio is extracted according to the audio feature extraction method provided by the embodiment, and the audio classification model obtained by training according to the audio feature can be used for detecting the classification audio and improving the accuracy of audio classification.
It should be noted that, when actually detecting an audio, the input audio is also subjected to audio feature extraction according to the audio feature extraction method provided in the above embodiment, then the audio features are subjected to normalization processing, and then the audio features are input into an audio classification model for processing and classification, so as to obtain a classification result. P > 1.
Preferably, the audio classification model of the present invention may be an SVM model, and if the audio classification model is an SVM model, during specific training, the training audio of the training sample set may be divided into K parts, and one part is selected for training each time, and K times of training are performed in total, so as to obtain a spatial hyperplane for audio classification, where P is 2.
Preferably, each of the training audios has corresponding audio features:
Figure BDA0002349466460000151
Ai=[ai1,ai2,…,aiq]
wherein X is an audio feature corresponding to the training audio; a. theiI is more than or equal to 1 and less than or equal to M for the characteristic quantity of the audio frame of the ith frame in the training audio; q is the number of elements of the characteristic quantity, q>1;
Then, the normalizing the audio features corresponding to each of the training audios specifically includes:
normalizing the audio features corresponding to each training audio according to the following formula:
Figure BDA0002349466460000152
Figure BDA0002349466460000153
wherein, X' is the audio frequency characteristic after the standardization treatment; a'iThe feature quantity of the audio frame of the ith frame in the training audio after the normalization processing is carried out; a isk-meanK is more than or equal to 1 and less than or equal to q and is the average value of the kth element in the characteristic quantity of the M frames of audio frames of the training audio; std (a)k) Is the variance of the kth element in the feature quantities of the M frames of audio of the training audio.
The present invention normalizes the audio features by the above equations (3) and (4).
It should be noted that the training method for the audio classification model provided in the embodiment of the present invention has the same or corresponding technical features as the audio feature extraction method provided in the above embodiment, and the working principles and beneficial effects of the two methods are similar, so that further description is omitted.
Example four
Fig. 4 is a schematic structural diagram of an electronic device according to a preferred embodiment of the present invention, and fig. 4 is a schematic structural diagram of an electronic device according to a preferred embodiment of the present invention; specifically, the electronic device includes a processor, a memory, and a computer program stored in the memory and configured to be executed by the processor, and the processor implements the audio feature extraction method according to any one of the embodiments provided above when executing the computer program.
Specifically, the electronic device may be one or more processors and memories, and may be a computer, a mobile phone, a tablet or other device capable of performing sound detection, or may be a baby monitor or other device when the audio features are used to detect the crying of a baby.
The electronic device of the embodiment includes: a processor, a memory, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, the steps in the audio feature extraction method provided in the foregoing embodiment are implemented, for example, in step S1 shown in fig. 1, the audio to be extracted is obtained according to a preset window length, and the audio to be extracted is divided into M audio frames according to a preset frame length, where M > 1. Or, the processor implements the functions of the modules in the apparatus embodiments when executing the computer program, for example, implements an audio frame obtaining module 11, configured to obtain an audio to be extracted according to a preset window length, and divide the audio to be extracted into M audio frames according to a preset frame length, where M > 1.
Illustratively, the computer program can be divided into one or more modules/units (e.g., computer program 1, computer program 2, shown in FIG. 4), which are stored in the memory and executed by the processor to implement the invention. The one or more modules/units may be a series of computer program instruction segments capable of performing specific functions, which are used to describe the execution process of the computer program in the electronic device. For example, the computer program may be divided into an audio frame obtaining module 11, a first calculating module 12, a fitting module 13, a spectral flatness calculating module 14, a second calculating module 15, a spectral contrast calculating module 16, a feature quantity obtaining module 17, and an extracting module 18, and each module has the following specific functions:
the audio frame obtaining module 11 is configured to obtain an audio to be extracted according to a preset window length, and divide the audio to be extracted into M audio frames according to a preset frame length, where M is greater than 1;
a first calculating module 12, configured to calculate a frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
the fitting module 13 is configured to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
the spectrum flatness calculation module 14 is configured to calculate and obtain a spectrum flatness of each spectrum according to the spectrum and a preset calculation formula;
the second calculating module 15 is configured to divide each spectrum into m segments of spectrum bands, and calculate to obtain a logarithmic spectrum corresponding to each segment of spectrum band; m is greater than 1;
the spectrum contrast calculation module 16 is configured to obtain a spectrum contrast of each spectrum according to the m segments of log spectrums corresponding to each spectrum;
a feature quantity obtaining module 17, configured to obtain a feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame;
and the extracting module 18 is configured to extract the audio features of the audio to be extracted according to the feature quantities of the M frames of the audio frames.
The Processor may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), an off-the-shelf Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like that is the control center for the electronic device and that connects the various parts of the overall electronic device using various interfaces and wires.
The memory may be used to store the computer programs and/or modules, and the processor may implement various functions of the electronic device by running or executing the computer programs and/or modules stored in the memory and calling data stored in the memory. The memory may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required by at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the cellular phone, and the like. In addition, the memory may include high speed random access memory, and may also include non-volatile memory, such as a hard disk, a memory, a plug-in hard disk, a Smart Media Card (SMC), a Secure Digital (SD) Card, a Flash memory Card (Flash Card), at least one magnetic disk storage device, a Flash memory device, or other volatile solid state storage device.
Wherein, the integrated module/unit of the electronic device can be stored in a computer readable storage medium if it is implemented in the form of software functional unit and sold or used as a stand-alone product. Based on such understanding, all or part of the processes in the audio feature extraction method provided by the above embodiments may be implemented by a computer program, which may be stored in a computer-readable storage medium, and when the computer program is executed by a processor, the steps of the audio feature extraction method provided by any of the above embodiments may be implemented. Wherein the computer program comprises computer program code, which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, usb disk, removable hard disk, magnetic disk, optical disk, computer Memory, Read-Only Memory (ROM), Random Access Memory (RAM), electrical carrier wave signals, telecommunications signals, software distribution medium, and the like. It should be noted that the computer readable medium may contain content that is subject to appropriate increase or decrease as required by legislation and patent practice in jurisdictions, for example, in some jurisdictions, computer readable media does not include electrical carrier signals and telecommunications signals as is required by legislation and patent practice.
It should be noted that the above-mentioned electronic device may include, but is not limited to, a processor and a memory, and those skilled in the art will understand that the structural diagram of fig. 4 is only an example of the above-mentioned electronic device, and does not constitute a limitation of the electronic device, and may include more or less components than those shown in the drawings, or may combine some components, or may be different components.
While the foregoing is directed to the preferred embodiment of the present invention, it will be understood by those skilled in the art that various changes and modifications may be made without departing from the spirit and scope of the invention.

Claims (10)

1. A method of audio feature extraction, the method comprising:
acquiring audio to be extracted according to a preset window length, and dividing the audio to be extracted into M audio frames according to a preset frame length, wherein M is greater than 1;
calculating a frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
calculating to obtain the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula;
dividing each frequency spectrum into m sections of frequency spectrum bands, and calculating to obtain a logarithmic frequency spectrum corresponding to each section of frequency spectrum band; m is greater than 1;
obtaining the spectrum contrast of each spectrum according to the m sections of log spectrums corresponding to each spectrum;
obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness and the spectral contrast of each audio frame;
and extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of the audio frames.
2. The audio feature extraction method according to claim 1, wherein the linear fitting algorithm is a linear least squares algorithm, and the obtaining of the fitting slope and the fitting intercept corresponding to each of the frequency spectrums based on the linear fitting algorithm according to the N frequency domain points of each of the frequency spectrums specifically includes:
selecting frequency domain points with corresponding frequencies within a preset frequency range from the N frequency domain points of each frequency spectrum;
and performing linear fitting on the selected frequency domain points with the corresponding frequencies within a preset frequency range based on a linear least square algorithm to obtain a fitting slope and a fitting intercept corresponding to each frequency spectrum.
3. The audio feature extraction method according to claim 1, wherein the dividing each spectrum into m spectral bands and obtaining a logarithmic spectrum corresponding to each spectral band by calculation specifically comprises:
dividing each frequency spectrum into m sections of frequency spectrum bands, and respectively carrying out K-L conversion processing on each section of frequency spectrum band;
obtaining a logarithmic spectrum corresponding to the spectrum band after each section of the spectrum band is subjected to the K-L conversion processing according to the following formula:
si(f″)=10×log10si(f′);
wherein s isi(f') is the ith spectral band, s, after K-L transformationi(f') is si(f') i is more than or equal to 1 and less than or equal to m.
4. The method for extracting audio features according to claim 1, wherein the obtaining the spectral contrast of each spectrum according to the m segments of log spectrums corresponding to each spectrum specifically comprises:
for each section of the log spectrum, acquiring a spectrum peak value and a spectrum valley value of the log spectrum, and calculating a peak-valley difference value between the spectrum peak value and the spectrum valley value;
and for each spectrum, calculating the average value of m peak-valley difference values of the corresponding m sections of log spectrum to obtain the spectrum contrast of the spectrum.
5. The audio feature extraction method of claim 1, wherein the calculation formula is:
Figure FDA0002349466450000021
wherein s (f) is the frequency spectrum; (s (f)) is the spectral Flatness corresponding to spectrum s (f); n is the number of frequency domain points included in the spectrum, and x (N) is the amplitude of the nth frequency domain point of the spectrum s (f).
6. The audio feature extraction method of claim 1, further comprising:
calculating to obtain a Mel cepstrum coefficient of each audio frame;
then, the obtaining the feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame specifically includes:
and obtaining the characteristic quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, the spectral contrast and the Mel cepstrum coefficient of each audio frame.
7. An audio feature extraction apparatus, characterized in that the apparatus comprises:
the audio frame acquisition module is used for acquiring audio to be extracted according to a preset window length and dividing the audio to be extracted into M audio frames according to a preset frame length, wherein M is greater than 1;
the first calculating module is used for calculating the frequency spectrum corresponding to each audio frame; wherein the frequency spectrum comprises N frequency domain points, N > 1;
the fitting module is used for obtaining a fitting slope and a fitting intercept corresponding to each frequency spectrum based on a linear fitting algorithm according to the N frequency domain points of each frequency spectrum;
the spectrum flatness calculation module is used for calculating and obtaining the spectrum flatness of each spectrum according to the spectrums and a preset calculation formula;
the second calculation module is used for dividing each spectrum into m sections of spectrum bands and calculating to obtain a logarithmic spectrum corresponding to each section of spectrum band; m is greater than 1;
the spectrum contrast calculation module is used for obtaining the spectrum contrast of each spectrum according to the m sections of log spectrums corresponding to each spectrum;
a feature quantity obtaining module, configured to obtain a feature quantity of each audio frame according to the fitting slope, the fitting intercept, the spectral flatness, and the spectral contrast of each audio frame;
and the extraction module is used for extracting the audio features of the audio to be extracted according to the feature quantity of the M frames of the audio frames.
8. A method for training an audio classification model, the method comprising:
constructing an audio classification initial model; the audio classification initial model is used for classifying the audio;
obtaining a plurality of training audios corresponding to each classification result; each training audio is pre-allocated with a classification identifier matched with the corresponding classification result;
taking the training audio as the audio to be extracted, and extracting the audio feature corresponding to each training audio according to the audio feature extraction method of any one of claims 1 to 6;
carrying out standardization processing on the audio features corresponding to each training audio, and constructing a training sample set according to each audio feature after standardization processing and the matched classification identification;
and training the audio classification initial model according to the training sample set to obtain an audio classification model.
9. The method for training an audio classification model according to claim 8, wherein the audio features corresponding to each of the training audios are:
Figure FDA0002349466450000041
Ai=[ai1,ai2,…,aiq]
wherein X is an audio feature corresponding to the training audio; a. theiI is more than or equal to 1 and less than or equal to M for the characteristic quantity of the audio frame of the ith frame in the training audio; q is the number of elements of the characteristic quantity, q>1;
Then, the normalizing the audio features corresponding to each of the training audios specifically includes:
normalizing the audio features corresponding to each training audio according to the following formula:
Figure FDA0002349466450000051
Figure FDA0002349466450000052
wherein, X' is the audio frequency characteristic after the standardization treatment; a'iThe feature quantity of the audio frame of the ith frame in the training audio after the normalization processing is carried out; a isk-meanK is more than or equal to 1 and less than or equal to q and is the average value of the kth element in the characteristic quantity of the M frames of audio frames of the training audio; std (a)k) Is the variance of the kth element in the feature quantities of the M frames of audio of the training audio.
10. An electronic device comprising a processor, a memory, and a computer program stored in the memory and configured to be executed by the processor, the processor implementing the audio feature extraction method of any one of claims 1 to 6 when executing the computer program.
CN201911409010.XA 2019-12-31 2019-12-31 Audio feature extraction method and device, training method and electronic equipment Pending CN111105812A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911409010.XA CN111105812A (en) 2019-12-31 2019-12-31 Audio feature extraction method and device, training method and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911409010.XA CN111105812A (en) 2019-12-31 2019-12-31 Audio feature extraction method and device, training method and electronic equipment

Publications (1)

Publication Number Publication Date
CN111105812A true CN111105812A (en) 2020-05-05

Family

ID=70424543

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911409010.XA Pending CN111105812A (en) 2019-12-31 2019-12-31 Audio feature extraction method and device, training method and electronic equipment

Country Status (1)

Country Link
CN (1) CN111105812A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113889077A (en) * 2021-09-22 2022-01-04 武汉普惠海洋光电技术有限公司 Voice recognition method, voice recognition device, electronic equipment and storage medium
CN113948108A (en) * 2021-10-09 2022-01-18 广州蓝仕威克软件开发有限公司 Method and system for automatically identifying physiological sound

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6567777B1 (en) * 2000-08-02 2003-05-20 Motorola, Inc. Efficient magnitude spectrum approximation
JP2005099509A (en) * 2003-09-25 2005-04-14 Yamaha Corp Voice processing device, voice processing method, and voice processing program
JP2006323008A (en) * 2005-05-17 2006-11-30 Sharp Corp Musical piece search system and musical piece search method
US20090299741A1 (en) * 2006-04-03 2009-12-03 Naren Chittar Detection and Use of Acoustic Signal Quality Indicators
CN106683687A (en) * 2016-12-30 2017-05-17 杭州华为数字技术有限公司 Abnormal voice classifying method and device
WO2017135127A1 (en) * 2016-02-01 2017-08-10 国立大学法人徳島大学 Bioacoustic extraction device, bioacoustic analysis device, bioacoustic extraction program, and computer-readable storage medium and stored device
CN109712641A (en) * 2018-12-24 2019-05-03 重庆第二师范学院 A kind of processing method of audio classification and segmentation based on support vector machines
CN110390942A (en) * 2019-06-28 2019-10-29 平安科技(深圳)有限公司 Mood detection method and its device based on vagitus

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6567777B1 (en) * 2000-08-02 2003-05-20 Motorola, Inc. Efficient magnitude spectrum approximation
JP2005099509A (en) * 2003-09-25 2005-04-14 Yamaha Corp Voice processing device, voice processing method, and voice processing program
JP2006323008A (en) * 2005-05-17 2006-11-30 Sharp Corp Musical piece search system and musical piece search method
US20090299741A1 (en) * 2006-04-03 2009-12-03 Naren Chittar Detection and Use of Acoustic Signal Quality Indicators
WO2017135127A1 (en) * 2016-02-01 2017-08-10 国立大学法人徳島大学 Bioacoustic extraction device, bioacoustic analysis device, bioacoustic extraction program, and computer-readable storage medium and stored device
CN106683687A (en) * 2016-12-30 2017-05-17 杭州华为数字技术有限公司 Abnormal voice classifying method and device
CN109712641A (en) * 2018-12-24 2019-05-03 重庆第二师范学院 A kind of processing method of audio classification and segmentation based on support vector machines
CN110390942A (en) * 2019-06-28 2019-10-29 平安科技(深圳)有限公司 Mood detection method and its device based on vagitus

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JEFFREY SCOTT等: "Analysis of Acoustic Features for Automated Multi-Track Mixing", 《12TH INTERNATIONAL SOCIETY FOR MUSIC INFORMATION RETRIEVAL CONFERENCE (ISMIR 2011)》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113889077A (en) * 2021-09-22 2022-01-04 武汉普惠海洋光电技术有限公司 Voice recognition method, voice recognition device, electronic equipment and storage medium
CN113948108A (en) * 2021-10-09 2022-01-18 广州蓝仕威克软件开发有限公司 Method and system for automatically identifying physiological sound

Similar Documents

Publication Publication Date Title
Briggs et al. The 9th annual MLSP competition: New methods for acoustic classification of multiple simultaneous bird species in a noisy environment
EP1662485B1 (en) Signal separation method, signal separation device, signal separation program, and recording medium
CN109272016B (en) Target detection method, device, terminal equipment and computer readable storage medium
US20120103166A1 (en) Signal Processing Device, Signal Processing Method, and Program
CN110428399B (en) Method, apparatus, device and storage medium for detecting image
CN111105812A (en) Audio feature extraction method and device, training method and electronic equipment
Nanni et al. Combining visual and acoustic features for bird species classification
US9218540B2 (en) Apparatus and computer readable medium for signal classification using spectrogram and templates
CN112990313A (en) Hyperspectral image anomaly detection method and device, computer equipment and storage medium
CN108037422A (en) A kind of photovoltaic system direct current arc fault feature extraction and recognition methods and device
CN112969134B (en) Microphone abnormality detection method, device, equipment and storage medium
CN111680642A (en) Terrain classification method and device
CN106250809A (en) Pretreatment includes the method for the image of bio information
CN111310571A (en) Hyperspectral image classification method and device based on spatial-spectral-dimensional filtering
CN110189767B (en) Recording mobile equipment detection method based on dual-channel audio
CN114155875B (en) Method and device for identifying voice scene tampering, electronic equipment and storage medium
CN112382302A (en) Baby cry identification method and terminal equipment
CN107481732B (en) Noise reduction method and device in spoken language evaluation and terminal equipment
US20150356164A1 (en) Method and device for clustering file
CN110534128B (en) Noise processing method, device, equipment and storage medium
Zhang et al. Discriminant WSRC for Large‐Scale Plant Species Recognition
CN114792374A (en) Image recognition method based on texture classification, electronic device and storage medium
CN108227750B (en) Ground target real-time tracking performance evaluation method and system
CN108964809B (en) Signal detection method and electronic device
CN113743387B (en) Video pedestrian re-identification method and device, electronic equipment and readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200505

RJ01 Rejection of invention patent application after publication