CN105448290A - Variable frame rate audio feature extraction method - Google Patents
Variable frame rate audio feature extraction method Download PDFInfo
- Publication number
- CN105448290A CN105448290A CN201510782814.XA CN201510782814A CN105448290A CN 105448290 A CN105448290 A CN 105448290A CN 201510782814 A CN201510782814 A CN 201510782814A CN 105448290 A CN105448290 A CN 105448290A
- Authority
- CN
- China
- Prior art keywords
- frames
- frame
- audio
- candidate
- feature vector
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000605 extraction Methods 0.000 title claims abstract description 16
- 239000013598 vector Substances 0.000 claims abstract description 81
- 238000000034 method Methods 0.000 claims abstract description 20
- 238000009432 framing Methods 0.000 claims description 11
- 230000000694 effects Effects 0.000 abstract description 6
- 230000004927 fusion Effects 0.000 abstract description 5
- 238000007500 overflow downdraw method Methods 0.000 description 4
- 238000001228 spectrum Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 230000007547 defect Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention provides a variable frame rate audio feature extraction method. The feature vector of a reference frame is obtained by the weight fusion of the feature vectors of a plurality of selected frames, and the Euclidean distance between the feature vector of the reference frame and the feature vectors of a plurality of candidate frames, and an audio frame which best represents the plurality of candidate frames is selected from the plurality of candidate frames to be subjected to audio retrieval. The method prevents the shielding effect of the selected frames on the candidate frames, facilitates the extraction of useful audio information, and improves the accuracy of audio retrieval.
Description
Technical Field
The invention relates to the technical field of digital audio processing, in particular to a variable frame rate audio feature extraction method.
Background
With the explosion of the internet, the data volume of audio media on the internet is increasing day by day, and the traditional audio search based on text labels cannot meet the increasing use requirements of people. In recent years, with the rise of cloud computing and big data, the audio retrieval technology based on contents is becoming the focus of attention of scholars at home and abroad.
The audio feature extraction is a key of the content-based audio retrieval technology, and the general processing process of the audio feature extraction needs to firstly perform framing processing on audio data. The framing process usually adopts an overlapping framing method to maintain continuity between frames, the length of a divided frame is called a frame length, and the non-overlapping part between frames is called a frame offset. There are two main ways to extract audio features at present: fixed frame rate and variable frame rate. The fixed frame rate, namely the frame length and the frame offset are kept unchanged in the framing process, and the method cannot be well adapted to the change characteristic of an audio frequency spectrum; the frame rate is changed, namely, dynamic frame offset is adopted for framing in the framing process, so that the defect that the fixed frame rate cannot reflect audio frequency spectrum change can be effectively overcome.
In the existing method for extracting the audio features of the variable frame rate, the fixed frame rate and the small frame offset are adopted for framing, then the similarity between a candidate frame and a selected frame is calculated, and the effect of the variable frame rate is achieved by frame loss. The method is easy to cause the shielding effect of the selected frame on the candidate frame, so that the audio frame which can show the audio characteristic most is discarded in the framing process, the audio information is lost, and the retrieval accuracy is low.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a variable frame rate audio feature extraction method. According to the method, the audio frame which can represent audio changes most is selected from the candidate frames, so that the shielding effect of the selected frame on the candidate frames is avoided, the effectiveness of audio feature extraction is improved, and the accuracy of audio retrieval is improved.
In order to solve the technical problems, the invention adopts the technical scheme that:
the invention provides a variable frame rate audio feature extraction method, which comprises the following specific processes:
step a, framing audio data according to a preset frame length and frame offset to obtain audio frames, and calculating characteristic vectors of the audio frames;
step b, sequentially selecting n audio frames from the audio frames as initial selected frames, and taking the subsequent h audio frames as candidate frames; n and h are preset natural numbers;
step c, calculating the reference frame feature vector of the selected frame according to the n selected frame feature vectors
Wherein v isiRepresenting the ith selected frame feature vector, wiRepresents the weight of the ith selected frame; i is the label of the selected frame and is a positive integer; m represents a total of m selected frames, and 1<n≤m;
Each audio frame feature vector is represented by a Q-dimensional component, the ith selected frame feature vector viAre respectively expressed as:wherein k is an integer and k is greater than or equal to 1 and less than or equal to Q, so that the reference frame feature vectorComponent of the k-th dimension ofThe calculation formula of (2) is as follows:
wherein,representing the kth dimension component of the ith selected frame feature vector,representing the weight of the kth dimensional component of the ith selected frame feature vector,the calculation formula of (2) is as follows:
and satisfies the following conditions:wherein,the calculation formula of (2) is as follows:
step d, according to the reference frame characteristic vectorCalculating Euclidean distance d between h candidate frame feature vectors and reference frame feature vectorl;
Euclidean distance dlThe calculation formula of (a) is as follows:
wherein l is the label of the candidate frame and is a positive integer, and l is more than or equal to 1 and less than or equal to h;
e, according to Euclidean distance d between h candidate frame characteristic vectors and reference frame characteristic vectorlSelecting candidate frames meeting the conditions, and adding the candidate frames into the selected frames, wherein the specific process is as follows:
step e-1, according to the Euclidean distanceFrom dlThe maximum Euclidean distance d is selectedmax;
Step e-2, determining the maximum Euclidean distance dmaxComparing the maximum Euclidean distance D with a threshold value D, and if the maximum Euclidean distance D exceeds the threshold valuemaxAdding the corresponding candidate frames into the selected frames, taking the h audio frames after the selected frames as new candidate frames, and if not, discarding all the h candidate frames and continuously using the subsequent h audio frames as candidate frames for selection;
and c, repeatedly executing the step d and the step e, sequentially selecting h audio frames from the rest audio frames as candidate frames, selecting candidate frames meeting the conditions from the h candidate frames to be added into the selected frames, if the number of the rest audio frames is less than h, taking the actually rest audio frames as the candidate frames, selecting the candidate frames meeting the conditions from the candidate frames by using the same method to be added into the selected frames until the number of the rest audio frames is zero, and taking the feature vectors of all the selected frames as the feature vectors of the audio data.
Has the advantages that: the audio feature extraction method of the variable frame rate provided by the invention obtains the reference frame feature vector by performing weighted fusion on the selected frame feature vectors, calculates Euclidean distances between the reference frame feature vector and the candidate frame feature vectors respectively, and selects the audio frame which can represent the candidate frames most from the candidate frames according to the Euclidean distances to perform audio retrieval, thereby avoiding the shielding effect of the selected frame on the candidate frames, being beneficial to extracting more useful audio information and improving the accuracy of the audio retrieval.
Drawings
FIG. 1 is a schematic diagram of the calculation of feature vectors of reference frames using a weighted fusion method;
FIG. 2 is a schematic view of a process for selecting qualified candidate frames from candidate frames and adding the qualified candidate frames to a selected frame according to a reference frame;
fig. 3 is a schematic flow chart illustrating an implementation of a variable frame rate audio feature extraction method.
Detailed Description
For more detailed description of the variable frame rate audio feature extraction method proposed by the present invention, the following examples are illustrated with reference to the accompanying drawings:
the invention mainly comprises three contents: firstly, weighting and fusing a plurality of selected frame feature vectors to obtain a reference frame feature vector; calculating Euclidean distances between the reference frame feature vector and a plurality of candidate frame feature vectors; thirdly, selecting the candidate frame meeting the conditions according to the Euclidean distances between the feature vector of the reference frame and the feature vectors of the plurality of candidate frames and adding the candidate frame into the selected frame.
1. The reference frame is calculated using a weighted fusion method:
FIG. 1 is a schematic diagram illustrating calculation of feature vectors of reference frames by using a weighted fusion method, where all m selected frame feature vectors are respectively denoted as v1,v2,v3,...,vmSetting to perform weighted fusion according to the last 4 selected frames, and respectively marking the feature vectors of the 4 selected frames as vm-3,vm-2,vm-1,vm. Each audio frame feature vector is expressed by Q-dimensional component, and the feature vector v of the ith selected frameiIs represented as:wherein k is an integer, and k is more than or equal to 1 and less than or equal to Q; the feature types extracted from each audio frame include, but are not limited to, time domain features, frequency domain features, or a combination thereof of the audio, such as autocorrelation coefficients, mel-frequency cepstral coefficients, and the like.
Each dimension component of the reference frame feature vector is obtained by weighting and fusing each dimension component of the 4 selected frame feature vectors, and the calculation formula is as follows:
wherein,representing the kth dimension component of the ith selected frame feature vector,a k-dimension component representing a feature vector of a reference frame;a weight representing the kth dimension component of the ith selected frame feature vector; i. k respectively represents the label of the selected frame and the label of each dimension component of each audio frame feature vector, and i and k are positive integers; m represents a total of m selected frames.
Weight ofThe quality of the reference frame is directly affected, and the selection of the candidate frame is further affected. In the schemeThe calculation formula of (2) is as follows:
and the weight coefficient satisfies:whereinRepresenting the absolute value of the difference between the kth dimension component of the ith and jth selected frame feature vector and the mean value of the kth dimension component of the nth selected frame feature vector,the calculation formula of (2) is as follows:
wherein j is the label of the selected frame and is a positive integer.
It should be noted that, in the embodiment, the reference frame feature vector is calculated by performing weighted fusion on the 4 selected frame feature vectors, and the reference frame feature vector may also be calculated by using other numbers of selected frame feature vectors according to the actual situation.
2. Calculating Euclidean distances between the reference frame feature vector and a plurality of candidate frame feature vectors:
supposing that selection is carried out from the subsequent 3 candidate frames, according to the obtained reference frame feature vector, the Euclidean distance d between the feature vector of the 3 candidate frames and the feature vector of the reference frame is calculatedlThe calculation formula is as follows:
the Euclidean distance between the reference frame feature vector and each candidate frame feature vector is used for measuring the similarity degree between the reference frame and each candidate frame, wherein l is the label of the candidate frame, is a positive integer, and is more than or equal to 1 and less than or equal to 3;
3. according to Euclidean distance d between the reference frame feature vector and the candidate frame feature vectorslSelecting candidate frames meeting the conditions to be added into the selected frames:
according to Euclidean distance dlA schematic flow chart of selecting a candidate frame meeting the condition from the candidate frames is shown in fig. 2. Of the 3 candidate frames, the candidate frame with the largest euclidean distance is least similar to the reference frame, and the candidate frame with the least similarity to the reference frame can reflect the audio change most. The specific process of selecting the candidate frame meeting the conditions from the candidate frames according to the Euclidean distance and adding the candidate frame into the selected frame is as follows:
step 3-1: from Euclidean distance dlSelecting the maximum Euclidean distance dmax;
Step 3-2: the maximum Euclidean distance dmaxComparing with a threshold value D, and if the threshold value is exceeded, determining the maximum Euclidean distance DmaxAnd adding the corresponding candidate frame into the selected frame, taking the 3 audio frames after the selected frame as new candidate frames, or else, completely discarding the 3 audio frames, and continuously using the subsequent 3 audio frames as candidate frames for selection.
The audio frames which can represent the candidate frames most are selected from the candidate frames, and the preset threshold value D is combined, so that more frames are selected in the place where the audio frequency spectrum changes greatly in the characteristic extraction process, less frames are selected in the place where the audio frequency spectrum changes slowly, and the shielding effect of the selected frames on the candidate frames can be avoided.
It should be noted that, in the embodiment, for example, a candidate frame meeting the condition is selected from the subsequent 3 candidate frames and added to the selected frame, and a candidate frame meeting the condition may also be selected from other number of candidate frames and added to the selected frame according to the actual situation.
Fig. 3 shows an execution flow diagram of the variable frame rate audio feature extraction method provided by the present invention, and the specific execution flow is as follows:
step 1, framing is carried out according to a preset frame length and frame offset, if the frame length is 30ms and the frame offset is 2ms, audio data are framed to obtain audio frames, and feature vectors of the audio frames are calculated;
step 2, according to preset n and h, sequentially selecting n audio frames from the audio frames as initial selected frames, and taking the subsequent h audio frames as candidate frames;
step 3, calculating the weight of each dimensional component of the n selected frame feature vectors by using a weighted fusion method, and performing weighted fusion to obtain a reference frame feature vector;
step 4, calculating Euclidean distances between the feature vectors of the reference frame and the feature vectors of the h candidate frames;
and 5, selecting the maximum Euclidean distance according to the Euclidean distances between the feature vectors of the reference frame and the h candidate frame feature vectors, comparing the maximum Euclidean distance with a threshold value, adding the candidate frame corresponding to the maximum Euclidean distance into the selected frame if the maximum Euclidean distance exceeds the threshold value, and discarding the group of candidate frames and selecting from the next group of candidate frames if the maximum Euclidean distance does not exceed the threshold value.
And repeating the step 3, the step 4 and the step 5, sequentially selecting h audio frames from the rest audio frames as candidate frames, selecting candidate frames meeting the conditions from the candidate frames to be added into the selected frames, if the number of the rest audio frames is less than h, taking the actually rest audio frames as the candidate frames, selecting the candidate frames meeting the conditions from the candidate frames to be added into the selected frames by using the same method until the number of the rest audio frames is zero, and taking the feature vectors of all the selected frames as the feature vectors of the audio data.
Claims (1)
1. A frame rate variable audio feature extraction method is characterized by comprising the following specific processes:
step a, framing audio data according to a preset frame length and frame offset to obtain audio frames, and calculating characteristic vectors of the audio frames;
step b, sequentially selecting n audio frames from the audio frames as initial selected frames, and taking the subsequent h audio frames as candidate frames; n and h are preset natural numbers;
step c, calculating the selected frame according to n selected frame feature vectorsReference frame feature vector of selected frame
Wherein v isiRepresenting the ith selected frame feature vector, wiRepresents the weight of the ith selected frame; i is the label of the selected frame and is a positive integer; m represents a total of m selected frames, and 1<n≤m;
Each audio frame feature vector is represented by a Q-dimensional component, the ith selected frame feature vector viAre respectively expressed as:wherein k is an integer and k is greater than or equal to 1 and less than or equal to Q, so that the reference frame feature vectorComponent of the k-th dimension ofThe calculation formula of (2) is as follows:
wherein,representing the kth dimension component of the ith selected frame feature vector,representing the weight of the kth dimensional component of the ith selected frame feature vector,the calculation formula of (2) is as follows:
and satisfies the following conditions:wherein,the calculation formula of (2) is as follows:
step d, according to the reference frame characteristic vectorCalculating Euclidean distance d between h candidate frame feature vectors and reference frame feature vectorl;
Euclidean distance dlThe calculation formula of (a) is as follows:
wherein l is the label of the candidate frame and is a positive integer, and l is more than or equal to 1 and less than or equal to h;
e, according to Euclidean distance d between h candidate frame characteristic vectors and reference frame characteristic vectorlSelecting candidate frames meeting the conditions, and adding the candidate frames into the selected frames, wherein the specific process is as follows:
step e-1, according to the Euclidean distance dlThe maximum Euclidean distance d is selectedmax;
Step e-2, determining the maximum Euclidean distance dmaxComparing the maximum Euclidean distance D with a threshold value D, and if the maximum Euclidean distance D exceeds the threshold valuemaxAdding the corresponding candidate frames into the selected frames, taking the h audio frames after the selected frames as new candidate frames, and if not, discarding all the h candidate frames and continuously using the subsequent h audio frames as candidate frames for selection;
c, repeating the step d and the step e, sequentially selecting h audio frames from the rest audio frames as candidate frames, and selecting candidate frames meeting the conditions from the h candidate frames to be added into the selected frames; and if the number of the remaining audio frames is less than h, taking the actually remaining audio frames as candidate frames, selecting the candidate frames meeting the conditions from the candidate frames by using the same method, adding the candidate frames into the selected frames until the number of the remaining audio frames is zero, and taking the feature vectors of all the selected frames as the feature vectors of the audio data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510782814.XA CN105448290B (en) | 2015-11-16 | 2015-11-16 | A kind of audio feature extraction methods becoming frame per second |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510782814.XA CN105448290B (en) | 2015-11-16 | 2015-11-16 | A kind of audio feature extraction methods becoming frame per second |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105448290A true CN105448290A (en) | 2016-03-30 |
CN105448290B CN105448290B (en) | 2019-03-01 |
Family
ID=55558397
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510782814.XA Active CN105448290B (en) | 2015-11-16 | 2015-11-16 | A kind of audio feature extraction methods becoming frame per second |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105448290B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108922556A (en) * | 2018-07-16 | 2018-11-30 | 百度在线网络技术(北京)有限公司 | sound processing method, device and equipment |
CN109643555A (en) * | 2016-07-04 | 2019-04-16 | 哈曼贝克自动系统股份有限公司 | Automatically correct the loudness level in the audio signal comprising voice signal |
CN114567780A (en) * | 2022-02-25 | 2022-05-31 | 杭州当虹科技股份有限公司 | Video frame alignment method |
CN117097909A (en) * | 2023-10-20 | 2023-11-21 | 深圳市星易美科技有限公司 | Distributed household audio and video processing method and system |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1505572A1 (en) * | 2002-05-06 | 2005-02-09 | Prous Science S.A. | Voice recognition method |
CN102332262A (en) * | 2011-09-23 | 2012-01-25 | 哈尔滨工业大学深圳研究生院 | Method for intelligently identifying songs based on audio features |
US20130166294A1 (en) * | 1999-12-10 | 2013-06-27 | At&T Intellectual Property Ii, L.P. | Frame Erasure Concealment Technique for a Bitstream-Based Feature Extractor |
WO2015092711A1 (en) * | 2013-12-18 | 2015-06-25 | Isis Innovation Ltd. | Method and apparatus for automatic speech recognition |
CN105046699A (en) * | 2015-07-09 | 2015-11-11 | 硅革科技(北京)有限公司 | Motion video superposition contrast method |
-
2015
- 2015-11-16 CN CN201510782814.XA patent/CN105448290B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130166294A1 (en) * | 1999-12-10 | 2013-06-27 | At&T Intellectual Property Ii, L.P. | Frame Erasure Concealment Technique for a Bitstream-Based Feature Extractor |
EP1505572A1 (en) * | 2002-05-06 | 2005-02-09 | Prous Science S.A. | Voice recognition method |
CN102332262A (en) * | 2011-09-23 | 2012-01-25 | 哈尔滨工业大学深圳研究生院 | Method for intelligently identifying songs based on audio features |
WO2015092711A1 (en) * | 2013-12-18 | 2015-06-25 | Isis Innovation Ltd. | Method and apparatus for automatic speech recognition |
CN105046699A (en) * | 2015-07-09 | 2015-11-11 | 硅革科技(北京)有限公司 | Motion video superposition contrast method |
Non-Patent Citations (2)
Title |
---|
严焰 等: "基于HMM的手势识别研究", 《华中师范大学学报(自然科学版)》 * |
杨涛 等: "基于量子粒子群优化算法的运动捕获数据关键帧提取", 《计算机应用研究》 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109643555A (en) * | 2016-07-04 | 2019-04-16 | 哈曼贝克自动系统股份有限公司 | Automatically correct the loudness level in the audio signal comprising voice signal |
CN109643555B (en) * | 2016-07-04 | 2024-01-30 | 哈曼贝克自动系统股份有限公司 | Automatic correction of loudness level in an audio signal containing a speech signal |
CN108922556A (en) * | 2018-07-16 | 2018-11-30 | 百度在线网络技术(北京)有限公司 | sound processing method, device and equipment |
CN108922556B (en) * | 2018-07-16 | 2019-08-27 | 百度在线网络技术(北京)有限公司 | Sound processing method, device and equipment |
CN114567780A (en) * | 2022-02-25 | 2022-05-31 | 杭州当虹科技股份有限公司 | Video frame alignment method |
CN114567780B (en) * | 2022-02-25 | 2024-08-23 | 杭州当虹科技股份有限公司 | Video frame alignment method |
CN117097909A (en) * | 2023-10-20 | 2023-11-21 | 深圳市星易美科技有限公司 | Distributed household audio and video processing method and system |
CN117097909B (en) * | 2023-10-20 | 2024-02-02 | 深圳市星易美科技有限公司 | Distributed household audio and video processing method and system |
Also Published As
Publication number | Publication date |
---|---|
CN105448290B (en) | 2019-03-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020173133A1 (en) | Training method of emotion recognition model, emotion recognition method, device, apparatus, and storage medium | |
CN107480143B (en) | Method and system for segmenting conversation topics based on context correlation | |
EP3309782B1 (en) | Method, device and system for noise suppression | |
CN106297776B (en) | A kind of voice keyword retrieval method based on audio template | |
CN106653056B (en) | Fundamental frequency extraction model and training method based on LSTM recurrent neural network | |
CN107193797B (en) | Hot topic detection and trend prediction method for Chinese microblog | |
CN103049501B (en) | Based on mutual information and the Chinese domain term recognition method of conditional random field models | |
CN105448290B (en) | A kind of audio feature extraction methods becoming frame per second | |
CN104599292B (en) | A kind of anti-noise moving object detection algorithm decomposed based on low-rank matrix | |
CN103559196B (en) | Video retrieval method based on multi-core canonical correlation analysis | |
CN101819638B (en) | Establishment method of pornographic detection model and pornographic detection method | |
CN104536983A (en) | Method and device for predicting advertisement click rate | |
CN110147760B (en) | Novel efficient electric energy quality disturbance image feature extraction and identification method | |
CN103943107A (en) | Audio/video keyword identification method based on decision-making level fusion | |
CN107123432A (en) | A kind of Self Matching Top N audio events recognize channel self-adapted method | |
CN112767922B (en) | Speech recognition method for contrast predictive coding self-supervision structure joint training | |
CN112488133A (en) | Video/picture-text cross-modal retrieval method | |
CN105138983A (en) | Pedestrian detection method based on weighted part model and selective search segmentation | |
CN110956222A (en) | Method for detecting network for underwater target detection | |
CN111724766B (en) | Language identification method, related equipment and readable storage medium | |
Pang et al. | Discrete Cosine Transformation and Temporal Adjacent Convolutional Neural Network‐Based Remaining Useful Life Estimation of Bearings | |
CN104867493B (en) | Multifractal Dimension end-point detecting method based on wavelet transformation | |
CN108564967B (en) | Mel energy voiceprint feature extraction method for crying detection system | |
CN103336765A (en) | Markov matrix off-line correction method of text keywords | |
CN107894837A (en) | Dynamic sentiment analysis model sample processing method and processing device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |