CN105448290A - Variable frame rate audio feature extraction method - Google Patents

Variable frame rate audio feature extraction method Download PDF

Info

Publication number
CN105448290A
CN105448290A CN201510782814.XA CN201510782814A CN105448290A CN 105448290 A CN105448290 A CN 105448290A CN 201510782814 A CN201510782814 A CN 201510782814A CN 105448290 A CN105448290 A CN 105448290A
Authority
CN
China
Prior art keywords
frames
frame
audio
candidate
feature vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510782814.XA
Other languages
Chinese (zh)
Other versions
CN105448290B (en
Inventor
张晖
刘宝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Post and Telecommunication University
Original Assignee
Nanjing Post and Telecommunication University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Post and Telecommunication University filed Critical Nanjing Post and Telecommunication University
Priority to CN201510782814.XA priority Critical patent/CN105448290B/en
Publication of CN105448290A publication Critical patent/CN105448290A/en
Application granted granted Critical
Publication of CN105448290B publication Critical patent/CN105448290B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention provides a variable frame rate audio feature extraction method. The feature vector of a reference frame is obtained by the weight fusion of the feature vectors of a plurality of selected frames, and the Euclidean distance between the feature vector of the reference frame and the feature vectors of a plurality of candidate frames, and an audio frame which best represents the plurality of candidate frames is selected from the plurality of candidate frames to be subjected to audio retrieval. The method prevents the shielding effect of the selected frames on the candidate frames, facilitates the extraction of useful audio information, and improves the accuracy of audio retrieval.

Description

Audio feature extraction method with variable frame rate
Technical Field
The invention relates to the technical field of digital audio processing, in particular to a variable frame rate audio feature extraction method.
Background
With the explosion of the internet, the data volume of audio media on the internet is increasing day by day, and the traditional audio search based on text labels cannot meet the increasing use requirements of people. In recent years, with the rise of cloud computing and big data, the audio retrieval technology based on contents is becoming the focus of attention of scholars at home and abroad.
The audio feature extraction is a key of the content-based audio retrieval technology, and the general processing process of the audio feature extraction needs to firstly perform framing processing on audio data. The framing process usually adopts an overlapping framing method to maintain continuity between frames, the length of a divided frame is called a frame length, and the non-overlapping part between frames is called a frame offset. There are two main ways to extract audio features at present: fixed frame rate and variable frame rate. The fixed frame rate, namely the frame length and the frame offset are kept unchanged in the framing process, and the method cannot be well adapted to the change characteristic of an audio frequency spectrum; the frame rate is changed, namely, dynamic frame offset is adopted for framing in the framing process, so that the defect that the fixed frame rate cannot reflect audio frequency spectrum change can be effectively overcome.
In the existing method for extracting the audio features of the variable frame rate, the fixed frame rate and the small frame offset are adopted for framing, then the similarity between a candidate frame and a selected frame is calculated, and the effect of the variable frame rate is achieved by frame loss. The method is easy to cause the shielding effect of the selected frame on the candidate frame, so that the audio frame which can show the audio characteristic most is discarded in the framing process, the audio information is lost, and the retrieval accuracy is low.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a variable frame rate audio feature extraction method. According to the method, the audio frame which can represent audio changes most is selected from the candidate frames, so that the shielding effect of the selected frame on the candidate frames is avoided, the effectiveness of audio feature extraction is improved, and the accuracy of audio retrieval is improved.
In order to solve the technical problems, the invention adopts the technical scheme that:
the invention provides a variable frame rate audio feature extraction method, which comprises the following specific processes:
step a, framing audio data according to a preset frame length and frame offset to obtain audio frames, and calculating characteristic vectors of the audio frames;
step b, sequentially selecting n audio frames from the audio frames as initial selected frames, and taking the subsequent h audio frames as candidate frames; n and h are preset natural numbers;
step c, calculating the reference frame feature vector of the selected frame according to the n selected frame feature vectors
v ~ m = Σ i = m - n + 1 m w i v i
Wherein v isiRepresenting the ith selected frame feature vector, wiRepresents the weight of the ith selected frame; i is the label of the selected frame and is a positive integer; m represents a total of m selected frames, and 1<n≤m;
Each audio frame feature vector is represented by a Q-dimensional component, the ith selected frame feature vector viAre respectively expressed as:wherein k is an integer and k is greater than or equal to 1 and less than or equal to Q, so that the reference frame feature vectorComponent of the k-th dimension ofThe calculation formula of (2) is as follows:
v ~ m k = &Sigma; i = m - n + 1 m w i k v i k
wherein,representing the kth dimension component of the ith selected frame feature vector,representing the weight of the kth dimensional component of the ith selected frame feature vector,the calculation formula of (2) is as follows:
w i k = 1 / ( &sigma; i k ) 2 &Sigma; j = m - n + 1 m 1 / ( &sigma; j k ) 2
and satisfies the following conditions:wherein,the calculation formula of (2) is as follows:
&sigma; i k = | v i k - 1 n &Sigma; j = m - n + 1 m v j k |
step d, according to the reference frame characteristic vectorCalculating Euclidean distance d between h candidate frame feature vectors and reference frame feature vectorl
Euclidean distance dlThe calculation formula of (a) is as follows:
d l = &Sigma; k = 1 Q ( v ~ m k - v l k ) 2
wherein l is the label of the candidate frame and is a positive integer, and l is more than or equal to 1 and less than or equal to h;
e, according to Euclidean distance d between h candidate frame characteristic vectors and reference frame characteristic vectorlSelecting candidate frames meeting the conditions, and adding the candidate frames into the selected frames, wherein the specific process is as follows:
step e-1, according to the Euclidean distanceFrom dlThe maximum Euclidean distance d is selectedmax
Step e-2, determining the maximum Euclidean distance dmaxComparing the maximum Euclidean distance D with a threshold value D, and if the maximum Euclidean distance D exceeds the threshold valuemaxAdding the corresponding candidate frames into the selected frames, taking the h audio frames after the selected frames as new candidate frames, and if not, discarding all the h candidate frames and continuously using the subsequent h audio frames as candidate frames for selection;
and c, repeatedly executing the step d and the step e, sequentially selecting h audio frames from the rest audio frames as candidate frames, selecting candidate frames meeting the conditions from the h candidate frames to be added into the selected frames, if the number of the rest audio frames is less than h, taking the actually rest audio frames as the candidate frames, selecting the candidate frames meeting the conditions from the candidate frames by using the same method to be added into the selected frames until the number of the rest audio frames is zero, and taking the feature vectors of all the selected frames as the feature vectors of the audio data.
Has the advantages that: the audio feature extraction method of the variable frame rate provided by the invention obtains the reference frame feature vector by performing weighted fusion on the selected frame feature vectors, calculates Euclidean distances between the reference frame feature vector and the candidate frame feature vectors respectively, and selects the audio frame which can represent the candidate frames most from the candidate frames according to the Euclidean distances to perform audio retrieval, thereby avoiding the shielding effect of the selected frame on the candidate frames, being beneficial to extracting more useful audio information and improving the accuracy of the audio retrieval.
Drawings
FIG. 1 is a schematic diagram of the calculation of feature vectors of reference frames using a weighted fusion method;
FIG. 2 is a schematic view of a process for selecting qualified candidate frames from candidate frames and adding the qualified candidate frames to a selected frame according to a reference frame;
fig. 3 is a schematic flow chart illustrating an implementation of a variable frame rate audio feature extraction method.
Detailed Description
For more detailed description of the variable frame rate audio feature extraction method proposed by the present invention, the following examples are illustrated with reference to the accompanying drawings:
the invention mainly comprises three contents: firstly, weighting and fusing a plurality of selected frame feature vectors to obtain a reference frame feature vector; calculating Euclidean distances between the reference frame feature vector and a plurality of candidate frame feature vectors; thirdly, selecting the candidate frame meeting the conditions according to the Euclidean distances between the feature vector of the reference frame and the feature vectors of the plurality of candidate frames and adding the candidate frame into the selected frame.
1. The reference frame is calculated using a weighted fusion method:
FIG. 1 is a schematic diagram illustrating calculation of feature vectors of reference frames by using a weighted fusion method, where all m selected frame feature vectors are respectively denoted as v1,v2,v3,...,vmSetting to perform weighted fusion according to the last 4 selected frames, and respectively marking the feature vectors of the 4 selected frames as vm-3,vm-2,vm-1,vm. Each audio frame feature vector is expressed by Q-dimensional component, and the feature vector v of the ith selected frameiIs represented as:wherein k is an integer, and k is more than or equal to 1 and less than or equal to Q; the feature types extracted from each audio frame include, but are not limited to, time domain features, frequency domain features, or a combination thereof of the audio, such as autocorrelation coefficients, mel-frequency cepstral coefficients, and the like.
Each dimension component of the reference frame feature vector is obtained by weighting and fusing each dimension component of the 4 selected frame feature vectors, and the calculation formula is as follows:
v ~ m k = &Sigma; i = m - 3 m w i k v i k - - - ( 1 )
wherein,representing the kth dimension component of the ith selected frame feature vector,a k-dimension component representing a feature vector of a reference frame;a weight representing the kth dimension component of the ith selected frame feature vector; i. k respectively represents the label of the selected frame and the label of each dimension component of each audio frame feature vector, and i and k are positive integers; m represents a total of m selected frames.
Weight ofThe quality of the reference frame is directly affected, and the selection of the candidate frame is further affected. In the schemeThe calculation formula of (2) is as follows:
w i k = 1 / ( &sigma; i k ) 2 &Sigma; j = m - 3 m 1 / ( &sigma; j k ) 2 - - - ( 2 )
and the weight coefficient satisfies:whereinRepresenting the absolute value of the difference between the kth dimension component of the ith and jth selected frame feature vector and the mean value of the kth dimension component of the nth selected frame feature vector,the calculation formula of (2) is as follows:
&sigma; i k = | v i k - 1 n &Sigma; j = m - 3 m v j k | - - - ( 3 )
wherein j is the label of the selected frame and is a positive integer.
It should be noted that, in the embodiment, the reference frame feature vector is calculated by performing weighted fusion on the 4 selected frame feature vectors, and the reference frame feature vector may also be calculated by using other numbers of selected frame feature vectors according to the actual situation.
2. Calculating Euclidean distances between the reference frame feature vector and a plurality of candidate frame feature vectors:
supposing that selection is carried out from the subsequent 3 candidate frames, according to the obtained reference frame feature vector, the Euclidean distance d between the feature vector of the 3 candidate frames and the feature vector of the reference frame is calculatedlThe calculation formula is as follows:
d l = &Sigma; k = 1 Q ( v ~ m k - v l k ) 2 - - - ( 4 )
the Euclidean distance between the reference frame feature vector and each candidate frame feature vector is used for measuring the similarity degree between the reference frame and each candidate frame, wherein l is the label of the candidate frame, is a positive integer, and is more than or equal to 1 and less than or equal to 3;
3. according to Euclidean distance d between the reference frame feature vector and the candidate frame feature vectorslSelecting candidate frames meeting the conditions to be added into the selected frames:
according to Euclidean distance dlA schematic flow chart of selecting a candidate frame meeting the condition from the candidate frames is shown in fig. 2. Of the 3 candidate frames, the candidate frame with the largest euclidean distance is least similar to the reference frame, and the candidate frame with the least similarity to the reference frame can reflect the audio change most. The specific process of selecting the candidate frame meeting the conditions from the candidate frames according to the Euclidean distance and adding the candidate frame into the selected frame is as follows:
step 3-1: from Euclidean distance dlSelecting the maximum Euclidean distance dmax
Step 3-2: the maximum Euclidean distance dmaxComparing with a threshold value D, and if the threshold value is exceeded, determining the maximum Euclidean distance DmaxAnd adding the corresponding candidate frame into the selected frame, taking the 3 audio frames after the selected frame as new candidate frames, or else, completely discarding the 3 audio frames, and continuously using the subsequent 3 audio frames as candidate frames for selection.
The audio frames which can represent the candidate frames most are selected from the candidate frames, and the preset threshold value D is combined, so that more frames are selected in the place where the audio frequency spectrum changes greatly in the characteristic extraction process, less frames are selected in the place where the audio frequency spectrum changes slowly, and the shielding effect of the selected frames on the candidate frames can be avoided.
It should be noted that, in the embodiment, for example, a candidate frame meeting the condition is selected from the subsequent 3 candidate frames and added to the selected frame, and a candidate frame meeting the condition may also be selected from other number of candidate frames and added to the selected frame according to the actual situation.
Fig. 3 shows an execution flow diagram of the variable frame rate audio feature extraction method provided by the present invention, and the specific execution flow is as follows:
step 1, framing is carried out according to a preset frame length and frame offset, if the frame length is 30ms and the frame offset is 2ms, audio data are framed to obtain audio frames, and feature vectors of the audio frames are calculated;
step 2, according to preset n and h, sequentially selecting n audio frames from the audio frames as initial selected frames, and taking the subsequent h audio frames as candidate frames;
step 3, calculating the weight of each dimensional component of the n selected frame feature vectors by using a weighted fusion method, and performing weighted fusion to obtain a reference frame feature vector;
step 4, calculating Euclidean distances between the feature vectors of the reference frame and the feature vectors of the h candidate frames;
and 5, selecting the maximum Euclidean distance according to the Euclidean distances between the feature vectors of the reference frame and the h candidate frame feature vectors, comparing the maximum Euclidean distance with a threshold value, adding the candidate frame corresponding to the maximum Euclidean distance into the selected frame if the maximum Euclidean distance exceeds the threshold value, and discarding the group of candidate frames and selecting from the next group of candidate frames if the maximum Euclidean distance does not exceed the threshold value.
And repeating the step 3, the step 4 and the step 5, sequentially selecting h audio frames from the rest audio frames as candidate frames, selecting candidate frames meeting the conditions from the candidate frames to be added into the selected frames, if the number of the rest audio frames is less than h, taking the actually rest audio frames as the candidate frames, selecting the candidate frames meeting the conditions from the candidate frames to be added into the selected frames by using the same method until the number of the rest audio frames is zero, and taking the feature vectors of all the selected frames as the feature vectors of the audio data.

Claims (1)

1. A frame rate variable audio feature extraction method is characterized by comprising the following specific processes:
step a, framing audio data according to a preset frame length and frame offset to obtain audio frames, and calculating characteristic vectors of the audio frames;
step b, sequentially selecting n audio frames from the audio frames as initial selected frames, and taking the subsequent h audio frames as candidate frames; n and h are preset natural numbers;
step c, calculating the selected frame according to n selected frame feature vectorsReference frame feature vector of selected frame
v ~ m = &Sigma; i = m - n + 1 m w i v i
Wherein v isiRepresenting the ith selected frame feature vector, wiRepresents the weight of the ith selected frame; i is the label of the selected frame and is a positive integer; m represents a total of m selected frames, and 1<n≤m;
Each audio frame feature vector is represented by a Q-dimensional component, the ith selected frame feature vector viAre respectively expressed as:wherein k is an integer and k is greater than or equal to 1 and less than or equal to Q, so that the reference frame feature vectorComponent of the k-th dimension ofThe calculation formula of (2) is as follows:
v ~ m k = &Sigma; i = m - n + 1 m w i k v i k
wherein,representing the kth dimension component of the ith selected frame feature vector,representing the weight of the kth dimensional component of the ith selected frame feature vector,the calculation formula of (2) is as follows:
w i k = 1 / ( &sigma; i k ) 2 &Sigma; j = m - n + 1 m 1 / ( &sigma; j k ) 2
and satisfies the following conditions:wherein,the calculation formula of (2) is as follows:
&sigma; i k = | v i k - 1 n &Sigma; j = m - n + 1 m v j k |
step d, according to the reference frame characteristic vectorCalculating Euclidean distance d between h candidate frame feature vectors and reference frame feature vectorl
Euclidean distance dlThe calculation formula of (a) is as follows:
d l = &Sigma; k = 1 Q ( v ~ m k - v l k ) 2
wherein l is the label of the candidate frame and is a positive integer, and l is more than or equal to 1 and less than or equal to h;
e, according to Euclidean distance d between h candidate frame characteristic vectors and reference frame characteristic vectorlSelecting candidate frames meeting the conditions, and adding the candidate frames into the selected frames, wherein the specific process is as follows:
step e-1, according to the Euclidean distance dlThe maximum Euclidean distance d is selectedmax
Step e-2, determining the maximum Euclidean distance dmaxComparing the maximum Euclidean distance D with a threshold value D, and if the maximum Euclidean distance D exceeds the threshold valuemaxAdding the corresponding candidate frames into the selected frames, taking the h audio frames after the selected frames as new candidate frames, and if not, discarding all the h candidate frames and continuously using the subsequent h audio frames as candidate frames for selection;
c, repeating the step d and the step e, sequentially selecting h audio frames from the rest audio frames as candidate frames, and selecting candidate frames meeting the conditions from the h candidate frames to be added into the selected frames; and if the number of the remaining audio frames is less than h, taking the actually remaining audio frames as candidate frames, selecting the candidate frames meeting the conditions from the candidate frames by using the same method, adding the candidate frames into the selected frames until the number of the remaining audio frames is zero, and taking the feature vectors of all the selected frames as the feature vectors of the audio data.
CN201510782814.XA 2015-11-16 2015-11-16 A kind of audio feature extraction methods becoming frame per second Active CN105448290B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510782814.XA CN105448290B (en) 2015-11-16 2015-11-16 A kind of audio feature extraction methods becoming frame per second

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510782814.XA CN105448290B (en) 2015-11-16 2015-11-16 A kind of audio feature extraction methods becoming frame per second

Publications (2)

Publication Number Publication Date
CN105448290A true CN105448290A (en) 2016-03-30
CN105448290B CN105448290B (en) 2019-03-01

Family

ID=55558397

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510782814.XA Active CN105448290B (en) 2015-11-16 2015-11-16 A kind of audio feature extraction methods becoming frame per second

Country Status (1)

Country Link
CN (1) CN105448290B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108922556A (en) * 2018-07-16 2018-11-30 百度在线网络技术(北京)有限公司 sound processing method, device and equipment
CN109643555A (en) * 2016-07-04 2019-04-16 哈曼贝克自动系统股份有限公司 Automatically correct the loudness level in the audio signal comprising voice signal
CN114567780A (en) * 2022-02-25 2022-05-31 杭州当虹科技股份有限公司 Video frame alignment method
CN117097909A (en) * 2023-10-20 2023-11-21 深圳市星易美科技有限公司 Distributed household audio and video processing method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1505572A1 (en) * 2002-05-06 2005-02-09 Prous Science S.A. Voice recognition method
CN102332262A (en) * 2011-09-23 2012-01-25 哈尔滨工业大学深圳研究生院 Method for intelligently identifying songs based on audio features
US20130166294A1 (en) * 1999-12-10 2013-06-27 At&T Intellectual Property Ii, L.P. Frame Erasure Concealment Technique for a Bitstream-Based Feature Extractor
WO2015092711A1 (en) * 2013-12-18 2015-06-25 Isis Innovation Ltd. Method and apparatus for automatic speech recognition
CN105046699A (en) * 2015-07-09 2015-11-11 硅革科技(北京)有限公司 Motion video superposition contrast method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130166294A1 (en) * 1999-12-10 2013-06-27 At&T Intellectual Property Ii, L.P. Frame Erasure Concealment Technique for a Bitstream-Based Feature Extractor
EP1505572A1 (en) * 2002-05-06 2005-02-09 Prous Science S.A. Voice recognition method
CN102332262A (en) * 2011-09-23 2012-01-25 哈尔滨工业大学深圳研究生院 Method for intelligently identifying songs based on audio features
WO2015092711A1 (en) * 2013-12-18 2015-06-25 Isis Innovation Ltd. Method and apparatus for automatic speech recognition
CN105046699A (en) * 2015-07-09 2015-11-11 硅革科技(北京)有限公司 Motion video superposition contrast method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
严焰 等: "基于HMM的手势识别研究", 《华中师范大学学报(自然科学版)》 *
杨涛 等: "基于量子粒子群优化算法的运动捕获数据关键帧提取", 《计算机应用研究》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109643555A (en) * 2016-07-04 2019-04-16 哈曼贝克自动系统股份有限公司 Automatically correct the loudness level in the audio signal comprising voice signal
CN109643555B (en) * 2016-07-04 2024-01-30 哈曼贝克自动系统股份有限公司 Automatic correction of loudness level in an audio signal containing a speech signal
CN108922556A (en) * 2018-07-16 2018-11-30 百度在线网络技术(北京)有限公司 sound processing method, device and equipment
CN108922556B (en) * 2018-07-16 2019-08-27 百度在线网络技术(北京)有限公司 Sound processing method, device and equipment
CN114567780A (en) * 2022-02-25 2022-05-31 杭州当虹科技股份有限公司 Video frame alignment method
CN114567780B (en) * 2022-02-25 2024-08-23 杭州当虹科技股份有限公司 Video frame alignment method
CN117097909A (en) * 2023-10-20 2023-11-21 深圳市星易美科技有限公司 Distributed household audio and video processing method and system
CN117097909B (en) * 2023-10-20 2024-02-02 深圳市星易美科技有限公司 Distributed household audio and video processing method and system

Also Published As

Publication number Publication date
CN105448290B (en) 2019-03-01

Similar Documents

Publication Publication Date Title
WO2020173133A1 (en) Training method of emotion recognition model, emotion recognition method, device, apparatus, and storage medium
CN107480143B (en) Method and system for segmenting conversation topics based on context correlation
EP3309782B1 (en) Method, device and system for noise suppression
CN106297776B (en) A kind of voice keyword retrieval method based on audio template
CN106653056B (en) Fundamental frequency extraction model and training method based on LSTM recurrent neural network
CN107193797B (en) Hot topic detection and trend prediction method for Chinese microblog
CN103049501B (en) Based on mutual information and the Chinese domain term recognition method of conditional random field models
CN105448290B (en) A kind of audio feature extraction methods becoming frame per second
CN104599292B (en) A kind of anti-noise moving object detection algorithm decomposed based on low-rank matrix
CN103559196B (en) Video retrieval method based on multi-core canonical correlation analysis
CN101819638B (en) Establishment method of pornographic detection model and pornographic detection method
CN104536983A (en) Method and device for predicting advertisement click rate
CN110147760B (en) Novel efficient electric energy quality disturbance image feature extraction and identification method
CN103943107A (en) Audio/video keyword identification method based on decision-making level fusion
CN107123432A (en) A kind of Self Matching Top N audio events recognize channel self-adapted method
CN112767922B (en) Speech recognition method for contrast predictive coding self-supervision structure joint training
CN112488133A (en) Video/picture-text cross-modal retrieval method
CN105138983A (en) Pedestrian detection method based on weighted part model and selective search segmentation
CN110956222A (en) Method for detecting network for underwater target detection
CN111724766B (en) Language identification method, related equipment and readable storage medium
Pang et al. Discrete Cosine Transformation and Temporal Adjacent Convolutional Neural Network‐Based Remaining Useful Life Estimation of Bearings
CN104867493B (en) Multifractal Dimension end-point detecting method based on wavelet transformation
CN108564967B (en) Mel energy voiceprint feature extraction method for crying detection system
CN103336765A (en) Markov matrix off-line correction method of text keywords
CN107894837A (en) Dynamic sentiment analysis model sample processing method and processing device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant