CN103077203A - Method for detecting repetitive audio/video clips - Google Patents
Method for detecting repetitive audio/video clips Download PDFInfo
- Publication number
- CN103077203A CN103077203A CN2012105802599A CN201210580259A CN103077203A CN 103077203 A CN103077203 A CN 103077203A CN 2012105802599 A CN2012105802599 A CN 2012105802599A CN 201210580259 A CN201210580259 A CN 201210580259A CN 103077203 A CN103077203 A CN 103077203A
- Authority
- CN
- China
- Prior art keywords
- video
- fragment
- audio frequency
- repeated
- coupling
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Landscapes
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
The invention discloses a method for detecting repetitive audio/video clips. The method comprises the following steps of: A, segmenting a video streaming in a to-be-detected audio/video clip to obtain video streaming clips, and extracting video features from each video streaming clip; B, detecting a video sequence of audio/video clips with the repeatability in the to-be-detected audio/video clip to obtain the repetitive audio/video clips, wherein the audio/video clips with the repeatability are respectively matched with the video streaming clips; C, accurately positioning respective starting time points and end time points of the repetitive audio/video clips by utilizing fine-grit video matching for the repetitive audio/video clips; and D, merging matching pairs by using a sequence-based method for the repetitive audio/video clips accurately positioned in the step C. According to the method for detecting the repetitive audio/video clips, specific audio/video clips in an audio/video can be accurately and efficiently detected.
Description
Technical field
The present invention relates to the detection technique field of video video segment, particularly relate to a kind of detection method of repeated audio frequency and video fragment.
Background technology
The automatic detection of audio frequency and video is exactly to utilize computing machine to look, automatically detect specific audio frequency and video fragment the sound stream and accurately locate the position of this specific audio frequency and video fragment from one section.At present the method for the automatic detection of common audio frequency and video be based on the method for rule, based on the method for sign, based on the method for identification etc.Rule-based method is that some rules are formulated in artificial going for the Some features of audio frequency and video.But the defective of art methods is to select to represent that the feature of audio frequency and video is not enough stablizing sometimes, so be difficult to set up unified detection system by these features.
Method based on sign, detect the audio frequency and video fragment such as the station symbol by TV station, but, at present a lot of TV stations do not conceal station symbol when intercutting specific audio frequency and video fragment (such as advertisement), and this phenomenon is more and more, so this method that detects the audio frequency and video fragment by station symbol had just lost efficacy.Based on the method for identification, then the database that the method requires to have in advance a very large very complete storage advertisement utilizes this database identification to be embedded in the audio frequency and video fragment of TV programme the inside, and the method can not detect non-existent audio frequency and video fragment in the database.Above-mentioned existing detection method nearly all is to realize the detection of audio frequency and video by video features simultaneously.Because video itself, these method desired data amounts are large, and the feature complicacy is high, and computing velocity is slow.
Summary of the invention
The object of the present invention is to provide a kind of detection method and system of repeated audio frequency and video fragment, it can detect the specific audio frequency and video fragment in the audio frequency and video more accurately and efficiently.
For realizing the detection method of a kind of repeated audio frequency and video fragment that the object of the invention provides, comprise the following steps:
Steps A is cut apart the video flowing in the audio frequency and video fragment to be detected and is obtained the video flowing fragment, and extracts video features from each video flowing fragment;
Step B, the video sequence of the audio frequency and video fragment with repeatability of detection and video flowing fragment match obtains repeated audio frequency and video fragment in the audio frequency and video fragment.
Step C to described repeated audio frequency and video fragment, utilizes fine-grained video matching accurately to locate the starting and ending time point of repeated audio frequency and video fragment;
Step D, to the repeated audio frequency and video fragment of location accurately among the step C, use method based on sequence to coupling to merging, obtain complete repeated audio frequency and video fragment.
Described steps A comprises the following steps:
Steps A 1 is divided into a plurality of video flowing fragments with the video flowing in the audio frequency and video fragment to be detected, and each video flowing fragment is for carrying out mark take a time unit as a unit;
Steps A 2 is extracted the video features parameter from the video flowing fragment.
Described video features parameter comprises the Mel-cepstrum coefficient, a parameter or more than one parameter combinations in zero-crossing rate and the short-time energy.
Described extraction Mel-cepstrum coefficient, zero-crossing rate and short-time energy comprise the following steps:
The video data volume that gathers take 40ms is as a frame, and adjacent video frames does not repeat, and extracts 12 Mel-cepstrum coefficients, zero-crossing rate and short-time energies totally 14 parameters, consists of the 14 frame characteristic parameters of tieing up.
The described video sequence that detects in audio/video flow with the audio frequency and video fragment with repeatability of video flowing fragment match is by utilizing Euclidean distance to realize as the distance measure of coarseness similarity matching.
The described Euclidean distance of utilizing comprises the following steps: as the distance measure of coarseness similarity matching
Utilize Euclidean distance to find the audio frequency and video small fragment with repeatability of all and its coupling as the distance measure of coarseness similarity matching, and stipulate that the part that repeats in two small fragments just is the coupling fragment sequence greater than half.
Among the described step D, coupling to merging, is comprised the following steps:
The repeated audio frequency and video fragment of the every a pair of exact matching that matches for detection, it is right less than all couplings of pre-set threshold value TT to search respectively with their time interval, new-found coupling to the coupling of being connected to connecting, and then the duplicate detection matching process, until new coupling right start and end time that the coupling that does not satisfy condition to existing, obtains is exactly the starting and ending time of a complete repeated audio frequency and video fragment.
The invention has the beneficial effects as follows: the detection method of repeated audio frequency and video fragment of the present invention, utilize repeated audio frequency and video fragment) repeatability be the most stable feature of repeated audio frequency and video fragment, more stable than its its feature and rule, so accuracy rate can be higher.The present invention adopts video features to detect repeated audio frequency and video fragment in addition, only just be enough to represent, distinguish repeated audio frequency and video fragment and normal program with video information, and video information have still less lower distinguishing feature and calculated amount still less and the more efficient computing velocity of data volume, complicacy than video information.
Description of drawings
Fig. 1 is the detection method process flow diagram of the present invention's repeatability audio frequency and video fragment;
Embodiment
Introduce in detail the detection method of the present invention's repeatability audio frequency and video fragment below in conjunction with above-mentioned target, comprise the following steps:
Step S100 is cut apart the video flowing in the audio frequency and video fragment to be detected and is obtained the video flowing fragment, and extracts video features from each video flowing fragment;
Step S110 is divided into a plurality of video flowing fragments with the video flowing in the audio frequency and video fragment to be detected, and each video flowing fragment is for carrying out mark take a time unit as a unit.
Video flowing in the audio frequency and video fragment to be detected being extracted, this section video flowing is divided into one by one small video clips, is the important preparation process that video detects.
Here so-called cut apart be not one large section video flowing real be divided into one by one n(such as n=5) second small fragment, but artificial make a mark every n fragment second, then process take n video flowing second as the video flowing of a unit.Its fundamental purpose is convenient extraction feature, more efficiently carries out various processing.
The present invention, is cut apart the video flowing in the audio frequency and video fragment to be detected as the elementary cell of divided video stream not have overlapping length as 10 seconds video flowing fragment.
Selecting 10 seconds is a detecting unit, be because the length of repeated audio frequency and video fragment generally all greater than 10 seconds, whether the repeated audio frequency and video fragment of repeated matching is not arranged so just do not spend in detecting these 10 seconds, and can also guarantee to find the repeated audio frequency and video fragment sequence of all repetitions.
Step S120 extracts the video features parameter from the video flowing fragment.
The video features parameter attribute extracts and refers to seek the raw video signal expression-form, and extraction can represent the data of original signal.
The video data volume that the present invention gathers take 40ms is as a frame, adjacent video frames do not repeat, be that frame moves also be 40ms, extract 12 Mel-cepstrum coefficients (Mel Frequency Cepstrum Coeficient, MFCC), zero-crossing rate and short-time energy totally 14 parameters, consist of the frame characteristic parameters of 14 dimensions.
1) extracts Mel-cepstrum coefficient (Mel-Frequency Cepstral Coefficients, MFCC).
The MFCC feature is the feature that often adopts in speech recognition and the Speaker Identification, and it is to utilize the triangular filter group that the filtering of Fourier transform energy coefficient is got, and its frequency domain is carried out Mei Er (Mel) change of scale, more meets human auditory properties.
2) extract zero-crossing rate (Zero-Crossing Rate, ZCR)
Zero-crossing rate refers to that signal value is by the number of times of null value in the unit interval.It has illustrated the average signal frequency to a certain extent.When discrete two adjacent samplings of time signal had different symbols, " zero passage " phenomenon just appearred.
Wherein, sgn[.] the is-symbol function, x (m) is the sampled value of vision signal.
3) extract short-time energy (Short Time Energy)
The energy spectrometer of vision signal is based on the vision signal energy sizable variation in time. and short-time average energy has illustrated the intensity of vision signal.
Wherein, x (m) is the sampled value of vision signal.
Step S200 utilizes Euclidean distance as the distance measure of coarseness similarity matching, and the video sequence of the audio frequency and video fragment with repeatability of detection and video flowing fragment match tentatively obtains repeated audio frequency and video fragment in audio/video flow;
It is exactly the starting and ending position of locating rapidly and accurately repeated audio frequency and video fragment in a large amount of audio/video flows that repeatability audio frequency and video fragment detects the most key stage.
In order to realize above-mentioned requirements, the present invention adopts video information to detect potential matching sequence.Selecting video information, is because video has the lower distinguishing feature of still less data volume, complicacy and calculated amount still less and more efficient computing velocity than video.And can reach the effect that video information detects.
Utilization cut apart good length be 10 seconds video flowing fragment as a probe, all radio and television stream of search in audio/video flow comprises TV feed and other TV feed at this 10 seconds places.
Preferably, utilize Euclidean distance to find the audio frequency and video small fragment with repeatability of all and its coupling as the distance measure of coarseness similarity matching, and stipulate that the part that repeats in two small fragments just is the coupling fragment sequence greater than half.
Utilize Euclidean distance as the distance measure of coarseness similarity matching, the method for employing is to calculate frame level Euclidean distance D1 every 10 frames, and such advantage is that calculated amount is to calculate frame by frame 1/10th of Euclidean distance, and almost not loss of degree of accuracy.
Wherein N is the frame characteristic parameter of 14 dimensions.
Obtain behind the distance matrix by comparing with pre-set matching threshold TD, if exist 7 or above frame number (being that length was greater than 2.10 seconds) less than pre-set matching threshold TD, just think that this audio frequency and video fragment is the video sequence of audio frequency and video fragment with repeatability with the video flowing fragment match, repeated audio frequency and video fragment similar, coupling is arranged in audio/video flow, tentatively obtain this repeatability audio frequency and video fragment, it is right namely to mate.
Step S300 to the repeated audio frequency and video fragment that tentatively obtains among the step S200, utilizes fine-grained video matching accurately to locate the starting and ending time point of repeated audio frequency and video fragment;
The coarseness similarity matching methods also has the limitation of two aspects: 1) because above-mentioned coupling is rough coupling, there are deviation to a certain extent in coupling border and real repeated audio frequency and video segment boundaries; 2) can only detect greater than half matching sequence of fragment length, be invalid for length less than half matching sequence.
The coupling that detects for the distance measure that utilizes Euclidean distance as the coarseness similarity matching is right, utilizes the Euclidean distance D2 of improvement type to redeterminate the starting and ending time point of repeated audio frequency and video fragment as the distance measure of fine granularity similarity matching.
Wherein N is the frame characteristic parameter of 14 dimensions.
Right for every a pair of 10 seconds couplings that find, they are labeled as A
2And B
2Then calculate respectively A
2B
2, A
2B
1, A
1B
2, A
1B
1, A
2B
3, A
3B
2, A
3B
3Improvement type Euclidean distance frame by frame obtains behind the distance matrix writing down respectively the position less than the point of threshold value TD by comparing with pre-set matching threshold TD.So just can accurately locate the time of beginning and the end of repeated audio frequency and video fragment.
Step S400, to the repeated audio frequency and video fragment of location accurately among the step S300, use method based on sequence to coupling to merging, obtain complete repeated audio frequency and video fragment.
Owing to being that audio frequency and video are divided into 10 seconds is a unit, so when the length of repeated audio frequency and video fragment during greater than 10 seconds, a complete repeated audio frequency and video fragment will be by excessive cutting apart.Therefore to the repeated audio frequency and video fragment of an over-segmentation be merged.
Utilization of the present invention merges a complete repeated audio frequency and video fragment based on the method for sequence.
10 seconds repeated audio frequency and video fragments of the every a pair of exact matching that matches for detection, it is right less than all couplings of pre-set threshold value TT to search respectively with their time interval, new-found coupling to the coupling of being connected to connecting, and then the duplicate detection matching process, until new coupling right start and end time that the coupling that does not satisfy condition to existing, obtains is exactly the starting and ending time of a complete repeated audio frequency and video fragment.
Below by being that advertisement in 10 minutes the broadcast TV program detects as an example to a segment length, introduce in detail the implementation process of the detection method of repeated broadcast TV program fragment of the present invention.
Whole process is divided into four-stage substantially: video flowing cut apart extraction with video features; The right detection of video sequence coupling with repeatability; Utilize fine-grained video matching to come the starting and ending time point of accurate positioning advertising; Use based on the method for sequence to coupling to merging.
Video flowing cut apart extraction stage with video features, this stage is divided into 10 minutes broadcast TV program fragment do not have 120 overlapping length to be 10 seconds small fragments, then respectively 120 small fragments are carried out feature extraction, the feature of extracting comprises: 12 MFCC, zero-crossing rate, short-time energy, the frame length that adopts is 40ms, frame moves and is 40ms, forms the proper vector of 14 dimensions.
Each small fragment has the proper vector of 125 14 dimensions.
For example this segment length is that 2 different advertisements are arranged in 10 minutes the TV programme: new * * *, the Shandong * * * *.Wherein new * * * position that occurs 2 times is respectively 10-25 second (the 3rd, 4,5 section), 123-138 second (the 25th, 26,27,28 section); The Shandong * * * * position that occurs 2 times is respectively 30-50 second (the 7th, 8,9,10 section), 155-175 second (the 32nd, 33,34,35 section).
Have the right detection-phase of video sequence coupling of repeatability, utilize above-mentioned proper vector to calculate 125 small fragment Euclidean distance D1 between any two, if the identical length of certain two small fragment was greater than half of fragment length, namely 2.5 seconds.These two small fragments of mark mate.
Utilize the proper vector of above-mentioned 125 small fragments to pass through formula
Wherein N is the frame characteristic parameter of 14 dimensions.
Calculate Euclidean distance between any two, if the identical length of certain two small fragment was greater than half of fragment length, namely 2.10 seconds.These two small fragments of mark mate.Then (3,26), (4,27), (5,28) are similar; (7,32), (8,33), (9,34), (10,35) are similar.
Utilize fine-grained video matching to locate accurately the starting and ending time point stage of repeated broadcast TV program fragment, to the above-mentioned coupling small fragment that finds pair, calculate the Euclidean distance of its improvement type, the start and end time point of these two small fragment couplings of accurate mark.
Wherein N is the frame characteristic parameter of 14 dimensions, calculates between any two and (3,25) (2,26) (1,25) (5,29) (6,28) (6,29), (7,31) (6,32) Euclidean distance of (6,31) (10,36) (11,35) (11,36) improvement type.The 0-2 that can obtain at last the 3rd section is similar second with the 25th section 3-10 second, and the 3rd section 2-10 is similar second with the 26th section 0-3 second, and the 4th section 0-2 is similar second with the 26th section 3-10 second, the 4th section 2-10 is similar second with the 27th section 0-3 second, and the 5th section 0-2 is similar second with the 27th section 3-10 second, and the 5th section 2-10 is similar second with the 28th section 0-3 second, and (7,32), (8,33), (9,34), (10,35) then are fully similar.
Use based on the method for sequence to coupling to carrying out merging phase, by the above-mentioned right sequence number of exact matching small fragment that finds, it is merged.
10 seconds fragments for the every a pair of exact matching that finds above, merge according to the method based on sequence, namely obtain the 3rd section and began in 0 second to finish by 10 seconds of the 5th section and the 25th section the end in the 3rd second that began to the 28th section in the 4th second, length be respectively 110 seconds be exactly that 10-210 second and 123-138 mate second.And (7,32), (8,33), (9,34), (10,35) they are to mate then to be 30-50 second and 155-1710 mates second fully.
In sum, the detection method of repeated audio frequency and video fragment of the present invention is by the extraction with video features cut apart of the video flowing in the renaturation audio frequency and video fragment; The right detection of video sequence coupling with repeatability; Utilize fine-grained video matching to locate accurately the starting and ending time point of repeated audio frequency and video fragment; Use based on the method for sequence to coupling to merging.So the method is the application of a kind of video content analysis and retrieval, it utilizes the video features in the audio frequency and video, detect automatically the repeated audio frequency and video fragment that repeats in the audio frequency and video, and can accurately locate the position of the repeated audio frequency and video fragment that all repeat with mark.
Claims (7)
1. the detection method of a repeated audio frequency and video fragment is characterized in that, comprises the following steps:
Steps A is cut apart the video flowing in the audio frequency and video fragment to be detected and is obtained the video flowing fragment, and extracts video features from each video flowing fragment;
Step B, the video sequence of the audio frequency and video fragment with repeatability of detection and video flowing fragment match obtains repeated audio frequency and video fragment in the audio frequency and video fragment.
Step C to described repeated audio frequency and video fragment, utilizes fine-grained video matching accurately to locate the starting and ending time point of repeated audio frequency and video fragment;
Step D, to the repeated audio frequency and video fragment of location accurately among the step C, use method based on sequence to coupling to merging, obtain complete repeated audio frequency and video fragment.
2. the detection method of repeated audio frequency and video fragment according to claim 1 is characterized in that, described steps A comprises the following steps:
Steps A 1 is divided into a plurality of video flowing fragments with the video flowing in the audio frequency and video fragment to be detected, and each video flowing fragment is for carrying out mark take a time unit as a unit;
Steps A 2 is extracted the video features parameter from the video flowing fragment.
3. the detection method of repeated audio frequency and video fragment according to claim 2 is characterized in that, described video features parameter comprises the Mel-cepstrum coefficient, a parameter or more than one parameter combinations in zero-crossing rate and the short-time energy.
4. the detection method of repeated audio frequency and video fragment according to claim 3 is characterized in that, described extraction Mel-cepstrum coefficient, and zero-crossing rate and short-time energy comprise the following steps:
The video data volume that gathers take 40ms is as a frame, and adjacent video frames does not repeat, and extracts 12 Mel-cepstrum coefficients, zero-crossing rate and short-time energies totally 14 parameters, consists of the 14 frame characteristic parameters of tieing up.
5. the detection method of repeated audio frequency and video fragment according to claim 1 and 2, it is characterized in that, the described video sequence that detects in audio/video flow with the audio frequency and video fragment with repeatability of video flowing fragment match is by utilizing Euclidean distance to realize as the distance measure of coarseness similarity matching.
6. the detection method of repeated audio frequency and video fragment according to claim 5 is characterized in that, the described Euclidean distance of utilizing comprises the following steps: as the distance measure of coarseness similarity matching
Utilize Euclidean distance to find the audio frequency and video small fragment with repeatability of all and its coupling as the distance measure of coarseness similarity matching, and stipulate that the part that repeats in two small fragments just is the coupling fragment sequence greater than half.
7. the detection method of repeated audio frequency and video fragment according to claim 1 is characterized in that, among the described step D, coupling to merging, is comprised the following steps:
The repeated audio frequency and video fragment of the every a pair of exact matching that matches for detection, it is right less than all couplings of pre-set threshold value TT to search respectively with their time interval, new-found coupling to the coupling of being connected to connecting, and then the duplicate detection matching process, until new coupling right start and end time that the coupling that does not satisfy condition to existing, obtains is exactly the starting and ending time of a complete repeated audio frequency and video fragment.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2012105802599A CN103077203A (en) | 2012-12-28 | 2012-12-28 | Method for detecting repetitive audio/video clips |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2012105802599A CN103077203A (en) | 2012-12-28 | 2012-12-28 | Method for detecting repetitive audio/video clips |
Publications (1)
Publication Number | Publication Date |
---|---|
CN103077203A true CN103077203A (en) | 2013-05-01 |
Family
ID=48153733
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2012105802599A Pending CN103077203A (en) | 2012-12-28 | 2012-12-28 | Method for detecting repetitive audio/video clips |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103077203A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103617233A (en) * | 2013-11-26 | 2014-03-05 | 烟台中科网络技术研究所 | Method and device for detecting repeated video based on semantic content multilayer expression |
CN107180056A (en) * | 2016-03-11 | 2017-09-19 | 阿里巴巴集团控股有限公司 | The matching process and device of fragment in video |
CN111246244A (en) * | 2020-02-04 | 2020-06-05 | 北京贝思科技术有限公司 | Method and device for rapidly analyzing and processing audio and video in cluster and electronic equipment |
CN111370022A (en) * | 2019-12-25 | 2020-07-03 | 厦门快商通科技股份有限公司 | Audio advertisement detection method and device, electronic equipment and medium |
CN116634226A (en) * | 2023-06-15 | 2023-08-22 | 北京柏睿数据技术股份有限公司 | Method and system for intelligent real-time processing of video stream data |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060221232A1 (en) * | 2003-04-30 | 2006-10-05 | Xuwen Yu | Equipment and method for detecting commercial automately |
CN101159834A (en) * | 2007-10-25 | 2008-04-09 | 中国科学院计算技术研究所 | Method and system for detecting repeatable video and audio program fragment |
CN101241552A (en) * | 2008-01-24 | 2008-08-13 | 北京六维世纪网络技术有限公司 | Image characteristic recognition method and device |
-
2012
- 2012-12-28 CN CN2012105802599A patent/CN103077203A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060221232A1 (en) * | 2003-04-30 | 2006-10-05 | Xuwen Yu | Equipment and method for detecting commercial automately |
CN101159834A (en) * | 2007-10-25 | 2008-04-09 | 中国科学院计算技术研究所 | Method and system for detecting repeatable video and audio program fragment |
CN101241552A (en) * | 2008-01-24 | 2008-08-13 | 北京六维世纪网络技术有限公司 | Image characteristic recognition method and device |
Non-Patent Citations (1)
Title |
---|
吴潇等: "基于多特征匹配的视频拷贝检测算法", 《计算机辅助设计与图形学学报》 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103617233A (en) * | 2013-11-26 | 2014-03-05 | 烟台中科网络技术研究所 | Method and device for detecting repeated video based on semantic content multilayer expression |
CN107180056A (en) * | 2016-03-11 | 2017-09-19 | 阿里巴巴集团控股有限公司 | The matching process and device of fragment in video |
CN107180056B (en) * | 2016-03-11 | 2020-11-06 | 阿里巴巴集团控股有限公司 | Method and device for matching segments in video |
CN111370022A (en) * | 2019-12-25 | 2020-07-03 | 厦门快商通科技股份有限公司 | Audio advertisement detection method and device, electronic equipment and medium |
CN111246244A (en) * | 2020-02-04 | 2020-06-05 | 北京贝思科技术有限公司 | Method and device for rapidly analyzing and processing audio and video in cluster and electronic equipment |
CN116634226A (en) * | 2023-06-15 | 2023-08-22 | 北京柏睿数据技术股份有限公司 | Method and system for intelligent real-time processing of video stream data |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN101159834B (en) | Method and system for detecting repeatable video and audio program fragment | |
CN100580693C (en) | Advertisement detecting and recognizing method and system | |
CN107393554B (en) | Feature extraction method for fusion inter-class standard deviation in sound scene classification | |
CN110012349B (en) | A kind of news program structural method end to end | |
CN103077203A (en) | Method for detecting repetitive audio/video clips | |
CN101247470A (en) | Method for detecting scene boundaries in genre independent videos | |
CN102799605A (en) | Method and system for monitoring advertisement broadcast | |
CN102855317B (en) | A kind of multi-mode indexing means and system based on demonstration video | |
CN105701470A (en) | Analog circuit fault characteristic extraction method based on optimal wavelet packet decomposition | |
CN102073636A (en) | Program climax search method and system | |
CN104409080A (en) | Voice end node detection method and device | |
CN102436483A (en) | Video advertisement detecting method based on explicit type sharing subspace | |
CN103207901B (en) | A kind of method and apparatus that IP address ownership place is obtained based on search engine | |
Johnson et al. | Spoken Document Retrieval for TREC-8 at Cambridge University. | |
CN104221079A (en) | Modified Mel filter bank structure using spectral characteristics for sound analysis | |
CN111370022A (en) | Audio advertisement detection method and device, electronic equipment and medium | |
Li et al. | Multi-level attention model with deep scattering spectrum for acoustic scene classification | |
CN113992944A (en) | Video cataloging method, device, equipment, system and medium | |
Kim et al. | Quick audio retrieval using multiple feature vectors | |
CN104167211A (en) | Multi-source scene sound abstracting method based on hierarchical event detection and context model | |
CN103294696A (en) | Audio and video content retrieval method and system | |
CN104731890B (en) | A kind of combination PLSA and AT audio event sorting technique | |
CN103065152A (en) | Identification method of digital clock in videos | |
CN105512272A (en) | System for comparing audio information and audio information comparison method | |
Zhao et al. | Fast commercial detection based on audio retrieval |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C02 | Deemed withdrawal of patent application after publication (patent law 2001) | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20130501 |