CN104113789A - On-line video abstraction generation method based on depth learning - Google Patents

On-line video abstraction generation method based on depth learning Download PDF

Info

Publication number
CN104113789A
CN104113789A CN201410326406.9A CN201410326406A CN104113789A CN 104113789 A CN104113789 A CN 104113789A CN 201410326406 A CN201410326406 A CN 201410326406A CN 104113789 A CN104113789 A CN 104113789A
Authority
CN
China
Prior art keywords
video
frame piece
frame
degree
dictionary
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410326406.9A
Other languages
Chinese (zh)
Other versions
CN104113789B (en
Inventor
李平
俞俊
李黎
徐向华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hangzhou Huicui Intelligent Technology Co ltd
Original Assignee
Hangzhou Dianzi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hangzhou Dianzi University filed Critical Hangzhou Dianzi University
Priority to CN201410326406.9A priority Critical patent/CN104113789B/en
Publication of CN104113789A publication Critical patent/CN104113789A/en
Application granted granted Critical
Publication of CN104113789B publication Critical patent/CN104113789B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Compression Or Coding Systems Of Tv Signals (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an on-line video abstraction generation method based on depth learning. An original video is subjected to the following operation: 1) cutting the video uniformly into a group of small frame blocks, extracting statistical characteristics of each frame image and forming corresponding vectorization expressions; 2) pre-training video frame multilayer depth network and obtaining the nonlinearity expression of each frame; 3) selecting the front m frame blocks being as an initial concise video, and carrying out reconstruction on the concise video through a group sparse coding algorithm to obtain an initial dictionary and reconstruction coefficients; 4) updating depth network parameters according to the next frame block, carrying out reconstruction and reconstruction error calculation on the frame block, and adding the frame block to the concise video and updating the dictionary if the error is larger than a set threshold; and 5) processing new frame blocks till the end in sequence on line according to the step 4), and the updated concise video being generated video abstraction. With the method, latent high-level semantic information of the video can be excavated deeply, the video abstraction can be generated quickly, time of users is saved, and visual experience is improved.

Description

A kind of video frequency abstract based on degree of depth study generates method online
Technical field
The invention belongs to the technical field that video frequency abstract generates, particularly the video frequency abstract based on degree of depth study generates method online.
Background technology
In recent years, day by day universal along with portable sets such as digital camera, smart mobile phone, palmtop PCs, the quantity of all kinds of videos is blowout formula and increases.For example, intelligent transportation, safety monitoring, public security deploy troops on garrison duty etc. the video capture device of social key areas in a medium-sized city up to Ji Wan road, the video data that these equipment produce reaches PB level.For lock onto target personage or vehicle, the personnel such as traffic police need to expend a large amount of time and retrieve for examination the video flowing that tedium is monitored, and this has greatly affected work efficiency, is unfavorable for the establishment of safe city.Therefore, effectively choose the frame of video that comprises key message from tediously long video flowing, i.e. video summarization technique, has been subject to the extensive concern of academia and industrial quarters.
Traditional video summarization technique is mainly for edited structuring video, and as a film can be divided into multiple scenes, multiple plots that each scene is occurred by same place form, and each plot is made up of a series of smooth continuous frame of video again.Be different from the structuring videos such as traditional film, TV play, news report, monitor video is generally the destructuring video without montage, and this application that is video summarization technique brings larger challenge.
At present, main video frequency abstract field has based on key frame method, creates new images, frame of video piece, turns the technology such as natural language processing.Method based on key frame comprises the strategies such as plot rim detection, frame of video cluster, color histogram, action stability; Create some successive frames generations that new images utilization comprises important content, the method is easily subject to the fuzzy factors impact between different frame; Frame of video block method utilizes the technology such as scene rim detection, dialog analysis in structuring video to carry out cutting to original, forms short and small subject movies; Turn natural language processing and refer to that the captions that utilize in video and voice messaging are converted into video frequency abstract the technology of text snippet, this technology is not suitable for processing the monitor video without captions or sound.
Produce continuously a large amount of destructuring videos for intelligent transportation, the security protection key areas such as deploy to ensure effective monitoring and control of illegal activities, traditional video summarization method can not meet the application requirements of online processing stream-type video.For this reason, in the urgent need to processing online video flowing, can choose the video summarization method that comprises key content by efficiently and accurately again.
Summary of the invention
Concentrate and simplify tedious video flowing for online efficiently and accurately, to save the visual effect of user time augmented video content, the present invention proposes a kind of video frequency abstract based on degree of depth study and generate online method, the method comprises the following steps:
1, obtain after original video data, carry out following operation:
1) be one group little frame piece by even video cutting, each frame piece comprises multiframe, extracts the statistical nature of each two field picture, forms corresponding vectorization and represents;
2) pre-training video frame multilayer degree of depth network, obtains the non-linear expression of each frame;
3) choose front m frame piece for initially simplifying video, and by group sparse coding algorithm, it is reconstructed, obtain initial dictionary and reconstruction coefficients;
4) upgrade degree of depth network parameter according to next frame piece, this frame piece is reconstructed and calculates reconstructed error simultaneously, if error is greater than setting threshold, this frame piece is added and simplifies in video and upgrade dictionary;
5) according to step 4) process new frame piece until finish online successively, the video frequency abstract of simplifying video and be generation of renewal.
Further, described step 1) described in the statistical nature of the each two field picture of extraction form corresponding vectorization and represent, specifically:
1) establish original video and be evenly divided into n frame piece, each frame piece comprises t two field picture (as t=80), and each two field picture is scaled to unified pixel size and keeps original vertical-horizontal proportion;
2) extract each two field picture global characteristics and the yardstick invariant features conversion (SIFT:Scale-Invariant Feature Transform) such as color histogram, color moment, edge orientation histogram, Gabor wavelet transformation, local binary patterns, accelerate the local features such as robust features (SURF:Speeded Up Robust Feature);
3) order connects the above-mentioned characteristics of image of each frame, and formation dimension is n fvectorization represent.
Further, described step 2) in pre-training video frame multilayer degree of depth network obtain the non-linear expression of each frame, specifically:
Utilize stacking denoising own coding device (SDA:Stacked Denoising Autoencoder) to train in advance multilayer degree of depth network (number of plies is less than 10);
A, at every one deck, each two field picture is proceeded as follows: first, be that the approach such as arbitrary value generate each frame noise image by adding less Gaussian noise, establishing input variable at random; Then, noise image is shone upon and is obtained its non-linear expression by own coding device (AE:Auto Encoder);
B, utilize stochastic gradient descent algorithm to adjust renewal to each layer parameter of degree of depth network;
Described step 3) in the group sparse coding algorithm that passes through be reconstructed initially simplifying video, specifically:
1) initially simplify video and form (m is less than 50 positive integer) by front m frame piece of original video, total n init=m × t two field picture, X kcorresponding k primitive frame piece; Obtaining corresponding non-linear table by pre-training degree of depth network is shown y kthe non-linear expression of corresponding k frame piece;
2) establish initial dictionary D by n dindividual atom composition, d jcorresponding j atom; If reconstruction coefficients is C, the corresponding frame number of its element number, the atom number of the corresponding dictionary of its dimension, c kcorresponding k frame piece coefficient, corresponding i two field picture;
3) utilize the group sparse coding target function of multiplier alternating direction implicit Optimal Regularization dictionary, can obtain respectively initial dictionary D and reconstruction coefficients C, solve
Wherein, symbol || || 2represent the l of variable 2normal form, regularization parameter λ is greater than 0 real number, function of many variables F (Y k, C k, D) be embodied as:
F ( Y k , C k , D ) = 1 2 n f Σ y i ∈ Y k , d j ∈ D | | y i - Σ j = 1 n d c j i d j | | 2 2 + γ Σ j = 1 n d | | c j | | 2 ,
Wherein, parameter γ is greater than 0 real number, symbol in mathematical expression subrepresentation use dictionary D to be reconstructed i two field picture.The multiplier alternating direction implicit is here specially: first preset parameter D, makes above-mentioned target function become the convex function for parameters C; Then preset parameter C, makes above-mentioned target function become the convex function for parameter D, and iteration is alternately upgraded two parameters.
Described step 4) in upgrade degree of depth network parameter this frame piece is reconstructed and calculates reconstructed error according to next frame piece, specifically:
1) each two field picture of this frame piece is done as follows successively:
A. utilize the parameter of last one deck in online gradient descent algorithm renewal degree of depth neural net, i.e. weights W and side-play amount b;
B. utilize Back Propagation Algorithm to upgrade the parameter of other layers in degree of depth neural net;
2) upgrade the non-linear expression of each two field picture according to new parameter;
3), based on existing dictionary D, utilization group sparse coding is reconstructed and error of calculation ε present frame piece, to present frame piece X knon-linear expression Y kbe reconstructed, concrete steps are: first minimize function of many variables F (Y k, C k, D) and obtain optimum reconstruction coefficients then bring into section 1 in and calculate its value and be current reconstructed error ε.
Described step 4) if in error be greater than setting threshold and present frame piece added and simplifies in video and upgrade dictionary, specifically:
1) if to present frame piece X knon-linear expression Y kthe reconstructed error ε calculating is greater than setting threshold θ (getting empirical value), present frame piece is added and is simplified in video,
2) if the current video of simplifying in contain q frame piece, the non-linear expression set of two field picture of upgrading dictionary is use so upgrade dictionary D and solve target function
Wherein, parameter lambda is to be greater than 0 real number, for regulating the impact of regularization term.
The video frequency abstract the present invention proposes based on degree of depth study generates method online, its advantage is: utilize degree of depth study to excavate the high-level semantics features in video, make to organize the degree that sparse coding can better reflect dictionary reconstruct current video frame piece, thereby the frame of video piece of tool amount of information forms the video frequency abstract that comprises region-of-interest and key person's event; The video frequency abstract of simplifying is that user has saved a large amount of time, has strengthened the visual experience of key content simultaneously.
Brief description of the drawings
Fig. 1 is method flow diagram of the present invention.
Embodiment
With reference to accompanying drawing 1, further illustrate the present invention:
1, obtain after original video data, carry out following operation:
1) be one group little frame piece by even video cutting, each frame piece comprises multiframe, extracts the statistical nature of each two field picture, forms corresponding vectorization and represents;
2) pre-training video frame multilayer degree of depth network, obtains the non-linear expression of each frame;
3) choose front m frame piece for initially simplifying video, and by group sparse coding algorithm, it is reconstructed, obtain initial dictionary and reconstruction coefficients;
4) upgrade degree of depth network parameter according to next frame piece, this frame piece is reconstructed and calculates reconstructed error simultaneously, if error is greater than setting threshold, this frame piece is added and simplifies in video and upgrade dictionary;
5) according to step 4) process new frame piece until finish online successively, the video frequency abstract of simplifying video and be generation of renewal.
Step 1) described in the statistical nature of the each two field picture of extraction form corresponding vectorization and represent, specifically:
1) establish original video and be evenly divided into n frame piece, each frame piece comprises t two field picture (as t=80), and each two field picture is scaled to unified pixel size and keeps original vertical-horizontal proportion;
2) extract each two field picture global characteristics and the yardstick invariant features conversion (SIFT:Scale-Invariant Feature Transform) such as color histogram, color moment, edge orientation histogram, Gabor wavelet transformation, local binary patterns, accelerate the local features such as robust features (SURF:Speeded Up Robust Feature);
3) order connects the above-mentioned characteristics of image of each frame, and the vectorization that formation dimension is nf represents.
Step 2) in pre-training video frame multilayer degree of depth network obtain the non-linear expression of each frame, specifically:
Utilize stacking denoising own coding device (SDA:Stacked Denoising Autoencoder) to train in advance multilayer degree of depth network (number of plies is less than 10);
A, at every one deck, each two field picture is proceeded as follows: first, be that the approach such as arbitrary value generate each frame noise image by adding less Gaussian noise, establishing input variable at random; Then, noise image is shone upon and is obtained its non-linear expression by own coding device (AE:Auto Encoder);
B, utilize stochastic gradient descent algorithm to adjust renewal to each layer parameter of degree of depth network;
Step 3) in the group sparse coding algorithm that passes through be reconstructed initially simplifying video, specifically:
1) initially simplify video and form (m is less than 50 positive integer) by front m frame piece of original video, total n init=m × t two field picture, X kcorresponding k primitive frame piece; Obtaining corresponding non-linear table by pre-training degree of depth network is shown y kthe non-linear expression of corresponding k frame piece;
2) establish initial dictionary D by n dindividual atom composition, d jcorresponding j atom; If reconstruction coefficients is C, the corresponding frame number of its element number, the atom number of the corresponding dictionary of its dimension, c kcorresponding k frame piece coefficient, corresponding i two field picture;
3) utilize the group sparse coding target function of multiplier alternating direction implicit Optimal Regularization dictionary, can obtain respectively initial dictionary D and reconstruction coefficients C, solve
Wherein, symbol || || 2represent the l of variable 2normal form, regularization parameter λ is greater than 0 real number, function of many variables F (Y k, C k, D) be embodied as:
F ( Y k , C k , D ) = 1 2 n f Σ y i ∈ Y k , d j ∈ D | | y i - Σ j = 1 n d c j i d j | | 2 2 + γ Σ j = 1 n d | | c j | | 2 ,
Wherein, parameter γ is greater than 0 real number, symbol in mathematical expression subrepresentation use dictionary D to be reconstructed i two field picture.The multiplier alternating direction implicit is here specially: first preset parameter D, makes above-mentioned target function become the convex function for parameters C; Then preset parameter C, makes above-mentioned target function become the convex function for parameter D, and iteration is alternately upgraded two parameters.
Step 4) in upgrade degree of depth network parameter this frame piece is reconstructed and calculates reconstructed error according to next frame piece, specifically:
1) each two field picture of this frame piece is done as follows successively:
A. utilize the parameter of last one deck in online gradient descent algorithm renewal degree of depth neural net, i.e. weights W and side-play amount b;
B. utilize Back Propagation Algorithm to upgrade the parameter of other layers in degree of depth neural net;
2) upgrade the non-linear expression of each two field picture according to new parameter;
3), based on existing dictionary D, utilization group sparse coding is reconstructed and error of calculation ε present frame piece, to present frame piece X knon-linear expression Y kbe reconstructed, concrete steps are: first minimize function of many variables F (Y k, C k, D) and obtain optimum reconstruction coefficients then bring into section 1 in and calculate its value and be current reconstructed error ε.
Step 4) if in error be greater than setting threshold and present frame piece added and simplifies in video and upgrade dictionary, specifically:
1) if to present frame piece X knon-linear expression Y kthe reconstructed error ε calculating is greater than setting threshold θ (getting empirical value), present frame piece is added and is simplified in video,
2) if the current video of simplifying in contain q frame piece, the non-linear expression set of two field picture of upgrading dictionary is use so upgrade dictionary D and solve target function
Wherein, parameter lambda is to be greater than 0 real number, for regulating the impact of regularization term.

Claims (6)

1. the video frequency abstract based on degree of depth study generates a method online, the method is characterized in that and obtains after original video, proceeds as follows:
1) be one group little frame piece by even video cutting, each frame piece comprises multiframe, extracts the statistical nature of each two field picture, forms corresponding vectorization and represents;
2) pre-training video frame multilayer degree of depth network, obtains the non-linear expression of each frame;
3) choose front m frame piece for initially simplifying video, and by group sparse coding algorithm, it is reconstructed, obtain initial dictionary and reconstruction coefficients;
4) upgrade degree of depth network parameter according to next frame piece, this frame piece is reconstructed and calculates reconstructed error simultaneously, if error is greater than setting threshold, this frame piece is added and simplifies in video and upgrade dictionary;
5) according to step 4) process new frame piece until finish online successively, the video frequency abstract of simplifying video and be generation of renewal.
2. the video frequency abstract based on degree of depth study as claimed in claim 1 generates method online, it is characterized in that: described step 1) in the statistical nature of the each two field picture of extraction form corresponding vectorization and represent, concrete steps are:
1.1) establish original video and be evenly divided into n frame piece, each frame piece comprises t two field picture, and each two field picture is scaled to unified pixel size and keeps original vertical-horizontal proportion;
1.2) extract global characteristics and the local feature of each two field picture;
Described global characteristics comprises color histogram, color moment, edge orientation histogram, Gabor wavelet transformation, local binary patterns;
Described local feature comprises: yardstick invariant features conversion SIFT, acceleration robust features SURF;
1.3) order connects the above-mentioned characteristics of image of each frame, and formation dimension is n fvectorization represent.
3. the video frequency abstract based on degree of depth study as claimed in claim 1 generates method online, it is characterized in that: described step 2) in pre-training video frame multilayer degree of depth network obtain the non-linear expression of each frame, specifically utilize stacking denoising own coding device SDA to train in advance multilayer degree of depth network, comprising:
A, at every one deck, each two field picture is proceeded as follows: first, be that arbitrary value generates each frame noise image by adding Gaussian noise or establishing at random input variable; Then, noise image is shone upon and is obtained its non-linear expression by own coding device AE;
B, utilize stochastic gradient descent algorithm to adjust renewal to each layer parameter of degree of depth network.
4. the video frequency abstract based on degree of depth study as claimed in claim 1 generates method online, it is characterized in that: described step 3) in the group sparse coding algorithm that passes through be reconstructed initially simplifying video, concrete steps are:
3.1) initially simplify video and formed by front m frame piece of original video, total n init=m × t two field picture, X kcorresponding k primitive frame piece; Obtaining corresponding non-linear table by pre-training degree of depth network is shown y kthe non-linear expression of corresponding k frame piece;
3.2) establish initial dictionary D by n dindividual atom composition, d jcorresponding j atom; If reconstruction coefficients is C, the corresponding frame number of its element number, the atom number of the corresponding dictionary of its dimension, c kcorresponding k frame piece coefficient, corresponding i two field picture;
3.3) utilize the group sparse coding target function of multiplier alternating direction implicit Optimal Regularization dictionary, can obtain respectively initial dictionary D and reconstruction coefficients C, solve
Wherein, symbol || || 2represent the l of variable 2normal form, regularization parameter λ is greater than 0 real number, function of many variables F (Y k, C k, D) be embodied as:
F ( Y k , C k , D ) = 1 2 n f Σ y i ∈ Y k , d j ∈ D | | y i - Σ j = 1 n d c j i d j | | 2 2 + γ Σ j = 1 n d | | c j | | 2 ,
Wherein, parameter γ is greater than 0 real number, symbol in mathematical expression subrepresentation use dictionary D to be reconstructed i two field picture; The multiplier alternating direction implicit is here specially: first preset parameter D, makes above-mentioned target function become the convex function for parameters C; Then preset parameter C, makes above-mentioned target function become the convex function for parameter D, and iteration is alternately upgraded two parameters.
5. the video frequency abstract based on degree of depth study as claimed in claim 1 generates method online, it is characterized in that: described step 4) in upgrade degree of depth network parameter and this frame piece be reconstructed and calculate reconstructed error according to next frame piece, concrete steps are:
4.1) each two field picture of this frame piece is done as follows successively:
4.1.1) utilize online gradient descent algorithm to upgrade the parameter of last one deck in degree of depth neural net, i.e. weights W and side-play amount b;
4.1.2) utilize Back Propagation Algorithm to upgrade the parameter of other layers in degree of depth neural net;
4.2) upgrade the non-linear expression of each two field picture according to new parameter;
4.3), based on existing dictionary D, utilization group sparse coding is reconstructed and error of calculation ε present frame piece, to present frame piece X knon-linear expression Y kbe reconstructed, be specially: first minimize function of many variables F (Y k, C k, D) and obtain optimum reconstruction coefficients then bring into section 1 in and calculate its value and be current reconstructed error ε.
6. the video frequency abstract based on degree of depth study as claimed in claim 1 generates method online, it is characterized in that: described step 4) if in error be greater than setting threshold and present frame piece added and simplifies in video and upgrade dictionary, specifically:
1) if to present frame piece X knon-linear expression Y kthe reconstructed error ε calculating is greater than setting threshold θ, present frame piece is added and is simplified in video,
2) if the current video of simplifying in contain q frame piece, the non-linear expression set of two field picture of upgrading dictionary is use so upgrade dictionary D and solve target function
Wherein, parameter lambda is to be greater than 0 real number, for regulating the impact of regularization term.
CN201410326406.9A 2014-07-10 2014-07-10 On-line video abstraction generation method based on depth learning Active CN104113789B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410326406.9A CN104113789B (en) 2014-07-10 2014-07-10 On-line video abstraction generation method based on depth learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410326406.9A CN104113789B (en) 2014-07-10 2014-07-10 On-line video abstraction generation method based on depth learning

Publications (2)

Publication Number Publication Date
CN104113789A true CN104113789A (en) 2014-10-22
CN104113789B CN104113789B (en) 2017-04-12

Family

ID=51710398

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410326406.9A Active CN104113789B (en) 2014-07-10 2014-07-10 On-line video abstraction generation method based on depth learning

Country Status (1)

Country Link
CN (1) CN104113789B (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104778659A (en) * 2015-04-15 2015-07-15 杭州电子科技大学 Single-frame image super-resolution reconstruction method on basis of deep learning
CN105279495A (en) * 2015-10-23 2016-01-27 天津大学 Video description method based on deep learning and text summarization
CN105930314A (en) * 2016-04-14 2016-09-07 清华大学 Text summarization generation system and method based on coding-decoding deep neural networks
CN106331433A (en) * 2016-08-25 2017-01-11 上海交通大学 Video denoising method based on deep recursive neural network
CN106502985A (en) * 2016-10-20 2017-03-15 清华大学 A kind of neural network modeling approach and device for generating title
CN106686403A (en) * 2016-12-07 2017-05-17 腾讯科技(深圳)有限公司 Video preview generation method, device, server and system
CN106778571A (en) * 2016-12-05 2017-05-31 天津大学 A kind of digital video feature extracting method based on deep neural network
CN107209941A (en) * 2014-12-30 2017-09-26 诺基亚技术有限公司 Mobile object detection in video
CN107679031A (en) * 2017-09-04 2018-02-09 昆明理工大学 Based on the advertisement blog article recognition methods for stacking the self-editing ink recorder of noise reduction
CN107729821A (en) * 2017-09-27 2018-02-23 浙江大学 A kind of video summarization method based on one-dimensional sequence study
CN107886109A (en) * 2017-10-13 2018-04-06 天津大学 It is a kind of based on have supervision Video segmentation video summarization method
CN107911755A (en) * 2017-11-10 2018-04-13 天津大学 A kind of more video summarization methods based on sparse self-encoding encoder
CN108388942A (en) * 2018-02-27 2018-08-10 四川云淞源科技有限公司 Information intelligent processing method based on big data
CN108417204A (en) * 2018-02-27 2018-08-17 四川云淞源科技有限公司 Information security processing method based on big data
CN108417206A (en) * 2018-02-27 2018-08-17 四川云淞源科技有限公司 High speed information processing method based on big data
CN108848422A (en) * 2018-04-19 2018-11-20 清华大学 A kind of video abstraction generating method based on target detection
CN109360436A (en) * 2018-11-02 2019-02-19 Oppo广东移动通信有限公司 A kind of video generation method, terminal and storage medium
CN109635777A (en) * 2018-12-24 2019-04-16 广东理致技术有限公司 A kind of video data editing recognition methods and device
CN109803067A (en) * 2017-11-16 2019-05-24 富士通株式会社 Video concentration method, video enrichment facility and electronic equipment
CN109905778A (en) * 2019-01-03 2019-06-18 上海大学 The method of the expansible breviary of single unstructured video based on group sparse coding
CN110110646A (en) * 2019-04-30 2019-08-09 浙江理工大学 A kind of images of gestures extraction method of key frame based on deep learning
CN110225368A (en) * 2019-06-27 2019-09-10 腾讯科技(深圳)有限公司 A kind of video locating method, device and electronic equipment
WO2019169996A1 (en) * 2018-03-05 2019-09-12 腾讯科技(深圳)有限公司 Video processing method and apparatus, video retrieval method and apparatus, storage medium and server
CN110366050A (en) * 2018-04-10 2019-10-22 北京搜狗科技发展有限公司 Processing method, device, electronic equipment and the storage medium of video data
CN110446067A (en) * 2019-08-30 2019-11-12 杭州电子科技大学 Video concentration method based on tensor resolution
CN111046887A (en) * 2018-10-15 2020-04-21 华北电力大学(保定) Method for extracting characteristics of image with noise
CN111246246A (en) * 2018-11-28 2020-06-05 华为技术有限公司 Video playing method and device
CN111563423A (en) * 2020-04-17 2020-08-21 西北工业大学 Unmanned aerial vehicle image target detection method and system based on depth denoising automatic encoder
CN106993240B (en) * 2017-03-14 2020-10-16 天津大学 Multi-video abstraction method based on sparse coding
CN113626641A (en) * 2021-08-11 2021-11-09 南开大学 Method for generating video abstract based on multi-mode data and aesthetic principle through neural network
US11295084B2 (en) 2019-09-16 2022-04-05 International Business Machines Corporation Cognitively generating information from videos
CN117725148A (en) * 2024-02-07 2024-03-19 湖南三湘银行股份有限公司 Question-answer word library updating method based on self-learning

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102930518A (en) * 2012-06-13 2013-02-13 上海汇纳网络信息科技有限公司 Improved sparse representation based image super-resolution method
CN103118220A (en) * 2012-11-16 2013-05-22 佳都新太科技股份有限公司 Keyframe pick-up algorithm based on multi-dimensional feature vectors
CN103167284A (en) * 2011-12-19 2013-06-19 中国电信股份有限公司 Video streaming transmission method and system based on picture super-resolution
CN103295242A (en) * 2013-06-18 2013-09-11 南京信息工程大学 Multi-feature united sparse represented target tracking method
CN103413125A (en) * 2013-08-26 2013-11-27 中国科学院自动化研究所 Horror video identification method based on discriminant instance selection and multi-instance learning
CN103531199A (en) * 2013-10-11 2014-01-22 福州大学 Ecological sound identification method on basis of rapid sparse decomposition and deep learning
CN103761531A (en) * 2014-01-20 2014-04-30 西安理工大学 Sparse-coding license plate character recognition method based on shape and contour features

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103167284A (en) * 2011-12-19 2013-06-19 中国电信股份有限公司 Video streaming transmission method and system based on picture super-resolution
CN102930518A (en) * 2012-06-13 2013-02-13 上海汇纳网络信息科技有限公司 Improved sparse representation based image super-resolution method
CN103118220A (en) * 2012-11-16 2013-05-22 佳都新太科技股份有限公司 Keyframe pick-up algorithm based on multi-dimensional feature vectors
CN103295242A (en) * 2013-06-18 2013-09-11 南京信息工程大学 Multi-feature united sparse represented target tracking method
CN103413125A (en) * 2013-08-26 2013-11-27 中国科学院自动化研究所 Horror video identification method based on discriminant instance selection and multi-instance learning
CN103531199A (en) * 2013-10-11 2014-01-22 福州大学 Ecological sound identification method on basis of rapid sparse decomposition and deep learning
CN103761531A (en) * 2014-01-20 2014-04-30 西安理工大学 Sparse-coding license plate character recognition method based on shape and contour features

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
尚丽: "基于稀疏编码的自然特征提起及去噪", 《系统仿真学报》 *

Cited By (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107209941A (en) * 2014-12-30 2017-09-26 诺基亚技术有限公司 Mobile object detection in video
CN104778659A (en) * 2015-04-15 2015-07-15 杭州电子科技大学 Single-frame image super-resolution reconstruction method on basis of deep learning
CN105279495A (en) * 2015-10-23 2016-01-27 天津大学 Video description method based on deep learning and text summarization
CN105279495B (en) * 2015-10-23 2019-06-04 天津大学 A kind of video presentation method summarized based on deep learning and text
CN105930314A (en) * 2016-04-14 2016-09-07 清华大学 Text summarization generation system and method based on coding-decoding deep neural networks
CN105930314B (en) * 2016-04-14 2019-02-05 清华大学 System and method is generated based on coding-decoding deep neural network text snippet
CN106331433B (en) * 2016-08-25 2020-04-24 上海交通大学 Video denoising method based on deep recurrent neural network
CN106331433A (en) * 2016-08-25 2017-01-11 上海交通大学 Video denoising method based on deep recursive neural network
CN106502985A (en) * 2016-10-20 2017-03-15 清华大学 A kind of neural network modeling approach and device for generating title
CN106778571B (en) * 2016-12-05 2020-03-27 天津大学 Digital video feature extraction method based on deep neural network
CN106778571A (en) * 2016-12-05 2017-05-31 天津大学 A kind of digital video feature extracting method based on deep neural network
CN106686403A (en) * 2016-12-07 2017-05-17 腾讯科技(深圳)有限公司 Video preview generation method, device, server and system
CN106686403B (en) * 2016-12-07 2019-03-08 腾讯科技(深圳)有限公司 A kind of video preview drawing generating method, device, server and system
CN106993240B (en) * 2017-03-14 2020-10-16 天津大学 Multi-video abstraction method based on sparse coding
CN107679031A (en) * 2017-09-04 2018-02-09 昆明理工大学 Based on the advertisement blog article recognition methods for stacking the self-editing ink recorder of noise reduction
CN107679031B (en) * 2017-09-04 2021-01-05 昆明理工大学 Advertisement and blog identification method based on stacking noise reduction self-coding machine
CN107729821A (en) * 2017-09-27 2018-02-23 浙江大学 A kind of video summarization method based on one-dimensional sequence study
CN107729821B (en) * 2017-09-27 2020-08-11 浙江大学 Video summarization method based on one-dimensional sequence learning
CN107886109A (en) * 2017-10-13 2018-04-06 天津大学 It is a kind of based on have supervision Video segmentation video summarization method
CN107886109B (en) * 2017-10-13 2021-06-25 天津大学 Video abstraction method based on supervised video segmentation
CN107911755B (en) * 2017-11-10 2020-10-20 天津大学 Multi-video abstraction method based on sparse self-encoder
CN107911755A (en) * 2017-11-10 2018-04-13 天津大学 A kind of more video summarization methods based on sparse self-encoding encoder
CN109803067A (en) * 2017-11-16 2019-05-24 富士通株式会社 Video concentration method, video enrichment facility and electronic equipment
CN108388942A (en) * 2018-02-27 2018-08-10 四川云淞源科技有限公司 Information intelligent processing method based on big data
CN108417206A (en) * 2018-02-27 2018-08-17 四川云淞源科技有限公司 High speed information processing method based on big data
CN108417204A (en) * 2018-02-27 2018-08-17 四川云淞源科技有限公司 Information security processing method based on big data
US11368705B2 (en) 2018-03-05 2022-06-21 Tencent Technology (Shenzhen) Company Limited Video feature extraction and video content understanding method, apparatus, storage medium and server
US11934454B2 (en) 2018-03-05 2024-03-19 Tencent Technology (Shenzhen) Company Limited Video processing method and apparatus, video retrieval method and apparatus, storage medium, and server
WO2019169996A1 (en) * 2018-03-05 2019-09-12 腾讯科技(深圳)有限公司 Video processing method and apparatus, video retrieval method and apparatus, storage medium and server
CN110366050A (en) * 2018-04-10 2019-10-22 北京搜狗科技发展有限公司 Processing method, device, electronic equipment and the storage medium of video data
CN108848422B (en) * 2018-04-19 2020-06-02 清华大学 Video abstract generation method based on target detection
CN108848422A (en) * 2018-04-19 2018-11-20 清华大学 A kind of video abstraction generating method based on target detection
CN111046887A (en) * 2018-10-15 2020-04-21 华北电力大学(保定) Method for extracting characteristics of image with noise
CN109360436A (en) * 2018-11-02 2019-02-19 Oppo广东移动通信有限公司 A kind of video generation method, terminal and storage medium
CN111246246A (en) * 2018-11-28 2020-06-05 华为技术有限公司 Video playing method and device
CN109635777B (en) * 2018-12-24 2022-09-13 广东理致技术有限公司 Video data editing and identifying method and device
CN109635777A (en) * 2018-12-24 2019-04-16 广东理致技术有限公司 A kind of video data editing recognition methods and device
CN109905778A (en) * 2019-01-03 2019-06-18 上海大学 The method of the expansible breviary of single unstructured video based on group sparse coding
CN109905778B (en) * 2019-01-03 2021-12-03 上海大学 Method for scalable compression of single unstructured video based on group sparse coding
CN110110646B (en) * 2019-04-30 2021-05-04 浙江理工大学 Gesture image key frame extraction method based on deep learning
CN110110646A (en) * 2019-04-30 2019-08-09 浙江理工大学 A kind of images of gestures extraction method of key frame based on deep learning
CN110225368A (en) * 2019-06-27 2019-09-10 腾讯科技(深圳)有限公司 A kind of video locating method, device and electronic equipment
CN110446067B (en) * 2019-08-30 2021-11-02 杭州电子科技大学 Tensor decomposition-based video concentration method
CN110446067A (en) * 2019-08-30 2019-11-12 杭州电子科技大学 Video concentration method based on tensor resolution
US11295084B2 (en) 2019-09-16 2022-04-05 International Business Machines Corporation Cognitively generating information from videos
CN111563423A (en) * 2020-04-17 2020-08-21 西北工业大学 Unmanned aerial vehicle image target detection method and system based on depth denoising automatic encoder
CN113626641A (en) * 2021-08-11 2021-11-09 南开大学 Method for generating video abstract based on multi-mode data and aesthetic principle through neural network
CN113626641B (en) * 2021-08-11 2023-09-01 南开大学 Method for generating video abstract based on neural network of multi-modal data and aesthetic principle
CN117725148A (en) * 2024-02-07 2024-03-19 湖南三湘银行股份有限公司 Question-answer word library updating method based on self-learning

Also Published As

Publication number Publication date
CN104113789B (en) 2017-04-12

Similar Documents

Publication Publication Date Title
CN104113789B (en) On-line video abstraction generation method based on depth learning
Ahmed et al. Image splicing detection using mask-RCNN
Kim et al. Median filtered image restoration and anti-forensics using adversarial networks
Thai et al. Image classification using support vector machine and artificial neural network
CN106530200B (en) Steganographic image detection method and system based on deep learning model
CN102393900B (en) Video copying detection method based on robust hash
CN101971190A (en) Real-time body segmentation system
CN102682298B (en) Video fingerprint method based on graph modeling
CN111597983B (en) Method for realizing identification of generated false face image based on deep convolutional neural network
CN111382305B (en) Video deduplication method, video deduplication device, computer equipment and storage medium
CN108564166A (en) Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link
Wang et al. Wavelet based region duplication forgery detection
CN106529492A (en) Video topic classification and description method based on multi-image fusion in view of network query
CN107092935A (en) A kind of assets alteration detection method
CN103606136A (en) Video super resolution method based on keyframes and non-local restriction
CN103530634A (en) Face characteristic extraction method
Sasirekha et al. Enhanced techniques for PDF image segmentation and text extraction
CN110807369B (en) Short video content intelligent classification method based on deep learning and attention mechanism
CN102547477B (en) Video fingerprint method based on contourlet transformation model
Wang et al. Versatile Backdoor Attack with Visible, Semantic, Sample-Specific, and Compatible Triggers
Zuo et al. A novel slant transform-based image feature extract algorithm
CN104751470A (en) Image quick-matching method
CN108205666A (en) A kind of face identification method based on depth converging network
CN102034102B (en) Image-based significant object extraction method as well as complementary significance graph learning method and system
Bashir et al. Towards deep learning-based image steganalysis: practices and open research issues

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20220809

Address after: Room 406, building 19, haichuangyuan, No. 998, Wenyi West Road, Yuhang District, Hangzhou City, Zhejiang Province

Patentee after: HANGZHOU HUICUI INTELLIGENT TECHNOLOGY CO.,LTD.

Address before: 310018 No. 2 street, Xiasha Higher Education Zone, Hangzhou, Zhejiang

Patentee before: HANGZHOU DIANZI University