CN103559196A - Video retrieval method based on multi-core canonical correlation analysis - Google Patents
Video retrieval method based on multi-core canonical correlation analysis Download PDFInfo
- Publication number
- CN103559196A CN103559196A CN201310438216.1A CN201310438216A CN103559196A CN 103559196 A CN103559196 A CN 103559196A CN 201310438216 A CN201310438216 A CN 201310438216A CN 103559196 A CN103559196 A CN 103559196A
- Authority
- CN
- China
- Prior art keywords
- video
- matrix
- frame
- feature
- correlation analysis
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
- G06F16/786—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using motion, e.g. object motion or camera motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/732—Query formulation
- G06F16/7328—Query by example, e.g. a complete video frame or video sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
Abstract
Disclosed is a video retrieval method based on multi-core canonical correlation analysis. The method includes grasping text descriptions corresponding to the video on internet, and then operating on the video: firstly dividing the video according to whether a shot is mutated or not, extracting key frames, extracting vision features of the key frames and moving features of the shot to form video feature vectors, and extracting word-frequency features from the text descriptions of each video; then utilizing the method of the multi-core canonical correlation analysis to obtain mapping matrixes and low-dimensional representation of the video features and the word-frequency features, and allowing the mapping matrixes and the low-dimensional representation to have the maximum correlation in low-dimensional space; finally, when a user inputs key words to perform video retrieval, acquiring the low-dimensional representation of the word-frequency features of the key words according to the mapping matrixes of the word-frequency features, and returning video retrieval results sequentially from large to small of the degrees of cosine similarity with the low-dimensional representation of the video features. The method has the advantages that the correlation of video content and the retrieval key words is enhanced, and the accuracy of retrieval by the user is improved.
Description
Technical field
The present invention relates to the technical field of video frequency searching, particularly the video retrieval method based on multinuclear canonical correlation analysis.
Background technology
Along with the develop rapidly of computer network, multimedia technology mechanics of communication in recent years, people can upload, watch and download all kinds of video informations by internet.Internet has become huge video warehouse gradually, and how more fast and effeciently the required video information of retrieval user has become information retrieval hot issue day by day.
Traditional video retrieval method is text based, it using video tab information as key word and video form matching relationship one by one, by key word being carried out to the operations such as feature extraction, pre-service, carry out Cluster Classification afterwards.The method efficiency of this manual mark video information that places one's entire reliance upon is lower, to the descriptive power deficiency of video information and experience that need to be certain.Therefore, text based search method can not meet the growing demand of user.From 20th century the eighties, content-based video frequency searching is progressively paid close attention to by people, more becomes in recent years study hotspot.The method can under the artificial situation participating in, automatically be extracted video features, and be not only to depend on label information not having.Particularly, the method camera lens cut apart with key frame extraction after extract the visual signature of key frame and the motion feature of camera lens, and typing video frequency searching database.During user search, video is retrieved according to the words-frequency feature of user search keyword, and according to similarity order from high to low, result for retrieval is returned to user.Adopt content-based video frequency searching can more objective, specifically and all sidedly describe video information, reduce subjectivity and the limitation of textual description information, and greatly improve the degree of accuracy of retrieval.
In machine learning field, content-based video retrieval method is not still a lot, but also obtained certain achievement, as the good joint classification feature optimal algorithm of performance of classifying at high dimensional data, the method can be selected and the maximally related union feature of classifying, and reduces in addition input dimension descent algorithm and the method based on rough set of descriptor data are carried out to the semantic dimensionality reduction descriptor index method keeping etc..
Summary of the invention
In order to help user to arrive the video of search by quick-searching, to improve the Experience Degree of watching video, the present invention proposes a kind of video retrieval method based on multinuclear canonical correlation analysis, the method comprises the following steps:
1, from internet captures video, for each video and textual description thereof, carry out following operation:
1) according to camera lens, whether suddenly change video is carried out to cutting, extract its key frame, and the motion feature that extracts visual signature in key frame and camera lens forms video feature vector, for the textual description of each video, extract words-frequency feature;
2) utilize multinuclear Canonical Correlation Analysis, obtain respectively the mapping matrix of video features and words-frequency feature, thus the low-dimensional that obtains both correspondences represent, make them maximum in the correlativity of low-dimensional data space;
3) when user inputs keyword and carries out video frequency searching, the low-dimensional that obtains keyword words-frequency feature according to the mapping matrix of words-frequency feature represents, the descending video frequency searching result of returning successively of cosine similarity representing according to itself and video features low-dimensional.
Further, the processing video described in step 1) and the method for corresponding textual description thereof, specifically:
1) when video lens cuts, adopt two comparative approach, establishing Tb is the threshold value of detector lens sudden change, Ts is the threshold value of detector lens gradual change, detects the difference of consecutive frame, if difference is greater than Tb by difference metric method, belong to lens mutation, should carry out shot segmentation, if difference is less than Tb, be greater than Ts, likely belong to gradual shot, now need this frame and subsequent frame to compare, if frame difference is less than Ts, but between consecutive frame, difference cumulative sum is greater than Tb, means and really belongs to gradual shot;
2) while extracting key frame, first first the head and the tail frame of camera lens is set as to key frame, headed by cause, frame is conventionally in order to attract spectators to represent the theme of camera lens, tail frame wishes that spectators' aftertaste often represents with a kind of feature, after head and the tail frame is determined, also to choose the frame that is changed significantly as key frame, each frame that is about to non-key frame in camera lens compares with key frame successively, if differed greatly, using it as key frame, relatively go down successively until compared all non-key frames in camera lens, also to choose key frame according to the frame method of average afterwards, on certain assigned address, calculate the mean value of all frame pixel values, key frame is the frame that pixel value approaches mean value most,
3) during key frame feature extraction, the Visual Feature Retrieval Process color of still image, shape, texture, the variation of camera motion feature extraction camera motion, movement locus, moving target size, the video features that forms video represents;
4) videotext is described and carried out participle, statistics word frequency, forms its words-frequency feature and represents.
Further, step 2) the multinuclear Canonical Correlation Analysis described in, specifically:
1) training sample video sum is made as to n, with m dimensional vector X
ithe video features that represents i video, by the vectorial X of n m dimension
ibe merged into matrix X, represent video feature matrix, in like manner, with k dimensional vector Y
jthe words-frequency feature that represents the textual description that j video is corresponding, by the vectorial Y of n k dimension
jbe merged into matrix Y, represent words-frequency feature matrix;
2) utilize gaussian kernel function formula
calculate the Gram matrix K of X
x, K wherein
xfor the matrix of the capable n row of n, ‖ ‖
2represent vectorial I
2norm, real parameters σ represents the bandwidth of gaussian kernel, by different σ is set, can obtain one group of gaussian kernel with different nonlinear organizations, in like manner, to Y application 2-d polynomial kernel function formula
calculate the Gram matrix K of Y
y, wherein ()
trepresent vector or transpose of a matrix, parameters C is to be more than or equal to 0 kernel functional parameter, by different C is set, can obtain one group of polynomial kernel with different nonlinear organizations;
3) by the K carrying out after kernel function conversion
x, K
ybe normalized;
4) the objective function expression formula of the video retrieval method based on multinuclear canonical correlation analysis is:
,
Solve above-mentioned objective function expression equivalence in solving following generalized eigenvalue problem:
(K
x+κI)
-1K
y(K
y+κI)
-1K
xW
x=λ
2W
x,
As required, get (K
x+ κ I)
-1k
y(K
y+ κ I)
-1k
xfront p maximum eigenwert characteristic of correspondence vector forms the matrix of the capable p row of n
as K
xcorresponding mapping matrix, p represents to adopt the dimension after the video retrieval method dimensionality reduction of multinuclear canonical correlation analysis, and κ is arithmetic number parameter, ()
-1inverting of representing matrix, I representation unit matrix, K
ycorresponding mapping matrix
can basis
with
between relation obtain:
5) according to new mapping matrix
structure K
x, K
yexpression under new space:
The matrix that is the capable p row of n, now both correlativitys are maximum.
Further, the search matching method described in step 3), specifically:
1) search key of submitting to according to user is converted into keyword words-frequency feature vector y
q, to y
qapplication laplace kernel function formula
calculate y
qn dimension Gram vector
y
jfor j sample in training data Y, parameters C is to be more than or equal to 0 2-d polynomial kernel functional parameter;
2) according to mapping matrix
convert,
the low-dimensional that obtains keyword words-frequency feature represents, wherein
for p dimensional vector;
3) will
vector respectively with obtain before
every a line of matrix, each sample carries out the calculating of cosine similarity successively,
each sample be p dimensional vector, cosine similarity shows that more greatly similarity is higher;
4) according to cosine similarity order from high to low, return to results for video.
The present invention proposes the video retrieval method based on multinuclear canonical correlation analysis, its advantage is: video is effectively retrieved, can be improved the precision of video frequency searching; Be applicable to all types of videos, without backstage manual operation, also can be used for helping domestic consumer to improve web page browsing quality.
Accompanying drawing explanation
Fig. 1 is method flow diagram of the present invention.
Embodiment
With reference to accompanying drawing, further illustrate the present invention:
A video retrieval method based on multinuclear canonical correlation analysis, the method comprises the following steps:
1, from internet captures video, for each video and textual description thereof, carry out following operation:
1) according to camera lens, whether suddenly change video is carried out to cutting, extract its key frame, and the motion feature that extracts visual signature in key frame and camera lens forms video feature vector, for the textual description of each video, extract words-frequency feature;
2) utilize multinuclear Canonical Correlation Analysis, obtain respectively the mapping matrix of video features and words-frequency feature, thus the low-dimensional that obtains both correspondences represent, make them maximum in the correlativity of low-dimensional data space;
3) when user inputs keyword and carries out video frequency searching, the low-dimensional that obtains keyword words-frequency feature according to the mapping matrix of words-frequency feature represents, the descending video frequency searching result of returning successively of cosine similarity representing according to itself and video features low-dimensional.
Processing video described in step 1) and the method for correspondence-textual description thereof, specifically:
1) when video lens cuts, adopt two comparative approach, establishing Tb is the threshold value of detector lens sudden change, Ts is the threshold value of detector lens gradual change, detects the difference of consecutive frame, if difference is greater than Tb by difference metric method, belong to lens mutation, should carry out shot segmentation, if difference is less than Tb, be greater than Ts, likely belong to gradual shot, now need this frame and subsequent frame to compare, if frame difference is less than Ts, but between consecutive frame, difference cumulative sum is greater than Tb, means and really belongs to gradual shot;
2) while extracting key frame, first first the head and the tail frame of camera lens is set as to key frame, headed by cause, frame is conventionally in order to attract spectators to represent the theme of camera lens, tail frame wishes that spectators' aftertaste often represents with a kind of feature, after head and the tail frame is determined, also to choose the frame that is changed significantly as key frame, each frame that is about to non-key frame in camera lens compares with key frame successively, if differed greatly, using it as key frame, relatively go down successively until compared all non-key frames in camera lens, also to choose key frame according to the frame method of average afterwards, on certain assigned address, calculate the mean value of all frame pixel values, key frame is the frame that pixel value approaches mean value most,
3) during key frame feature extraction, the Visual Feature Retrieval Process color of still image, shape, texture, the variation of camera motion feature extraction camera motion, movement locus, moving target size, the video features that forms video represents;
4) videotext is described and carried out participle, statistics word frequency, forms its words-frequency feature and represents.
Step 2) the multinuclear Canonical Correlation Analysis described in, specifically:
1) training sample video sum is made as to n, with m dimensional vector X
ithe video features that represents i video, by the vectorial X of n m dimension
ibe merged into matrix X, represent video feature matrix, in like manner, with k dimensional vector Y
jthe words-frequency feature that represents the textual description that j video is corresponding, by the vectorial Y of n k dimension
jbe merged into matrix Y, represent words-frequency feature matrix;
2) utilize gaussian kernel function formula
calculate the Gram matrix K of X
x, K wherein
xfor the matrix of the capable n row of n, ‖ ‖
2represent vectorial I
2norm, real parameters σ represents the bandwidth of gaussian kernel, by different σ is set, can obtain one group of gaussian kernel with different nonlinear organizations, in like manner, to Y application 2-d polynomial kernel function formula
calculate the Gram matrix K of Y
y, wherein ()
trepresent vector or transpose of a matrix, parameters C is to be more than or equal to 0 kernel functional parameter, by different C is set, can obtain one group of polynomial kernel with different nonlinear organizations;
3) by the K carrying out after kernel function conversion
x, K
ybe normalized;
4) the objective function expression formula of the video retrieval method based on multinuclear canonical correlation analysis is:
,
Solve above-mentioned objective function expression equivalence in solving following generalized eigenvalue problem:
(K
x+κI)
-1K
y(K
y+κI)
-1K
xW
x=λ
2W
x,
As required, get (K
x+ κ I)
-1k
y(K
y+ κ I)
-1k
xfront p maximum eigenwert characteristic of correspondence vector forms the matrix of the capable p row of n
as
corresponding mapping matrix, p represents to adopt the dimension after the video retrieval method dimensionality reduction of multinuclear canonical correlation analysis, and κ is arithmetic number parameter, ()
-1inverting of representing matrix, I representation unit matrix, K
ycorresponding mapping matrix
can basis
with
between relation obtain:
5) according to new mapping matrix
structure K
x, K
yexpression under new space:
The matrix that is the capable p row of n, now both correlativitys are maximum.
Search matching method described in step 3), specifically:
1) search key of submitting to according to user is converted into keyword words-frequency feature vector y
q, to y
qapplication laplace kernel function formula
calculate y
qn dimension Gram vector
y
jfor j sample in training data Y, parameters C is to be more than or equal to 0 2-d polynomial kernel functional parameter;
2) according to mapping matrix
convert,
the low-dimensional that obtains keyword words-frequency feature represents, wherein
for p dimensional vector;
3) will
vector respectively with obtain before
every a line of matrix, each sample carries out the calculating of cosine similarity successively,
each sample be p dimensional vector, cosine similarity shows that more greatly similarity is higher;
4) according to cosine similarity order from high to low, return to results for video.
Content described in this instructions embodiment is only enumerating the way of realization of inventive concept; protection scope of the present invention should not be regarded as only limiting to the concrete form that embodiment states, protection scope of the present invention is also and in those skilled in the art, according to the present invention, conceive the equivalent technologies means that can expect.
Claims (4)
1. the video retrieval method based on multinuclear canonical correlation analysis, the method is characterized in that from internet captures video, for each video, carries out following operation:
1) according to camera lens, whether suddenly change video is carried out to cutting, extract its key frame, and the motion feature that extracts visual signature in key frame and camera lens forms video feature vector, for the textual description of each video, extract words-frequency feature;
2) utilize multinuclear Canonical Correlation Analysis, obtain respectively the mapping matrix of video features and words-frequency feature, thus the low-dimensional that obtains both correspondences represent, make them maximum in the correlativity of low-dimensional data space;
3) when user inputs keyword and carries out video frequency searching, the low-dimensional that obtains keyword words-frequency feature according to the mapping matrix of words-frequency feature represents, the descending video frequency searching result of returning successively of cosine similarity representing according to itself and video features low-dimensional.
2. the video retrieval method based on multinuclear canonical correlation analysis as claimed in claim 1, is characterized in that: the processing video described in described step 1) and the method for corresponding textual description thereof, specifically:
1) when video lens cuts, adopt two comparative approach, establishing Tb is the threshold value of detector lens sudden change, Ts is the threshold value of detector lens gradual change, detects the difference of consecutive frame, if difference is greater than Tb by difference metric method, belong to lens mutation, should carry out shot segmentation, if difference is less than Tb, be greater than Ts, likely belong to gradual shot, now need this frame and subsequent frame to compare, if frame difference is less than Ts, but between consecutive frame, difference cumulative sum is greater than Tb, means and really belongs to gradual shot;
2) while extracting key frame, first first the head and the tail frame of camera lens is set as to key frame, headed by cause, frame is conventionally in order to attract spectators to represent the theme of camera lens, tail frame wishes that spectators' aftertaste often represents with a kind of feature, after head and the tail frame is determined, also to choose the frame that is changed significantly as key frame, each frame that is about to non-key frame in camera lens compares with key frame successively, if differed greatly, using it as key frame, relatively go down successively until compared all non-key frames in camera lens, also to choose key frame according to the frame method of average afterwards, on certain assigned address, calculate the mean value of all frame pixel values, key frame is the frame that pixel value approaches mean value most,
3) during key frame feature extraction, the Visual Feature Retrieval Process color of still image, shape, texture, the variation of camera motion feature extraction camera motion, movement locus, moving target size, the video features that forms video represents;
4) videotext is described and carried out participle, statistics word frequency, forms its words-frequency feature and represents.
3. the video retrieval method based on multinuclear canonical correlation analysis as claimed in claim 2, is characterized in that: the multinuclear Canonical Correlation Analysis described step 2), specifically:
1) training sample video sum is made as to n, with m dimensional vector X
ithe video features that represents i video, by the vectorial X of n m dimension
ibe merged into matrix X, represent video feature matrix, in like manner, with k dimensional vector Y
jthe words-frequency feature that represents the textual description that j video is corresponding, by the vectorial Y of n k dimension
jbe merged into matrix Y, represent words-frequency feature matrix;
2) utilize gaussian kernel function formula
calculate the Gram matrix K of X
x, K wherein
xfor the matrix of the capable n row of n, ‖ ‖
2represent vectorial I
2norm, real parameters σ represents the bandwidth of gaussian kernel, by different σ is set, can obtain one group of gaussian kernel with different nonlinear organizations, in like manner, to Y application 2-d polynomial kernel function formula
calculate the Gram matrix K of Y
y, wherein ()
trepresent vector or transpose of a matrix, parameters C is to be more than or equal to 0 kernel functional parameter, by different C is set, can obtain one group of polynomial kernel with different nonlinear organizations;
3) by the K carrying out after kernel function conversion
x, K
ybe normalized;
4) the objective function expression formula of the video retrieval method based on multinuclear canonical correlation analysis is:
,
Solve above-mentioned objective function expression equivalence in solving following generalized eigenvalue problem:
(K
x+κI)
-1K
y(K
y+κ
1)
-1K
xW
x=λ
2W
x,
As required, get (K
x+ κ I)
-1k
y(K
y+ κ I)
-1k
xfront p maximum eigenwert characteristic of correspondence vector forms the matrix of the capable p row of n
as K
xcorresponding mapping matrix, p represents to adopt the dimension after the video retrieval method dimensionality reduction of multinuclear canonical correlation analysis, and κ is arithmetic number parameter, ()
-1inverting of representing matrix, I representation unit matrix, K
ycorresponding mapping matrix
can basis
with
between relation obtain:
4. the video retrieval method based on multinuclear canonical correlation analysis as claimed in claim 3, is characterized in that: the search matching method described in described step 3), specifically:
1) search key of submitting to according to user is converted into keyword words-frequency feature vector y
q, to y
qapplication laplace kernel function formula
calculate y
qn dimension Gram vector
y
jfor j sample in training data Y, parameters C is to be more than or equal to 0 2-d polynomial kernel functional parameter;
2) according to mapping matrix
convert,
the low-dimensional that obtains keyword words-frequency feature represents, wherein
for p dimensional vector;
3) will
vector respectively with obtain before
every a line of matrix, each sample carries out the calculating of cosine similarity successively,
each sample be p dimensional vector, cosine similarity shows that more greatly similarity is higher;
4) according to cosine similarity order from high to low, return to results for video.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310438216.1A CN103559196B (en) | 2013-09-23 | 2013-09-23 | Video retrieval method based on multi-core canonical correlation analysis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310438216.1A CN103559196B (en) | 2013-09-23 | 2013-09-23 | Video retrieval method based on multi-core canonical correlation analysis |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103559196A true CN103559196A (en) | 2014-02-05 |
CN103559196B CN103559196B (en) | 2017-02-22 |
Family
ID=50013443
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201310438216.1A Active CN103559196B (en) | 2013-09-23 | 2013-09-23 | Video retrieval method based on multi-core canonical correlation analysis |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103559196B (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104700410A (en) * | 2015-03-14 | 2015-06-10 | 西安电子科技大学 | Collaborative filtering-based teaching video labeling method |
CN104699844A (en) * | 2015-03-31 | 2015-06-10 | 北京奇艺世纪科技有限公司 | Method and device for determining video tags for advertisements |
CN105389326A (en) * | 2015-09-16 | 2016-03-09 | 中国科学院计算技术研究所 | Image annotation method based on weak matching probability canonical correlation model |
CN105938561A (en) * | 2016-04-13 | 2016-09-14 | 南京大学 | Canonical-correlation-analysis-based computer data attribute reduction method |
CN106294454A (en) * | 2015-05-29 | 2017-01-04 | 中兴通讯股份有限公司 | Video retrieval method and device |
CN106570196A (en) * | 2016-11-18 | 2017-04-19 | 广州视源电子科技股份有限公司 | Video program searching method and apparatus |
CN106570165A (en) * | 2016-11-07 | 2017-04-19 | 北京航空航天大学 | Content-based video retrieval method and apparatus |
CN106682108A (en) * | 2016-12-06 | 2017-05-17 | 浙江大学 | Video retrieval method based on multi-modal convolutional neural network |
CN106708929A (en) * | 2016-11-18 | 2017-05-24 | 广州视源电子科技股份有限公司 | Video program search method and device |
CN107122423A (en) * | 2017-04-06 | 2017-09-01 | 深圳Tcl数字技术有限公司 | Video display promotion method and device |
CN107748750A (en) * | 2017-08-30 | 2018-03-02 | 百度在线网络技术(北京)有限公司 | Similar video lookup method, device, equipment and storage medium |
CN108182684A (en) * | 2017-12-22 | 2018-06-19 | 河南师范大学 | A kind of image partition method and its device based on weighting kernel fuzzy cluster |
CN110222594A (en) * | 2019-05-20 | 2019-09-10 | 厦门能见易判信息科技有限公司 | Pirate video recognition methods and system |
CN110598049A (en) * | 2019-09-26 | 2019-12-20 | 北京字节跳动网络技术有限公司 | Method, apparatus, electronic device and computer readable medium for retrieving video |
CN110704655A (en) * | 2019-10-18 | 2020-01-17 | 中国科学技术大学 | Online multi-quantization image retrieval method |
CN110991470A (en) * | 2019-07-03 | 2020-04-10 | 北京市安全生产科学技术研究院 | Data dimension reduction method, portrait construction method and system and readable storage medium |
CN112182292A (en) * | 2020-09-30 | 2021-01-05 | 百度(中国)有限公司 | Training method and device for video retrieval model, electronic equipment and storage medium |
CN113221674A (en) * | 2021-04-25 | 2021-08-06 | 广东电网有限责任公司东莞供电局 | Video stream key frame extraction system and method based on rough set reduction and SIFT |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7143434B1 (en) * | 1998-11-06 | 2006-11-28 | Seungyup Paek | Video description system and method |
CN1851710A (en) * | 2006-05-25 | 2006-10-25 | 浙江大学 | Embedded multimedia key frame based video search realizing method |
CN101021855B (en) * | 2006-10-11 | 2010-04-07 | 北京新岸线网络技术有限公司 | Video searching system based on content |
CN103279473A (en) * | 2013-04-10 | 2013-09-04 | 深圳康佳通信科技有限公司 | Method, system and mobile terminal for searching massive amounts of video content |
-
2013
- 2013-09-23 CN CN201310438216.1A patent/CN103559196B/en active Active
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104700410B (en) * | 2015-03-14 | 2017-09-22 | 西安电子科技大学 | Instructional video mask method based on collaborative filtering |
CN104700410A (en) * | 2015-03-14 | 2015-06-10 | 西安电子科技大学 | Collaborative filtering-based teaching video labeling method |
CN104699844A (en) * | 2015-03-31 | 2015-06-10 | 北京奇艺世纪科技有限公司 | Method and device for determining video tags for advertisements |
CN104699844B (en) * | 2015-03-31 | 2019-03-15 | 北京奇艺世纪科技有限公司 | The method and device of video tab is determined for advertisement |
CN106294454A (en) * | 2015-05-29 | 2017-01-04 | 中兴通讯股份有限公司 | Video retrieval method and device |
CN105389326A (en) * | 2015-09-16 | 2016-03-09 | 中国科学院计算技术研究所 | Image annotation method based on weak matching probability canonical correlation model |
CN105389326B (en) * | 2015-09-16 | 2018-08-31 | 中国科学院计算技术研究所 | Image labeling method based on weak matching probability typical relevancy models |
CN105938561A (en) * | 2016-04-13 | 2016-09-14 | 南京大学 | Canonical-correlation-analysis-based computer data attribute reduction method |
CN106570165A (en) * | 2016-11-07 | 2017-04-19 | 北京航空航天大学 | Content-based video retrieval method and apparatus |
CN106570165B (en) * | 2016-11-07 | 2019-09-13 | 北京航空航天大学 | A kind of content based video retrieval system method and device |
CN106570196B (en) * | 2016-11-18 | 2020-06-05 | 广州视源电子科技股份有限公司 | Video program searching method and device |
WO2018090468A1 (en) * | 2016-11-18 | 2018-05-24 | 广州视源电子科技股份有限公司 | Method and device for searching for video program |
CN106570196A (en) * | 2016-11-18 | 2017-04-19 | 广州视源电子科技股份有限公司 | Video program searching method and apparatus |
CN106708929A (en) * | 2016-11-18 | 2017-05-24 | 广州视源电子科技股份有限公司 | Video program search method and device |
CN106682108A (en) * | 2016-12-06 | 2017-05-17 | 浙江大学 | Video retrieval method based on multi-modal convolutional neural network |
CN106682108B (en) * | 2016-12-06 | 2022-07-12 | 浙江大学 | Video retrieval method based on multi-mode convolutional neural network |
CN107122423A (en) * | 2017-04-06 | 2017-09-01 | 深圳Tcl数字技术有限公司 | Video display promotion method and device |
CN107748750A (en) * | 2017-08-30 | 2018-03-02 | 百度在线网络技术(北京)有限公司 | Similar video lookup method, device, equipment and storage medium |
US10853416B2 (en) | 2017-08-30 | 2020-12-01 | Baidu Online Network Technology (Beijing) Co., Ltd. | Similar video lookup method and apparatus, device and storage medium |
CN108182684A (en) * | 2017-12-22 | 2018-06-19 | 河南师范大学 | A kind of image partition method and its device based on weighting kernel fuzzy cluster |
CN108182684B (en) * | 2017-12-22 | 2021-06-25 | 河南师范大学 | Image segmentation method and device based on weighted kernel function fuzzy clustering |
CN110222594A (en) * | 2019-05-20 | 2019-09-10 | 厦门能见易判信息科技有限公司 | Pirate video recognition methods and system |
CN110222594B (en) * | 2019-05-20 | 2021-11-16 | 厦门能见易判信息科技有限公司 | Pirated video identification method and system |
CN110991470A (en) * | 2019-07-03 | 2020-04-10 | 北京市安全生产科学技术研究院 | Data dimension reduction method, portrait construction method and system and readable storage medium |
CN110598049A (en) * | 2019-09-26 | 2019-12-20 | 北京字节跳动网络技术有限公司 | Method, apparatus, electronic device and computer readable medium for retrieving video |
CN110704655A (en) * | 2019-10-18 | 2020-01-17 | 中国科学技术大学 | Online multi-quantization image retrieval method |
CN110704655B (en) * | 2019-10-18 | 2022-05-13 | 中国科学技术大学 | Online multi-quantization image retrieval method |
CN112182292A (en) * | 2020-09-30 | 2021-01-05 | 百度(中国)有限公司 | Training method and device for video retrieval model, electronic equipment and storage medium |
CN112182292B (en) * | 2020-09-30 | 2024-02-09 | 百度(中国)有限公司 | Training method and device for video retrieval model, electronic equipment and storage medium |
CN113221674A (en) * | 2021-04-25 | 2021-08-06 | 广东电网有限责任公司东莞供电局 | Video stream key frame extraction system and method based on rough set reduction and SIFT |
Also Published As
Publication number | Publication date |
---|---|
CN103559196B (en) | 2017-02-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103559196A (en) | Video retrieval method based on multi-core canonical correlation analysis | |
CN111858954B (en) | Task-oriented text-generated image network model | |
Wei et al. | Should we encode rain streaks in video as deterministic or stochastic? | |
Gould et al. | Decomposing a scene into geometric and semantically consistent regions | |
CN106649490B (en) | Image retrieval method and device based on depth features | |
CN108898145A (en) | A kind of image well-marked target detection method of combination deep learning | |
CN107944035B (en) | Image recommendation method integrating visual features and user scores | |
Feng et al. | Image tag completion by noisy matrix recovery | |
CN107943990B (en) | Multi-video abstraction method based on prototype analysis technology with weight | |
CN103258037A (en) | Trademark identification searching method for multiple combined contents | |
CN110083729B (en) | Image searching method and system | |
CN102799637A (en) | Method for automatically generating main character abstract in television program | |
CN103488664B (en) | A kind of image search method | |
CN110851731B (en) | Collaborative filtering recommendation method for user attribute coupling similarity and interest semantic similarity | |
CN106529492A (en) | Video topic classification and description method based on multi-image fusion in view of network query | |
CN109272440A (en) | A kind of reduced graph generating method and system for combining text and picture material | |
Zhang et al. | Retargeting semantically-rich photos | |
Dong et al. | An adult image detection algorithm based on Bag-of-Visual-Words and text information | |
CN102831161B (en) | For the semi-supervised sequence learning method based on manifold regularization of image retrieval | |
CN109934852B (en) | Video description method based on object attribute relation graph | |
Baghel et al. | Image conditioned keyframe-based video summarization using object detection | |
Chen et al. | Image retrieval via improved relevance ranking | |
CN108717436B (en) | Commodity target rapid retrieval method based on significance detection | |
Li et al. | Action recognition with spatio-temporal augmented descriptor and fusion method | |
CN108628999B (en) | Video recommendation method based on explicit and implicit information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |