CN103559196A - Video retrieval method based on multi-core canonical correlation analysis - Google Patents

Video retrieval method based on multi-core canonical correlation analysis Download PDF

Info

Publication number
CN103559196A
CN103559196A CN201310438216.1A CN201310438216A CN103559196A CN 103559196 A CN103559196 A CN 103559196A CN 201310438216 A CN201310438216 A CN 201310438216A CN 103559196 A CN103559196 A CN 103559196A
Authority
CN
China
Prior art keywords
video
matrix
frame
feature
correlation analysis
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201310438216.1A
Other languages
Chinese (zh)
Other versions
CN103559196B (en
Inventor
卜佳俊
高珊
李平
陈纯
何占盈
宋明黎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201310438216.1A priority Critical patent/CN103559196B/en
Publication of CN103559196A publication Critical patent/CN103559196A/en
Application granted granted Critical
Publication of CN103559196B publication Critical patent/CN103559196B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7847Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
    • G06F16/786Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using motion, e.g. object motion or camera motion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/732Query formulation
    • G06F16/7328Query by example, e.g. a complete video frame or video sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7847Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content

Abstract

Disclosed is a video retrieval method based on multi-core canonical correlation analysis. The method includes grasping text descriptions corresponding to the video on internet, and then operating on the video: firstly dividing the video according to whether a shot is mutated or not, extracting key frames, extracting vision features of the key frames and moving features of the shot to form video feature vectors, and extracting word-frequency features from the text descriptions of each video; then utilizing the method of the multi-core canonical correlation analysis to obtain mapping matrixes and low-dimensional representation of the video features and the word-frequency features, and allowing the mapping matrixes and the low-dimensional representation to have the maximum correlation in low-dimensional space; finally, when a user inputs key words to perform video retrieval, acquiring the low-dimensional representation of the word-frequency features of the key words according to the mapping matrixes of the word-frequency features, and returning video retrieval results sequentially from large to small of the degrees of cosine similarity with the low-dimensional representation of the video features. The method has the advantages that the correlation of video content and the retrieval key words is enhanced, and the accuracy of retrieval by the user is improved.

Description

A kind of video retrieval method based on multinuclear canonical correlation analysis
Technical field
The present invention relates to the technical field of video frequency searching, particularly the video retrieval method based on multinuclear canonical correlation analysis.
Background technology
Along with the develop rapidly of computer network, multimedia technology mechanics of communication in recent years, people can upload, watch and download all kinds of video informations by internet.Internet has become huge video warehouse gradually, and how more fast and effeciently the required video information of retrieval user has become information retrieval hot issue day by day.
Traditional video retrieval method is text based, it using video tab information as key word and video form matching relationship one by one, by key word being carried out to the operations such as feature extraction, pre-service, carry out Cluster Classification afterwards.The method efficiency of this manual mark video information that places one's entire reliance upon is lower, to the descriptive power deficiency of video information and experience that need to be certain.Therefore, text based search method can not meet the growing demand of user.From 20th century the eighties, content-based video frequency searching is progressively paid close attention to by people, more becomes in recent years study hotspot.The method can under the artificial situation participating in, automatically be extracted video features, and be not only to depend on label information not having.Particularly, the method camera lens cut apart with key frame extraction after extract the visual signature of key frame and the motion feature of camera lens, and typing video frequency searching database.During user search, video is retrieved according to the words-frequency feature of user search keyword, and according to similarity order from high to low, result for retrieval is returned to user.Adopt content-based video frequency searching can more objective, specifically and all sidedly describe video information, reduce subjectivity and the limitation of textual description information, and greatly improve the degree of accuracy of retrieval.
In machine learning field, content-based video retrieval method is not still a lot, but also obtained certain achievement, as the good joint classification feature optimal algorithm of performance of classifying at high dimensional data, the method can be selected and the maximally related union feature of classifying, and reduces in addition input dimension descent algorithm and the method based on rough set of descriptor data are carried out to the semantic dimensionality reduction descriptor index method keeping etc..
Summary of the invention
In order to help user to arrive the video of search by quick-searching, to improve the Experience Degree of watching video, the present invention proposes a kind of video retrieval method based on multinuclear canonical correlation analysis, the method comprises the following steps:
1, from internet captures video, for each video and textual description thereof, carry out following operation:
1) according to camera lens, whether suddenly change video is carried out to cutting, extract its key frame, and the motion feature that extracts visual signature in key frame and camera lens forms video feature vector, for the textual description of each video, extract words-frequency feature;
2) utilize multinuclear Canonical Correlation Analysis, obtain respectively the mapping matrix of video features and words-frequency feature, thus the low-dimensional that obtains both correspondences represent, make them maximum in the correlativity of low-dimensional data space;
3) when user inputs keyword and carries out video frequency searching, the low-dimensional that obtains keyword words-frequency feature according to the mapping matrix of words-frequency feature represents, the descending video frequency searching result of returning successively of cosine similarity representing according to itself and video features low-dimensional.
Further, the processing video described in step 1) and the method for corresponding textual description thereof, specifically:
1) when video lens cuts, adopt two comparative approach, establishing Tb is the threshold value of detector lens sudden change, Ts is the threshold value of detector lens gradual change, detects the difference of consecutive frame, if difference is greater than Tb by difference metric method, belong to lens mutation, should carry out shot segmentation, if difference is less than Tb, be greater than Ts, likely belong to gradual shot, now need this frame and subsequent frame to compare, if frame difference is less than Ts, but between consecutive frame, difference cumulative sum is greater than Tb, means and really belongs to gradual shot;
2) while extracting key frame, first first the head and the tail frame of camera lens is set as to key frame, headed by cause, frame is conventionally in order to attract spectators to represent the theme of camera lens, tail frame wishes that spectators' aftertaste often represents with a kind of feature, after head and the tail frame is determined, also to choose the frame that is changed significantly as key frame, each frame that is about to non-key frame in camera lens compares with key frame successively, if differed greatly, using it as key frame, relatively go down successively until compared all non-key frames in camera lens, also to choose key frame according to the frame method of average afterwards, on certain assigned address, calculate the mean value of all frame pixel values, key frame is the frame that pixel value approaches mean value most,
3) during key frame feature extraction, the Visual Feature Retrieval Process color of still image, shape, texture, the variation of camera motion feature extraction camera motion, movement locus, moving target size, the video features that forms video represents;
4) videotext is described and carried out participle, statistics word frequency, forms its words-frequency feature and represents.
Further, step 2) the multinuclear Canonical Correlation Analysis described in, specifically:
1) training sample video sum is made as to n, with m dimensional vector X ithe video features that represents i video, by the vectorial X of n m dimension ibe merged into matrix X, represent video feature matrix, in like manner, with k dimensional vector Y jthe words-frequency feature that represents the textual description that j video is corresponding, by the vectorial Y of n k dimension jbe merged into matrix Y, represent words-frequency feature matrix;
2) utilize gaussian kernel function formula
Figure 2013104382161100002DEST_PATH_IMAGE001
calculate the Gram matrix K of X x, K wherein xfor the matrix of the capable n row of n, ‖ ‖ 2represent vectorial I 2norm, real parameters σ represents the bandwidth of gaussian kernel, by different σ is set, can obtain one group of gaussian kernel with different nonlinear organizations, in like manner, to Y application 2-d polynomial kernel function formula
Figure 2013104382161100002DEST_PATH_IMAGE002
calculate the Gram matrix K of Y y, wherein () trepresent vector or transpose of a matrix, parameters C is to be more than or equal to 0 kernel functional parameter, by different C is set, can obtain one group of polynomial kernel with different nonlinear organizations;
3) by the K carrying out after kernel function conversion x, K ybe normalized;
4) the objective function expression formula of the video retrieval method based on multinuclear canonical correlation analysis is:
max α , β W x T K x K y W x ( W x T K x 2 W x + κW x T K x W x ) · ( W y T K y 2 W y + κW y T K y W y ) ,
subjectto ( W x T K x 2 W x + κW x T K x W x ) = 1 , ( W y T K y 2 W y + κW y T K y W y ) = 1
Solve above-mentioned objective function expression equivalence in solving following generalized eigenvalue problem:
(K x+κI) -1K y(K y+κI) -1K xW x=λ 2W x
As required, get (K x+ κ I) -1k y(K y+ κ I) -1k xfront p maximum eigenwert characteristic of correspondence vector forms the matrix of the capable p row of n
Figure 2013104382161100002DEST_PATH_IMAGE005
as K xcorresponding mapping matrix, p represents to adopt the dimension after the video retrieval method dimensionality reduction of multinuclear canonical correlation analysis, and κ is arithmetic number parameter, () -1inverting of representing matrix, I representation unit matrix, K ycorresponding mapping matrix
Figure 2013104382161100002DEST_PATH_IMAGE006
can basis
Figure 2013104382161100002DEST_PATH_IMAGE007
with
Figure 2013104382161100002DEST_PATH_IMAGE008
between relation obtain:
W y p = ( K y + κI ) - 1 K x W x p λ ,
Wherein,
Figure 2013104382161100002DEST_PATH_IMAGE010
also be the matrix of the capable p row of n;
5) according to new mapping matrix
Figure 2013104382161100002DEST_PATH_IMAGE011
structure K x, K yexpression under new space: K x new = K x W x p , K y new = K y W y p , K x new , K y new The matrix that is the capable p row of n, now both correlativitys are maximum.
Further, the search matching method described in step 3), specifically:
1) search key of submitting to according to user is converted into keyword words-frequency feature vector y q, to y qapplication laplace kernel function formula
Figure 2013104382161100002DEST_PATH_IMAGE014
calculate y qn dimension Gram vector y jfor j sample in training data Y, parameters C is to be more than or equal to 0 2-d polynomial kernel functional parameter;
2) according to mapping matrix
Figure 2013104382161100002DEST_PATH_IMAGE016
convert,
Figure 2013104382161100002DEST_PATH_IMAGE017
the low-dimensional that obtains keyword words-frequency feature represents, wherein
Figure 2013104382161100002DEST_PATH_IMAGE018
for p dimensional vector;
3) will
Figure 2013104382161100002DEST_PATH_IMAGE019
vector respectively with obtain before every a line of matrix, each sample carries out the calculating of cosine similarity successively,
Figure BDA0000385806800000063
each sample be p dimensional vector, cosine similarity shows that more greatly similarity is higher;
4) according to cosine similarity order from high to low, return to results for video.
The present invention proposes the video retrieval method based on multinuclear canonical correlation analysis, its advantage is: video is effectively retrieved, can be improved the precision of video frequency searching; Be applicable to all types of videos, without backstage manual operation, also can be used for helping domestic consumer to improve web page browsing quality.
Accompanying drawing explanation
Fig. 1 is method flow diagram of the present invention.
Embodiment
With reference to accompanying drawing, further illustrate the present invention:
A video retrieval method based on multinuclear canonical correlation analysis, the method comprises the following steps:
1, from internet captures video, for each video and textual description thereof, carry out following operation:
1) according to camera lens, whether suddenly change video is carried out to cutting, extract its key frame, and the motion feature that extracts visual signature in key frame and camera lens forms video feature vector, for the textual description of each video, extract words-frequency feature;
2) utilize multinuclear Canonical Correlation Analysis, obtain respectively the mapping matrix of video features and words-frequency feature, thus the low-dimensional that obtains both correspondences represent, make them maximum in the correlativity of low-dimensional data space;
3) when user inputs keyword and carries out video frequency searching, the low-dimensional that obtains keyword words-frequency feature according to the mapping matrix of words-frequency feature represents, the descending video frequency searching result of returning successively of cosine similarity representing according to itself and video features low-dimensional.
Processing video described in step 1) and the method for correspondence-textual description thereof, specifically:
1) when video lens cuts, adopt two comparative approach, establishing Tb is the threshold value of detector lens sudden change, Ts is the threshold value of detector lens gradual change, detects the difference of consecutive frame, if difference is greater than Tb by difference metric method, belong to lens mutation, should carry out shot segmentation, if difference is less than Tb, be greater than Ts, likely belong to gradual shot, now need this frame and subsequent frame to compare, if frame difference is less than Ts, but between consecutive frame, difference cumulative sum is greater than Tb, means and really belongs to gradual shot;
2) while extracting key frame, first first the head and the tail frame of camera lens is set as to key frame, headed by cause, frame is conventionally in order to attract spectators to represent the theme of camera lens, tail frame wishes that spectators' aftertaste often represents with a kind of feature, after head and the tail frame is determined, also to choose the frame that is changed significantly as key frame, each frame that is about to non-key frame in camera lens compares with key frame successively, if differed greatly, using it as key frame, relatively go down successively until compared all non-key frames in camera lens, also to choose key frame according to the frame method of average afterwards, on certain assigned address, calculate the mean value of all frame pixel values, key frame is the frame that pixel value approaches mean value most,
3) during key frame feature extraction, the Visual Feature Retrieval Process color of still image, shape, texture, the variation of camera motion feature extraction camera motion, movement locus, moving target size, the video features that forms video represents;
4) videotext is described and carried out participle, statistics word frequency, forms its words-frequency feature and represents.
Step 2) the multinuclear Canonical Correlation Analysis described in, specifically:
1) training sample video sum is made as to n, with m dimensional vector X ithe video features that represents i video, by the vectorial X of n m dimension ibe merged into matrix X, represent video feature matrix, in like manner, with k dimensional vector Y jthe words-frequency feature that represents the textual description that j video is corresponding, by the vectorial Y of n k dimension jbe merged into matrix Y, represent words-frequency feature matrix;
2) utilize gaussian kernel function formula
Figure 2013104382161100002DEST_PATH_IMAGE020
calculate the Gram matrix K of X x, K wherein xfor the matrix of the capable n row of n, ‖ ‖ 2represent vectorial I 2norm, real parameters σ represents the bandwidth of gaussian kernel, by different σ is set, can obtain one group of gaussian kernel with different nonlinear organizations, in like manner, to Y application 2-d polynomial kernel function formula calculate the Gram matrix K of Y y, wherein () trepresent vector or transpose of a matrix, parameters C is to be more than or equal to 0 kernel functional parameter, by different C is set, can obtain one group of polynomial kernel with different nonlinear organizations;
3) by the K carrying out after kernel function conversion x, K ybe normalized;
4) the objective function expression formula of the video retrieval method based on multinuclear canonical correlation analysis is:
max α , β W x T K x K y W x ( W x T K x 2 W x + κW x T K x W x ) · ( W y T K y 2 W y + κW y T K y W y ) ,
subjectto ( W x T K x 2 W x + κW x T K x W x ) = 1 , ( W y T K y 2 W y + κW y T K y W y ) = 1
Solve above-mentioned objective function expression equivalence in solving following generalized eigenvalue problem:
(K x+κI) -1K y(K y+κI) -1K xW x=λ 2W x
As required, get (K x+ κ I) -1k y(K y+ κ I) -1k xfront p maximum eigenwert characteristic of correspondence vector forms the matrix of the capable p row of n as
Figure BDA0000385806800000094
corresponding mapping matrix, p represents to adopt the dimension after the video retrieval method dimensionality reduction of multinuclear canonical correlation analysis, and κ is arithmetic number parameter, () -1inverting of representing matrix, I representation unit matrix, K ycorresponding mapping matrix
Figure BDA0000385806800000095
can basis
Figure BDA0000385806800000096
with
Figure BDA0000385806800000097
between relation obtain:
W y p = ( K y + κI ) - 1 K x W x p λ ,
Wherein,
Figure BDA0000385806800000099
also be the matrix of the capable p row of n;
5) according to new mapping matrix
Figure BDA00003858068000000910
structure K x, K yexpression under new space: K x new = K x W x p , K y new = K y W y p , K x new , K y new The matrix that is the capable p row of n, now both correlativitys are maximum.
Search matching method described in step 3), specifically:
1) search key of submitting to according to user is converted into keyword words-frequency feature vector y q, to y qapplication laplace kernel function formula calculate y qn dimension Gram vector
Figure BDA0000385806800000102
y jfor j sample in training data Y, parameters C is to be more than or equal to 0 2-d polynomial kernel functional parameter;
2) according to mapping matrix
Figure BDA0000385806800000103
convert, the low-dimensional that obtains keyword words-frequency feature represents, wherein
Figure 2013104382161100002DEST_PATH_IMAGE028
for p dimensional vector;
3) will vector respectively with obtain before
Figure BDA0000385806800000108
every a line of matrix, each sample carries out the calculating of cosine similarity successively,
Figure 431129DEST_PATH_IMAGE002
each sample be p dimensional vector, cosine similarity shows that more greatly similarity is higher;
4) according to cosine similarity order from high to low, return to results for video.
Content described in this instructions embodiment is only enumerating the way of realization of inventive concept; protection scope of the present invention should not be regarded as only limiting to the concrete form that embodiment states, protection scope of the present invention is also and in those skilled in the art, according to the present invention, conceive the equivalent technologies means that can expect.

Claims (4)

1. the video retrieval method based on multinuclear canonical correlation analysis, the method is characterized in that from internet captures video, for each video, carries out following operation:
1) according to camera lens, whether suddenly change video is carried out to cutting, extract its key frame, and the motion feature that extracts visual signature in key frame and camera lens forms video feature vector, for the textual description of each video, extract words-frequency feature;
2) utilize multinuclear Canonical Correlation Analysis, obtain respectively the mapping matrix of video features and words-frequency feature, thus the low-dimensional that obtains both correspondences represent, make them maximum in the correlativity of low-dimensional data space;
3) when user inputs keyword and carries out video frequency searching, the low-dimensional that obtains keyword words-frequency feature according to the mapping matrix of words-frequency feature represents, the descending video frequency searching result of returning successively of cosine similarity representing according to itself and video features low-dimensional.
2. the video retrieval method based on multinuclear canonical correlation analysis as claimed in claim 1, is characterized in that: the processing video described in described step 1) and the method for corresponding textual description thereof, specifically:
1) when video lens cuts, adopt two comparative approach, establishing Tb is the threshold value of detector lens sudden change, Ts is the threshold value of detector lens gradual change, detects the difference of consecutive frame, if difference is greater than Tb by difference metric method, belong to lens mutation, should carry out shot segmentation, if difference is less than Tb, be greater than Ts, likely belong to gradual shot, now need this frame and subsequent frame to compare, if frame difference is less than Ts, but between consecutive frame, difference cumulative sum is greater than Tb, means and really belongs to gradual shot;
2) while extracting key frame, first first the head and the tail frame of camera lens is set as to key frame, headed by cause, frame is conventionally in order to attract spectators to represent the theme of camera lens, tail frame wishes that spectators' aftertaste often represents with a kind of feature, after head and the tail frame is determined, also to choose the frame that is changed significantly as key frame, each frame that is about to non-key frame in camera lens compares with key frame successively, if differed greatly, using it as key frame, relatively go down successively until compared all non-key frames in camera lens, also to choose key frame according to the frame method of average afterwards, on certain assigned address, calculate the mean value of all frame pixel values, key frame is the frame that pixel value approaches mean value most,
3) during key frame feature extraction, the Visual Feature Retrieval Process color of still image, shape, texture, the variation of camera motion feature extraction camera motion, movement locus, moving target size, the video features that forms video represents;
4) videotext is described and carried out participle, statistics word frequency, forms its words-frequency feature and represents.
3. the video retrieval method based on multinuclear canonical correlation analysis as claimed in claim 2, is characterized in that: the multinuclear Canonical Correlation Analysis described step 2), specifically:
1) training sample video sum is made as to n, with m dimensional vector X ithe video features that represents i video, by the vectorial X of n m dimension ibe merged into matrix X, represent video feature matrix, in like manner, with k dimensional vector Y jthe words-frequency feature that represents the textual description that j video is corresponding, by the vectorial Y of n k dimension jbe merged into matrix Y, represent words-frequency feature matrix;
2) utilize gaussian kernel function formula
Figure 2013104382161100001DEST_PATH_IMAGE001
calculate the Gram matrix K of X x, K wherein xfor the matrix of the capable n row of n, ‖ ‖ 2represent vectorial I 2norm, real parameters σ represents the bandwidth of gaussian kernel, by different σ is set, can obtain one group of gaussian kernel with different nonlinear organizations, in like manner, to Y application 2-d polynomial kernel function formula
Figure 2013104382161100001DEST_PATH_IMAGE002
calculate the Gram matrix K of Y y, wherein () trepresent vector or transpose of a matrix, parameters C is to be more than or equal to 0 kernel functional parameter, by different C is set, can obtain one group of polynomial kernel with different nonlinear organizations;
3) by the K carrying out after kernel function conversion x, K ybe normalized;
4) the objective function expression formula of the video retrieval method based on multinuclear canonical correlation analysis is:
Figure 2013104382161100001DEST_PATH_IMAGE003
Figure 2013104382161100001DEST_PATH_IMAGE004
Solve above-mentioned objective function expression equivalence in solving following generalized eigenvalue problem:
(K x+κI) -1K y(K y1) -1K xW x=λ 2W x
As required, get (K x+ κ I) -1k y(K y+ κ I) -1k xfront p maximum eigenwert characteristic of correspondence vector forms the matrix of the capable p row of n
Figure FDA0000385806790000032
as K xcorresponding mapping matrix, p represents to adopt the dimension after the video retrieval method dimensionality reduction of multinuclear canonical correlation analysis, and κ is arithmetic number parameter, () -1inverting of representing matrix, I representation unit matrix, K ycorresponding mapping matrix
Figure FDA0000385806790000033
can basis
Figure FDA0000385806790000034
with
Figure FDA0000385806790000035
between relation obtain:
Figure 2013104382161100001DEST_PATH_IMAGE005
Wherein,
Figure FDA0000385806790000037
also be the matrix of the capable p row of n;
5) according to new mapping matrix
Figure 2013104382161100001DEST_PATH_IMAGE006
structure K x, K yexpression under new space:
Figure 2013104382161100001DEST_PATH_IMAGE007
the matrix that is the capable p row of n, now both correlativitys are maximum.
4. the video retrieval method based on multinuclear canonical correlation analysis as claimed in claim 3, is characterized in that: the search matching method described in described step 3), specifically:
1) search key of submitting to according to user is converted into keyword words-frequency feature vector y q, to y qapplication laplace kernel function formula
Figure 2013104382161100001DEST_PATH_IMAGE009
calculate y qn dimension Gram vector
Figure FDA00003858067900000311
y jfor j sample in training data Y, parameters C is to be more than or equal to 0 2-d polynomial kernel functional parameter;
2) according to mapping matrix
Figure FDA00003858067900000312
convert, the low-dimensional that obtains keyword words-frequency feature represents, wherein for p dimensional vector;
3) will vector respectively with obtain before
Figure 2013104382161100001DEST_PATH_IMAGE013
every a line of matrix, each sample carries out the calculating of cosine similarity successively,
Figure 2013104382161100001DEST_PATH_IMAGE014
each sample be p dimensional vector, cosine similarity shows that more greatly similarity is higher;
4) according to cosine similarity order from high to low, return to results for video.
CN201310438216.1A 2013-09-23 2013-09-23 Video retrieval method based on multi-core canonical correlation analysis Active CN103559196B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310438216.1A CN103559196B (en) 2013-09-23 2013-09-23 Video retrieval method based on multi-core canonical correlation analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310438216.1A CN103559196B (en) 2013-09-23 2013-09-23 Video retrieval method based on multi-core canonical correlation analysis

Publications (2)

Publication Number Publication Date
CN103559196A true CN103559196A (en) 2014-02-05
CN103559196B CN103559196B (en) 2017-02-22

Family

ID=50013443

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310438216.1A Active CN103559196B (en) 2013-09-23 2013-09-23 Video retrieval method based on multi-core canonical correlation analysis

Country Status (1)

Country Link
CN (1) CN103559196B (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104700410A (en) * 2015-03-14 2015-06-10 西安电子科技大学 Collaborative filtering-based teaching video labeling method
CN104699844A (en) * 2015-03-31 2015-06-10 北京奇艺世纪科技有限公司 Method and device for determining video tags for advertisements
CN105389326A (en) * 2015-09-16 2016-03-09 中国科学院计算技术研究所 Image annotation method based on weak matching probability canonical correlation model
CN105938561A (en) * 2016-04-13 2016-09-14 南京大学 Canonical-correlation-analysis-based computer data attribute reduction method
CN106294454A (en) * 2015-05-29 2017-01-04 中兴通讯股份有限公司 Video retrieval method and device
CN106570196A (en) * 2016-11-18 2017-04-19 广州视源电子科技股份有限公司 Video program searching method and apparatus
CN106570165A (en) * 2016-11-07 2017-04-19 北京航空航天大学 Content-based video retrieval method and apparatus
CN106682108A (en) * 2016-12-06 2017-05-17 浙江大学 Video retrieval method based on multi-modal convolutional neural network
CN106708929A (en) * 2016-11-18 2017-05-24 广州视源电子科技股份有限公司 Video program search method and device
CN107122423A (en) * 2017-04-06 2017-09-01 深圳Tcl数字技术有限公司 Video display promotion method and device
CN107748750A (en) * 2017-08-30 2018-03-02 百度在线网络技术(北京)有限公司 Similar video lookup method, device, equipment and storage medium
CN108182684A (en) * 2017-12-22 2018-06-19 河南师范大学 A kind of image partition method and its device based on weighting kernel fuzzy cluster
CN110222594A (en) * 2019-05-20 2019-09-10 厦门能见易判信息科技有限公司 Pirate video recognition methods and system
CN110598049A (en) * 2019-09-26 2019-12-20 北京字节跳动网络技术有限公司 Method, apparatus, electronic device and computer readable medium for retrieving video
CN110704655A (en) * 2019-10-18 2020-01-17 中国科学技术大学 Online multi-quantization image retrieval method
CN110991470A (en) * 2019-07-03 2020-04-10 北京市安全生产科学技术研究院 Data dimension reduction method, portrait construction method and system and readable storage medium
CN112182292A (en) * 2020-09-30 2021-01-05 百度(中国)有限公司 Training method and device for video retrieval model, electronic equipment and storage medium
CN113221674A (en) * 2021-04-25 2021-08-06 广东电网有限责任公司东莞供电局 Video stream key frame extraction system and method based on rough set reduction and SIFT

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7143434B1 (en) * 1998-11-06 2006-11-28 Seungyup Paek Video description system and method
CN1851710A (en) * 2006-05-25 2006-10-25 浙江大学 Embedded multimedia key frame based video search realizing method
CN101021855B (en) * 2006-10-11 2010-04-07 北京新岸线网络技术有限公司 Video searching system based on content
CN103279473A (en) * 2013-04-10 2013-09-04 深圳康佳通信科技有限公司 Method, system and mobile terminal for searching massive amounts of video content

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104700410B (en) * 2015-03-14 2017-09-22 西安电子科技大学 Instructional video mask method based on collaborative filtering
CN104700410A (en) * 2015-03-14 2015-06-10 西安电子科技大学 Collaborative filtering-based teaching video labeling method
CN104699844A (en) * 2015-03-31 2015-06-10 北京奇艺世纪科技有限公司 Method and device for determining video tags for advertisements
CN104699844B (en) * 2015-03-31 2019-03-15 北京奇艺世纪科技有限公司 The method and device of video tab is determined for advertisement
CN106294454A (en) * 2015-05-29 2017-01-04 中兴通讯股份有限公司 Video retrieval method and device
CN105389326A (en) * 2015-09-16 2016-03-09 中国科学院计算技术研究所 Image annotation method based on weak matching probability canonical correlation model
CN105389326B (en) * 2015-09-16 2018-08-31 中国科学院计算技术研究所 Image labeling method based on weak matching probability typical relevancy models
CN105938561A (en) * 2016-04-13 2016-09-14 南京大学 Canonical-correlation-analysis-based computer data attribute reduction method
CN106570165A (en) * 2016-11-07 2017-04-19 北京航空航天大学 Content-based video retrieval method and apparatus
CN106570165B (en) * 2016-11-07 2019-09-13 北京航空航天大学 A kind of content based video retrieval system method and device
CN106570196B (en) * 2016-11-18 2020-06-05 广州视源电子科技股份有限公司 Video program searching method and device
WO2018090468A1 (en) * 2016-11-18 2018-05-24 广州视源电子科技股份有限公司 Method and device for searching for video program
CN106570196A (en) * 2016-11-18 2017-04-19 广州视源电子科技股份有限公司 Video program searching method and apparatus
CN106708929A (en) * 2016-11-18 2017-05-24 广州视源电子科技股份有限公司 Video program search method and device
CN106682108A (en) * 2016-12-06 2017-05-17 浙江大学 Video retrieval method based on multi-modal convolutional neural network
CN106682108B (en) * 2016-12-06 2022-07-12 浙江大学 Video retrieval method based on multi-mode convolutional neural network
CN107122423A (en) * 2017-04-06 2017-09-01 深圳Tcl数字技术有限公司 Video display promotion method and device
CN107748750A (en) * 2017-08-30 2018-03-02 百度在线网络技术(北京)有限公司 Similar video lookup method, device, equipment and storage medium
US10853416B2 (en) 2017-08-30 2020-12-01 Baidu Online Network Technology (Beijing) Co., Ltd. Similar video lookup method and apparatus, device and storage medium
CN108182684A (en) * 2017-12-22 2018-06-19 河南师范大学 A kind of image partition method and its device based on weighting kernel fuzzy cluster
CN108182684B (en) * 2017-12-22 2021-06-25 河南师范大学 Image segmentation method and device based on weighted kernel function fuzzy clustering
CN110222594A (en) * 2019-05-20 2019-09-10 厦门能见易判信息科技有限公司 Pirate video recognition methods and system
CN110222594B (en) * 2019-05-20 2021-11-16 厦门能见易判信息科技有限公司 Pirated video identification method and system
CN110991470A (en) * 2019-07-03 2020-04-10 北京市安全生产科学技术研究院 Data dimension reduction method, portrait construction method and system and readable storage medium
CN110598049A (en) * 2019-09-26 2019-12-20 北京字节跳动网络技术有限公司 Method, apparatus, electronic device and computer readable medium for retrieving video
CN110704655A (en) * 2019-10-18 2020-01-17 中国科学技术大学 Online multi-quantization image retrieval method
CN110704655B (en) * 2019-10-18 2022-05-13 中国科学技术大学 Online multi-quantization image retrieval method
CN112182292A (en) * 2020-09-30 2021-01-05 百度(中国)有限公司 Training method and device for video retrieval model, electronic equipment and storage medium
CN112182292B (en) * 2020-09-30 2024-02-09 百度(中国)有限公司 Training method and device for video retrieval model, electronic equipment and storage medium
CN113221674A (en) * 2021-04-25 2021-08-06 广东电网有限责任公司东莞供电局 Video stream key frame extraction system and method based on rough set reduction and SIFT

Also Published As

Publication number Publication date
CN103559196B (en) 2017-02-22

Similar Documents

Publication Publication Date Title
CN103559196A (en) Video retrieval method based on multi-core canonical correlation analysis
CN111858954B (en) Task-oriented text-generated image network model
Wei et al. Should we encode rain streaks in video as deterministic or stochastic?
Gould et al. Decomposing a scene into geometric and semantically consistent regions
CN106649490B (en) Image retrieval method and device based on depth features
CN108898145A (en) A kind of image well-marked target detection method of combination deep learning
CN107944035B (en) Image recommendation method integrating visual features and user scores
Feng et al. Image tag completion by noisy matrix recovery
CN107943990B (en) Multi-video abstraction method based on prototype analysis technology with weight
CN103258037A (en) Trademark identification searching method for multiple combined contents
CN110083729B (en) Image searching method and system
CN102799637A (en) Method for automatically generating main character abstract in television program
CN103488664B (en) A kind of image search method
CN110851731B (en) Collaborative filtering recommendation method for user attribute coupling similarity and interest semantic similarity
CN106529492A (en) Video topic classification and description method based on multi-image fusion in view of network query
CN109272440A (en) A kind of reduced graph generating method and system for combining text and picture material
Zhang et al. Retargeting semantically-rich photos
Dong et al. An adult image detection algorithm based on Bag-of-Visual-Words and text information
CN102831161B (en) For the semi-supervised sequence learning method based on manifold regularization of image retrieval
CN109934852B (en) Video description method based on object attribute relation graph
Baghel et al. Image conditioned keyframe-based video summarization using object detection
Chen et al. Image retrieval via improved relevance ranking
CN108717436B (en) Commodity target rapid retrieval method based on significance detection
Li et al. Action recognition with spatio-temporal augmented descriptor and fusion method
CN108628999B (en) Video recommendation method based on explicit and implicit information

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant