CN105956603A - Video sequence classifying method based on tensor time domain association model - Google Patents

Video sequence classifying method based on tensor time domain association model Download PDF

Info

Publication number
CN105956603A
CN105956603A CN201610237984.4A CN201610237984A CN105956603A CN 105956603 A CN105956603 A CN 105956603A CN 201610237984 A CN201610237984 A CN 201610237984A CN 105956603 A CN105956603 A CN 105956603A
Authority
CN
China
Prior art keywords
tensor
video sequence
sigma
video
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610237984.4A
Other languages
Chinese (zh)
Inventor
张静
徐传忠
苏育挺
井佩光
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201610237984.4A priority Critical patent/CN105956603A/en
Publication of CN105956603A publication Critical patent/CN105956603A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/245Classification techniques relating to the decision surface
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/06Topological mapping of higher dimensional structures onto lower dimensional surfaces

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

A video sequence classifying method based on a tensor time domain association model comprises the steps of representing an original video sequence in a three-grade video tensor manner; performing tensor Tucker decomposition on the three-grade video tensor for obtaining a latent kernel tensor; applying an autoregression model on the time domain of the obtained latent kernel tensor for establishing relevance between adjacent time slices; dynamically studying the front process, updating the result until an algorithm is convergent, and obtaining an optimal result. The video sequence classifying method ensures time domain relevance and dependence of the video sequence after dimension reduction through limiting the time domain of the video sequence. The video sequence classifying method has advantages of sufficiently utilizing latent useful information in the video, eliminating redundant information in the video, ensuring high continuity of the video sequence in time domain, reducing classification difficulty of the video sequence, and improving classification accuracy of the video sequence. The video sequence classifying method is better than a traditional video sequence classification method and greatly improves classification precision.

Description

A kind of video sequence sorting technique based on tensor time domain correlation model
Technical field
The present invention relates to a kind of video sequence sorting technique.Particularly relate to one resolution of tensor technology associated with time domain Model combines, tensor video sequence is carried out that spatial domain dimensionality reduction obtains that the video sequence of potential low-dimensional represents based on tensor time The video sequence sorting technique of territory correlation model.
Background technology
Flourish recently as video capture device, the quantity of video data has obtained significant increase.In order to Being analyzed these video datas, video sequence classification has obtained great concern.Video sequence classification is widely used in Video frequency abstract, video frequency searching, action recognition etc..The gesture of human body, action video are ingredients important in video data, its In the transmission of deaf mute's information, the transmission of noisy environmental information, computer is the most medium has a wide range of applications.
Learning method based on subspace had obtained the very big concern in field in recent years, according to the sign shape that data are different Formula, learning method based on subspace is broadly divided into two classes: 1) traditional learning method based on subspace;2) based on multi-thread The sub-space learning method of property algebraically.
Video sequence is generally transformed into the form of vector or matrix, this kind by traditional learning method based on subspace The structural information that processing mode is easily destroyed between adjacent data causes the loss of time-space domain information and easily causes dimension disaster. Method based on polyteny sub-space learning uses tensor to model video sequence, so can preservation video sequence naturally In intrinsic structural information, do not result in the loss of structural information between data.But, mostly this method is to use isotropic Processing mode.Owing to video sequence is typical time series, in this sequence, data show in time domain and associate on the time Property, the method for isotropic can weaken between video sequence characteristic in time domain, affect classifying quality.
Owing to video sequence is inherently the time series data of higher-dimension, therefore in it relativity of time domain and dependence Property is most important to video content analysis.A few thing was had to start to focus on higher-dimension time series analysis, by polyteny in recent years Algebraic operation combines with time domain smoothness constraint and processes the time series data of high-order, but this method is not considered The statistical distribution rule that video time sequence is intrinsic, it is impossible to the architectural feature of enough practical coupling video sequence data.Simultaneously Such method typically requires and manual arranges many important parameters, is more suitable for longer video sequence classification problem.
Summary of the invention
The technical problem to be solved is to provide a kind of redundancy eliminated in video, maintains video The sequence successional based on tensor time domain correlation model video sequence sorting technique in time domain.
The technical solution adopted in the present invention is: a kind of video sequence sorting technique based on tensor time domain correlation model, Comprise the steps:
1) original video sequence is characterized as the form of three rank video tensors;
2) three rank video tensors are carried out tensor Tucker and decomposes the core tensor that acquisition is potential;
3) the time domain application autoregression model to the potential core tensor obtained sets up the association between adjacent time section Property;
4) dynamic learning step 2)~step 3) updating result until algorithmic statement, result reaches optimum.
Step 1) it is given a series of video sequence X={X1,…,Xi..., XN, and described video sequence is represented Become the form of three rank video tensors, wherein:
It is three rank video tensors, wherein I1、I2With T be expressed as a video sequence width, Height and the length of time shaft, N is the quantity of video sequence;Definition It it is one of tensor Isochronous surface.
Step 2) including: by finding two mapping matrixes in the framework that tensor Tucker decomposesWithBy three initial rank video tensorsMigration is a potential core tensorMoving J is kept during shifting1<I1, J2<I2, the potential core tensor Y that obtainediIt is three initial rank video tensor XiA kind of low-dimensional The representation of intrinsic, described transition process is expressed as:
Yi=Xi×1U1×2U2 (1)
Wherein, I1、I2With the length that T is expressed as the width of a video sequence, height and time shaft;J1And J2Also It is expressed as width and the height of a video sequence.
Step 3) comprise the steps:
(1) application autoregression model sets up isochronous surfaceWith this isochronous surfaceIsochronous surface aboveRelatedness model, by relatedness model with m isochronous surfaceGo to predict next Individual isochronous surfaceDescribed relatedness model is as follows:
Y t i = &Sigma; k = 1 m a k Y t - k i + &epsiv; t i - - - ( 2 )
WhereinIt is white Gaussian noise, meetsN is Gauss distribution,It is that the covariance of white noise is opened Amount, meetsVar is variance,It it is the model coefficient of autoregression model;
(2) model parameter in relatedness modelMaximal possibility estimation with Yule-Waler side based on tensor Journey obtains:
&Sigma; k = 1 m a k ( &Sigma; t = 1 T < Y t - k i , Y t - l i > ) = &Sigma; t = 1 T < Y t i , Y t - l i > - - - ( 3 )
(3) by Yi=Xi×1U1×2U2It is applied to relatedness modelIn, and rearrange and obtain Following noise covariance tensor expression:
&epsiv; t i = Y t i - &Sigma; k = 1 m a k Y t - k i = ( X t i - &Sigma; k = 1 m a k Y t - k i ) &times; U 1 1 &times; U 2 2 - - - ( 4 )
(4) as long as having estimated model parameterNoise covariance between actual value and the predictive value of isochronous surface The accumulation norm of tensor is obtained by following formula:
&Sigma; t = 2 T || &epsiv; t i || F 2 = &Sigma; t = 2 T || ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 || F 2 - - - ( 5 )
(5) object function is obtained
By minimizing the accumulation norm of noise covariance tensor, obtain goal of function:
min U 1 , U 2 , { a k } &Sigma; i = 1 N &Sigma; t = 2 T || ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 || F 2 s . t . U 1 U 1 T = E U 2 U 2 T = E - - - ( 6 )
Wherein, E representation unit orthogonal matrix;
(6) object function is solved
Alternately descent algorithm is used to go to solve object function, first random initializtion mapping matrix U1And U2Go to obtain potential Core tensor Yi, then with potential core tensor YiGo to calculate the model parameter of relatedness modelFixing mapping square therewith Battle array U2With potential core tensor Yi, update the mapping matrix U of object function1:
min U 1 &Sigma; i = 1 N &Sigma; t = 2 T || ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; { U } ( - 1 ) &times; U 1 1 || F 2 s . t . U 1 U 1 T = E - - - ( 7 )
OrderObtain the equivalent form of value of object function:
max U 1 t r ( U 1 ( &Sigma; i = 1 N &Sigma; t = 2 T G t 1 i G t 1 i T ) U 1 T ) s . t . U 1 U 1 T = E - - - ( 8 )
Minimize the equivalent form of value of object function by method of Lagrange multipliers and Eigenvalues Decomposition method, then object function is relative In U1Partial derivative calculated by following formula:
&Sigma; i = 1 N &Sigma; t = 2 T G t 1 i G t 1 i T u 1 j T = &lambda; 1 j u 1 j T - - - ( 9 )
U in the formula1jIt it is matrixGeneralized eigenvector, λ1jIt it is corresponding eigenvalue;
Equally, fixing mapping matrix U1, then object function is relative to mapping matrix U2Partial derivative calculated by following formula:
&Sigma; i = 1 N &Sigma; t = 2 T G t 2 i G t 2 i T u 2 j T = &lambda; 2 j u 2 j T - - - ( 10 )
U in formula2jIt it is matrixGeneralized eigenvector, λ2jIt it is corresponding eigenvalue.
Step 4) be by step 2)~step 3) be incorporated in a dynamic learning framework make step 2)~step 3) Habit process is updated over time, until meeting the condition of convergenceWherein, β represents forecast error.
A kind of based on tensor time domain correlation model the video sequence sorting technique of the present invention, core is tensor video Sequence carries out the video sequence of the potential low-dimensional of spatial domain dimensionality reduction acquisition and represents.In this process by the time domain of video sequence is entered Row limits still has time domain relatedness and dependency with the video sequence after ensureing dimensionality reduction.The present invention sufficiently utilizes in video latent Useful information, eliminate the redundancy in video, maintain video sequence seriality in time domain.Reduce video The difficulty of sequence classification, promotes the accuracy rate of video sequence classification.The method of the present invention is better than traditional video sequence classification side Method, precision is greatly improved.
Accompanying drawing explanation
The flow chart of a kind of based on tensor time domain correlation model the video sequence sorting technique that Fig. 1 is.
Detailed description of the invention
Below in conjunction with embodiment and accompanying drawing, a kind of based on tensor time domain correlation model the video sequence of the present invention is classified Method is described in detail.
As it is shown in figure 1, a kind of based on tensor time domain correlation model the video sequence sorting technique of the present invention, including as follows Step:
1) original video sequence is characterized as the form of three rank video tensors;Specifically:
Given a series of video sequence X={X1,…,Xi..., XN, and described video sequence is expressed as one three The form of rank video tensor, wherein:
It is three rank video tensors, wherein I1、I2With T be expressed as a video sequence width, Height and the length of time shaft, N is the quantity of video sequence;Definition It it is one of tensor Isochronous surface.Compared with the method for each frame visual signature in traditional extraction video sequence, video sequence is configured to three The method of rank video tensor can preserve the structural information of original video data and not cause the loss of information, and tensor and tensor Decomposition technique is the most highly developed, also haves laid a good foundation for follow-up algorithm design.
2) three rank video tensors are carried out tensor Tucker and decomposes the core tensor that acquisition is potential;Including:
By finding two mapping matrixes in the framework that tensor Tucker decomposesWithWill be Three first rank video tensorsMigration is a potential core tensorJ is kept in transition process1< I1, J2<I2, the potential core tensor Y that obtainediIt is three initial rank video tensor XiThe representation of a kind of low-dimensional intrinsic, Described transition process is expressed as:
Yi=Xi×1U1×2U2 (1)
Wherein, I1、I2With the length that T is expressed as the width of a video sequence, height and time shaft;J1And J2Also It is expressed as width and the height of a video sequence.
3) the time domain application autoregression model to the potential core tensor obtained sets up the association between adjacent time section Property;Owing to autoregression model shows powerful advantage when sequence analysis time, he can go by method statistically Model the seriality relation between each time point in time series.All have due to frame each in video sequence in the present invention The time domain continuity of height and time dependence, preserve initial video sequence tensor to obtain the character representation of robust simultaneously Time domain continuity and time dependence, we apply autoregression model to be modeled the internal relation of video sequence time domain. Specifically
Comprise the steps:
(1) in order to keep video sequence seriality in time domain between each time point, application autoregression model (AR) Set up isochronous surfaceWith this isochronous surfaceIsochronous surface aboveRelatedness model, by association Property model is with m isochronous surfaceGo to predict next isochronous surfaceDescribed relatedness model is as follows:
Y t i = &Sigma; k = 1 m a k Y t - k i + &epsiv; t i - - - ( 2 )
WhereinIt is white Gaussian noise, meetsN is Gauss distribution,It is the covariance tensor of white noise, MeetVar is variance,It it is the model coefficient of autoregression model;
(2) model parameter in relatedness modelMaximal possibility estimation with Yule-Waler side based on tensor Journey obtains:
&Sigma; k = 1 m a k ( &Sigma; t = 1 T < Y t - k i , Y t - l i > ) = &Sigma; t = 1 T < Y t i , Y t - l i > - - - ( 3 )
(3) by Yi=Xi×1U1×2U2It is applied to relatedness modelIn, and rearrange and obtain Following noise covariance tensor expression:
&epsiv; t i = Y t i - &Sigma; k = 1 m a k Y t - k i = ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 - - - ( 4 )
(4) as long as having estimated model parameterNoise covariance between actual value and the predictive value of isochronous surface The accumulation norm of tensor is obtained by following formula:
&Sigma; t = 2 T || &epsiv; t i || F 2 = &Sigma; t = 2 T || ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 || F 2 - - - ( 5 )
(5) object function is obtained
By minimizing the accumulation norm of noise covariance tensor, obtain goal of function:
min U 1 , U 2 , { a k } &Sigma; i = 1 N &Sigma; t = 2 T || ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 || F 2 s . t . U 1 U 1 T = E U 2 U 2 T = E - - - ( 6 )
Wherein, E representation unit orthogonal matrix;
In order to preserve more information as far as possible and control the yardstick of known variables, we add in object function The restrictive condition of orthogonality.This object function shows that the new video sequence tensor obtained should subtract on time dimension as far as possible Few forecast error is to ensure the seriality on time dimension, and eliminates the noise in original video sequence and redundancy on spatial domain Information.
(6) object function is solved
In order to effectively be solved scheme, typical alternately descent algorithm is used to go to solve object function, the most at random Initialize mapping matrix U1And U2Go to obtain potential core tensor Yi, then with potential core tensor YiGo to calculate relatedness model Model parameterFix mapping matrix U therewith2With potential core tensor Yi, update the mapping matrix U of object function1:
min U 1 &Sigma; i = 1 N &Sigma; t = 2 T || ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; { U } ( - 1 ) &times; U 1 1 || F 2 s . t . U 1 U 1 T = E - - - ( 7 )
OrderObtain the equivalent form of value of object function:
max U 1 t r ( U 1 ( &Sigma; i = 1 N &Sigma; t = 2 T G t 1 i G t 1 i T ) U 1 T ) s . t . U 1 U 1 T = E - - - ( 8 )
Minimize the equivalent form of value of object function by method of Lagrange multipliers and Eigenvalues Decomposition method, then object function is relative In U1Partial derivative calculated by following formula:
&Sigma; i = 1 N &Sigma; t = 2 T G t 1 i G t 1 i T u 1 j T = &lambda; 1 j u 1 j T - - - ( 9 )
U in the formula1jIt it is matrixGeneralized eigenvector, λ1jIt it is corresponding eigenvalue;
Equally, fixing mapping matrix U1, then object function is relative to mapping matrix U2Partial derivative calculated by following formula:
&Sigma; i = 1 N &Sigma; t = 2 T G t 2 i G t 2 i T u 2 j T = &lambda; 2 j u 2 j T - - - ( 10 )
U in formula2jIt it is matrixGeneralized eigenvector, λ2jIt it is corresponding eigenvalue.
4) dynamic learning step 2)~step 3) updating result until algorithmic statement, result reaches optimum.Specifically will step Rapid 2)~step 3) be incorporated in a dynamic learning framework make step 2)~step 3) learning process carry out more over time Newly, until meeting the condition of convergenceWherein, β represents forecast error.
After being modeled processing to the method for the original video sequence present invention and obtaining new video sequence, we adopt With method (Y.M.Lui, J.R.Beveridge, the and M.Kirby, " Action classification on of product manifold product manifolds,”In Proceedings of the International Conference on Computer Vision and Pattern Recognition, 2010, pp.833 839.) the new video sequence obtained is classified. In product manifold, each tensor time series can be mapped as a point in product manifold, can be by calculating survey between points Ground linear distance goes to weigh the distance between tensor.
The effectiveness of the inventive method is verified below with a concrete experiment:
Experiment uses Cambridge gesture database, and this data base comprises the gesture that 9 classes are different altogether, and every class gesture has 100 a total of 900 videos of video.Every kind of gesture is all capture under two kinds of different main body backgrounds, be respectively 5 kinds not Same illumination and 10 kinds of actions.5 subsets it are divided into again according to this data base of different illumination conditions, Set1, Set2, Set3, Set4, Set5.In an experiment we by 20 frames in the middle of each video extraction the size of all video sequences Be adjusted to 20 × 20 × 20, using the Set5 of shooting under normal lighting conditions as training set, other as test set.In addition What Set5 was the most random has been divided into training set and checking collection.Experimental result such as table 1.
Wherein, KAHM, RLPP, DCCA, GDA are based on traditional sub-space learning method;TB, TCCA, PM are based on many The method of linear algebra sub-space learning;MLDS, BPTF are to tie mutually with relativity of time domain based on multilinear algebra sub-space learning The method closed.By table 1 it can be seen that the method for the present invention is significantly better than traditional video sequence sorting technique, precision obtains Arrive the biggest raising.
Table 1

Claims (5)

1. a video sequence sorting technique based on tensor time domain correlation model, it is characterised in that comprise the steps:
1) original video sequence is characterized as the form of three rank video tensors;
2) three rank video tensors are carried out tensor Tucker and decomposes the core tensor that acquisition is potential;
3) the time domain application autoregression model to the potential core tensor obtained sets up the relatedness between adjacent time section;
4) dynamic learning step 2)~step 3) updating result until algorithmic statement, result reaches optimum.
A kind of video sequence sorting technique based on tensor time domain correlation model the most according to claim 1, its feature exists In, step 1) it is given a series of video sequence X={X1,…,Xi..., XN, and described video sequence is expressed as one The form of individual three rank video tensors, wherein:
It is three rank video tensors, wherein I1、I2With the width, highly that T is expressed as a video sequence With the length of time shaft, N is the quantity of video sequence;Definition It it is the time of tensor Section.
A kind of video sequence sorting technique based on tensor time domain correlation model the most according to claim 1, its feature exists In, step 2) including: by finding two mapping matrixes in the framework that tensor Tucker decomposesWithBy three initial rank video tensorsMigration is a potential core tensorMoving J is kept during shifting1<I1, J2<I2, the potential core tensor Y that obtainediIt is three initial rank video tensor XiA kind of low-dimensional The representation of intrinsic, described transition process is expressed as:
Yi=Xi×1U1×2U2 (1)
Wherein, I1、I2With the length that T is expressed as the width of a video sequence, height and time shaft;J1And J2Also table is distinguished It is shown as width and the height of a video sequence.
A kind of video sequence sorting technique based on tensor time domain correlation model the most according to claim 1, its feature exists In, step 3) comprise the steps:
(1) application autoregression model sets up isochronous surfaceWith this isochronous surfaceIsochronous surface above Relatedness model, by relatedness model with m isochronous surfaceGo to predict next isochronous surface Described relatedness model is as follows:
Y t i = &Sigma; k = 1 m a k Y t - k i + &epsiv; t i - - - ( 2 )
WhereinIt is white Gaussian noise, meetsN is Gauss distribution,It is the covariance tensor of white noise, meetsVar is variance,It it is the model coefficient of autoregression model;
(2) model parameter in relatedness modelMaximal possibility estimation obtain with Yule-Waler equation based on tensor Arrive:
&Sigma; k = 1 m a k ( &Sigma; t = 1 T < Y t - k i , Y t - l i > ) = &Sigma; t = 1 T < Y t i , Y t - l i > - - - ( 3 )
(3) by Yi=Xi×1U1×2U2It is applied to relatedness modelIn, and rearrange obtain following Noise covariance tensor expression:
&epsiv; t i = Y t i - &Sigma; k = 1 m a k Y t - k i = ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 - - - ( 4 )
(4) as long as having estimated model parameterNoise covariance tensor between actual value and the predictive value of isochronous surface Accumulation norm obtained by following formula:
&Sigma; t = 2 T | | &epsiv; t i | | F 2 = &Sigma; t = 2 T | | ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 | | F 2 - - - ( 5 )
(5) object function is obtained
By minimizing the accumulation norm of noise covariance tensor, obtain goal of function:
m i n U 1 , U 2 , { a k } &Sigma; i = 1 N &Sigma; t = 2 T | | ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; U 1 1 &times; U 2 2 | | F 2 - - - ( 6 )
s . t . U 1 U 1 T = E , U 2 U 2 T = E
Wherein, E representation unit orthogonal matrix;
(6) object function is solved
Alternately descent algorithm is used to go to solve object function, first random initializtion mapping matrix U1And U2Go to obtain potential core Tensor Yi, then with potential core tensor YiGo to calculate the model parameter of relatedness modelFix mapping matrix U therewith2 With potential core tensor Yi, update the mapping matrix U of object function1:
m i n U 1 &Sigma; i = 1 N &Sigma; t = 2 T | | ( X t i - &Sigma; k = 1 m a k X t - k i ) &times; { U } ( - 1 ) &times; U 1 1 | | F 2 - - - ( 7 )
s . t . U 1 U 1 T = E
OrderObtain the equivalent form of value of object function:
m a x U 1 t r ( U 1 ( &Sigma; i = 1 N &Sigma; t = 2 T G t 1 i G t 1 i T ) U 1 T ) - - - ( 8 )
s . t . U 1 U 1 T = E
Minimize the equivalent form of value of object function by method of Lagrange multipliers and Eigenvalues Decomposition method, then object function is relative to U1 Partial derivative calculated by following formula:
&Sigma; i = 1 N &Sigma; t = 2 T G t 1 i G t 1 i T u 1 j T = &lambda; 1 j u 1 j T - - - ( 9 )
U in the formula1jIt it is matrixGeneralized eigenvector, λ1jIt it is corresponding eigenvalue;
Equally, fixing mapping matrix U1, then object function is relative to mapping matrix U2Partial derivative calculated by following formula:
&Sigma; i = 1 N &Sigma; t = 2 T G t 2 i G t 2 i T u 2 j T = &lambda; 2 j u 2 j T - - - ( 10 )
U in formula2jIt it is matrixGeneralized eigenvector, λ2jIt it is corresponding eigenvalue.
A kind of video sequence sorting technique based on tensor time domain correlation model the most according to claim 1, its feature exists In, step 4) be by step 2)~step 3) be incorporated in a dynamic learning framework make step 2)~step 3) learning process It is updated over time, until meeting the condition of convergenceWherein, β represents forecast error.
CN201610237984.4A 2016-04-15 2016-04-15 Video sequence classifying method based on tensor time domain association model Pending CN105956603A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610237984.4A CN105956603A (en) 2016-04-15 2016-04-15 Video sequence classifying method based on tensor time domain association model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610237984.4A CN105956603A (en) 2016-04-15 2016-04-15 Video sequence classifying method based on tensor time domain association model

Publications (1)

Publication Number Publication Date
CN105956603A true CN105956603A (en) 2016-09-21

Family

ID=56918057

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610237984.4A Pending CN105956603A (en) 2016-04-15 2016-04-15 Video sequence classifying method based on tensor time domain association model

Country Status (1)

Country Link
CN (1) CN105956603A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106485212A (en) * 2016-09-26 2017-03-08 天津大学 A kind of resolution of tensor method for non-isometric video gesture identification
CN106503659A (en) * 2016-10-24 2017-03-15 天津大学 Action identification method based on sparse coding tensor resolution
CN106529435A (en) * 2016-10-24 2017-03-22 天津大学 Action recognition method based on sensor quantization
CN107122745A (en) * 2017-04-28 2017-09-01 深圳市茁壮网络股份有限公司 The method and device of personage track in a kind of identification video
CN111639243A (en) * 2020-06-04 2020-09-08 东北师范大学 Space-time data progressive multi-dimensional mode extraction and anomaly detection visual analysis method
CN111797849A (en) * 2019-04-09 2020-10-20 Oppo广东移动通信有限公司 User activity identification method and device, storage medium and electronic equipment
CN112069161A (en) * 2020-09-01 2020-12-11 上海佰贝科技发展股份有限公司 Data cleaning method and device
CN112801142A (en) * 2021-01-08 2021-05-14 北京工业大学 Tensor autoregressive moving average model-based video classification method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104778482A (en) * 2015-05-05 2015-07-15 西安电子科技大学 Hyperspectral image classifying method based on tensor semi-supervised scale cutting dimension reduction
CN104992187A (en) * 2015-07-14 2015-10-21 西安电子科技大学 Aurora video classification method based on tensor dynamic texture model

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104778482A (en) * 2015-05-05 2015-07-15 西安电子科技大学 Hyperspectral image classifying method based on tensor semi-supervised scale cutting dimension reduction
CN104992187A (en) * 2015-07-14 2015-10-21 西安电子科技大学 Aurora video classification method based on tensor dynamic texture model

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Y SONG ET AL.: "Describing trajectory of surface patch for human action recognition on rgb and depth videos", 《IEEE SIGNAL PROCESSING LETTERS》 *
郭炜炜: "基于张量表示的多维信息处理方法研究", 《中国博士学位论文全文数据库 信息科技辑》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106485212B (en) * 2016-09-26 2020-01-24 天津大学 Tensor decomposition method for non-isometric video gesture recognition
CN106485212A (en) * 2016-09-26 2017-03-08 天津大学 A kind of resolution of tensor method for non-isometric video gesture identification
CN106503659A (en) * 2016-10-24 2017-03-15 天津大学 Action identification method based on sparse coding tensor resolution
CN106529435A (en) * 2016-10-24 2017-03-22 天津大学 Action recognition method based on sensor quantization
CN106529435B (en) * 2016-10-24 2019-10-15 天津大学 Action identification method based on tensor quantization
CN106503659B (en) * 2016-10-24 2019-10-15 天津大学 Action identification method based on sparse coding tensor resolution
CN107122745A (en) * 2017-04-28 2017-09-01 深圳市茁壮网络股份有限公司 The method and device of personage track in a kind of identification video
CN107122745B (en) * 2017-04-28 2020-10-20 深圳市茁壮网络股份有限公司 Method and device for identifying person track in video
CN111797849A (en) * 2019-04-09 2020-10-20 Oppo广东移动通信有限公司 User activity identification method and device, storage medium and electronic equipment
CN111639243A (en) * 2020-06-04 2020-09-08 东北师范大学 Space-time data progressive multi-dimensional mode extraction and anomaly detection visual analysis method
CN111639243B (en) * 2020-06-04 2021-03-09 东北师范大学 Space-time data progressive multi-dimensional mode extraction and anomaly detection visual analysis method
CN112069161A (en) * 2020-09-01 2020-12-11 上海佰贝科技发展股份有限公司 Data cleaning method and device
CN112069161B (en) * 2020-09-01 2023-11-03 上海佰贝科技发展股份有限公司 Data cleaning method and device
CN112801142A (en) * 2021-01-08 2021-05-14 北京工业大学 Tensor autoregressive moving average model-based video classification method
CN112801142B (en) * 2021-01-08 2024-05-28 北京工业大学 Video classification method based on tensor autoregressive moving average model

Similar Documents

Publication Publication Date Title
CN105956603A (en) Video sequence classifying method based on tensor time domain association model
CN108229444B (en) Pedestrian re-identification method based on integral and local depth feature fusion
Zhu et al. Intelligent logging lithological interpretation with convolution neural networks
Feng et al. Hyperspectral band selection based on trivariate mutual information and clonal selection
CN107132516B (en) A kind of Radar range profile&#39;s target identification method based on depth confidence network
Krasnopolsky et al. Using ensemble of neural networks to learn stochastic convection parameterizations for climate and numerical weather prediction models from data simulated by a cloud resolving model
CN110503128A (en) The spectrogram that confrontation network carries out Waveform composition is generated using convolution
CN107798349B (en) Transfer learning method based on depth sparse self-coding machine
CN108133188A (en) A kind of Activity recognition method based on motion history image and convolutional neural networks
CN110321926A (en) A kind of moving method and system based on depth residual GM network
CN108182260B (en) Multivariate time sequence classification method based on semantic selection
CN108121975A (en) A kind of face identification method combined initial data and generate data
CN105701480A (en) Video semantic analysis method
CN105678261B (en) Based on the direct-push Method of Data with Adding Windows for having supervision figure
CN106886793B (en) Hyperspectral image waveband selection method based on discrimination information and manifold information
CN109034228A (en) A kind of image classification method based on difference privacy and level relevance propagation
CN106503659A (en) Action identification method based on sparse coding tensor resolution
CN109101717A (en) Solid propellant rocket Reliability Prediction Method based on reality with the study of fuzzy data depth integration
CN105354532A (en) Hand motion frame data based gesture identification method
CN110096976A (en) Human behavior micro-Doppler classification method based on sparse migration network
CN106021402A (en) Multi-modal multi-class Boosting frame construction method and device for cross-modal retrieval
CN106803105B (en) Image classification method based on sparse representation dictionary learning
Zhang et al. Multi-source information fused generative adversarial network model and data assimilation based history matching for reservoir with complex geologies
Li et al. Soil seismic response modeling of KiK-net downhole array sites with CNN and LSTM networks
CN114399661A (en) Instance awareness backbone network training method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
AD01 Patent right deemed abandoned
AD01 Patent right deemed abandoned

Effective date of abandoning: 20191022