CN100570640C - A kind of method for expressing of people's motion - Google Patents
A kind of method for expressing of people's motion Download PDFInfo
- Publication number
- CN100570640C CN100570640C CNB2007101753713A CN200710175371A CN100570640C CN 100570640 C CN100570640 C CN 100570640C CN B2007101753713 A CNB2007101753713 A CN B2007101753713A CN 200710175371 A CN200710175371 A CN 200710175371A CN 100570640 C CN100570640 C CN 100570640C
- Authority
- CN
- China
- Prior art keywords
- motion
- people
- model
- dimensional
- linear
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000000034 method Methods 0.000 title claims abstract description 28
- 230000009467 reduction Effects 0.000 claims abstract description 16
- 239000011159 matrix material Substances 0.000 claims abstract description 12
- 230000002123 temporal effect Effects 0.000 claims abstract description 12
- 238000013507 mapping Methods 0.000 claims abstract description 8
- 230000008569 process Effects 0.000 claims abstract description 4
- 239000013598 vector Substances 0.000 claims description 9
- 238000011160 research Methods 0.000 description 5
- 238000012549 training Methods 0.000 description 4
- 241001123248 Arma Species 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000005021 gait Effects 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000036624 brainpower Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
Landscapes
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a kind of method for expressing of people's motion, will come on the embedded space of people's Motion mapping to a low-dimensional by Nonlinear Dimension Reduction; With the linear temporal model modeling of the data behind the dimensionality reduction.Wherein will come on the configuration space of the embedding of people's Motion mapping to a low-dimensional by Nonlinear Dimension Reduction; Be embodied as: z=(x
1+ jy
1, x
2+ jy
2..., x
M+ jy
M)
TWith the linear temporal model modeling of the data behind the dimensionality reduction; Performing step: based on the motion modeling of linear session series model; The autoregressive model (AR) of setting up p rank comprises; The autoregressive model on p rank (AR) AR (p) has following parameter: coefficient matrices A
k∈ R
M * m, be the parameter v that non-zero is introduced in order to guarantee the dynamic process average, the covariance matrix Q of white Gaussian noise establishes A
kBeing diagonal matrix, is separate between each component of z (t) then; Given two autoregressive models (AR) A=[v, A
1, A
2..., A
p] and A '=[v ', A
1', A
2' ..., A
p'], distance metric D (A, A ')=‖ A-A ' ‖ then
F, ‖ ‖ wherein
FThe F-norm of representing matrix.
Description
Technical field
The present invention relates to a kind of method for expressing of people's motion, belong to computer vision field and video content intellectual analysis field.
Background technology
In computer vision field and the video content intellectual analysis field, human motion analysis has become the research topic in an extremely important and forward position
[1-7]In human motion analysis, people's motion detection, tracking belong to the processing of low level in the vision, and people's motion is expressed and understanding belongs to high-level processing.People's motion is expressed and is understood and play crucial effects in applications such as video brainpower watch and control.
In the more than ten years in past, a lot of movement representation methods have been emerged about the people
[1-6]Wherein, most research work movable information of extracting static information in every two field picture or consecutive frame is represented people's motion.People such as Efros
[1]Adopt light stream to come the movable information of expressing human, and remote motion is analyzed.Arie
[2]Adopt 3D column model to come volume modeling Deng the people, in every frame, extract angle and angular velocity thereof between health various piece and the transverse axis for the people.These method for expressing have only reflected the spatial character of motion in every frame, do not embody the dynamic perfromance of motion in a period of time well, and the data volume that expression needs is very big.In recent years, there was a spot of research work to adopt the linear temporal model to study the dynamic perfromance of motion.Liu and Ahuja
[4]Propose a kind of dynamic shape model and represented the variation of target in a period of time, the profile employing Fourier descriptors (FD) of target has been represented, adopted autoregression (AR) model to come the dynamic perfromance of evaluating objects motion then.People such as Veeraraghavan
[5], Jin and Nokhtarian
[6]Adopt the Kendall shape theory to come the profile of expressing human, adopt autoregression (AR) model and autoregressive moving average (ARMA) model to come people's gait and action are analyzed then.The shortcoming of these methods is that the data dimension is high especially before adopting the linear temporal model analysis, need more training data when modeling, and the estimation of parameter is also not accurate enough.
The movable information that static information in every two field picture or consecutive frame are extracted in most research work is represented people's motion.People such as Efros [1] adopt light stream to come the movable information of expressing human, and remote motion is analyzed.People such as Arie [2] adopt 3 dimension column models to come the volume modeling for the people, extract angle and angular velocity thereof between health various piece and the transverse axis in every frame.
These method for expressing have only reflected the spatial character of motion in every frame, do not embody the dynamic perfromance of motion in a period of time well, and the data volume that expression needs is very big.
In recent years, there was a spot of research work to adopt the linear temporal model to study the dynamic perfromance of motion.Liu and Ahuja[4] propose a kind of dynamic shape model and represented the variation of target in a period of time, adopt Fourier descriptors (FD) to represent the profile of target, adopt the AR model to come the dynamic perfromance of evaluating objects motion then.People such as Veeraraghavan [5], Jin and Nokhtarian[6] adopt the Kendall shape theory to come the profile of expressing human, adopt AR model and arma modeling to come people's gait and action are analyzed then.
The shortcoming of these methods is that the data dimension is high especially before adopting the linear temporal model analysis, need more training data when modeling, and the estimation of parameter is also not accurate enough.
Summary of the invention
Fundamental purpose of the present invention is, by a kind of method for expressing of people's motion is provided, to solve at present in computer vision field and video content intellectual analysis field, the data dimension is high especially before the employing linear temporal model analysis that exists, when modeling, need more training data, and the also not accurate enough problem of the estimation of parameter.
In people's the motion, each attitude is represented by the vector of a higher-dimension that all the distribution of these vectors in feature space presents severe nonlinear.For the data that in higher dimensional space, present nonlinear Distribution, adopt the Nonlinear Dimension Reduction method to come the inner relation that exists of learning data to seem extremely important.This patent has proposed a kind of compact expression of people's motion.We at first adopt a kind of method of Nonlinear Dimension Reduction---and local linear algorithm (LLE) algorithm [8] that embeds is mapped to the embedded space of low-dimensional with the feature space of higher-dimension, and this helps the immanent structure that the finder moves.In lower dimensional space, come motion is analyzed then by linear temporal model AR model and arma modeling.
The present invention adopts following technological means to realize:
A kind of method for expressing of people's motion: will come on the embedded space of people's Motion mapping to a low-dimensional by Nonlinear Dimension Reduction; With the linear temporal model modeling of the data behind the dimensionality reduction;
Wherein saidly will be on the configuration space of the embedding of people's Motion mapping to a low-dimensional come by Nonlinear Dimension Reduction; Realize by following steps:
The expression of described configuration space; May further comprise the steps:
Adopt M gauge point P={p
1, p
2..., p
MPeople's outline is described, then each profile can be represented with complex vector z:
z=(x
1+jy
1,x
2+jy
2,…,x
M+jy
M)
T
Wherein: x
iAnd y
iRepresent i gauge point p respectively
iHorizontal ordinate and ordinate; The expression of this people's attitude need have unchangeability for position and isotropic dimensional variation, and vector Z is normalized to z '; The real part of described z ' and imaginary part constitute profile;
The immanent structure that adopts the local linear motion feature that embeds algorithm (LLE) finder in high-dimensional feature space, to hide; Suppose that the input sample set is Z, then the point set Y of Dui Ying low-dimensional embedded space can obtain by the local linear algorithm (LLE) that embeds;
Determine the dimension e of embedded space; Wherein said with the linear temporal model modeling of the data behind the dimensionality reduction; Realize by following steps: based on the motion modeling of linear session series model; The autoregressive model (AR) of setting up p rank comprises; The autoregressive model on p rank (AR) AR (p) has following parameter: coefficient matrices A
k∈ R
M * m, be the parameter v that non-zero is introduced in order to guarantee the dynamic process average, the covariance matrix Q of white Gaussian noise establishes A
kBeing diagonal matrix, is separate between each component of z (t) then; Given two autoregressive models (AR) A=[v, A
1, A
2..., A
p] and A '=[v ', A
1', A
2' ..., A
p'], distance metric then:
D (A, A ')=‖ A-A ' ‖
F, ‖ ‖ wherein
FThe F-norm of representing matrix.
Aforesaid motion modeling based on the linear session series model, can also adopt following steps to realize:
Set up autoregressive moving-average model (ARMA):
A given observation sequence z (1), z (2) ..., z (τ) obtains the model parameter of limited sample sequence correspondence by the method for maximal possibility estimation;
Distance between the autoregressive moving-average model can be measured with the subspace angle; The angle of two autoregressive moving-average model models is θ
i(i=1,2 ..., n); Usually there are three distances can be used for calculating: Martin distance (d
M), gap distance (d
g) and Frobenius distance (d
F).
The present invention compared with prior art has following remarkable advantages and beneficial effect:
The present invention has promptly reduced and is used for representing the data dimension that the people moves, and has given expression to the dynamic perfromance of people's motion again, has great significance for the classification and the understanding of motion.
Embodiment
Down to specific embodiments of the invention are illustrated:
Step 1: to the expression of configuration space: the gesticulate that at first is based on profile;
Profile is a kind of method of good expression people attitude, and it is insensitive for the color of the variation on people surface such as clothes and texture etc.Adopt M gauge point P={p
1, p
2..., p
MPeople's outline is described, then each profile can be represented with complex vector z: z=(x
1+ jy
1, x
2+ jy
2..., x
M+ jy
M)
T, x wherein
iAnd y
iRepresent i gauge point p respectively
iHorizontal ordinate and ordinate.The expression of this people's attitude need have unchangeability for position and isotropic dimensional variation, so Z is normalized to z '.The representation that constitutes profile by real part and the imaginary part of complex vector z ' then.
The Nonlinear Mapping of profile below is described:
Local linear embedding algorithm (LLE) is a kind of good Nonlinear Dimension Reduction method, and it can make the neighborhood relationships of dimensionality reduction front and back data remain unchanged.Attitude in people's the motion in a certain frame usually and the attitude in adjacent several frames have substantial connection, and and other attitude relation not tight, so the local linear motion that algorithm (LLE) is very suitable for the analyst that embeds.
This patent adopts the local linear algorithm (LLE) that embeds to come the immanent structure that finder's motion feature hides and learn its stream shape at the low-dimensional embedded space in high-dimensional feature space.Suppose that the input sample set is Z, then the point set Y of Dui Ying low-dimensional embedded space can learn to obtain by the local linear algorithm (LLE) that embeds.
After obtaining the local linear result who embeds algorithm (LLE) on the training set, it is necessary learning a mapping in from the higher dimensional space to the lower dimensional space again.There is certain methods to address this problem at present.
Local linear the embedding in the algorithm (LLE) has an important parameters---and the dimension e of embedded space need determine.Occurred a lot of methods at present, this patent adopts the method for Eigenvalue Analysis to solve.
Step 2: based on the motion modeling of linear session series model:
How people's attitude is changed show very important.If learn a model that can reflect people's attitude variation exactly, we just can obtain a good movement representation.Because people's motion more complicated, and become when being, be difficult to it and set up accurate, unified analytic model.It is motion modeling that this patent adopts the linear session model approximation, and its parameter can reflect the dynamic perfromance that the people moves preferably.This patent adopts following two kinds of methods to analyze.
A) autoregressive model (AR)
The autoregressive model on p rank (AR) AR (p) has following parameter: coefficient matrices A
k∈ R
M * m, be the parameter v that non-zero is introduced in order to guarantee the dynamic process average, the covariance matrix Q of white Gaussian noise.In order to simplify complexity, we suppose that Ak is a diagonal matrix, that is to say that between each component of z (t) be separate.
This patent adopts Neumaier and Schneider algorithm to learn autoregressive model (AR), and this algorithm can guarantee the uniqueness of parameter estimation.Given two autoregressive models (AR) A=[v, A
1, A
2..., A
p] and A '=[v ', A
1', A
2' ..., A
p'], this patent adopts following distance metric:
D (A, A ')=‖ A-A ' ‖
F, ‖ ‖ wherein
FThe F-norm of representing matrix.
B) autoregressive moving-average model (ARMA)
Autoregressive moving-average model (ARMA) and autoregressive model (AR) are similar.A given observation sequence z (1), z (2) ..., z (τ) obtains the model parameter of limited sample sequence correspondence by the method for maximal possibility estimation.
Distance between the autoregressive moving-average model (ARMA) can be measured with the subspace angle.The angle of two autoregressive moving-average model models is θ
i(i=1,2 ..., n).Usually there are three distances can be used for calculating: Martin distance (d
M), gap distance (d
g) and Frobenius distance (d
F).
It should be noted that at last: above embodiment only in order to the explanation the present invention and and unrestricted technical scheme described in the invention; Therefore, although this instructions has been described in detail the present invention with reference to each above-mentioned embodiment,, those of ordinary skill in the art should be appreciated that still and can make amendment or be equal to replacement the present invention; And all do not break away from the technical scheme and the improvement thereof of the spirit and scope of invention, and it all should be encompassed in the middle of the claim scope of the present invention.
Claims (2)
1, a kind of method for expressing of people's motion is characterized in that;
To come on the embedding configuration space of people's Motion mapping to a low-dimensional by Nonlinear Dimension Reduction;
With the linear temporal model modeling of the data behind the dimensionality reduction;
Wherein saidly will be on the embedding configuration space of people's Motion mapping to a low-dimensional come, realize by following two steps by Nonlinear Dimension Reduction:
The 1st step was based on the gesticulate of profile, comprised following two steps:
At first, adopt M gauge point P={p
1, p
2..., p
MPeople's outline is described, each profile can be represented with complex vector z:
z=(x
1+jy
1,x
2+jy
2,…,x
M+jy
M)
T
X wherein
iAnd y
iRepresent i gauge point p respectively
iHorizontal ordinate and ordinate;
Secondly, vectorial z is normalized to complex vector z ', makes the expression of people's attitude have unchangeability for position and isotropic dimensional variation, the real part of described complex vector z ' and imaginary part constitute profile;
The 2nd step was non-linear dimensionality reduction, the immanent structure that adopts the local linear motion feature that embeds algorithm (LLE) finder to hide in high-dimensional feature space; Suppose that the input sample set is Z, the point set Y that then corresponding low-dimensional embeds configuration space can obtain by the local linear algorithm (LLE) that embeds; Determine the dimension e of the embedding configuration space of low-dimensional;
Wherein said with the linear temporal model modeling of the data behind the dimensionality reduction, adopt the linear session series model to be approximately motion modeling, specifically realize by following steps:
Set up the autoregressive model on p rank, the autoregressive model AR (p) on one of them p rank has following parameter:
Coefficient matrices A
k∈ R
M * m, be parameter v, the covariance matrix Q of white Gaussian noise that non-zero is introduced in order to guarantee the dynamic process average;
If A
kBe diagonal matrix, then separate between each component among the observation sequence z (t);
Given two autoregressive models (AR) A=[v, A
1, A
2..., A
p] and A '=[v ', A
1', A
2' ..., A
p'], it is apart from adopting following tolerance:
D (A, A ')=|| A-A ' ||
F, wherein || ‖
FThe F-norm of representing matrix; This distance is used for measuring two similaritys between motion, thereby the motion of determining the people is expressed.
2, the method for expressing of a kind of people's motion according to claim 1 is characterized in that described employing linear session series model is approximately motion modeling and can also adopts following steps to realize:
Set up autoregressive moving-average model (ARMA):
A given observation sequence z (1), z (2) ..., z (τ) obtains the model parameter of limited sample sequence correspondence by the method for maximal possibility estimation;
Distance between the autoregressive moving-average model can be measured with the subspace angle; The angle of two autoregressive moving-average models is θ
i(i=1,2 ..., n); Usually there are three distances can be used for calculating: Martin distance (d
M), gap distance (d
g) and Frobenius distance (d
F).
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNB2007101753713A CN100570640C (en) | 2007-09-29 | 2007-09-29 | A kind of method for expressing of people's motion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNB2007101753713A CN100570640C (en) | 2007-09-29 | 2007-09-29 | A kind of method for expressing of people's motion |
Publications (2)
Publication Number | Publication Date |
---|---|
CN101123004A CN101123004A (en) | 2008-02-13 |
CN100570640C true CN100570640C (en) | 2009-12-16 |
Family
ID=39085329
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CNB2007101753713A Expired - Fee Related CN100570640C (en) | 2007-09-29 | 2007-09-29 | A kind of method for expressing of people's motion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN100570640C (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110598733A (en) * | 2019-08-05 | 2019-12-20 | 南京智谷人工智能研究院有限公司 | Multi-label distance measurement learning method based on interactive modeling |
-
2007
- 2007-09-29 CN CNB2007101753713A patent/CN100570640C/en not_active Expired - Fee Related
Non-Patent Citations (2)
Title |
---|
基于视觉的人的运动识别综述. 杜友田,陈峰,徐文立,李永彬.电子学报,第35卷第1期. 2007 * |
流形学习概述. 徐容,姜峰,姚鸿勋.智能系统学报,第1卷第1期. 2006 * |
Also Published As
Publication number | Publication date |
---|---|
CN101123004A (en) | 2008-02-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104182718B (en) | A kind of man face characteristic point positioning method and device | |
Kishore et al. | Optical flow hand tracking and active contour hand shape features for continuous sign language recognition with artificial neural networks | |
CN106778628A (en) | A kind of facial expression method for catching based on TOF depth cameras | |
CN107229920B (en) | Behavior identification method based on integration depth typical time warping and related correction | |
CN114463677B (en) | Safety helmet wearing detection method based on global attention | |
CN110827312B (en) | Learning method based on cooperative visual attention neural network | |
CN104091350B (en) | A kind of object tracking methods of utilization motion blur information | |
CN104821010A (en) | Binocular-vision-based real-time extraction method and system for three-dimensional hand information | |
CN108470178B (en) | Depth map significance detection method combined with depth credibility evaluation factor | |
CN110060286A (en) | A kind of monocular depth estimation method | |
Amrutha et al. | Human Body Pose Estimation and Applications | |
Zhang et al. | A Gaussian mixture based hidden Markov model for motion recognition with 3D vision device | |
Ling et al. | Stableface: Analyzing and improving motion stability for talking face generation | |
Kan et al. | Background modeling method based on improved multi-Gaussian distribution | |
Kang et al. | Event Camera-Based Pupil Localization: Facilitating Training With Event-Style Translation of RGB Faces | |
Nagori et al. | Communication interface for deaf-mute people using microsoft kinect | |
CN100570640C (en) | A kind of method for expressing of people's motion | |
Zhang et al. | Tcdm: Transformational complexity based distortion metric for perceptual point cloud quality assessment | |
CN110852335B (en) | Target tracking system based on multi-color feature fusion and depth network | |
Schachner et al. | Extracting Quantitative Descriptions of Pedestrian Pre-crash Postures from Real-World AccidentVideos | |
Jiang et al. | Observation-oriented silhouette-aware fast full body tracking with Kinect | |
CN115220574A (en) | Pose determination method and device, computer readable storage medium and electronic equipment | |
Tian et al. | Robust facial marker tracking based on a synthetic analysis of optical flows and the YOLO network | |
An et al. | MTAtrack: Multilevel transformer attention for visual tracking | |
Kumar Bhunia et al. | Person Image Synthesis via Denoising Diffusion Model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
C17 | Cessation of patent right | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20091216 Termination date: 20110929 |