CN105046195A - Human behavior identification method based on asymmetric generalized Gaussian distribution model (AGGD) - Google Patents

Human behavior identification method based on asymmetric generalized Gaussian distribution model (AGGD) Download PDF

Info

Publication number
CN105046195A
CN105046195A CN201510313321.1A CN201510313321A CN105046195A CN 105046195 A CN105046195 A CN 105046195A CN 201510313321 A CN201510313321 A CN 201510313321A CN 105046195 A CN105046195 A CN 105046195A
Authority
CN
China
Prior art keywords
video
formula
sigma
space
delta
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510313321.1A
Other languages
Chinese (zh)
Other versions
CN105046195B (en
Inventor
李俊峰
方建良
张飞燕
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University of Technology ZJUT
Original Assignee
Zhejiang University of Technology ZJUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University of Technology ZJUT filed Critical Zhejiang University of Technology ZJUT
Priority to CN201510313321.1A priority Critical patent/CN105046195B/en
Publication of CN105046195A publication Critical patent/CN105046195A/en
Application granted granted Critical
Publication of CN105046195B publication Critical patent/CN105046195B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a human behavior identification method based on an asymmetric generalized Gaussian distribution model (AGGD). First of all, a training video is preprocessed, a spatio-temporal interest point of the video is detected, then by taking the interest point as a center, a video block is extracted, optical flow information and gradient information are calculated, a corresponding histogram is drafted according to the obtained optical flow information and the gradient information, then, the corresponding histogram is fitted by use of the AGGD, and AGGD parameters of the optical flow information and the gradient information are taken as features so as to form a feature matrix for training the video. According to the invention, the above processing is also performed on a test video in the same way so that a feature matrix of the test video is obtained. Finally, the mahalanobis distance between the feature matrix of the training video and the feature matrix of the test video is calculated, and then the behavior of the test video is identified according to a nearest neighbor principle. The method provided by the invention improves the accuracy of the behavior of a video to be identified to a quite large degree.

Description

Based on the Human bodys' response method of asymmetric generalized gaussian model
Technical field
The present invention relates to a kind of method of Human bodys' response, belong to computer vision and machine learning field, is a kind of Human bodys' response algorithm specifically.
Background technology
In recent years, the research and apply of Intelligent Video Surveillance Technology enjoys the concern of people.As its base conditioning part, Activity recognition is a very active research direction, belongs to the important research content of computer vision field.
Two classes can be divided into: based on the Activity recognition method of global characteristics and local feature according to current research method.Global characteristics adopts the information such as edge, light stream, outline profile to be described the whole interested human region detected usually, more responsive to noise, visual angle change, partial occlusion.Adaptive updates background model is carried out as utilized mixed Gauss model, after sport foreground is extracted to video sequence, zone marker is carried out to prospect, then employing Lucas-Kanade optical flow computation method obtains the Optic flow information in moving region, uses the weighting direction histogram based on amplitude to describe behavior; Double-background model is utilized to carry out adaptive updates background model in addition, after extracting the sport foreground of video sequence, adopt Lucas-Kanade optical flow computation method to calculate Optic flow information to the minimum adjacent rectangular area in prospect, utilize the unit weighting light stream energy of moving target to carry out Activity recognition; Some scholar proposes the Optic flow information first extracted in video, then an empirical covariance matrix is utilized to obtain a covariance descriptor to its dimensionality reduction, being mapped to vector space by taking the logarithm to it, utilizing the logarithm covariance descriptor finally obtained to carry out Activity recognition; Be extracted one in addition and carry out Describing Motion behavior, without any need for human body segmentation and background subtraction during this feature extraction towards the histogrammic feature of Optical-flow Feature; Proposing a sampling system Activity recognition method based on depth image 3D articulation point Sampling characters bag, describing human body behavior by extracting the 3D articulation point characterizing human body attitude from range image sequence; What have is then propose a kind of motor pattern analytical approach that uses to carry out the detection of abnormal behaviour, it generates a kind of motion model by the light stream calculating video and defines track, then utilize space time information to carry out hierarchical cluster to learn this statistics motor pattern to track, finally utilize this statistical model to carry out abnormality detection; .
Local feature is described after extracting interested piece or point in human body, do not need accurately to locate human body and follow the tracks of, and local feature is insensitive to screening part gear, visual angle change etc.Therefore in Activity recognition, local feature frequency of utilization is higher.From compression of video sequence, such as extract quantization parameter and motion vector as feature; Utilize 3D-HOG characteristic sum Optical-flow Feature to describe video behavior; Some is then extracted 3D-SIFT feature in video sequence; And the space-time cube describing together and extract from video sequence that HOG characteristic sum HOF integrate features is got up had; From video, extract space-time word feature bag, then utilize the latent Di Li Cray apportion model of mark as sorter to carry out Activity recognition; Propose a kind of fast density track Activity recognition method, by the region of interesting extraction density trace feature in frame of video, then utilize time pyramid to realize the velocity adaptive mechanism of different action to carry out Activity recognition; More accurate Optical-flow Feature is calculated utilize Harris Corner Detection Algorithm detected image point of interest after removing image background through pre-service after; Light stream detection method is first used to detect position and the direction of motion, the method of being known together by random sampling is further located and motion the most outstanding in the framework that identifies, then according to the mean difference of position, point of interest horizontal and vertical direction in optical flow field and a little rectangular area of standard deviation location human motion, this little rectangular area is divided into several pieces, light stream is calculated frame by frame according to point of interest, synthesize a matrix again, on average represent the behavior again after the matrix of identical behavior is added, finally utilize simple sorter to carry out Activity recognition.
How obtain from image sequence the feature of effective expression body motion information to be the key of Human bodys' response.Optical-flow Feature is reasonable space-time characteristic, is also a kind of motion feature often used during motion identifies.In said method, be all after extraction video motion prospect, foreground moving region is marked, then optical flow computation is carried out to it; Some is then calculate light stream to regular-shape motion area dividing again after detecting whole human motion region.For the various different behavior of human body, the Optic flow information of the unconspicuous body part that moves is negligible, and said method needs the light stream calculating whole human region, which not only adds calculated amount, but also can reduce accuracy of identification.Simultaneously for space-time characteristic, space-time characteristic descriptor constructs word bag code book again after carrying out PCA dimensionality reduction, carry out cluster calculation generation " dictionary " again after namely sampling to training data, and this method makes training sample not to be fully utilized; And in order to ensure certain average recognition rate, even if dimensionality reduction, sample data amount is still too high, and cluster speed is slow.In addition, may there is certain similarity in the characteristic of all directions, and all directions together cluster will reduce descriptive to behavior of different directions feature.
Summary of the invention
The technical problem to be solved in the present invention is to provide a kind of Human bodys' response method based on asymmetric generalized gaussian model.
In order to solve the problems of the technologies described above, the invention provides a kind of Human bodys' response method based on asymmetric generalized gaussian model, be realized by training video storehouse and test video; Comprise the following steps: step one, respectively point of interest detection is carried out for given training video storehouse and test video; Step 2, extracts video block centered by point of interest; Step 3, respectively the video block information of calculation training video and test video, and obtain respective X, Y, Z tri-direction gradient feature data and Optical-flow Feature u, v two component data; Step 4, draws gradient three direction histogram and light stream two direction histogram respectively to above-mentioned data; Step 5, carrys out the corresponding histogram of matching with asymmetric generalized gaussian model; Step 6, the parameter extracting asymmetric generalized gaussian model forms the eigenmatrix of each behavior of training video and the eigenmatrix of test video as feature; Step 7, calculates the mahalanobis distance between test video eigenmatrix and each behavioural characteristic matrix of training video; Step 8, carries out Activity recognition according to nearest neighbouring rule.
Improvement as to the Human bodys' response method based on asymmetric generalized gaussian model of the present invention: in described step, point of interest detects as follows: the image sequence f (x, y, t) video being regarded as multiple image composition; Defined function f: f obtains L after adding linear space yardstick: by image sequence f through separate space variable be with time variable be gaussian function carry out convolutional filtering and obtain, expression formula is as follows:
Gauss's window in time-space domain be defined as:
g ( x , y , t ; σ l 2 ; τ l 2 ) = 1 ( 2 π ) 3 σ l 4 τ l 2 × exp ( - ( x 2 + y 2 ) 2 2 σ l 2 - t 2 2 τ l 2 ) - - - ( 10 )
σ in formula lfor space scale variable, τ lfor time scale variable, t is time dimension; Definition response function R is:
R(x,y,t)=(I*g*h ev) 2+(I*g*h od) 2(11)
In formula, * is convolution algorithm symbol, and I is video image, and g is dimensional Gaussian smoothing kernel, h evand h odit is one dimension Gabor filter orthogonal in spatial domain; h evand h odbe defined as:
h e v ( t ; τ ; ω ) = - c o s ( 2 π t ω ) e - t 2 - τ 2 - - - ( 12 )
h o d ( t ; τ ; ω ) = - s i n ( 2 π t ω ) e - t 2 - τ 2 - - - ( 13 )
In formula (12), formula (13), σ and τ is respectively the detection yardstick in spatial domain and time domain, gets σ=2 and τ=3; Gaussian smoothing filter yardstick is 2; The adjacent domain of each maximum point of response function R contains the body local movable information in I (x, y, t).
As to the Human bodys' response further improvements in methods based on asymmetric generalized gaussian model of the present invention: in described step, Optical-flow Feature u, v two component data comprise as follows: the extraction of space-time characteristic point: after point of interest detection is carried out to image sequence, obtain space-time interest points, centered by space-time interest points, define a space-time cube, extract the pixel of this space-time cube to construct space-time characteristic; If space-time cube is I (x, y, t), the then gradient G of its X-axis, Y-axis, Z-direction x, G y, G zcan be defined as respectively:
G x(x,y,t)=L(x+1,y,t)-L(x-1,y,t),(14)
G y(x,y,t)=L(x,y+1,t)-L(x,y-1,t),(15)
G z(x,y,t)=L(x,y,t+1)-L(x,y,t-1),(16)
In formula, L (x+1, y, t) is the gradient of this Gaussian function after convolutional filtering;
The extraction of Optical-flow Feature: adopt Lucas-Kanade method to calculate light stream: be located at moment t place, pixel (x, y) is at position 1 place, gray-scale value is herein I (x, y, t), in (the t+ △ t) moment, preimage vegetarian refreshments moves to position 2 place, now its change in location is (x+ △ x, y+ △ y), and new gray-scale value is I (x+ △ x, y+ △ y, t+ △ t); According to image consistency hypothesis, meet d I ( x , y , t ) d t = 0 , Then:
I(x,y,t)=I(x+△x,y+△y,t+△t)(17)
If u and v is respectively the component of light stream vector along x and y both direction of pixel (x, y), by formula (17) Taylor expansion be:
I ( x + Δ x , y + Δ y , t + Δ t ) = I ( x , y , t ) + ∂ I ∂ x Δ x + ∂ I ∂ y Δ y + ∂ I ∂ t Δ t + ϵ - - - ( 18 )
After high-order term ε more than second order is ignored, then meet:
∂ I ∂ x Δ x + ∂ I ∂ y Δ y + ∂ I ∂ t Δ t = 0 - - - ( 19 )
∵△t→0
∂ I ∂ x d x d t + ∂ I ∂ y d y d t + ∂ I ∂ t = 0
That is: I xu+I yv+I t=0 (20)
In formula (20), I x, I y, I tfor pixel (x, y) is along the partial derivative in x, y, t tri-directions; Can express with the vector expression of following formula:
▿ I · U + I t = 0 - - - ( 21 )
In formula (21), for gradient direction, U=(u, v) trepresent light stream; Suppose that the light stream in the window of specifying in a size keeps constant, the optical flow constraint equation in this window can be asked thus to obtain light stream (u, v) that size is the characteristic window of x × x, that is:
I x 1 I y 1 I y 2 I y 2 . . . . . . I x i I y i u v = - I t 1 I t 2 . . . I t i - - - ( 22 )
In formula (22), i is the number of pixels i=(x × x) in characteristic window, I xand I yfor the spatial gradient of image, I tit is time gradient; Solve formula (22) can obtain:
u v = Σ I 2 x i Σ I x i I y i ΣI x i I y i ΣI 2 y i - 1 - Σ I x i I t i - ΣI x i I t i - - - ( 23 ) .
As to the Human bodys' response further improvements in methods based on asymmetric generalized gaussian model of the present invention: in described step, the parameter attribute based on asymmetric generalized gaussian model extracts: the expression formula of asymmetric generalized gaussian model is as follows:
f ( x ; v , &sigma; l 2 , &sigma; r 2 ) = v ( &beta; l + &beta; r ) &Gamma; ( 1 v ) exp ( - ( - x - u &beta; l ) v ) , x < 0 v ( &beta; l + &beta; r ) &Gamma; ( 1 v ) exp ( - ( - x - u &beta; l ) v ) , x &GreaterEqual; 0 - - - ( 24 )
In formula (24), with wherein Γ () is gamma function, and expression formula is:
&Gamma; ( &alpha; ) = &Integral; 0 &infin; t &alpha; - 1 e - t d t , &alpha; > 0 - - - ( 25 )
After asymmetric generalized gaussian model matching is carried out to characteristic, extract its five parameters (α, β l, β r, v, u), by five parameters (α, β l, β r, v, u) and as feature.
First training video is carried out pre-service and detects the space-time interest points of this video by the present invention, centered by point of interest, extract its video block again and calculate its Optic flow information and gradient information, corresponding histogram is drawn according to the Optic flow information obtained and gradient information, then carry out the corresponding histogram of matching with asymmetric generalized gaussian model (AGGD), form the eigenmatrix of training video using the parameter of the AGGD of Optic flow information and gradient information as feature.The eigenmatrix that above-mentioned all process obtain test video is carried out equally for test video.Mahalanobis distance between last calculation training video and the eigenmatrix of test video, then the behavior of test video is identified according to nearest neighbouring rule.Method of the present invention improves the accuracy rate of video behavior to be identified to a great extent.
Accompanying drawing explanation
Below in conjunction with accompanying drawing, the specific embodiment of the present invention is described in further detail.
Fig. 1 is training video treatment scheme of the present invention;
Fig. 2 is Activity recognition flow process;
Fig. 3 angle point;
Fig. 4 elliptic function schematic diagram;
Fig. 5 eigenwert and angle point graph of a relation;
The extraction of the AGGD parameter of Fig. 6 Gradient Features;
The AGGD parameter attribute Activity recognition of Fig. 7 Gradient Features;
The Activity recognition rate of Fig. 8 Weizmann storehouse Gradient Features AGGD parameter attribute;
The Activity recognition rate of Fig. 9 KTH storehouse Gradient Features AGGD parameter attribute;
Figure 10 Optical-flow Feature AGGD parameter attribute extracts figure;
Figure 11 light stream AGGD parameter attribute Activity recognition;
Figure 12 Weizmann storehouse Optical-flow Feature AGGD parameter attribute Activity recognition rate;
Figure 13 KTH storehouse light stream AGGD parameter attribute Activity recognition rate;
Figure 14 merges the extraction of gradient and light stream parameter attribute;
Figure 15 merges gradient and light stream AGGD parameter attribute Activity recognition;
Gradient and light stream AGGD parameter attribute Activity recognition rate are merged in Figure 16 Weizmann storehouse;
Gradient and light stream AGGD parameter attribute Activity recognition rate are merged in Figure 17 KTH storehouse.
Embodiment
Embodiment 1, Fig. 1 ~ Figure 17 give a kind of Human bodys' response method based on asymmetric generalized gaussian model, comprise the following steps:
Step one, to extract by carrying out gradient and Optical-flow Feature data to training video storehouse, form each characteristic direction (characteristic direction is here 3 directions of Gradient Features and 2 components of Optical-flow Feature) characteristic set.
Step 2, histogram description is carried out for the characteristic of above-mentioned each characteristic direction.
Step 3, carry out the above-mentioned histogram of matching with asymmetric generalized gaussian model (AGGD), using the parameter of AGGD as feature, form the parameter attribute matrix of each behavior.
Step 4, for test video, we also carry out the extraction of gradient and Optical-flow Feature, form the characteristic set of each characteristic direction.
Step 5, histogram description is carried out for the characteristic of above-mentioned steps four.
Step 6, we also carry out matching histogram with AGGD equally, using the parameter of AGGD as feature, form the parameter attribute matrix of test video.
Step 7, calculate mahalanobis distance between test video eigenmatrix and the eigenmatrix of each behavior in training video storehouse respectively.
Step 8, judge the behavior of test video according to nearest neighbouring rule.
In above step 1 ~ 8, detect mainly through point of interest, the extraction of unique point and description, to extract based on the parameter attribute of asymmetric generalized gaussian model (AGGD), based on the Activity recognition of AGGD parameter.
1, point of interest detects as follows:
In the present invention, in order to the space-time interest points of image sequence I (x, y, t) effectively can be detected, adopt following method:
First we define the intersection point that image angle point is two edges, or can be understood as the unique point simultaneously in neighborhood with two principal directions, the turning in similar road and house.The neighborhood at general angle point place is normally stable in image and have the region of bulk information, and these regions have the characteristics such as affine-invariant features, scale invariability, rotational invariance.Human visual is completed by the region of a local or a wicket usually to the identification of angle point, as shown in Figure 3.If when being moved towards all directions by this specific wicket, meanwhile in moving window region, grey scale change is comparatively large, so just can judge there is angle point in this window, as shown in Figure 3.If when being moved towards all directions by this specific wicket, gray scale does not change in moving window region, so there is no angle point in this window, as shown in Figure 3.If when being moved towards some directions by this specific wicket, in moving window region, grey scale change is larger, and when moving towards another direction in window area gray scale constant, may be straight line so in this window, as shown in Figure 3.
According to autocorrelation function, the self-similarity expression formula of image I (x, y) after point (x, y) place translation (△ x, △ y) can be provided:
c ( x , y , &Delta; x , &Delta; y ) = &Sigma; ( u , v ) &Element; W ( x , y ) &omega; ( u , v ) ( I ( u , v ) - I ( u + &Delta; x , v + &Delta; y ) ) 2 - - - ( 1 )
In formula (1), ω (u, v) is weighting function, and can get constant also can be gaussian weighing function; W (x, y) is the window centered by point (x, y).
According to Taylor expansion, after point (x, y) place translation (△ x, △ y), first approximation is carried out to image I (x, y) and obtains:
I ( u + &Delta; x , v + &Delta; y ) = ( u , v ) + &lsqb; I x ( u , v ) , I y ( u , v ) ] &Delta; x &Delta; x - - - ( 2 )
In formula (2), I xand I yit is the partial derivative of I (x, y).
Then formula (2) can be approximately:
c ( x , y ; &Delta; x , &Delta; y ) = &Delta; x &Delta; y M ( x , y ) &Delta; x &Delta; y - - - ( 3 )
In formula (3), M ( x , y ) = &Sigma; &omega; I x ( u , v ) 2 &Sigma; &omega; I x ( u , v ) I y ( u , v ) &Sigma; &omega; I x ( u , v ) I y ( u , v ) &Sigma; &omega; I y ( u , v ) 2 , Namely the autocorrelation function of image I (x, y) after point (x, y) place translation (△ x, △ y) can be approximated to be quadratic term function.
Quadratic term function can see in fact an elliptic function as, and as shown in Figure 4, the ellipticity of elliptic function and size are by the eigenvalue λ of M (x, y) 1, λ 2determine, direction is determined by the eigenvector of M (x, y), and its equation is:
&Delta; x &Delta; y M ( x , y ) &Delta; x &Delta; y = 1 - - - ( 4 )
The angle point in the image in window, edge (straight line) and plane can be judged, as shown in Figure 5 according to the size of the eigenwert of quadratic term function.Work as λ 1< < λ 2or λ 1>> λ 2, namely the value of autocorrelation function is only larger on some directions, time smaller in other directions, can be judged as straight line in window; Work as λ 1≈ λ 2, and λ 1and λ 2all smaller, namely the value of autocorrelation function is all smaller in all directions, can be judged as plane in window; Work as λ 1≈ λ 2, and λ 1and λ 2all larger, namely the value of autocorrelation function is all larger in all directions, can be judged as angle point in window.
In fact, differentiate that angle point does not need to calculate concrete eigenwert, calculate its value after only need defining an angle point response function to judge angle point.Definition response function R is:
R=detM-α(traceM) 2(5)
M (x, y) in formula (3) is reduced to M ( x , y ) = A D D B , DetM and traceM then in formula (5) is respectively the determinant of M (x, y) and straight mark, and wherein α is empirical value, generally gets 0.04-0.06.
Detected by our the Harris point of interest of extending out of define method of angle point above, the thinking of Harris point of interest detection method finds image f spthe position all had significant change in all directions.Then the detection method of Harris point of interest can be described as: defining a sub-picture is f sp: f spl is obtained after linear filtering sp: its expression formula is as follows:
L s p ( x , y ; &sigma; l 2 ) = g s p ( x , y ; &sigma; l 2 ) * f s p ( x , y ) - - - ( 6 )
In formula (6), g spimage f spcarry out the gaussian kernel function of convolutional filtering, for its dimensional variation factor.
To the given observing result of formula (6) yardstick is utilized to be second-order matrix with Gauss's window finds point of interest, and expression formula is as follows:
In formula (7), * is convolution symbol, with being yardstick is gradient on Gaussian function x and y,
A second moment descriptor can be regarded as the framing covariance matrix of a some near zone Two dimensional Distribution.So, matrix μ speigenvalue λ 1and λ 21≤ λ 2) constitute f spchange descriptor in image both direction, and λ 1and λ 2just point of interest is there is when being all and being worth greatly.Harris and Stephens proposes the maximum value calculation method of a Corner Detection function based on this, and expression formula is as follows:
H sp=det(μ sp)-k×trace 2sp)=λ 1λ 2-k(λ 12) 2(8)
In the position that point of interest exists, the ratio α=λ of eigenwert 2/ λ 1value can be larger.From formula (8) known H spget positive maximum, the ratio α of eigenwert meets k≤α/(1+ α) 2if, definition k=0.25, now α=1, λ 12, H gets positive maximum, and point of interest has desirable isotropy.
What will detect due to this patent is point of interest in video (image sequence), video can be regarded as the image sequence f (x, y, t) of multiple image composition.Defined function f: f obtains L after adding linear space yardstick: by image sequence f through separate space variable be with time variable be gaussian function carry out convolutional filtering and obtain, expression formula is as follows:
Gauss's window in time-space domain be defined as:
g ( x , y , t ; &sigma; l 2 ; &tau; l 2 ) = 1 ( 2 &pi; ) 2 &sigma; l 4 &tau; l 2 &times; exp ( - ( x 2 + y 2 ) 2 2 &sigma; l 2 - t 2 2 &tau; l 2 ) - - - ( 10 )
σ in formula lfor space scale variable, τ lfor time scale variable, t is time dimension.
The point of interest detection method that this patent is used, space dimension is continued to use the point of interest method in above-mentioned image, and time dimension then adopts the Gabor filter that Dollar proposes, then define response function R to be:
R(x,y,t)=(I*g*h ev) 2+(I*g*h od) 2(11)
In formula, * is convolution algorithm symbol, and I is video image, and g is dimensional Gaussian smoothing kernel, h evand h odit is one dimension Gabor filter orthogonal in spatial domain.
H evand h odbe defined as:
h e v ( t ; &tau; ; &omega; ) = - c o s ( 2 &pi; t &omega; ) e - t 2 - &tau; 2 - - - ( 12 )
h o d ( t ; &tau; ; &omega; ) = - s i n ( 2 &pi; t &omega; ) e - t 2 - &tau; 2 - - - ( 13 )
In formula (12), formula (13), σ and τ is respectively the detection yardstick in spatial domain and time domain, gets σ=2 and τ=3 in the present invention; Gaussian smoothing filter yardstick is 2.
The adjacent domain of each maximum point of response function R contains the body local movable information in I (x, y, t).
2, feature extraction and be described below:
The extraction of 2.1 space-time characteristics:
After point of interest detection is carried out to image sequence, just a series of space-time interest points can be obtained, but only human body behavior cannot be effectively described by these points of interest.The present invention defines a space-time cube centered by space-time interest points, and extract the pixel of this space-time cube to construct space-time characteristic, the cubical length of side gets six times of its place yardstick.This space-time cube contains and is conducive to most of points that response function gets maximum value.
Describe the method for space-time cube have cube expansion value is described as a vector, pixel normalization description and histogram description etc.Because human body is when moving, the brightness of image change of point of interest near zone is very violent, and when the motor behavior of human body is different, the brightness of image change of point of interest near zone is also different.Therefore, the brightness of image of point of interest near zone can be utilized to change the point of interest describing different human body behavior.The brightness of image change of the point of interest near zone of different human body behavior can be reflected by the gradient of the brightness of space-time cube along X-axis, Y-axis and Z axis (i.e. time shaft) direction, and this patent extracts these gradients and carries out Human bodys' response as feature.
If space-time cube is I (x, y, t), the then gradient G of its X-axis, Y-axis, Z-direction x, G y, G zcan be defined as respectively:
G x(x,y,t)=L(x+1,y,t)-L(x-1,y,t),(14)
G y(x,y,t)=L(x,y+1,t)-L(x,y-1,t),(15)
G z(x,y,t)=L(x,y,t+1)-L(x,y,t-1),(16)
In formula, L (x+1, y, t) is the gradient of this Gaussian function after convolutional filtering.
2.2, the extraction of Optical-flow Feature:
Optical flow field is the how time dependent vector field of energy Description Image sequence, and containing the transient motion velocity information of pixel, is reasonable space-time characteristic.But optical flow computation amount is comparatively large, in order to reduce calculated amount by only calculating the light stream of the video block extracted, text selects Lucas-Kanade method to calculate light stream.
Optical flow computation principle:
Be located at moment t place, pixel (x, y) is at position 1 place, gray-scale value is herein I (x, y, t), in (the t+ △ t) moment, preimage vegetarian refreshments moves to position 2 place, now its change in location is (x+ △ x, y+ △ y), and new gray-scale value is I (x+ △ x, y+ △ y, t+ △ t).According to image consistency hypothesis, meet then:
I(x,y,t)=I(x+△x,y+△y,t+△t)(17)
If u and v is respectively the component of light stream vector along x and y both direction of pixel (x, y), by formula (17) Taylor expansion be:
I ( x + &Delta; x , y + &Delta; y , t + &Delta; t ) = I ( x , y , t ) + &part; I &part; x &Delta; x + &part; I &part; y &Delta; y + &part; I &part; t &Delta; t + &epsiv; - - - ( 18 )
After high-order term ε more than second order is ignored, then meet:
&part; I &part; x &Delta; x + &part; I &part; y &Delta; y + &part; I &part; t &Delta; t = 0 - - - ( 19 )
∵△t→0
&part; I &part; x d x d t + &part; I &part; y d y d t + &part; I &part; t = 0
That is: I xu+I yv+I t=0 (20)
In formula (20), I x, I y, I tfor pixel (x, y) is along the partial derivative in x, y, t tri-directions.Can express with the vector expression of following formula:
&dtri; I &CenterDot; U + I t = 0 - - - ( 21 )
In formula (21), for gradient direction, U=(u, v) trepresent light stream.
Lucas-Kanade optical flow method: Lucas-Kanade selected by text [25]optical flow method calculates light stream.Suppose that the light stream in the window of specifying in a size keeps constant, the optical flow constraint equation in this window can be asked thus to obtain light stream (u, v) that size is the characteristic window of x × x, that is:
I x 1 I y 1 I y 2 I y 2 . . . . . . I x i I y i u v = - I t 1 I t 2 . . . I t i - - - ( 22 )
In formula (22), i is the number of pixels i=(x × x) in characteristic window, I xand I yfor the spatial gradient of image, I tit is time gradient.Solve formula (22) can obtain:
u v = &Sigma; I 2 x i &Sigma; I x i I y i &Sigma; I x i I y i &Sigma; I 2 y i - 1 - &Sigma; I x i I t i - &Sigma; I x i I t i - - - ( 23 ) .
3, the parameter attribute based on asymmetric generalized gaussian model (AGGD) extracts:
Although characteristic compares after curve be close to Gaussian distribution, they are not strictly symmetrical, choose asymmetric generalized Gaussian density distribution (AGGD) carry out matching to two category feature data according to this feature this patent.
The expression formula of AGGD is as follows:
f ( x ; v , &sigma; l 2 , &sigma; r 2 ) = v ( &beta; l + &beta; r ) &Gamma; ( 1 v ) exp ( - ( - x - u &beta; l ) v ) , x < 0 v ( &beta; l + &beta; r ) &Gamma; ( 1 v ) exp ( - ( - x - u &beta; l ) v ) , x &GreaterEqual; 0 - - - ( 24 )
In formula (24), with wherein Γ () is gamma function, and expression formula is:
&Gamma; ( &alpha; ) = &Integral; 0 &infin; t &alpha; - 1 e - t d t , &alpha; > 0 - - - ( 25 )
After AGGD models fitting is carried out to characteristic, extract its five parameters (α, β l, β r, v, u).
4, based on the Activity recognition of AGGD parameter:
4.1, based on the Activity recognition of Gradient Features AGGD parameter:
4.1.1, Human bodys' response algorithm:
Gradient Features AGGD parameter based on most of behavior spatially can make a distinction this characteristic with other behavior, and this patent chooses five parameters (α, β of the asymmetric Generalized Gaussian Distribution Model of Gradient Features l, β r, v, u) and as feature, if table 4.1 is feature lexical or textual analysis table.
Table 4.1 feature lexical or textual analysis table
This patent is according to the movement characteristic of different behavior, the space-time cube of respective numbers is extracted respectively after point of interest is carried out to each pretreated behavior video, gradient is carried out to space-time cube and describes the feature (flow process as shown in Figure 6) shown in rear extraction table 4.1, then the mahalanobis distance between calculation training collection behavior video feature matrix and test set behavior video feature matrix, nearest neighbor classifier is finally utilized to carry out Activity recognition, as shown in Figure 7.
Nearest neighbor classifier be using training data concentrate each sample as distinguishing rule, in training set, find the sample nearest apart from sample to be sorted, then classify on this basis.Suppose that training set has N number of sample { x 1, x 2..., x n, be divided into y classification, then sample x to be sorted is to training sample x idistance d (x, x i) be:
d(x,x i)=||x-x i||(26)
If d is (x, x k) meet following formula: d ( x , x k ) = min i = 1 , 2 , ... , N { d ( x , x i ) } , x k &Element; &omega; j , Then x ∈ ω j.
4.1.1, Weizmann database experiment:
Can observe from Fig. 8, the recognition accuracy of jump with side behavior is compared with the recognition accuracy of other behavior, and it relatively low, is respectively 0.82 and 0.844.Wherein, the False Rate that jump behavior is mainly mistaken for side behavior is the highest, be 0.072, main cause is that jump behavior is more similar with the leg action of side behavior, side behavior has the action of similar jump when action, thus the Gradient Features numerical value of the Gradient Features numerical value of jump behavior and side behavior relatively, cause accurately identifying; And side behavior is mainly mistaken for walk behavior, False Rate is 0.048, reason is side behavior is horizontal walking when action, and walk behavior is forward walking when action, the footwork of two behaviors is similar, just movement range is slightly different with speed, and the Gradient Features numerical value of therefore side behavior and walk behavior has certain similarity, thus misjudged likelihood ratio is higher.
4.1.3, KTH database experiment:
From in Fig. 9, compared with other behavior, the False Rate of handclap and handwave behavior is higher, and handclap behavior is mainly mistaken for box and handwave behavior, and False Rate is respectively 0.056 and 0.048; Handwave behavior is mainly mistaken for box behavior, and False Rate is 0.072.Main cause is that the part hand behavior when action of this three behaviors is more similar, the action that they have hand to stretch out forward when action, and just the speed of action is different with amplitude.Therefore, their Gradient Features numerical value has certain similarity, thus causes misjudged.
4.2, based on the Human bodys' response of the AGGD parameter attribute of Optical-flow Feature:
4.2.1, Human bodys' response algorithm:
Similar as above, this patent utilizes AGGD model to carry out matching to Optical-flow Feature data, then, extracts its five parameters (α, β l, β r, v, u) and as feature, feature lexical or textual analysis is as shown in table 4.2.
The AGGD parameter attribute lexical or textual analysis table of table 4.2 Optical-flow Feature
With gradient parameter characteristic behavior identification experimental technique is consistent above, according to the movement characteristic of different behavior, the space-time cube of respective numbers is extracted respectively after point of interest is carried out to each pretreated behavior video, light stream is carried out to space-time cube and describes the feature (as shown in Figure 10) shown in rear extraction table 4.2, then Activity recognition is carried out by the mahalanobis distance between calculation training collection behavior video feature matrix and test set behavior video feature matrix, nearest neighbor classifier is finally utilized to carry out Activity recognition, as shown in figure 11.
4.2.2, Weizmann database experiment:
Can observe out by Figure 12, discrimination and other behavior of side and wave1 behavior are relatively low, and discrimination is respectively 0.856 and 0.76.Wherein, side behavior is mistaken for the probability of jump behavior up to 0.06, and this may be comparatively similar relevant with the part footwork of jump behavior with side behavior, thus cause the part Optical-flow Feature numerical value of these two kinds of behaviors close and accurately identify; And wave1 behavior is mainly mistaken for wave2, jack behavior, False Rate is respectively 0.096,0.084, reason is these three behaviors have hand wobbling action when action, amplitude when just swinging is different with speed, therefore the Optical-flow Feature numerical value of these three behaviors has certain similarity, thus causes erroneous judgement.
4.2.3, KTH database experiment:
Relatively in Figure 13, the recognition accuracy of each behavior is known, and the False Rate of jog and handwave behavior is higher.Wherein, jog behavior is mainly mistaken for run, False Rate is 0.096, reason may be that jog behavior and run behavior have running action when action, just (jog behavior speed is slow for speed difference, run behavior speed), thus cause the Optical-flow Feature numeric ratio of these two behaviors more similar and cannot accurately identify; Wherein, handwave behavior is mistaken for the probability of walk up to being 0.048, and reason is that the partial act of these two behavior hands is similar, and just action is different with amplitude.
4.3, based on the Human bodys' response of AGGD parameter attribute merging gradient and Optical-flow Feature:
4.3.1, Human bodys' response algorithm:
The Activity recognition rate of the AGGD parameter attribute of two databases, two kinds of features is known from the above, Weizmann database discrimination difference in different characteristic situation is smaller, and the Activity recognition rate of the AGGD parameter attribute of KTH database light stream is more much higher than the discrimination of the AGGD parameter attribute of gradient, proposes the AGGD parameter attribute of two kinds of features to be combined into according to the order of sequence a fusion AGGD parameter attribute based on this this patent and carry out Activity recognition.Fusion gradient and the lexical or textual analysis of light stream AGGD parameter attribute are as shown in table 4.3.1.
Table 4.3 merges gradient and light stream AGGD parameter attribute lexical or textual analysis table
Consistent with above-mentioned Activity recognition method, feature (as shown in figure 14) shown in table 4.3.1 is extracted respectively to the behavior video of training set and test set, then Activity recognition is carried out by the mahalanobis distance between calculation training collection behavior video feature matrix and test set behavior video feature matrix, nearest neighbor classifier is finally utilized to carry out Activity recognition, as shown in figure 15.
4.3.2, Weizmann database experiment:
What discrimination was minimum as can be known from Fig. 16 is jump and wave1 behavior, and discrimination is respectively 0.856 and 0.844.Wherein, jump behavior is mistaken for the probability of walk behavior up to 0.084, and reason is that jump behavior is similar to the part footwork of walk behavior, causes the AGGD parameter differences of fusion feature data little and cannot accurately identify; And wave1 behavior is mainly mistaken for wave2, jack behavior, False Rate is respectively 0.084,0.072, reason is these three behaviors have hand wobbling action when action, amplitude when just swinging is different with speed, therefore the AGGD parameter values of the fusion feature data of wave1 behavior and wave2, jack behavior has certain similarity, causes erroneous judgement.
4.3.3, KTH database experiment:
Adopt experiment group technology of the same race above to test, the average behavior discrimination obtaining KTH storehouse is 95.2%.Figure 17 is the confusion matrix of the Activity recognition rate in KTH storehouse, wherein, the discrimination of jog behavior is low compared with other behavior, its probability being mistaken for run behavior is up to 0.088, main cause may be jog and run is all running behavior, and just jog jogs, and run hurries up, both have a difference in speed, cause the fusion feature parameter value of these two behaviors more similar and cannot accurately identify.
4.3.4, the Activity recognition rate of two different AGGD parameter attributes of database:
Following table is the discrimination of three kinds of AGGD parameter attributes in Weizmann database and KTH database.Can observe in Weizmann database from table, the discrimination of the AGGD parameter attribute of light stream is minimum, and being that the discrimination of 90.16%, compound AGGD parameter attribute is the highest, is 93.16%; In KTH database, the discrimination of the AGGD parameter attribute of gradient is minimum, and being that the discrimination of 88.40%, compound AGGD parameter attribute is the highest, is 95.20%.
As can be seen from the Activity recognition rate of the different parameters of above-mentioned two databases, the Activity recognition rate merging light stream and gradient parameter is all higher than the discrimination respectively based on gradient and light stream parameter.
The Activity recognition rate of the different AGGD parameter attribute of table 4.4 two databases
Finally, it is also to be noted that what enumerate above is only a specific embodiment of the present invention.Obviously, the invention is not restricted to above embodiment, many distortion can also be had.All distortion that those of ordinary skill in the art can directly derive from content disclosed by the invention or associate, all should think protection scope of the present invention.

Claims (4)

1., based on a Human bodys' response method for asymmetric generalized gaussian model, realized by training video storehouse and test video; It is characterized in that: comprise the following steps:
Step one, carries out point of interest detection respectively for given training video storehouse and test video;
Step 2, extracts video block centered by point of interest;
Step 3, respectively the video block information of calculation training video and test video, and obtain respective X, Y, Z tri-direction gradient feature data and Optical-flow Feature u, v two component data;
Step 4, draws gradient three direction histogram and light stream two direction histogram respectively to above-mentioned data;
Step 5, carrys out the corresponding histogram of matching with asymmetric generalized gaussian model;
Step 6, extracts h evthe parameter of asymmetric generalized gaussian model forms the eigenmatrix of each behavior of training video and the eigenmatrix of test video as feature;
Step 7, calculates the mahalanobis distance between test video eigenmatrix and each behavioural characteristic matrix of training video;
Step 8, carries out Activity recognition according to nearest neighbouring rule.
2. as to the Human bodys' response method based on asymmetric generalized gaussian model according to claim 1, it is characterized in that: in described step, point of interest detects as follows:
Video is regarded as the image sequence f (x, y, t) of multiple image composition;
Defined function f: f obtains L after adding linear space yardstick: by image sequence f through separate space variable be with time variable be gaussian function carry out convolutional filtering and obtain, expression formula is as follows:
L ( &CenterDot; ; &sigma; l 2 ; &tau; l 2 ) = g ( &CenterDot; ; &sigma; l 2 ; &tau; l 2 ) * f ( . ) - - - ( 9 )
Gauss's window in time-space domain be defined as:
g ( x , y , t ; &sigma; l 2 ; &tau; l 2 ) = 1 ( 2 &pi; ) 3 &sigma; l 4 &tau; l 2 &times; exp ( - ( x 2 + y 2 ) 2 2 &sigma; 1 2 - t 2 2 &tau; l 2 ) - - - ( 10 )
σ in formula lfor space scale variable, τ lfor time scale variable, t is time dimension;
Definition response function R is:
R(x,y,t)=(I*g*h ev) 2+(I*g*h od) 2(11)
In formula, * is convolution algorithm symbol, and I is video image, and g is dimensional Gaussian smoothing kernel, h evand h odit is one dimension Gabor filter orthogonal in spatial domain;
H evand h odbe defined as:
h e v ( t ; &tau; ; &omega; ) = - c o s ( 2 &pi; t &omega; ) e - t 2 - &tau; 2 - - - ( 12 )
h o d ( t ; &tau; ; &omega; ) = - s i n ( 2 &pi; t &omega; ) e - t 2 - &tau; 2 - - - ( 13 )
In formula (12), formula (13), σ and τ is respectively the detection yardstick in spatial domain and time domain, gets σ=2 and τ=3; Gaussian smoothing filter yardstick is 2;
The adjacent domain of each maximum point of response function R contains the body local movable information in I (x, y, t).
3., as to the Human bodys' response method based on asymmetric generalized gaussian model according to claim 2, it is characterized in that: in described step, Optical-flow Feature u, v two component data comprise as follows:
The extraction of space-time characteristic point:
After point of interest detection is carried out to image sequence, obtain space-time interest points, centered by space-time interest points, define a space-time cube, extract the pixel of this space-time cube to construct space-time characteristic;
If space-time cube is I (x, y, t), the then gradient G of its X-axis, Y-axis, Z-direction x, G y, G zcan be defined as respectively:
G x(x,y,t)=L(x+1,y,t)-L(x-1,y,t),(14)
G y(x,y,t)=L(x,y+1,t)-L(x,y-1,t),(15)
G z(x,y,t)=L(x,y,t+1)-L(x,y,t-1),(16)
In formula, L (x+1, y, t) is the gradient of this Gaussian function after convolutional filtering;
The extraction of Optical-flow Feature:
Lucas-Kanade method is adopted to calculate light stream:
Be located at moment t place, pixel (x, y) is at position 1 place, gray-scale value is herein I (x, y, t), in (the t+ △ t) moment, preimage vegetarian refreshments moves to position 2 place, now its change in location is (x+ △ x, y+ △ y), and new gray-scale value is I (x+ △ x, y+ △ y, t+ △ t); According to image consistency hypothesis, meet then:
I(x,y,t)=I(x+△x,y+△y,t+△t)(17)
If u and v is respectively the component of light stream vector along x and y both direction of pixel (x, y), by formula (17) Taylor expansion be:
I ( x + &Delta; x , y + &Delta; y , t + &Delta; t ) = I ( x , y , t ) + &part; I &part; x &Delta; x + &part; I &part; y &Delta; y + &part; I &part; t &Delta; t + &epsiv; - - - ( 18 )
After high-order term ε more than second order is ignored, then meet:
&part; I &part; x &Delta; x + &part; I &part; y &Delta; y + &part; I &part; t &Delta; t = 0 - - - ( 19 )
∵△t→0
&part; I &part; x d x d t + &part; I &part; y d y d t + &part; I &part; t = 0
That is: I xu+I yv+I t=0 (20)
In formula (20), I x, I y, I tfor pixel (x, y) is along the partial derivative in x, y, t tri-directions; Can express with the vector expression of following formula:
&dtri; I &CenterDot; U + I t = 0 - - - ( 21 )
In formula (21), for gradient direction, U=(u, v) trepresent light stream;
Suppose that the light stream in the window of specifying in a size keeps constant, the optical flow constraint equation in this window can be asked thus to obtain light stream (u, v) that size is the characteristic window of x × x, that is:
I x 1 I y 1 I y 2 I y 2 . . . . . . I x i I y i u v = - I t 1 I t 2 . . . I t i - - - ( 22 )
In formula (22), i is the number of pixels i=(x × x) in characteristic window, I xand I yfor the spatial gradient of image, I tit is time gradient;
Solve formula (22) can obtain:
u v = &Sigma; I 2 x i &Sigma; I x i I y i &Sigma; I x i I y i &Sigma; I 2 y i - 1 &Sigma; I x i I t i &Sigma; I x i I t i - - - ( 23 ) .
4. as to the Human bodys' response method based on asymmetric generalized gaussian model according to claim 3, it is characterized in that: in described step, the parameter attribute based on asymmetric generalized gaussian model extracts:
The expression formula of asymmetric generalized gaussian model is as follows:
f ( x ; v , &sigma; l 2 , &sigma; r 2 ) = v ( &beta; l + &beta; r ) &Gamma; ( 1 v ) exp ( - ( - x - u &beta; l ) v ) , x < 0 v ( &beta; l + &beta; r ) &Gamma; ( 1 v ) exp ( - ( - x - u &beta; l ) v ) , x &GreaterEqual; 0 - - - ( 24 )
In formula (24), with wherein Γ () is gamma function, and expression formula is:
&Gamma; ( &alpha; ) = &Integral; 0 &infin; t &alpha; - 1 e - t d t &alpha; > 0 - - - ( 25 )
After asymmetric generalized gaussian model matching is carried out to characteristic, extract its five parameters (α, β l, β r, v, u), by five parameters (α, β l, β r, v, u) and as feature.
CN201510313321.1A 2015-06-09 2015-06-09 Human bodys' response method based on asymmetric generalized gaussian model Expired - Fee Related CN105046195B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510313321.1A CN105046195B (en) 2015-06-09 2015-06-09 Human bodys' response method based on asymmetric generalized gaussian model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510313321.1A CN105046195B (en) 2015-06-09 2015-06-09 Human bodys' response method based on asymmetric generalized gaussian model

Publications (2)

Publication Number Publication Date
CN105046195A true CN105046195A (en) 2015-11-11
CN105046195B CN105046195B (en) 2018-11-02

Family

ID=54452724

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510313321.1A Expired - Fee Related CN105046195B (en) 2015-06-09 2015-06-09 Human bodys' response method based on asymmetric generalized gaussian model

Country Status (1)

Country Link
CN (1) CN105046195B (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956606A (en) * 2016-04-22 2016-09-21 中山大学 Method for re-identifying pedestrians on the basis of asymmetric transformation
CN106250870A (en) * 2016-08-16 2016-12-21 电子科技大学 A kind of pedestrian's recognition methods again combining local and overall situation similarity measurement study
CN106485253A (en) * 2016-09-14 2017-03-08 同济大学 A kind of pedestrian of maximum particle size structured descriptor discrimination method again
CN106778595A (en) * 2016-12-12 2017-05-31 河北工业大学 The detection method of abnormal behaviour in crowd based on gauss hybrid models
CN106815600A (en) * 2016-12-27 2017-06-09 浙江工业大学 For the depth co-ordinative construction and structural chemistry learning method of human behavior identification
CN107403182A (en) * 2017-05-26 2017-11-28 深圳大学 The detection method and device of space-time interest points based on 3D SIFT frameworks
CN108230375A (en) * 2017-12-27 2018-06-29 南京理工大学 Visible images and SAR image registration method based on structural similarity fast robust
CN108241849A (en) * 2017-08-28 2018-07-03 北方工业大学 Human body interactive action recognition methods based on video
CN110135352A (en) * 2019-05-16 2019-08-16 南京砺剑光电技术研究院有限公司 A kind of tactical operation appraisal procedure based on deep learning
CN110837770A (en) * 2019-08-30 2020-02-25 深圳大学 Video behavior self-adaptive segmentation method and device based on multiple Gaussian models
CN111507275A (en) * 2020-04-20 2020-08-07 北京理工大学 Video data time sequence information extraction method and device based on deep learning
CN112926514A (en) * 2021-03-26 2021-06-08 哈尔滨工业大学(威海) Multi-target detection and tracking method, system, storage medium and application
CN114299602A (en) * 2021-11-09 2022-04-08 北京九州安华信息安全技术有限公司 Micro-amplitude motion image processing method
CN114399539A (en) * 2022-01-14 2022-04-26 合肥英睿系统技术有限公司 Method, apparatus and storage medium for detecting moving object
CN114614797A (en) * 2022-05-12 2022-06-10 之江实验室 Adaptive filtering method and system based on generalized maximum asymmetric correlation entropy criterion

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101719216A (en) * 2009-12-21 2010-06-02 西安电子科技大学 Movement human abnormal behavior identification method based on template matching
CN104036287A (en) * 2014-05-16 2014-09-10 同济大学 Human movement significant trajectory-based video classification method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101719216A (en) * 2009-12-21 2010-06-02 西安电子科技大学 Movement human abnormal behavior identification method based on template matching
CN104036287A (en) * 2014-05-16 2014-09-10 同济大学 Human movement significant trajectory-based video classification method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
张飞燕 等: "基于光流速度分量加权的人体行为识别", 《浙江理工大学学报》 *
李俊峰 等: "基于局部时空特征方向加权的人体行为识别", 《中国图象图形学报》 *

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956606A (en) * 2016-04-22 2016-09-21 中山大学 Method for re-identifying pedestrians on the basis of asymmetric transformation
CN105956606B (en) * 2016-04-22 2019-09-10 中山大学 A kind of pedestrian's identification method again based on asymmetry transformation
CN106250870A (en) * 2016-08-16 2016-12-21 电子科技大学 A kind of pedestrian's recognition methods again combining local and overall situation similarity measurement study
CN106250870B (en) * 2016-08-16 2019-10-08 电子科技大学 A kind of pedestrian's recognition methods again of joint part and global similarity measurement study
CN106485253A (en) * 2016-09-14 2017-03-08 同济大学 A kind of pedestrian of maximum particle size structured descriptor discrimination method again
CN106485253B (en) * 2016-09-14 2019-05-14 同济大学 A kind of pedestrian of maximum particle size structured descriptor discrimination method again
CN106778595A (en) * 2016-12-12 2017-05-31 河北工业大学 The detection method of abnormal behaviour in crowd based on gauss hybrid models
CN106778595B (en) * 2016-12-12 2020-04-07 河北工业大学 Method for detecting abnormal behaviors in crowd based on Gaussian mixture model
CN106815600B (en) * 2016-12-27 2019-07-30 浙江工业大学 Depth co-ordinative construction and structural chemistry learning method for human behavior identification
CN106815600A (en) * 2016-12-27 2017-06-09 浙江工业大学 For the depth co-ordinative construction and structural chemistry learning method of human behavior identification
CN107403182A (en) * 2017-05-26 2017-11-28 深圳大学 The detection method and device of space-time interest points based on 3D SIFT frameworks
CN108241849A (en) * 2017-08-28 2018-07-03 北方工业大学 Human body interactive action recognition methods based on video
CN108241849B (en) * 2017-08-28 2021-09-07 北方工业大学 Human body interaction action recognition method based on video
CN108230375A (en) * 2017-12-27 2018-06-29 南京理工大学 Visible images and SAR image registration method based on structural similarity fast robust
CN108230375B (en) * 2017-12-27 2022-03-22 南京理工大学 Registration method of visible light image and SAR image based on structural similarity rapid robustness
CN110135352A (en) * 2019-05-16 2019-08-16 南京砺剑光电技术研究院有限公司 A kind of tactical operation appraisal procedure based on deep learning
CN110135352B (en) * 2019-05-16 2023-05-12 南京砺剑光电技术研究院有限公司 Tactical action evaluation method based on deep learning
CN110837770A (en) * 2019-08-30 2020-02-25 深圳大学 Video behavior self-adaptive segmentation method and device based on multiple Gaussian models
CN110837770B (en) * 2019-08-30 2022-11-04 深圳大学 Video behavior self-adaptive segmentation method and device based on multiple Gaussian models
CN111507275A (en) * 2020-04-20 2020-08-07 北京理工大学 Video data time sequence information extraction method and device based on deep learning
CN111507275B (en) * 2020-04-20 2023-10-10 北京理工大学 Video data time sequence information extraction method and device based on deep learning
CN112926514A (en) * 2021-03-26 2021-06-08 哈尔滨工业大学(威海) Multi-target detection and tracking method, system, storage medium and application
CN114299602A (en) * 2021-11-09 2022-04-08 北京九州安华信息安全技术有限公司 Micro-amplitude motion image processing method
CN114399539A (en) * 2022-01-14 2022-04-26 合肥英睿系统技术有限公司 Method, apparatus and storage medium for detecting moving object
CN114614797A (en) * 2022-05-12 2022-06-10 之江实验室 Adaptive filtering method and system based on generalized maximum asymmetric correlation entropy criterion
CN114614797B (en) * 2022-05-12 2022-09-30 之江实验室 Adaptive filtering method and system based on generalized maximum asymmetric correlation entropy criterion

Also Published As

Publication number Publication date
CN105046195B (en) 2018-11-02

Similar Documents

Publication Publication Date Title
CN105046195A (en) Human behavior identification method based on asymmetric generalized Gaussian distribution model (AGGD)
Jia et al. Visual tracking via adaptive structural local sparse appearance model
CN104616316B (en) Personage&#39;s Activity recognition method based on threshold matrix and Fusion Features vision word
CN104978561A (en) Gradient and light stream characteristics-fused video motion behavior identification method
CN105719285A (en) Pedestrian detection method based on directional chamfering distance characteristics
CN101388080B (en) Passerby gender classification method based on multi-angle information fusion
CN103279768B (en) A kind of video face identification method based on incremental learning face piecemeal visual characteristic
Xiao et al. Multimodal fusion based on LSTM and a couple conditional hidden Markov model for Chinese sign language recognition
CN105956560A (en) Vehicle model identification method based on pooling multi-scale depth convolution characteristics
CN101894276A (en) Training method of human action recognition and recognition method
CN103065158B (en) The behavior recognition methods of the ISA model based on relative gradient
CN106295532B (en) A kind of human motion recognition method in video image
CN103605986A (en) Human motion recognition method based on local features
Wang et al. Improving human action recognition by non-action classification
CN105138983B (en) The pedestrian detection method divided based on weighting block model and selective search
CN105825233B (en) A kind of pedestrian detection method based on on-line study random fern classifier
CN105654139A (en) Real-time online multi-target tracking method adopting temporal dynamic appearance model
CN111046732A (en) Pedestrian re-identification method based on multi-granularity semantic analysis and storage medium
Naghavi et al. Integrated real-time object detection for self-driving vehicles
CN106204651A (en) A kind of method for tracking target based on the judgement improved with generation conjunctive model
CN113850221A (en) Attitude tracking method based on key point screening
CN105809713A (en) Object tracing method based on online Fisher discrimination mechanism to enhance characteristic selection
CN103020614A (en) Human movement identification method based on spatio-temporal interest point detection
Zhang et al. Trajectory series analysis based event rule induction for visual surveillance
CN104978569A (en) Sparse representation based incremental face recognition method

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20181102

Termination date: 20190609

CF01 Termination of patent right due to non-payment of annual fee