CN105956543A - Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features - Google Patents

Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features Download PDF

Info

Publication number
CN105956543A
CN105956543A CN201610272162.XA CN201610272162A CN105956543A CN 105956543 A CN105956543 A CN 105956543A CN 201610272162 A CN201610272162 A CN 201610272162A CN 105956543 A CN105956543 A CN 105956543A
Authority
CN
China
Prior art keywords
prime
time
local
operand
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610272162.XA
Other languages
Chinese (zh)
Inventor
王智文
蒋联源
王宇航
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangxi University of Science and Technology
Original Assignee
Guangxi University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangxi University of Science and Technology filed Critical Guangxi University of Science and Technology
Priority to CN201610272162.XA priority Critical patent/CN105956543A/en
Publication of CN105956543A publication Critical patent/CN105956543A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • G06V20/42Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items of sport video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • G06V20/53Recognition of crowd images, e.g. recognition of crowd congestion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/44Event detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a multiple athletes behavior detection method based on scale adaptation local spatiotemporal features, and the method comprises the following steps: S1, a normalized Laplace operand is adopted to estimate a local scale and establish a scale adaptation local spatiotemporal feature detection algorithm, and a Harris spatiotemporal interest point detection operand and the Laplace operand are combined to infer a Harris-Laplace spatiotemporal interest point detection operand; S2, the normalized Laplace operand is generalized to a spatiotemporal domain; S3, local spatiotemporal feature descriptors are generalized to three-dimensional soccer match video images; then before a K-means clustering algorithm is adopted to generate a spatiotemporal codebook, each spatiotemporal interest point is normalized to ensure that the each spatiotemporal interest point is invariable under scaling and shifting.

Description

A kind of person's of doing more physical exercises behavioral value method based on dimension self-adaption local space time feature
Technical field
The invention belongs to Activity recognition field, be specifically related to a kind of special based on dimension self-adaption local space time The person's of the doing more physical exercises behavioral value method levied.
Background technology
The quality of the person's of the doing more physical exercises behavior representation in section of football match video directly affects the accurate of Activity recognition Rate.During the person's of doing more physical exercises behavior analysis in football match and identification, the feature chosen is the most, Description to behavior is the most abundant.But choose too much feature and can cause the redundancy mistake between data Greatly, the dimension of characteristic vector is too high, and the essential laws of data distribution is not easy to be found, and trains behavior Identify that the data volume needed for model is excessive, algorithm computationally intensive, it is unfavorable for the real-time of Activity recognition Process.Therefore, what the person's of the doing more physical exercises Activity recognition in section of football match video was studied key challenge is how The feature effectively describing extraction can characterize behavior well.
Behavior representation is conducted extensive research by past researcher, wherein has than more typical research: BOBICK and DAVIS utilizes background subtraction to derive the expression of time template.The letter of behavior method for expressing Single, but the most affected by noise.WANG uses adaptive multi-thresholding to select Optical-flow Feature as behavior Represent and identify the team's behavior in section of football match video.But study limitation is in processing three class team row For, and easily by noise jamming.Existing employing seasonal effect in time series cause effect relation is described as behavior, And utilize spatial behavior path matching to identify behavior of men, but it is difficult to be generalized to the row of the person of doing more physical exercises For identifying, with four complete dimensional objects-real-time, interactive tensor, team's behavioral pattern is described, Reduce kernel by learning and optimizing tensor product, make it agglomerate to a diacritic space-time interaction square In Zhen.
The expression of target recognition based on model and probability Plan recognition constitutes four main assumptions: 1. in intelligence When can be engaged in team appointments between body, single intelligent body target is the natural former of the time-space relationship specified Subrepresentation unit;2. during the Activity recognition of the multiple agent of highly structural, the time of behavior The high-level description of structure uses less low order time-space relationship collection and logic to limit and just be enough to express intelligence Relation between energy body;3. the multiple source that Bayesian network is uncertain visually-perceptible feature provides A kind of suitable syncretizing mechanism;When 4. can merge uncertain with the Bayesian network automatically generated State information and calculating object trajectory data set are the probabilities of the behavior of a specific multiple agent.By Easily cause the biasing problem of labelling in the restriction assumed, and be difficult to.An energy is introduced bright for this Dynamic team member really encodes and adapts to the plan of new multiple agent of Plan recognition form Representation.The local time's dependency extracted from the plan representation method of multiple agent can be repaiied significantly Cut the hypothesis collection of potential effective team plan.Pruning process is relatively time-consuming, it is impossible to meet real-time football The requirement of match Activity recognition.
Hereafter during unsupervised time division, it is proposed that a kind of Gauss based on Dynamic Time Series The Activity recognition method of mixed model coupling behavior.The method suppose behavioral data be one can be abundant The how unidirectional label described, make use of space-time characteristic by assuming that the priority orders of behavior identifies many People's behavior.But in the character subset of team's Activity recognition, there will be multiple simultaneous events.Special In the case of, the behavior that the priority that is previously set is high is probably unessential behavior, so have impact on many The accuracy of athlete's Activity recognition.LAPTEV is succinct expression of video data to propose space-time interest Point, and inquired into the advantage of the behavior utilizing space-time interest points to describe people.
Use for reference the thought of LAPTEV, need a kind of new detection method to ensure that its zooming and panning not Degeneration.
Summary of the invention
For the deficiencies in the prior art, it is an object of the invention to provide a kind of based on dimension self-adaption local time The person's of doing more physical exercises behavioral value method of empty feature has been capable of before cluster generates code book, to each Space-time interest points has all carried out normalization, to ensure its zooming and panning invariance.
A kind of person's of doing more physical exercises behavioral value method based on dimension self-adaption local space time feature, described side Method comprises the following steps:
S1, uses normalized Laplace operand to estimate local scale, sets up dimension self-adaption
Local space time's feature detection algorithm, by Harris space-time interest points detection operand and Laplace
Operand combines and derives Harris-Laplace space-time interest points detection operand;
S2, is generalized to normalized Laplace operand in time-space domain;
S3, is generalized to local space time's Feature Descriptor in three-dimensional section of football match video image.
Preferably, described S1 mesoscale adaptive local space-time characteristic detection algorithm is specially
H = λ 1 3 ( α β - k ( 1 + α + β ) ) - - - ( 1 )
Wherein, the positive local maximum of H corresponds to λ1、λ2And λ3High level point, λ1< λ2< λ3, Ratio, α=λ21, β=λ31.From requiring that H >=0 can obtain k≤α β/(1+ alpha+beta)3
Preferably, local space time's Feature Descriptor is generalized to three-dimensional section of football match video figure by described S3 Computational methods in Xiang include: histogram calculation, the calculating of direction quantization and average gradient calculation.
Preferably, described histogram calculation and direction quantify calculating particularly as follows:
For a given cube c=(xc,yc,tc,lc,wc,hc), it is classified as S × S × S sub-block bi。 For each sub-block, utilize formula (4) to calculate its average gradient, then willIt is quantified as qbi.For Each region c, the average gradient vector after being quantified by summation calculates its rectangular histogram hc
h c = Σ i = 1 S 3 q b i - - - ( 3 )
For a given positive n face body, its center is exactly the initial point of three-dimensional system of coordinate, 3D gradient vector Being quantified by the translation of coordinate points relative to the direction of initial point, each point in matrix p is polyhedron The center p of each1,p2,…,pn, p=(p1,p2,…,pn)T, wherein, pi=(xi,yi,ti)T, then in 20 Heart point is:
Wherein, golden section
?On projection can pass through formula (5) and calculate;
q ^ b = [ q ^ b 1 , q ^ b 2 , ... , q ^ b n ] T = p g ‾ b | | g ‾ b | | 2 - - - ( 5 ) ,
WillThresholding is processed asThen formula (6) is utilized to carry out quantification treatment;
q b = | | g b | | 2 q ^ b t h | | q b t h | | 2 - - - ( 6 ) .
Preferably, described average gradient calculate particularly as follows:
By section of football match video sequence v, (x, y, t) along x, the partial derivative in y, t direction is divided into and being designated asThen Formula (7) can be used to calculate
iv ∂ x ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ x ( x ′ , y ′ , t ′ ) iv ∂ y ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ y ( x ′ , y ′ , t ′ ) iv ∂ t ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ t ( x ′ , y ′ , t ′ ) - - - ( 7 )
For a 3D cube b=(x, y, t, l, w, h), wherein, (x, y, t)TAnd l, w, h represent cube respectively The position of body and length and width and height;
Formula (8) can be used quickly to calculate its average gradient vector'sPoint Amount;
g ‾ b ∂ x = [ iv ∂ x ( x + w , y + h , t + l ) - iv ∂ x ( x , y + h , t + l ) - iv ∂ x ( x + w , y , t + l ) + iv ∂ x ( x , y , t + l ) ] - [ iv ∂ x ( x + w , y + h , t ) - iv ∂ x ( x , y + h , t ) - iv ∂ x ( x + w , y , t ) + iv ∂ x ( x , y , t ) ] - - - ( 8 )
In like manner, can quickly calculate
Technical scheme has the advantages that
The present invention provides a kind of person's of doing more physical exercises behavioral value based on dimension self-adaption local space time feature Method, utilizes space-time interest points to carry out behavior knowledge to the person's of the doing more physical exercises behavior representing in section of football match video Not.The person's of doing more physical exercises behavior in section of football match video sequence is regarded as the space-time interest in three dimensions The set of point.Use rectangular histogram quantification technique that space-time interest points set is quantified as the Nogata that dimension is fixing Figure (instant null word), uses K-means clustering algorithm to generate space-time code book.At cluster generated code Before Ben, each space-time interest points is carried out normalization, to ensure its zooming and panning invariance. Test result indicate that, method in this paper can greatly reduce the amount of calculation of algorithm, with based on yardstick The person's of doing more physical exercises behavior representation in the section of football match video of adaptive local space-time characteristic is done more physical exercises The accuracy rate of identification can be significantly improved during member's Activity recognition.
Accompanying drawing explanation
Below by drawings and Examples, technical scheme is described in further detail.
Fig. 1 a is the present invention person of doing more physical exercises behavioral value side based on dimension self-adaption local space time feature Whole description of method;
Fig. 1 b is the present invention person of doing more physical exercises behavioral value side based on dimension self-adaption local space time feature The histogram calculation of method;
Fig. 1 c is the present invention person of doing more physical exercises behavioral value side based on dimension self-adaption local space time feature The direction gradient of method quantifies;
Fig. 1 d is the present invention person of doing more physical exercises behavioral value side based on dimension self-adaption local space time feature The average gradient of method calculates.
Detailed description of the invention
In order to have a clear understanding of technical scheme, its detailed knot be will be set forth in the description that follows Structure.Obviously, concrete execution the deficiency of the embodiment of the present invention is limited to those skilled in the art and is familiar with Specific details.The preferred embodiments of the present invention are described in detail as follows, except these described in detail are implemented Exception, it is also possible to there is other embodiments.
With embodiment, the present invention is described in further details below in conjunction with the accompanying drawings.
Use for reference LINDEBERG herein about the adaptively selected method of local scale feature in space, will Harris detection operand is generalized in section of football match video time-space domain, it is proposed that dimension self-adaption local Space-time characteristic detection algorithm (1), can be simultaneously in time and space at defined in local time spatial domain one Obtain the difference operation number of maximum in dimension, use normalized Laplace operand to estimate local Yardstick.Formula (1) is utilized to be tied mutually with Laplace operand by Harris space-time interest points detection operand Harris-Laplace space-time interest points detection operand is derived in conjunction.
H = λ 1 3 ( α β - k ( 1 + α + β ) ) - - - ( 1 )
Wherein, the positive local maximum of H corresponds to λ1、λ2And λ3High level point, λ1< λ2< λ3, Ratio, α=λ21, β=λ31.From requiring that H >=0 can obtain k≤α β/(1+ alpha+beta)3.Assuming that α=β=1, The maximum value possible of k is 1/27.When k value is sufficiently large, point corresponding to the positive local maximum of H along time Between and direction in space image value change acutely.Working as the maximum of α and β in spatial domain especially is 23 Time, k ≈ 0.005.Therefore, football ratio can be detected by the positive local space time's maximum in detection H Space-time interest points in match sequence of video images v.
Formula (2) is utilized to be generalized in time-space domain by normalized Laplace operand.
▿ n o r m 2 r = r x 2 + r y 2 + r t 2 = σ r 2 τ r 1 / 2 ( r x 2 + r y 2 ) + σ r τ r 3 / 2 r t 2 - - - ( 2 )
Dimension self-adaption space-time after formula 1 describes popularization in detail detects specifically performing of operand Journey.
In section of football match video, the expression of regional area is an open problem.Utilization orientation ladder herein Degree rectangular histogram (Histogram Oriented Gradient, HoG) describes son by section of football match video It is considered as " Space Time " cuboid, and HoG is described son is generalized in three-dimensional section of football match video image.
For local space time's area-of-interest, local description represents this district by a characteristic vector Territory.Image or whole section of football match video sequence are represented by one group under different yardsticks and position Set of eigenvectors.In order to be able to effectively utilize local feature vectors to carry out in section of football match video The identification of the person's of doing more physical exercises behavior, objectively requires that these Feature Descriptors have stronger differentiation energy Power, simultaneously the most by illumination, the impact of the interference factors such as slight deformation.HoG is described son be generalized to The computational methods of 3D are as it is shown in figure 1, crucial calculating theing be directed to has: average gradient calculates, side The calculating of vectorization, histogram calculation.These three described in detail below calculates, in conjunction with Fig. 1 a to Fig. 1 d. Histogram calculation: the rectangular histogram of direction gradient needs to calculate in the set of a gradient vector. Circular is as follows: for a given cube c=(xc,yc,tc,lc,wc,hc), it is classified as S × S × S sub-block bi.For each sub-block, utilize formula (4) to calculate its average gradient, then will It is quantified as qbi.For each region c, it is straight that the average gradient vector after being quantified by summation calculates it Side figure hc
h c = Σ i = 1 S 3 q b i - - - ( 3 )
Direction gradient quantifies: in static 2D image, the histograms of oriented gradients of a n-bin is permissible Seeing as and surrounded by the regular polygon on n limit, each limit of regular polygon be equivalent in rectangular histogram One bin.In the 3 d space, regular polygon is extended to regular polygon (such as Fig. 1 (c)), often The limit number of the regular polygon seen has 4,6,8,12,20.Therefore, the polyhedron limit number used herein It is 20.For a given positive n face body, it is assumed that its center is exactly the initial point of three-dimensional system of coordinate, 3D Gradient vector is quantified by the translation (i.e. multiplication of matrices) of coordinate points relative to the direction of initial point. The center p that each point is each of polyhedron in matrix p1,p2,…,pn, p=(p1,p2,…,pn)T, its In, pi=(xi,yi,ti)T.Then 20 central points are:
Wherein, golden section?On projection can pass through formula (5) and calculate.
q ^ b = [ q ^ b 1 , q ^ b 2 , ... , q ^ b n ] T = p g ‾ b | | g ‾ b | | 2 - - - ( 5 )
It follows that willThresholding is processed asThen formula (6) is utilized to carry out quantification treatment.
q b = | | g b | | 2 q ^ b t h | | q b t h | | 2 - - - ( 6 )
The calculating of average gradient: in order to calculate 3D histograms of oriented gradients it may first have to the most effective Calculate the gradient vector (such as Fig. 1 (d)) of image.Due to must take into different space scales and Time scale, an effective solution uses space-time pyramid exactly, precalculates each different Direction gradient under yardstick.Although but this method reduces time complexity, substantially increases Space complexity.Calendar year 2001, Viola etc. proposes integrogram concept, and using integrogram as Harr The intermediate representation method of feature calculation is applied to face detection system, substantially increases computational efficiency.Ginseng The method examining the proposition such as Viola, by section of football match video sequence v (x, y, t) along x, the partial derivative in y, t direction It is divided into and being designated asFormula (7) then can be used to calculate
iv ∂ x ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ x ( x ′ , y ′ , t ′ ) iv ∂ y ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ y ( x ′ , y ′ , t ′ ) iv ∂ t ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ t ( x ′ , y ′ , t ′ ) - - - ( 7 )
For a 3D cube b=(x, y, t, l, w, h), wherein, (x, y, t)TAnd l, w, h represent cube respectively The position of body and length and width and height.
Formula (8) can be used quickly to calculate its average gradient vector'sPoint Amount.
g ‾ b ∂ x = [ iv ∂ x ( x + w , y + h , t + l ) - iv ∂ x ( x , y + h , t + l ) - iv ∂ x ( x + w , y , t + l ) + iv ∂ x ( x , y , t + l ) ] - [ iv ∂ x ( x + w , y + h , t ) - iv ∂ x ( x , y + h , t ) - iv ∂ x ( x + w , y , t ) + iv ∂ x ( x , y , t ) ] - - - ( 8 )
In like manner, can quickly calculateSo, by using for reference the thought of integrogram, Ke Yi great The big space-time interest points improving the person's of doing more physical exercises behavior represents
Space-time interest points refers to the point that temporally and spatially Strength Changes is bigger.Space-time interest points table Show that method is the new frame sequence low-level image feature method for expressing of a kind of comparison.If section of football match video space-time diagram As sequence isBy v with there is different spaces varianceAnd time varianceEach to Anisotropic Gaussian core convolution constructs linear-scale space representation
R ( · ; σ r 2 , τ r 2 ) = g ( . ; σ r 2 , τ r 2 ) * v ( · ) - - - ( 9 )
The separable gaussian kernel of space-time is defined by formula (10):
g ( x , y , t ; σ r 2 , τ r 2 ) = exp ( - ( x 2 + y 2 ) / 2 σ r 2 - t 2 / 2 τ r 2 ) ( 2 π ) 3 σ r 4 τ r 2 - - - ( 10 )
Time domain and spatial domain are used separately single scale parameter it is critical that, because event time Between and spatial component be in general independent.And, detect event with space-time interest point operators and depend on In the time observed and space scale, accordingly, it would be desirable to separately process corresponding scale parameterWith
Consider with 3 × 3 space-time second-order moments matrixes being made up of single order room and time yardstick, and with height This weighting function is averaging.
g ( . ; σ r 2 , τ r 2 ) μ = g ( . ; σ r 2 , τ r 2 ) * r x 2 r x r y r x r t r x r y r y 2 r y r t r x r t r y r t r t 2 - - - ( 11 )
UtilizeWithBy the scale parameter in formula (9)WithIt is fused to local scale parameterWithIn, and define first derivative and be: the calculating speed of gradient in 3d space
r x ( x , y , t ; σ r 2 , τ r 2 ) = ∂ x ( g * v ) r y ( x , y , t ; σ r 2 , τ r 2 ) = ∂ y ( g * v ) r t ( x , y , t ; σ r 2 , τ r 2 ) = ∂ t ( g * v ) - - - ( 12 )
In order to detect space-time interest points, search in μ in video v and there is marked feature value λ1、λ2And λ3District Territory.In the distinct methods of region of search, determinant and tracking expanded definition in conjunction with μ are for spatial domain Harris Angle function be time-space domain Harris Angle function:
H=λ1λ2λ3-k(λ123)3,(λ1≤λ2≤λ3) (13)
The present invention provides a kind of person of doing more physical exercises behavioral value side based on dimension self-adaption local space time feature Method, utilizes space-time interest points to carry out behavior knowledge to the person's of the doing more physical exercises behavior representing in section of football match video Not.The person's of doing more physical exercises behavior in section of football match video sequence is regarded as the space-time interest in three dimensions The set of point.Use rectangular histogram quantification technique that space-time interest points set is quantified as the Nogata that dimension is fixing Figure (instant null word), uses K-means clustering algorithm to generate space-time code book.At cluster generated code Before Ben, each space-time interest points is carried out normalization, to ensure its zooming and panning invariance. Test result indicate that, method in this paper can greatly reduce the amount of calculation of algorithm, with based on yardstick The person's of doing more physical exercises behavior representation in the section of football match video of adaptive local space-time characteristic is done more physical exercises The accuracy rate of identification can be significantly improved during member's Activity recognition.
Finally should be noted that: above example is only in order to illustrate technical scheme rather than right It limits, although the present invention being described in detail with reference to above-described embodiment, and art general The detailed description of the invention of the present invention still can be modified or equivalent by logical technical staff, this A bit without departing from any amendment or the equivalent of spirit and scope of the invention, the power all awaited the reply in application Within the scope of profit is claimed.

Claims (5)

1. the person's of doing more physical exercises behavioral value method based on dimension self-adaption local space time feature, its It is characterised by, said method comprising the steps of:
S1, uses normalized Laplace operand to estimate local scale, sets up dimension self-adaption Local space time's feature detection algorithm, by Harris space-time interest points detection operand and Laplace Operand combines and derives Harris-Laplace space-time interest points detection operand;
S2, is generalized to normalized Laplace operand in time-space domain;
S3, is generalized to local space time's Feature Descriptor in three-dimensional section of football match video image.
The person of doing more physical exercises based on dimension self-adaption local space time feature the most according to claim 1 Behavioral value method, it is characterised in that the detection of described S1 mesoscale adaptive local space-time characteristic is calculated Method is specially
H = λ 1 3 ( α β - k ( 1 + α + β ) ) - - - ( 1 )
Wherein, the positive local maximum of H corresponds to λ1、λ2And λ3High level point, λ123, Ratio, α=λ21, β=λ31.From requiring that H >=0 can obtain k≤α β/(1+ alpha+beta)3
The person of doing more physical exercises based on dimension self-adaption local space time feature the most according to claim 1 Behavioral value method, it is characterised in that local space time's Feature Descriptor is generalized to three-dimensional by described S3 Computational methods in section of football match video image include: the calculating peace that histogram calculation, direction quantify All gradient calculation.
The person of doing more physical exercises based on dimension self-adaption local space time feature the most according to claim 3 Behavioral value method, it is characterised in that calculating that described histogram calculation and direction quantify particularly as follows:
For a given cube c=(xc,yc,tc,lc,wc,hc), it is classified as S × S × S sub-block bi; For each sub-block, utilize formula (4) to calculate its average gradient, then willIt is quantified as qbi.For Each region c, the average gradient vector after being quantified by summation calculates its rectangular histogram hc
h c = Σ i = 1 S 3 q b i - - - ( 3 )
For a given positive n face body, its center is exactly the initial point of three-dimensional system of coordinate, 3D gradient vector Being quantified by the translation of coordinate points relative to the direction of initial point, each point in matrix p is polyhedron The center p of each1,p2,…,pn, p=(p1,p2,…,pn)T, wherein, pi=(xi,yi,ti)T, then in 20 Heart point is:
Wherein, golden section
?On projection can pass through formula (5) and calculate;
q ^ b = [ q ^ b 1 , q ^ b 2 , ... , q ^ b n ] T = p g ‾ b | | g ‾ b | | 2 - - - ( 5 ) ,
WillThresholding is processed asThen formula (6) is utilized to carry out quantification treatment;
q b = | | g b | | 2 q ^ bth | | q bth | | 2 - - - ( 6 ) .
The person of doing more physical exercises based on dimension self-adaption local space time feature the most according to claim 4 Behavioral value method, it is characterised in that described average gradient calculate particularly as follows:
By section of football match video sequence v, (x, y, t) along x, the partial derivative in y, t direction is divided into and being designated asThen Formula (7) can be used to calculate
iv ∂ x ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ x ( x ′ , y ′ , t ′ ) iv ∂ y ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ y ( x ′ , y ′ , t ′ ) iv ∂ t ( x , y , t ) = Σ x ′ ≤ x , y ′ ≤ y , t ′ ≤ t v ∂ t ( x ′ , y ′ , t ′ ) - - - ( 7 )
For a 3D cube b=(x, y, t, l, w, h), wherein, (x, y, t)TAnd l, w, h represent cube respectively The position of body and length and width and height;
Formula (8) can be used quickly to calculate its average gradient vector'sPoint Amount;
g ‾ b ∂ x = [ iv ∂ x ( x + w , y + h , t + l ) - iv ∂ x ( x , y + h , t + l ) - iv ∂ x ( x + w , y , t + l ) + iv ∂ x ( x , y , t + l ) ] - [ iv ∂ x ( x + w , y + h , t ) - iv ∂ x ( x , y + h , t ) - iv ∂ x ( x + w , y , t ) + iv ∂ x ( x , y , t ) ] - - - ( 8 )
In like manner, can quickly calculate
CN201610272162.XA 2016-04-27 2016-04-27 Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features Pending CN105956543A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610272162.XA CN105956543A (en) 2016-04-27 2016-04-27 Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610272162.XA CN105956543A (en) 2016-04-27 2016-04-27 Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features

Publications (1)

Publication Number Publication Date
CN105956543A true CN105956543A (en) 2016-09-21

Family

ID=56916052

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610272162.XA Pending CN105956543A (en) 2016-04-27 2016-04-27 Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features

Country Status (1)

Country Link
CN (1) CN105956543A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107025433A (en) * 2017-03-03 2017-08-08 深圳大学 Video Events class people's concept learning method and device
CN109858390A (en) * 2019-01-10 2019-06-07 浙江大学 The Activity recognition method of human skeleton based on end-to-end space-time diagram learning neural network

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1886759A (en) * 2003-11-24 2006-12-27 皇家飞利浦电子股份有限公司 Detection of local visual space-time details in a video signal
CN105160287A (en) * 2015-07-02 2015-12-16 中国人民解放军国防科学技术大学 Method for detecting characteristics of space-time interest point under condition of no camera movement

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1886759A (en) * 2003-11-24 2006-12-27 皇家飞利浦电子股份有限公司 Detection of local visual space-time details in a video signal
CN105160287A (en) * 2015-07-02 2015-12-16 中国人民解放军国防科学技术大学 Method for detecting characteristics of space-time interest point under condition of no camera movement

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
GAURAV SHARMA 等: "Expanded Parts Model for Human Attribute and Action Recognition in Still Images", 《2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
王智文: "足球比赛视频中的多运动员的行为识别方法研究", 《厦门大学博士学位论文》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107025433A (en) * 2017-03-03 2017-08-08 深圳大学 Video Events class people's concept learning method and device
CN107025433B (en) * 2017-03-03 2021-03-16 深圳大学 Video event human concept learning method and device
CN109858390A (en) * 2019-01-10 2019-06-07 浙江大学 The Activity recognition method of human skeleton based on end-to-end space-time diagram learning neural network
CN109858390B (en) * 2019-01-10 2020-11-24 浙江大学 Human skeleton behavior identification method based on end-to-end space-time diagram learning neural network

Similar Documents

Publication Publication Date Title
Masita et al. Deep learning in object detection: A review
Redmon et al. YOLO9000: better, faster, stronger
Wei et al. Learning to segment with image-level annotations
Yang et al. Layered object models for image segmentation
Girshick et al. Rich feature hierarchies for accurate object detection and semantic segmentation
Wang et al. Spatially encoding temporal correlations to classify temporal data using convolutional neural networks
CN110073369A (en) The unsupervised learning technology of time difference model
Sun et al. Novel hybrid CNN-SVM model for recognition of functional magnetic resonance images
CN102289685B (en) Behavior identification method for rank-1 tensor projection based on canonical return
Qiu et al. Hallucinating visual instances in total absentia
CN106228027A (en) A kind of semi-supervised feature selection approach of various visual angles data
CN109800768A (en) Hash feature representation learning method of semi-supervised GAN
Yu et al. Exemplar-based recursive instance segmentation with application to plant image analysis
Narayanan et al. Yoga pose detection using deep learning techniques
Zhang et al. Multiresolution attention extractor for small object detection
Wang et al. Building correlations between filters in convolutional neural networks
Alashban et al. Single convolutional neural network with three layers model for crowd density estimation
Ouadiay et al. Simultaneous object detection and localization using convolutional neural networks
Liu et al. Double Mask R‐CNN for Pedestrian Detection in a Crowd
Zhu et al. Graph-based reasoning attention pooling with curriculum design for content-based image retrieval
CN105956543A (en) Multiple athletes behavior detection method based on scale adaptation local spatiotemporal features
Xu et al. Representative feature alignment for adaptive object detection
Kuang et al. Multi-label Image Classification with Multi-scale Global-Local Semantic Graph Network
Lei et al. Continuous action recognition based on hybrid CNN-LDCRF model
Ullah et al. A strict pyramidal deep neural network for action recognition

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20160921