CN107085716A - Across the visual angle gait recognition method of confrontation network is generated based on multitask - Google Patents

Across the visual angle gait recognition method of confrontation network is generated based on multitask Download PDF

Info

Publication number
CN107085716A
CN107085716A CN201710373017.5A CN201710373017A CN107085716A CN 107085716 A CN107085716 A CN 107085716A CN 201710373017 A CN201710373017 A CN 201710373017A CN 107085716 A CN107085716 A CN 107085716A
Authority
CN
China
Prior art keywords
msub
mrow
gait
visual angle
exemplary feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710373017.5A
Other languages
Chinese (zh)
Other versions
CN107085716B (en
Inventor
何逸炜
张军平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fudan University
Original Assignee
Fudan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fudan University filed Critical Fudan University
Priority to CN201710373017.5A priority Critical patent/CN107085716B/en
Publication of CN107085716A publication Critical patent/CN107085716A/en
Application granted granted Critical
Publication of CN107085716B publication Critical patent/CN107085716B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/49Segmenting video sequences, i.e. computational techniques such as parsing or cutting the sequence, low-level clustering or determining units such as shots or scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/285Selection of pattern recognition techniques, e.g. of classifiers in a multi-classifier system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components

Abstract

The invention belongs to computer vision and machine learning field, specially a kind of across visual angle gait recognition method that confrontation network is generated based on multitask.Present invention mainly solves Gait Recognition the problem of big visual angle change drag Generalization Capability is reduced.For original pedestrian's sequence of frames of video, every two field picture is pre-processed first, and extracts gait exemplary feature;Then the hidden expression of gait is encoded to by neutral net, and carries out in latent space angular transformation;The gait exemplary feature that confrontation network reconfiguration goes out other visual angles is generated by multitask again;Finally it is identified using the hidden expression of gait.Compared to based on classification or other methods based on reconstruct, the present invention has stronger interpretation, and can lift recognition performance.

Description

Across the visual angle gait recognition method of confrontation network is generated based on multitask
Technical field
The invention belongs to computer vision, machine learning techniques field, and in particular to across the visual angle gait based on video is known Other method.
Background technology
One of the problem of across visual angle Gait Recognition problem based on video is computer vision and machine learning area research. Gait sequence of frames of video under given different visual angles, it is desirable to which gait frame sequence is judged according to computer vision or machine learning algorithm Whether the main body of row is same object.There have been many previous works in the current field, and it is big that its main method can be divided into three Class:Method based on reconstruct, the method based on subspace and the method based on deep learning.The following is some of this three classes method Bibliography:
[1]W.Kusakunniran,Q.Wu,J.Zhang,and H.Li,“Support vector regression for multi-view gait recognition based on local motion feature selection,”in Conference on Computer Vision and Pattern Recognition,pp.974–981,2010.
[2]M.Hu,Y.Wang,Z.Zhang,J.J.Little,and D.Huang,“View-invariant discriminative projection for multi-view gait-based human identification,” IEEE Transactions on Information Forensics and Security,vol.8,no.12,pp.2034– 2045,2013.
[3]W.Kusakunniran,Q.Wu,J.Zhang,H.Li,and L.Wang,“Recognizing gaits across views through correlated motion co-clustering,”IEEE Transactions on Image Processing,vol.23,no.2,pp.696–709,2014.
[4]S.Yu,H.Chen,Q.Wang,L.Shen,and Y.Huang,“Invariant feature extraction for gait recognition using only one uniform model,”Neurocomputing, vol.239,pp.81–93,2017.
[5]Z.Wu,Y.Huang,L.Wang,X.Wang,and T.Tan,“A comprehensive study on cross-view gait based human identification with deep CNNs,”IEEE Transactions on Pattern Analysis and Machine Intelligence,vol.39,no.2,pp.209–226,2017.
[6]Y.Feng,Y.Li,and J.Luo,“Learning effective gait features using LSTM,”in International Conference on Pattern Recognition,pp.320–325,2016.
[7]H.Iwama,M.Okumura,Y.Makihara,and Y.Yagi,“The OU-ISIR gait database comprising the large population dataset and performance evaluation of gait recognition,”IEEE Transactions on Information Forensics and Security,vol.7, no.5,pp.1511–1521,2012.
[8]Y.Makihara,R.Sagawa,Y.Mukaigawa,T.Echigo,and Y.Yagi,“Gait recognition using a view transformation model in the frequency domain,”in European Conference on Computer Vision,pp.151–163,Springer,2006.
[9]W.Kusakunniran,Q.Wu,H.Li,and J.Zhang,“Multiple views gait recognition using view transformation model based on optimized gait energy image,”in International Conference on Computer Vision,pp.1058–1064,2009.
[10]X.Xing,K.Wang,T.Yan,and Z.Lv,“Complete canonical correlation analysis with application to multi-view gait recognition,”Pattern Recognition,vol.50,pp.107–117,2016.。
The first method based on reconstruct is main based on reconstructing gait masterplate, such as the VTM moulds in [1,3,8,9] Type, this method predicts the value of each pixel in exemplary feature under aspect, this method by training multiple models, respectively Big, the low problem of recognition accuracy with computing cost.In order to reduce computing cost, the method [4] based on self-encoding encoder is come same When reconstruct each pixel of exemplary feature, and extract the feature representation of unchanged view angle.
The gait feature of different visual angles is projected to a public subspace by the method [2,10] based on subspace, and in public affairs Altogether similarity is calculated in subspace.But subspace method is typically only capable to, the linear dependence modeling feature, have ignored non- Linear dependence, discrimination is relatively low.
Recently, the model based on deep neural network has obtained larger application in computer vision field.[5,6] will be deep Spending network is used for Gait Recognition, and the feature representation of unchanged view angle has been extracted automatically.Although compared with previous methods on recognition accuracy Larger lifting has been obtained, but depth model lacks interpretation.
The content of the invention
It is an object of the invention to provide across the visual angle Gait Recognition based on video that a kind of discrimination height, computing cost are saved Method.
The relevant substance of the present invention is introduced first.
First, periodical energy figure
(a) in order to set up more effective gait exemplary feature, it is necessary to be pre-processed to original pedestrian's video sequence.It is right Each frame in video sequence, the separation of background, extracts gait profile diagram, afterwards occupies gait profile diagram before carrying out first In and be aligned to picture center, the gait profile diagram standardized;
(b) the gait profile graphic sequence based on standardization, is obtained often first by the cycle detection technology in Gait Recognition The period position of the standardization of one frame.A gait cycle is defined between two adjacent Local Minimums standardization period positions Time-domain.Afterwards, profile graphic sequence is divided into one or more gait cycles, and discarding can not constitute a complete gait week The frame of phase.According to the gait profile diagram sequence B of centralizationtWith the period position r of corresponding standardizationt, construction is with ncIndividual passage Periodical energy figure (PEI) exemplary feature.Value in periodical energy figure PEI at (x, y) coordinate of k-th of passage is counted as the following formula Calculate:
Wherein:
M represents the time-domain window size that each passage is covered, and T (k) determines the time-domain of each passage covering Scope.By this method, the spatial information of profile graphic sequence can be divided to by n according to the period position of each framecThe individual time On domain, to encode time and the spatial information of gait simultaneously with this.
2nd, multitask generation confrontation network
The multitask generation confrontation network for Gait Recognition of the present invention is as shown in figure 1, main become by encoder, visual angle Change layer, maker and the part of arbiter four composition.Wherein:
(a) encoder E:In order to obtain the hidden expression of certain viewing angles, using convolutional neural networks as model in coding Device.The input x of encoderuIt is the PEI masterplates that visual angle proposed by the present invention is u.The scale size of input is 64 × 64 × nc。PEI Each passage of masterplate use time pond and is extracted effectively by independent feeding encoder, afterwards after full articulamentum The hidden expression z of featureuIt is used as the output of encoder.For time pond, the mode used is average pond;
(b) view transformation layer V:It is assumed that gait data is as the change profile at visual angle is in the manifold of a higher-dimension.Manifold In sample the change at visual angle can be realized under conditions of identity information is kept towards specific direction movement.Given latent space Hidden expression zu, view transformation can be described as:
Wherein, hiRepresent the conversion vector by visual angle i-1 to visual angle i.zvIt is the hidden expression under the v of visual angle;
(c) maker G:By hidden expression zvAs input, the gait masterplate under the v of visual angle can be generated.Maker It is made up of five warp laminations, and uses ReLU as excitation function.Because directly reconstruct PEI masterplates have certain difficulty, Therefore, a passage of random selection PEI masterplates is reconstructed, and ensures hidden to represent that the letter of all passages can be preserved with this Breath.The input for defining maker is [zv, c], c represents for the one-hot coding of passage;
(d) arbiter D:Dimension and encoder of the framework of arbiter except not using the operation of time pondization and output layer Different outer, remaining structure is identical with encoder.The output dimension for defining arbiter is nv+nc+nd, wherein nvIt is the quantity at visual angle, ncIt is the quantity of PEI passages, ndIt is the quantity of different identity in training set.The output correspondence n of arbiterv+nc+ndHeight differentiates The parameter in addition to last layer is shared between device, sub- arbiter.Also, each arbiter is responsible for different differentiation tasks Judge whether that the sample of generation belongs to specific distribution.For example, in a model, preceding nv+ncIndividual sub- arbiter enables maker Generation meets the distribution of certain viewing angles and passage, then ndIndividual sub- arbiter can make maker generation belong to dividing for specific identity Sample corresponding to cloth.
3rd, loss function
The present invention trains multitask generation confrontation network using two kinds of loss functions.
(a) lose pixel-by-pixel:In order to strengthen the hidden ability for representing to keep identity information, generation template is minimized first and true Loss pixel-by-pixel between real masterplate.According to exemplary feature xvWith pseudo- exemplary featureCalculating loss pixel-by-pixel is:
(b) multitask confrontation loss:According to exemplary feature xvWith pseudo- exemplary featureCalculate multitask confrontation loss:
Wherein, E represents the expectation for correspondence sample set, | | | |1Represent L1 norms, vectorial s be identity, passage with The one-hot coding of angle information.D () represents the output of arbiter, and the dimension of arbiter output is identical with vector s dimension, to Nonzero element in amount s determines the distribution that pseudo- exemplary feature should belong to;
Defining final loss function is:
L=Lp+αLa (6)
α is used for weighing loss pixel-by-pixel as hyper parameter and lost with multitask confrontation, defines after final loss function, Use the parameter of Back Propagation Algorithm alternating more new encoder, view transformation layer, maker and arbiter.
Across the visual angle gait recognition method proposed by the present invention that confrontation network is generated based on multitask, is concretely comprised the following steps:
(1) pedestrian's sequence of frames of video under input different visual angles, constructs gait exemplary feature:
Vector xiIt is the gait exemplary feature under the i of visual angle, nvIt is the quantity at all visual angles;
(2) for any different visual angle u, v, corresponding gait exemplary feature x is encoded using convolutional neural networksuIt is extremely hidden Space, and obtain in latent space hidden be expressed as zu
(3) to hidden expression z in latent spaceuView transformation is carried out, angle v is converted into, obtains hidden expression zv
(4) by hidden expression zvAnd the one-hot coding of passage generates the life in confrontation network as input by multitask Pseudo- exemplary feature after into network under output angle v
(5) according to exemplary feature xuWith pseudo- exemplary featureL is lost in calculating pixel-by-pixelp
(6) the differentiation network in confrontation network is generated using multitask, according to exemplary feature xuWith pseudo- exemplary featureMeter Calculate multitask confrontation loss La
(7) loss resists loss with multitask pixel-by-pixel for weighting, according to total losses L=Lp+αLaTrain multitask generation pair Anti- network, α is used for weighing loss pixel-by-pixel as hyper parameter and lost with multitask confrontation.
In the present invention, the construction step of described gait masterplate is:
(1) preceding background separation is carried out to each frame of original video frame sequence, extracts gait profile diagram;And by gait wheel Wide figure translation zooms to picture centre, obtains the gait profile graphic sequence { B of centralization1,B2,B3,…,Bn};
(2) to each frame, the period position r of normalized in the gait profile graphic sequence of centralizationt, represent centralization Gait profile diagram t frames standardization period position;
(3) according to the gait profile diagram sequence B of centralizationtWith the period position r of corresponding standardizationt, construction is with ncIt is individual The periodical energy figure exemplary feature of passage;Value in periodical energy figure PEI at (x, y) coordinate of k-th of passage is calculated as follows:
Wherein:
M represents the time-domain window size that each passage is covered, and T (k) determines the time-domain of each passage covering Scope.
In the present invention, described view transformation step is:
Hidden expression z under given visual angle uu, transforming to visual angle v calculating process can be described as:
Wherein, hiRepresent the conversion vector by visual angle i-1 to visual angle i.
In the present invention, described loss LpWith LaCalculation procedure be:
(1) according to exemplary feature xvWith pseudo- exemplary featureCalculating is lost pixel-by-pixel:
(2) according to exemplary feature xvWith pseudo- exemplary featureCalculate multitask confrontation loss:
Wherein, E represents the expectation for correspondence sample set, | | | |1Represent L1 norms, vectorial s be identity, passage with The one-hot coding of angle information.D () represents the output of arbiter.
The inventive method uses the Nonlinear Modeling ability of convolutional neural networks, is extracted the specific hidden expression in visual angle.It is logical Cross and view transformation is carried out in latent space, reduce computing cost;And the ability of distribution modeling is carried based on generation confrontation network The feature with more expressiveness has been taken out, recognition efficiency is greatly improved.
Brief description of the drawings
Fig. 1:The detailed model flow chart of the present invention.
Fig. 2:OU-ISIR, CASIA-B, USF data set sample are shown.
Fig. 3:Average Accuracy under different walking states on CASIA-B data sets.
Embodiment
After the specific steps and model that describe the present invention, test of the invention in several gait data collection is shown below Effect.
Experiment employs three data sets, including OU-ISIR data sets, CASIA-B data sets and USF data sets.Fig. 2 Illustrate some samples of these three data sets.
OU-ISIR data sets one have 4007 different people, wherein male 2135, women 1872, age distribution Scope was from 1 years old to 94.The data have 4 different angles.Respectively 55 °, 65 °, 75 °, 85 °.Carried using the method for invention PEI masterplates are taken out, and set port number to be 3, after the picture of PEI interpolation to 64 × 64 pixel sizes, multitask generation are put into It is trained in confrontation network.
CASIA-B data sets one have 124 different people, 11 different visual angles.Wherein, everyone is at each visual angle On have 6 groups of gait sequences normally walked, 2 groups are carried the gait sequence that packet rows are walked, 2 groups of gait sequences of overcoat for wearing. CASIA-B has bigger angular field of view compared to OU-ISIR, but number is relatively fewer.We set PEI port number to be 10, The Gait Recognition accuracy rate tested respectively under different walking states.
USF is another conventional gait data collection, and one has 122 different people, everyone have 5 kinds it is various forms of Gait sequence, close to real scene.In an experiment, we are only tested using the gait sequence under different visual angles.I Set PEI port number be 10.
Experiment uses Rank-1 recognition accuracies as performance indications.By being used most in the latent space of view transformation layer Nearest Neighbor Classifier is identified.
Experimental example 1:The recognition performance of multitask generation confrontation network
This part Experiment illustrates different models, across recognition accuracy under visual angle.Method as a comparison, we select Self-encoding encoder, canonical correlation analysis, linear discriminant analysis, convolutional neural networks and local tensors discrimination model.Table 1 is illustrated Method of the invention and other method compare on three data sets.It can be seen that, the present invention has very big compared to other method Lifting.
Experimental example 2:Influence of the different loss functions to model performance
When table 2 is illustrated using different loss functions, performance change of the model on CASIA-B data sets.It can be seen that, Lost with reference to multitask confrontation and loss can be with the recognition performance of lift scheme pixel-by-pixel;And different loss functions are used alone When, the performance of model can be reduced.
Experimental example 3:Influence of the different walking states to model performance
Fig. 3 shows across the visual angle recognition accuracy on CASIA-B under different walking states, and one has three kinds of different rows Walk state:Normal walking, carries packet rows and walks, and wears overcoat walking.From the figure, it can be seen that accuracy rate during normal walking is most Height, the gait sequence for wearing overcoat is more notable compared to reduction of the gait sequence of bag to model performance is carried.
Experimental example 4:Influence of different masterplate and the PEI port numbers to recognition accuracy
Table 3 illustrates the influence to recognition accuracy using different masterplates and PEI port numbers.We use gait energy Scheme (GEI) and time step morphotype version (CGI), and analyze average recognition accuracy when PEI port numbers distinguish 3,5,10.Can be with See, the PEI masterplates that we invent compare 6EI and C6I, there is higher recognition accuracy, and with the increasing of PEI masterplate numbers Plus, recognition accuracy can be lifted further.
Table 1:Recognition accuracy (%) under distinct methods
Table 2:Model Identification accuracy rate (%) under different loss functions
54° 90° 126°
Loss+multitask confrontation loss pixel-by-pixel 82.4 73.1 81.3
Lose pixel-by-pixel 81.5 71.7 83.5
Multitask confrontation loss 74.6 68.6 75.4
The influence of table 3 different masterplates and PEI port numbers to recognition accuracy

Claims (5)

1. across the visual angle gait recognition method of confrontation network is generated based on multitask, it is characterised in that concretely comprise the following steps:
(1) pedestrian's sequence of frames of video under input different visual angles, constructs gait exemplary feature:
<mrow> <mi>X</mi> <mo>=</mo> <mo>{</mo> <msub> <mi>x</mi> <mn>1</mn> </msub> <mo>,</mo> <msub> <mi>x</mi> <mn>2</mn> </msub> <mo>,</mo> <msub> <mi>x</mi> <mn>3</mn> </msub> <mo>,</mo> <mo>...</mo> <mo>,</mo> <msub> <mi>x</mi> <msub> <mi>n</mi> <mi>v</mi> </msub> </msub> <mo>}</mo> </mrow>
Vector xiIt is the gait exemplary feature under the i of visual angle, nvIt is the quantity at all visual angles;
(2) for any different visual angle u, v, corresponding gait exemplary feature x is encoded using convolutional neural networksuTo latent space, And obtain in latent space hidden be expressed as zu
(3) to hidden expression z in latent spaceuView transformation is carried out, angle v is converted into, obtains hidden expression zv
(4) by hidden expression zvAnd the one-hot coding of passage generates the generation network in confrontation network as input by multitask Pseudo- exemplary feature under output angle v afterwards
(5) according to exemplary feature xuWith pseudo- exemplary featureL is lost in calculating pixel-by-pixelp
(6) the differentiation network in confrontation network is generated using multitask, according to exemplary feature xuWith pseudo- exemplary featureCalculate many Task confrontation loss La
(7) loss resists loss with multitask pixel-by-pixel for weighting, according to total losses L=Lp+αLaTrain multitask generation confrontation net Network.
2. gait recognition method according to claim 1, it is characterised in that the construction step of described gait masterplate is:
(1) preceding background separation is carried out to each frame of original video frame sequence, extracts gait profile diagram;And by gait profile diagram Translation zooms to picture centre, obtains the gait profile graphic sequence { B of centralization1,B2,B3,…,Bn};
(2) to each frame, the period position r of normalized in the gait profile graphic sequence of centralizationt, represent the step of centralization The period position of the standardization of the t frames of state profile diagram;
(3) according to the gait profile diagram sequence B of centralizationtWith the period position r of corresponding standardizationt, construction is with ncIndividual passage Periodical energy figure (PEI) exemplary feature;Value in periodical energy figure at (x, y) coordinate of k-th of passage is calculated as follows:
<mrow> <msub> <mi>PEI</mi> <mi>k</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mrow> <msub> <mi>r</mi> <mi>t</mi> </msub> <mo>&amp;Element;</mo> <mi>T</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </mrow> </munder> <msub> <mi>B</mi> <mi>t</mi> </msub> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>y</mi> <mo>)</mo> </mrow> </mrow>
Wherein:
<mrow> <mi>T</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> <mo>=</mo> <mo>&amp;lsqb;</mo> <mfrac> <mrow> <mi>k</mi> <mo>-</mo> <mn>1</mn> </mrow> <mrow> <msub> <mi>n</mi> <mi>c</mi> </msub> <mo>+</mo> <mn>1</mn> </mrow> </mfrac> <mfrac> <mrow> <mi>k</mi> <mo>-</mo> <mn>1</mn> </mrow> <mrow> <msub> <mi>n</mi> <mi>c</mi> </msub> <mo>+</mo> <mn>1</mn> </mrow> </mfrac> <mo>+</mo> <mi>m</mi> <mo>&amp;rsqb;</mo> </mrow>
M represents the time-domain window size that each passage is covered, and T (k) determines the scope of the time-domain of each passage covering.
3. gait recognition method according to claim 1, it is characterised in that described view transformation step is:
Hidden expression z under given visual angle uu, transform to the hidden expression z under the v of visual anglevCalculating process be:
<mrow> <msub> <mi>z</mi> <mi>v</mi> </msub> <mo>=</mo> <msub> <mi>z</mi> <mi>u</mi> </msub> <mo>+</mo> <munderover> <mo>&amp;Sigma;</mo> <mrow> <mi>i</mi> <mo>=</mo> <mi>u</mi> </mrow> <mrow> <mi>v</mi> <mo>-</mo> <mn>1</mn> </mrow> </munderover> <msub> <mi>h</mi> <mi>i</mi> </msub> </mrow>
Wherein, hiRepresent the conversion vector by visual angle i-1 to visual angle i.
4. gait recognition method according to claim 1, it is characterised in that described loss LpWith LaCalculation procedure be:
(1) according to exemplary feature xvWith pseudo- exemplary featureCalculating is lost pixel-by-pixel:
<mrow> <msub> <mi>L</mi> <mi>p</mi> </msub> <mo>=</mo> <mi>E</mi> <mo>|</mo> <mo>|</mo> <msub> <mi>x</mi> <mi>u</mi> </msub> <mo>-</mo> <msub> <mover> <mi>x</mi> <mo>^</mo> </mover> <mi>v</mi> </msub> <mo>|</mo> <msub> <mo>|</mo> <mn>1</mn> </msub> </mrow>
(2) according to exemplary feature xvWith pseudo- exemplary featureCalculate multitask confrontation loss:
<mrow> <msub> <mi>L</mi> <mi>a</mi> </msub> <mo>=</mo> <mi>E</mi> <mo>&amp;lsqb;</mo> <msup> <mi>s</mi> <mi>T</mi> </msup> <mi>log</mi> <mi> </mi> <mi>D</mi> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>v</mi> </msub> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> <mo>+</mo> <mi>E</mi> <mo>&amp;lsqb;</mo> <msup> <mi>s</mi> <mi>T</mi> </msup> <mi>l</mi> <mi>o</mi> <mi>g</mi> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <mi>D</mi> <mo>(</mo> <msub> <mover> <mi>x</mi> <mo>^</mo> </mover> <mi>v</mi> </msub> <mo>)</mo> <mo>)</mo> </mrow> <mo>&amp;rsqb;</mo> </mrow>
Wherein, E represents the expectation for correspondence sample set, | | | |1L1 norms are represented, vectorial s is identity, passage and angle The one-hot coding of information.D () represents the output of arbiter.
5. gait recognition method according to claim 1, it is characterised in that the multitask generation confrontation network it is main by Encoder, view transformation layer, maker and the part of arbiter four composition;Wherein:
(a) encoder:Encoder in using convolutional neural networks as model, the input x of encoderuIt is the PEI that visual angle is u Masterplate;The scale size of input is 64 × 64 × nc;Each passage of PEI masterplates is made afterwards by independent feeding encoder Extract that effective feature is hidden to represent z with time pond and after full articulamentumuIt is used as the output of encoder;
(b) view transformation layer:It is assumed that gait data is as the change profile at visual angle is in the manifold of a higher-dimension, the sample in manifold This is towards specific direction movement to keep realizing under conditions of identity information the change at visual angle;The hidden expression of given latent space zu, view transformation can be described as:
<mrow> <msub> <mi>z</mi> <mi>v</mi> </msub> <mo>=</mo> <msub> <mi>z</mi> <mi>u</mi> </msub> <mo>+</mo> <msubsup> <mi>&amp;Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mi>u</mi> </mrow> <mrow> <mi>v</mi> <mo>-</mo> <mn>1</mn> </mrow> </msubsup> <msub> <mi>h</mi> <mi>i</mi> </msub> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>3</mn> <mo>)</mo> </mrow> </mrow>
Wherein, hiRepresent the conversion vector by visual angle i-1 to visual angle i, zvIt is the hidden expression under the v of visual angle;
(c) maker:By hidden expression zvAs input, the gait masterplate under the v of visual angle is generated;Maker is by five warps Lamination is constituted, and uses ReLU as excitation function;One passage of random selection PEI masterplates is reconstructed, and is ensured with this Hidden expression can preserve the information of all passages;The input for defining maker is [zv, c], c represents for the one-hot coding of passage;
(d) arbiter:Except not using, time pondization is operated the framework of arbiter and the dimension of output layer is different with encoder Outside, remaining structure is identical with encoder;The output dimension for defining arbiter is nv+nc+nd, wherein nvIt is the quantity at visual angle, ncIt is The quantity of PEI passages, ndIt is the quantity of different identity in training set;The output correspondence n of arbiterv+nc+ndIndividual sub- arbiter, son The parameter in addition to last layer is shared between arbiter;Also, each arbiter is responsible for different differentiation tasks to judge The sample whether generated belongs to specific distribution.
CN201710373017.5A 2017-05-24 2017-05-24 Cross-view gait recognition method based on multi-task generation countermeasure network Active CN107085716B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710373017.5A CN107085716B (en) 2017-05-24 2017-05-24 Cross-view gait recognition method based on multi-task generation countermeasure network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710373017.5A CN107085716B (en) 2017-05-24 2017-05-24 Cross-view gait recognition method based on multi-task generation countermeasure network

Publications (2)

Publication Number Publication Date
CN107085716A true CN107085716A (en) 2017-08-22
CN107085716B CN107085716B (en) 2021-06-04

Family

ID=59607349

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710373017.5A Active CN107085716B (en) 2017-05-24 2017-05-24 Cross-view gait recognition method based on multi-task generation countermeasure network

Country Status (1)

Country Link
CN (1) CN107085716B (en)

Cited By (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107563428A (en) * 2017-08-25 2018-01-09 西安电子科技大学 Classification of Polarimetric SAR Image method based on generation confrontation network
CN107609587A (en) * 2017-09-11 2018-01-19 浙江工业大学 A kind of multi-class multi views data creation method that confrontation network is generated based on depth convolution
CN107993210A (en) * 2017-11-30 2018-05-04 北京小米移动软件有限公司 Image repair method, device and computer-readable recording medium
CN108009568A (en) * 2017-11-14 2018-05-08 华南理工大学 A kind of pedestrian detection method based on WGAN models
CN108171701A (en) * 2018-01-15 2018-06-15 复旦大学 Conspicuousness detection method based on U networks and confrontation study
CN108256627A (en) * 2017-12-29 2018-07-06 中国科学院自动化研究所 The mutual generating apparatus of audio-visual information and its training system that generation network is fought based on cycle
CN108491380A (en) * 2018-03-12 2018-09-04 苏州思必驰信息科技有限公司 Confrontation multitask training method for speech understanding
CN108596026A (en) * 2018-03-16 2018-09-28 中国科学院自动化研究所 Across the visual angle Gait Recognition device and training method of confrontation network are generated based on double fluid
CN108629823A (en) * 2018-04-10 2018-10-09 北京京东尚科信息技术有限公司 The generation method and device of multi-view image
CN108846355A (en) * 2018-06-11 2018-11-20 腾讯科技(深圳)有限公司 Image processing method, face identification method, device and computer equipment
CN109543546A (en) * 2018-10-26 2019-03-29 复旦大学 The gait age estimation method returned based on the distribution of depth sequence
CN109583298A (en) * 2018-10-26 2019-04-05 复旦大学 Across visual angle gait recognition method based on set
CN109697389A (en) * 2017-10-23 2019-04-30 北京京东尚科信息技术有限公司 Personal identification method and device
CN109726654A (en) * 2018-12-19 2019-05-07 河海大学 A kind of gait recognition method based on generation confrontation network
CN110119780A (en) * 2019-05-10 2019-08-13 西北工业大学 Based on the hyperspectral image super-resolution reconstruction method for generating confrontation network
WO2019196626A1 (en) * 2018-04-12 2019-10-17 腾讯科技(深圳)有限公司 Media processing method and related apparatus
CN110381268A (en) * 2019-06-25 2019-10-25 深圳前海达闼云端智能科技有限公司 method, device, storage medium and electronic equipment for generating video
CN110516525A (en) * 2019-07-01 2019-11-29 杭州电子科技大学 SAR image target recognition method based on GAN and SVM
CN110659586A (en) * 2019-08-31 2020-01-07 电子科技大学 Cross-view gait recognition method based on identity maintenance cyclic generation type countermeasure network
CN111144170A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Gait information registration method, system and storage medium
CN111144165A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Gait information identification method, system and storage medium
CN111144167A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Gait information identification optimization method, system and storage medium
CN111149104A (en) * 2017-09-21 2020-05-12 诺基亚技术有限公司 Apparatus, method and computer program product for biometric identification
CN111144171A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Abnormal crowd information identification method, system and storage medium
CN111639580A (en) * 2020-05-25 2020-09-08 浙江工商大学 Gait recognition method combining feature separation model and visual angle conversion model
CN112001254A (en) * 2020-07-23 2020-11-27 浙江大华技术股份有限公司 Pedestrian identification method and related device
CN112488984A (en) * 2019-09-11 2021-03-12 中信戴卡股份有限公司 Method and device for acquiring defect picture generation network and defect picture generation method
CN112733704A (en) * 2021-01-07 2021-04-30 浙江大学 Image processing method, electronic device, and computer-readable storage medium
CN112975968A (en) * 2021-02-26 2021-06-18 同济大学 Mechanical arm simulation learning method based on third visual angle variable main body demonstration video
CN113112572A (en) * 2021-04-13 2021-07-13 复旦大学 Hidden space search-based image editing method guided by hand-drawn sketch
CN113111797A (en) * 2021-04-19 2021-07-13 杭州电子科技大学 Cross-view gait recognition method combining self-encoder and view transformation model
CN113657463A (en) * 2021-07-28 2021-11-16 浙江大华技术股份有限公司 Gait recognition model training method, gait recognition method and related device
JP2022514935A (en) * 2019-01-04 2022-02-16 ソニー コーポレイション オブ アメリカ Multiple prediction network
CN114120076A (en) * 2022-01-24 2022-03-01 武汉大学 Cross-view video gait recognition method based on gait motion estimation

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426645A (en) * 2011-08-30 2012-04-25 北京航空航天大学 Multi-view and multi-state gait recognition method
CN104299012A (en) * 2014-10-28 2015-01-21 中国科学院自动化研究所 Gait recognition method based on deep learning
CN105574510A (en) * 2015-12-18 2016-05-11 北京邮电大学 Gait identification method and device
CN106096532A (en) * 2016-06-03 2016-11-09 山东大学 A kind of based on tensor simultaneous discriminant analysis across visual angle gait recognition method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426645A (en) * 2011-08-30 2012-04-25 北京航空航天大学 Multi-view and multi-state gait recognition method
CN104299012A (en) * 2014-10-28 2015-01-21 中国科学院自动化研究所 Gait recognition method based on deep learning
CN105574510A (en) * 2015-12-18 2016-05-11 北京邮电大学 Gait identification method and device
CN106096532A (en) * 2016-06-03 2016-11-09 山东大学 A kind of based on tensor simultaneous discriminant analysis across visual angle gait recognition method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ZIFENG WU ET AL: "A Comprehensive Study on Cross-View Gait Based Human Identification with Deep CNNs", 《IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE》 *
汪青: "基于深度学习的多视角步态识别算法设计与研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107563428A (en) * 2017-08-25 2018-01-09 西安电子科技大学 Classification of Polarimetric SAR Image method based on generation confrontation network
CN107609587A (en) * 2017-09-11 2018-01-19 浙江工业大学 A kind of multi-class multi views data creation method that confrontation network is generated based on depth convolution
CN107609587B (en) * 2017-09-11 2020-08-18 浙江工业大学 Multi-class multi-view data generation method for generating countermeasure network based on deep convolution
CN111149104A (en) * 2017-09-21 2020-05-12 诺基亚技术有限公司 Apparatus, method and computer program product for biometric identification
CN111149104B (en) * 2017-09-21 2023-10-27 诺基亚技术有限公司 Apparatus, method and computer readable storage medium for biometric identification
CN109697389A (en) * 2017-10-23 2019-04-30 北京京东尚科信息技术有限公司 Personal identification method and device
CN109697389B (en) * 2017-10-23 2021-10-01 北京京东尚科信息技术有限公司 Identity recognition method and device
CN108009568A (en) * 2017-11-14 2018-05-08 华南理工大学 A kind of pedestrian detection method based on WGAN models
CN107993210A (en) * 2017-11-30 2018-05-04 北京小米移动软件有限公司 Image repair method, device and computer-readable recording medium
CN108256627A (en) * 2017-12-29 2018-07-06 中国科学院自动化研究所 The mutual generating apparatus of audio-visual information and its training system that generation network is fought based on cycle
CN108171701A (en) * 2018-01-15 2018-06-15 复旦大学 Conspicuousness detection method based on U networks and confrontation study
CN108171701B (en) * 2018-01-15 2021-06-22 复旦大学 Significance detection method based on U network and counterstudy
CN108491380A (en) * 2018-03-12 2018-09-04 苏州思必驰信息科技有限公司 Confrontation multitask training method for speech understanding
CN108596026B (en) * 2018-03-16 2020-06-30 中国科学院自动化研究所 Cross-view gait recognition device and training method based on double-flow generation countermeasure network
CN108596026A (en) * 2018-03-16 2018-09-28 中国科学院自动化研究所 Across the visual angle Gait Recognition device and training method of confrontation network are generated based on double fluid
CN108629823A (en) * 2018-04-10 2018-10-09 北京京东尚科信息技术有限公司 The generation method and device of multi-view image
CN108629823B (en) * 2018-04-10 2022-09-06 北京京东尚科信息技术有限公司 Method and device for generating multi-view image
WO2019196626A1 (en) * 2018-04-12 2019-10-17 腾讯科技(深圳)有限公司 Media processing method and related apparatus
US11335127B2 (en) 2018-04-12 2022-05-17 Tencent Technology (Shenzhen) Company Ltd Media processing method, related apparatus, and storage medium
US11403876B2 (en) * 2018-06-11 2022-08-02 Tencent Technology (Shenzhen) Company Limited Image processing method and apparatus, facial recognition method and apparatus, and computer device
WO2019237846A1 (en) * 2018-06-11 2019-12-19 腾讯科技(深圳)有限公司 Image processing method and apparatus, face recognition method and apparatus, and computer device
CN108846355A (en) * 2018-06-11 2018-11-20 腾讯科技(深圳)有限公司 Image processing method, face identification method, device and computer equipment
CN109543546B (en) * 2018-10-26 2022-12-20 复旦大学 Gait age estimation method based on depth sequence distribution regression
CN109583298A (en) * 2018-10-26 2019-04-05 复旦大学 Across visual angle gait recognition method based on set
CN109583298B (en) * 2018-10-26 2023-05-02 复旦大学 Cross-view gait recognition method based on set
CN109543546A (en) * 2018-10-26 2019-03-29 复旦大学 The gait age estimation method returned based on the distribution of depth sequence
CN111144165B (en) * 2018-11-02 2024-04-12 银河水滴科技(宁波)有限公司 Gait information identification method, system and storage medium
CN111144171A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Abnormal crowd information identification method, system and storage medium
CN111144167A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Gait information identification optimization method, system and storage medium
CN111144165A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Gait information identification method, system and storage medium
CN111144170A (en) * 2018-11-02 2020-05-12 银河水滴科技(北京)有限公司 Gait information registration method, system and storage medium
CN109726654A (en) * 2018-12-19 2019-05-07 河海大学 A kind of gait recognition method based on generation confrontation network
JP2022514935A (en) * 2019-01-04 2022-02-16 ソニー コーポレイション オブ アメリカ Multiple prediction network
JP7379494B2 (en) 2019-01-04 2023-11-14 ソニー コーポレイション オブ アメリカ multiple prediction network
CN110119780A (en) * 2019-05-10 2019-08-13 西北工业大学 Based on the hyperspectral image super-resolution reconstruction method for generating confrontation network
CN110381268B (en) * 2019-06-25 2021-10-01 达闼机器人有限公司 Method, device, storage medium and electronic equipment for generating video
CN110381268A (en) * 2019-06-25 2019-10-25 深圳前海达闼云端智能科技有限公司 method, device, storage medium and electronic equipment for generating video
CN110516525B (en) * 2019-07-01 2021-10-08 杭州电子科技大学 SAR image target recognition method based on GAN and SVM
CN110516525A (en) * 2019-07-01 2019-11-29 杭州电子科技大学 SAR image target recognition method based on GAN and SVM
CN110659586A (en) * 2019-08-31 2020-01-07 电子科技大学 Cross-view gait recognition method based on identity maintenance cyclic generation type countermeasure network
CN110659586B (en) * 2019-08-31 2022-03-15 电子科技大学 Gait recognition method based on identity-preserving cyclic generation type confrontation network
CN112488984A (en) * 2019-09-11 2021-03-12 中信戴卡股份有限公司 Method and device for acquiring defect picture generation network and defect picture generation method
CN111639580A (en) * 2020-05-25 2020-09-08 浙江工商大学 Gait recognition method combining feature separation model and visual angle conversion model
CN112001254A (en) * 2020-07-23 2020-11-27 浙江大华技术股份有限公司 Pedestrian identification method and related device
CN112733704A (en) * 2021-01-07 2021-04-30 浙江大学 Image processing method, electronic device, and computer-readable storage medium
CN112975968B (en) * 2021-02-26 2022-06-28 同济大学 Mechanical arm imitation learning method based on third visual angle variable main body demonstration video
CN112975968A (en) * 2021-02-26 2021-06-18 同济大学 Mechanical arm simulation learning method based on third visual angle variable main body demonstration video
CN113112572A (en) * 2021-04-13 2021-07-13 复旦大学 Hidden space search-based image editing method guided by hand-drawn sketch
CN113111797A (en) * 2021-04-19 2021-07-13 杭州电子科技大学 Cross-view gait recognition method combining self-encoder and view transformation model
CN113111797B (en) * 2021-04-19 2024-02-13 杭州电子科技大学 Cross-view gait recognition method combining self-encoder and view transformation model
CN113657463A (en) * 2021-07-28 2021-11-16 浙江大华技术股份有限公司 Gait recognition model training method, gait recognition method and related device
CN114120076B (en) * 2022-01-24 2022-04-29 武汉大学 Cross-view video gait recognition method based on gait motion estimation
CN114120076A (en) * 2022-01-24 2022-03-01 武汉大学 Cross-view video gait recognition method based on gait motion estimation

Also Published As

Publication number Publication date
CN107085716B (en) 2021-06-04

Similar Documents

Publication Publication Date Title
CN107085716A (en) Across the visual angle gait recognition method of confrontation network is generated based on multitask
CN109344736B (en) Static image crowd counting method based on joint learning
CN108537743B (en) Face image enhancement method based on generation countermeasure network
Liu et al. Attribute-aware face aging with wavelet-based generative adversarial networks
CN105469034B (en) Face identification method based on Weighting type distinctive sparse constraint Non-negative Matrix Factorization
CN104268593B (en) The face identification method of many rarefaction representations under a kind of Small Sample Size
CN104616316B (en) Personage&#39;s Activity recognition method based on threshold matrix and Fusion Features vision word
CN105825183B (en) Facial expression recognizing method based on partial occlusion image
CN102938070B (en) A kind of behavior recognition methods based on action subspace and weight behavior model of cognition
Zhu et al. Deep learning multi-view representation for face recognition
CN111090764B (en) Image classification method and device based on multitask learning and graph convolution neural network
CN107145836A (en) Hyperspectral image classification method based on stack boundary discrimination self-encoding encoder
CN104915658B (en) A kind of emotion component analyzing method and its system based on emotion Distributed learning
CN106897669A (en) A kind of pedestrian based on consistent iteration various visual angles transfer learning discrimination method again
CN109389045A (en) Micro- expression recognition method and device based on mixing space-time convolution model
CN104298974A (en) Human body behavior recognition method based on depth video sequence
CN105868711B (en) Sparse low-rank-based human behavior identification method
CN104408731B (en) Region graph and statistic similarity coding-based SAR (synthetic aperture radar) image segmentation method
CN102592150B (en) Gait identification method of bidirectional two-dimensional principal component analysis based on fuzzy decision theory
CN107590427A (en) Monitor video accident detection method based on space-time interest points noise reduction
CN108460400A (en) A kind of hyperspectral image classification method of combination various features information
CN102184384A (en) Face identification method based on multiscale local phase quantization characteristics
CN102867171B (en) Label propagation and neighborhood preserving embedding-based facial expression recognition method
CN108985161A (en) A kind of low-rank sparse characterization image feature learning method based on Laplace regularization
Lim et al. Pose transforming network: Learning to disentangle human posture in variational auto-encoded latent space

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant