CN113505719B - Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm - Google Patents

Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm Download PDF

Info

Publication number
CN113505719B
CN113505719B CN202110824459.3A CN202110824459A CN113505719B CN 113505719 B CN113505719 B CN 113505719B CN 202110824459 A CN202110824459 A CN 202110824459A CN 113505719 B CN113505719 B CN 113505719B
Authority
CN
China
Prior art keywords
convolution
layer
network
pooling
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110824459.3A
Other languages
Chinese (zh)
Other versions
CN113505719A (en
Inventor
单彩峰
宋旭
陈宇
黄岩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University of Science and Technology
Original Assignee
Shandong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University of Science and Technology filed Critical Shandong University of Science and Technology
Priority to CN202110824459.3A priority Critical patent/CN113505719B/en
Publication of CN113505719A publication Critical patent/CN113505719A/en
Application granted granted Critical
Publication of CN113505719B publication Critical patent/CN113505719B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a gait recognition model compression system and method based on a local-integral combined knowledge distillation algorithm. The system designs a compact lightweight gait model network by adopting deep separation convolution, wherein the convolution network in the model network only keeps a backbone convolution network and simplifies all convolution modules in the backbone convolution network; the model network adopts a 16-layer lightweight full-connection network, so that model parameters are greatly compressed, calculation of an identification model is simplified, and identification efficiency is improved; the method of the invention utilizes the local feature vector output by the convolution network and the global feature vector output by the full-connection network in the teacher model and the student model to carry out combined knowledge distillation, thereby not only retaining the local feature of the gait of the pedestrian by convolution operation, but also extracting the global feature of the gait of the pedestrian by full-connection operation, increasing the information quantity of knowledge distillation and improving the recognition effect of the gait of the pedestrian.

Description

Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm
Technical Field
The invention belongs to the technical field of image recognition, and particularly relates to a gait recognition model compression system and method based on a local-integral combined knowledge distillation algorithm.
Background
Gait recognition is an emerging biological feature recognition technology, and aims to find and extract feature changes among different pedestrians from a series of walking gestures so as to realize automatic recognition of the pedestrians. Gait recognition has many advantages over other biometric technologies, such as: the method has the advantages of no need of active matching of the identified objects, low requirement on image pixels, no limitation of viewing angles, easy identification of camouflage, long identification distance and the like. Based on the above, gait recognition technology is widely applied in the fields of video monitoring, intelligent security and the like.
At present, gait recognition technology is mostly designed based on a standard convolutional neural network, and a recognition model is trained by collecting gait video samples containing pedestrian labels, so that the model learns useful gait appearance and action characteristics from the samples, and can be recognized according to the characteristics. Existing gait recognition techniques can be divided into model-based methods and appearance-based methods depending on the presence or absence of a human model build.
The method based on the model needs to extract gait characteristics by building a human skeleton or a gesture structure, and has large calculation amount and redundant network structure. The appearance-based method is the most commonly used method at present, can directly extract gait features from the original video captured by a camera, and can be subdivided into feature template-based, gait video-based and set-based methods.
The method based on the characteristic template (such as a gait energy diagram) is simple to realize, but the time sequence characteristics are easy to lose by extracting gait characteristics in the characteristic template for identification; the gait video sequence-based method effectively improves gait space-time characteristics through a three-dimensional standard convolutional neural network for recognition, but the model is large in scale and difficult to train; based on the method of aggregation, the single-frame gait outline features and the aggregate-pooling structure aggregate gait space-time features in the gait aggregate are extracted through the two-dimensional standard convolutional neural network, so that the efficient identification performance is realized, and the model scale is still large.
In conclusion, the current gait recognition method is mainly realized by adopting a large-capacity neural network model, has the advantages of multiple model parameters, long training time and difficult application and popularization, and is not suitable for practical application occasions with high real-time requirements.
While the conventional model compression method can reduce the capacity of the model and reduce the parameters of the model to a certain extent, the conventional model compression method is simpler, key information in the model cannot be reserved, and the recognition performance of the model is seriously reduced, so that the conventional model compression method is not suitable for solving the compression problem of the gait recognition model.
Disclosure of Invention
The invention aims to provide a gait recognition model compression method based on a local-whole combined knowledge distillation algorithm, which can effectively ensure the gait recognition accuracy of a model while reducing the scale of model parameters.
In order to achieve the above purpose, the invention adopts the following technical scheme:
gait recognition model compression system based on local-global joint knowledge distillation algorithm includes:
comprises a teacher model Mt and a student model M s The method comprises the steps of carrying out a first treatment on the surface of the Wherein:
the teacher model Mt consists of a convolution network, a collection pooling structure, a horizontal pyramid pooling structure and a full-connection network;
the convolution network consists of a backbone network and a plurality of layers of global channels;
the backbone network consists of a first convolution module, a second convolution module and a third convolution module;
the first convolution module is composed of three layers, wherein:
The first layer is a standard convolution layer, and a convolution kernel of 5×5 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the third layer is a pooling layer, and adopts a maximum pooling layer with pooling core of 2 multiplied by 2 and step length of 2;
the second convolution module consists of three layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the third layer is a pooling layer, and the largest pooling layer with pooling core of 2 multiplied by 2 and step length of 2 is used;
the third convolution module consists of two layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used;
the multi-layer global channel consists of a fourth convolution module and a fifth convolution module;
the fourth convolution module consists of three layers, wherein: the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the third layer is a pooling layer, and a pooling core of 2 multiplied by 2 is adopted;
the fifth convolution module consists of two layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used;
four aggregation pooling structures are arranged in the teacher model Mt, two horizontal pyramid pooling structures are arranged, and one full-connection network is arranged;
Defining four collection pooling structures as a first collection pooling structure, a second collection pooling structure and a third collection pooling structure and a fourth collection pooling structure respectively; defining two horizontal pyramid pooling structures as a first horizontal pyramid pooling structure and a second horizontal pyramid pooling structure respectively;
the fully-connected network comprises a first fully-connected sub-network and a second fully-connected sub-network;
the output of the first convolution module is connected with the input of the second convolution module;
the output of the first convolution module is also connected with the input of the fourth convolution module through a first collection pooling structure;
the output of the second convolution module is connected with the input of the third convolution module;
the output of the second convolution module is connected with the input of the second aggregation pooling structure, and the output of the second aggregation pooling structure is connected with the input of the fifth convolution module after corresponding position addition is carried out on the output of the fourth convolution module;
the output of the third convolution module is connected with the input of the third aggregation pooling structure, and the output of the third aggregation pooling structure is connected with the input of the first horizontal pyramid pooling structure after corresponding position addition is carried out on the output of the fifth convolution module;
the output of the third convolution module is also connected with the input of the second horizontal pyramid pooling structure through the fourth aggregate pooling structure;
The output of the first horizontal pyramid pooling structure is connected with the input of the first fully-connected subnetwork;
the output of the second horizontal pyramid pooling structure is connected with the input of the second fully-connected subnetwork;
the outputs of the first fully-connected subnetwork and the second fully-connected subnetwork are used as a teacher model M t An output of (2);
the first and second horizontal pyramid pooling structures have five scales;
the first full-connection sub-network and the second full-connection sub-network respectively comprise 31 independent full-connection neural network layers;
student model M s The method comprises the steps of simplifying a convolution network, assembling a pooling structure, simplifying a horizontal pyramid pooling structure and simplifying a full-connection network, wherein:
the simplified convolution network consists of a sixth convolution module, a seventh convolution module and an eighth convolution module;
the sixth convolution module consists of two layers, wherein: the first layer is a standard convolution layer, and a convolution kernel of 5×5 is used; the second layer is a pooling layer, and the largest pooling layer with pooling core of 2 multiplied by 2 and step length of 2 is used;
the seventh convolution module consists of three layers, wherein:
the first layer is a depth separation convolution layer, and a convolution kernel of 5×5 is used; the second layer is a point convolution layer, and a convolution kernel of 1 multiplied by 1 is used; the third layer is a pooling layer, and adopts a maximum pooling layer with pooling core of 2 multiplied by 2 and step length of 2;
The eighth convolution module consists of two layers, wherein:
the first layer uses a 3 x 3 convolution kernel for the depth-separated convolution layer; the second layer is a point convolution layer, and a convolution kernel of 1 multiplied by 1 is used;
the sixth convolution module, the seventh convolution module and the eighth convolution module are connected in sequence;
definition of student model M s The medium aggregate pooling structure is a fifth aggregate pooling structure;
the output of the eighth convolution module is connected with the input of the simplified horizontal pyramid pooling structure through the fifth aggregate pooling structure;
the output of the simplified horizontal pyramid pooling structure is connected with the input of the simplified fully-connected network;
the simplified horizontal pyramid pooling structure has only one scale; the simplified fully-connected network comprises 16 independent fully-connected neural network layers; output of simplified fully-connected network as student model M s Is provided.
In addition, the invention also provides a gait recognition model compression method based on the local-integral combined knowledge distillation algorithm, which is based on the gait recognition model compression system, and the specific technical scheme is as follows:
a gait recognition model compression method based on a local-whole combined knowledge distillation algorithm comprises the following steps:
step 1, extracting gait contour sequences in a gait video by using a background subtraction method, uniformly cutting the gait contour sequences into picture sets with the same size to form a data set X, and dividing the data set X into a training set X train And test set X test
Step 2, training set X train Training teacher model M t Setting a learning rate and iteration times, wherein an optimizer adopts an Adam optimizer, and a loss function adopts a triplet loss function L shown in a formula (1) tri
Wherein N is tri+ Representing the total number of sample pairs consisting of two samples of a training sample subset for which all Euclidean distances are not 0; the training sample subset is the secondary training set X when each training train A set formed by a plurality of sample images selected randomly;
n represents the number of fully connected neural network layers of the teacher network, and t represents the serial number of the fully connected neural network layers of the teacher network;
p represents the number of pedestrians contained in each training sample subset;
i and p respectively represent the serial numbers of the pedestrian samples to be trained in each training sample subset;
k represents the number of video sequences for each pedestrian in each training sample subset;
a. j and k respectively represent the sequence numbers of the pedestrian video sequences in each training sample subset;
m represents the boundary threshold of the loss function;
representing a j-th sample to be trained in the i-th training sample subset;
representing the sum +.>Any sample with the same identity of the represented pedestrian;
representing the sum +. >Any sample with different identities of the represented pedestrians;
sign | I 2 Representing the 2-norm of the matrix;
[] + representing ReLU operation, the calculation method is as follows: [ x ]] + =max {0, x }, max being the maximum value operation;
step 3, training set X train Input to a trained teacher model M t And untrained student model M s Respectively obtain the same data set in the teacher model M t Multi-dimensional feature matrix of convolution network output of (a)In student model M s Multi-dimensional characteristic matrix of the output of the reduced convolution network ∈>In teacher model M t Is a multi-dimensional feature matrix of the full-connection network output>In student model M s Multi-dimensional characteristic matrix of the output of the simplified fully-connected network>
Multi-dimensional feature matrixAnd multidimensional feature matrix->Is b x s x c x h x w; multidimensional feature matrix->And multidimensional feature matrix->Is b x n x d; wherein b represents the number of samples in each training sample subset; s represents the number of frames; c represents the number of the output feature matrixes of the convolution layer, h represents the height of the output feature matrixes of the convolution network, w represents the width of the output feature matrixes of the convolution network, and d represents the dimension of the feature matrixes output by the full-connection network;
step 4, making the difference measurement function L c_dis Calculating a multi-dimensional feature matrixAnd multidimensional feature matrix->Differences between; wherein,difference metric function L c_dis The calculation formula of (2) is as follows:
in the difference measurement function L c_dis Representing the partial distillation loss, the sign is 2 F Representing the F-norm of the matrix;
step 5, respectively calculating the multi-dimensional feature matrix by using the formula (3)And multidimensional feature matrix->The difference between each sample of (a) is used as a result of +.>And->A representation;
wherein,representing the distance between all the samples of the same class in the feature matrix output by the teacher model; />Representing the distance between all different types of samples in the feature matrix output by the teacher model; />Representing the distance between all the samples of the same class in the feature matrix output by the student model; />Representing the distance between all different types of samples in the feature matrix output by the student model;
when the teacher model is trained, the jth sample to be trained with the same category in the ith training sample subset is represented;
when the teacher model is trained, the kth sample to be trained with the same category in the kth training sample subset is represented; />Representing the jth sample to be trained with different categories in the ith training sample subset when training the teacher model;
representing a kth sample to be trained with different categories in a p-th training sample subset when training a teacher model; / >Representing the jth sample to be trained with the same category in the ith training sample subset when training the student model;
representing the kth sample to be trained with the same category in the kth training sample subset when training the student model; />Representing the jth sample to be trained with different categories in the ith training sample subset when training the student model;
representing the kth sample to be trained with different categories in the p-th training sample subset when training the student model;
step 6, using the triplet loss function L shown in the formula (1) tri Respectively calculateAnd->Triple loss->And->The specific calculation formulas are shown in formulas (4) and (5);
wherein,representing the total number of sample pairs consisting of two samples with all Euclidean distances other than 0 in one training sample subset in the teacher model;
wherein,representing the total number of sample pairs consisting of two samples with all Euclidean distances other than 0 in one training sample subset in the student model;
calculating the triplet loss by using the Smooth L1 loss function in the formula (6)And->Total distillation loss L of (2) f_dis
Step 7, partial distillation loss L c_dis Total distillation loss L f_dis Loss of tripletAnd->Integrating and calculating to obtain total loss L total The specific calculation formula is as follows:
Wherein alpha is a distillation loss weight;
step 8, setting a student model M s The iteration times are counted, the optimizer is an Adam optimizer, and the loss value L is reduced total Transferring the knowledge of the teacher model to the student model;
step 9. Test set X test The pedestrian video sequence is input to the student network M s And (3) performing recognition to obtain a recognition result.
The invention has the following advantages:
as described above, the present invention describes a gait recognition model compression system based on a local-global joint knowledge distillation algorithm, which designs a compact lightweight gait model network (i.e., student model) based on deep separation convolution, wherein the convolution network in the model network only retains a backbone convolution network, and simplifies each convolution module in the backbone convolution network, specifically, adopts a 3×3 deep separation convolution layer and a 1×1 point convolution layer, instead of the standard convolution layer in the existing scheme; in addition, the model network adopts a 16-layer lightweight full-connection network to replace a 31-layer full-connection network in the existing scheme, so that model parameters are greatly compressed, calculation of an identification model is simplified, and identification efficiency is improved. In addition, the invention also provides a local-whole combined knowledge distillation algorithm suitable for gait recognition tasks on the basis of the gait recognition model compression system, and compared with the existing method, the local-whole combined knowledge distillation method designed by the invention simultaneously utilizes the local feature vector output by a convolution network and the global feature vector output by a full-connection network to carry out combined knowledge distillation, thereby not only retaining the local feature of the gait of the pedestrian by convolution operation, but also extracting the global feature of the gait of the pedestrian by full-connection operation, increasing the information quantity of knowledge distillation, improving the effect of the gait recognition of the pedestrian, namely ensuring the gait recognition accuracy of the model.
Drawings
FIG. 1 is a block diagram of a gait recognition model compression system based on a local-global joint knowledge distillation algorithm of the present invention;
FIG. 2 is a schematic diagram of a structure of pooling in an embodiment of the present invention;
FIG. 3 is a schematic diagram of a horizontal pyramid pooling structure in an embodiment of the present invention;
FIG. 4 is a schematic diagram of a simplified horizontal pyramid pooling structure in accordance with an embodiment of the present invention;
FIG. 5 is a schematic diagram of a multi-layer global channel according to an embodiment of the present invention;
FIG. 6 is a schematic diagram of a depth separation convolution module according to an embodiment of the present disclosure;
fig. 7 is a flow chart of a gait recognition model compression method based on a local-global joint knowledge distillation algorithm.
Detailed Description
The invention is described in further detail below with reference to the attached drawings and detailed description:
example 1
The embodiment describes a gait recognition model compression system of a local-global joint knowledge distillation algorithm.
As shown in fig. 1, the gait recognition model compression system constructs two recognition models based on a deep neural network, namely a high-capacity teacher model Mt and a light student model M s
Wherein the symbol "# indicates addition by element" in fig. 1.
The teacher model Mt consists of a convolution network, a collection pooling structure, a horizontal pyramid pooling structure and a full-connection network.
The convolutional network consists of a backbone network and multiple layers of global channels.
The backbone network is composed of a first convolution module, a second convolution module and a third convolution module.
The first convolution module is composed of three layers, wherein:
the first layer is a standard convolution layer, and uses a convolution kernel of 5×5, and has s (frame number) ×1×64×44 data as input and s (frame number) ×32×64×44 feature map as output;
the second layer is a standard convolution layer, and uses a convolution kernel of 3×3, and inputs data of s (frame number) ×32×64×44 and outputs a feature map of s (frame number) ×32×64×44;
the third layer is a pooling layer, and adopts a maximum pooling layer with a pooling core of 2×2 and a step length of 2, and the input of the third layer is s (frame number) ×32×64×44, and the output is s (frame number) ×32×32×22.
The second convolution module consists of three layers, wherein:
the first layer is a standard convolution layer, and uses a convolution kernel of 3×3, and has s (frame number) ×32×32×22 data as input and outputs a feature map of s (frame number) ×64×32×22;
the second layer is a standard convolution layer, and uses a convolution kernel of 3×3, and the input of the second layer is s (frame number) ×64×32×22 data, and a feature map of s (frame number) ×64×32×22 is output;
The third layer was a pooling layer, and the largest pooling layer with a pooling kernel of 2×2 and a step size of 2 was used, and the input of the third layer was (frame number) ×32×32×22, and the output was s (frame number) ×32×16×11.
The third convolution module consists of two layers, wherein:
the first layer is a standard convolution layer, and uses a convolution kernel of 3×3, and has s (frame number) ×64×32×22 data as input and s (frame number) ×128×16×11 feature map as output;
the second layer is a standard convolution layer, and uses a convolution kernel of 3×3, and has s (frame number) ×128×16×11 data as input, and outputs a feature map of s (frame number) ×128×16×11.
As shown in fig. 5, the multi-layer global channel is composed of a fourth convolution module and a fifth convolution module.
The fourth convolution module consists of three layers, wherein:
the first layer is a standard convolution layer, and uses a convolution kernel of 3×3, and the input of the first layer is 32×32×22 data, and a 64×32×32 feature map is output;
the second layer is a standard convolution layer, and uses a convolution kernel of 3×3, and the input of the second layer is 64×32×22 data, and a characteristic diagram of 64×32×22 is output;
the third layer is a pooling layer, and uses a 2×2 pooling core, inputs 64×32×22 data, and outputs a 64×16×11 feature map.
The fifth convolution module consists of two layers, wherein:
the first layer is a standard convolution layer, and uses a convolution kernel of 3×3, and the input of the first layer is 64×16×11 data, and the output of the first layer is a feature map of 128×16×11;
the second layer is a standard convolution layer, and uses a convolution kernel of 3×3, and has 128×16×11 data as an input and 128×16×11 feature maps as an output.
Four aggregation pooling structures are arranged in the teacher model Mt, two horizontal pyramid pooling structures are arranged, and one full-connection network is arranged.
The four aggregate pooling are defined as a first, second, third, and fourth aggregate pooling structure, respectively.
Taking one of the aggregate pooling structures as an example, as shown in fig. 2.
The input of a set pool is a feature matrix corresponding to s video frames, each dimension is 128 multiplied by 16 multiplied by 11, the output is a processed single feature matrix, the dimension is 128 multiplied by 16 multiplied by 11, and the value of each element in the feature matrix is the maximum value element extracted from the corresponding positions of a plurality of feature matrices through maximum value operation.
The pooling is characterized in that the sequence of the feature matrixes corresponding to the plurality of input video frames can be arranged randomly.
The two horizontal pyramid pooling structures are respectively defined as a first horizontal pyramid pooling structure and a second horizontal pyramid pooling structure, and one of the two horizontal pyramid pooling structures is taken as an example, as shown in fig. 3.
The input of one horizontal pyramid pool is a 128×16×11 feature matrix, the matrix is decomposed according to 5 scales to obtain an intermediate feature matrix, and the intermediate feature matrix is respectively 1 feature matrix with 128×16×11 dimensions, 2 feature matrices with 128×8×11 dimensions, 4 feature matrices with 128×4×11 dimensions, 8 feature matrices with 128×2×11 dimensions, and 16 feature matrices with 128×1×11 dimensions, and the total is 31 feature matrices.
The invention adopts global maximum pooling operation and global average pooling operation to compress the second and third dimensions of each intermediate feature matrix, and changes the second and third dimensions into 128-dimension vectors. The above procedure is illustrated:
when global maximum pooling operation is carried out on the 128 multiplied by 16 multiplied by 11 feature matrix, the matrix is decomposed into 128 multiplied by 11 submatrices, the maximum value of each 16 multiplied by 11 submatrices is calculated, 128 maximum value calculation results are obtained in total, and the 128 maximum value calculation results are combined into an output feature vector with 128 dimensions; similarly, when global average pooling is performed on a 128×16×11 feature matrix, the matrix is decomposed into 128 16×11 sub-matrices, the average value of each 16×11 sub-matrix is calculated, 128 average value calculation results are obtained in total, and the 128 average value calculation results are combined into an output feature vector of 128 dimensions.
The final output of one horizontal pyramid pooling structure is 31 128-dimensional vectors.
The fully-connected network comprises a first fully-connected sub-network and a second fully-connected sub-network; the first full-connection sub-network and the second full-connection sub-network respectively comprise 31 independent full-connection neural network layers.
The fully connected network contains a total of 62 fully connected layers, each layer inputting 128 data and outputting 256 features.
The outputs of the first fully-connected subnetwork and the second fully-connected subnetwork serve as outputs of the teacher model Mt.
The output of the first convolution module is connected to the input of the second convolution module. The output of the first convolution module is also connected to the input of the fourth convolution module through a first set pooling structure.
The output of the second convolution module is connected to the input of the third convolution module. The output of the second convolution module is connected with the input of the second aggregation pooling structure, and the output of the second aggregation pooling structure is connected with the input of the fifth convolution module after corresponding position addition is carried out on the output of the fourth convolution module.
The output of the third convolution module is connected with the input of the third aggregation pooling structure, and the output of the third aggregation pooling structure is connected with the input of the first horizontal pyramid pooling structure after corresponding position addition is carried out on the output of the fifth convolution module.
The output of the third convolution module is also connected to the input of the second horizontal pyramid pooling structure through a fourth aggregate pooling structure.
The output of the first horizontal pyramid pooling structure is connected with the input of the first fully-connected subnetwork; the output of the second horizontal pyramid pooling structure is connected to the input of the second fully connected subnetwork.
Student model M s The method consists of a simplified convolution network, a collection pooling structure, a simplified horizontal pyramid pooling structure and a simplified full-connection network.
The simplified convolutional network only retains one backbone convolutional network, and specifically, as shown in fig. 1, the simplified convolutional network is composed of a sixth convolutional module, a seventh convolutional module and an eighth convolutional module.
Compared with a first convolution module, a second convolution module and a third convolution module in a backbone network in a teacher model, the sixth convolution module, the seventh convolution module and the eighth convolution module are respectively simplified.
Wherein the sixth convolution module eliminates a standard convolution layer having a convolution kernel of 3 x 3 as compared to the first convolution module.
The seventh convolution module replaces the two standard convolution layers in the second convolution module with a 3 x 3 depth separation convolution layer and a 1 x 1 point convolution, respectively.
Similarly, the eighth convolution module replaces the two standard convolution layers in the third convolution module with a 3×3 depth separation convolution layer and a 1×1 point convolution, respectively.
Specifically, the sixth convolution module is composed of two layers, wherein:
the first layer is a standard convolution layer, and uses a convolution kernel of 5×5, and has s (frame number) ×1×64×44 data as input and s (frame number) ×32×64×44 feature map as output;
the second layer is a pooling layer, and the largest pooling layer with a pooling kernel of 2×2 and a step size of 2 is used, and the second layer has 32×64×44 data as input and 32×32×22 feature map as output.
The seventh convolution module consists of three layers, wherein:
the first layer is a depth-separated convolution layer, and uses a convolution kernel of 5×5, and has s (frame number) ×32×32×22 data as input and outputs a feature map of s (frame number) ×32×32×22;
the second layer is a point convolution layer, and uses a convolution kernel of 1×1, and has s (frame number) ×64×32×22 data as input and outputs a feature map of s (frame number) ×64×32×22;
the third layer is a pooling layer, adopts a maximum pooling layer with a pooling core of 2 multiplied by 2 and a step length of 2, and has 64 multiplied by 32 multiplied by 22 data input and 64 multiplied by 16 multiplied by 11 feature images output;
The eighth convolution module consists of two layers, wherein:
the first layer is a depth-separated convolution layer, and uses a convolution kernel of 3×3, and has s (frame number) ×64×16×11 data as input and outputs a feature map of s (frame number) ×64×16×11;
the second layer is a point convolution layer, and uses a convolution kernel of 1×1, and has s (frame number) ×64×16×11 data as input, and outputs a feature map of s (frame number) ×128×16×11.
The structure of the depth-separated convolution layer is shown in fig. 6, and this structure is a known structure and will not be described in detail.
The sixth convolution module, the seventh convolution module and the eighth convolution module are connected in sequence.
In a preferred embodiment, a dot convolution layer is added in front of the first layer of the depth separation convolution layer of the seventh convolution module, and similarly, a dot convolution layer is added in front of the first layer of the depth separation convolution layer of the eighth convolution module.
The design realizes the improvement of the network performance of the lightweight chemical green model on the premise of almost not changing the capacity of the model.
Definition of student model M s The medium aggregate pooling structure is a fifth aggregate pooling structure; the output of the eighth convolution module is connected to the input of the reduced horizontal pyramid pooling structure through a fifth aggregate pooling structure.
The fifth aggregate pooling structure is also composed of a statistical (maximum) function, which inputs a feature matrix of s (frame number) ×128×16×11, and outputs a feature matrix of 128×16×11.
The reduced horizontal pyramid pooling structure consists of global maximum pooling and global average pooling, and is structured as shown in fig. 4.
The input of the reduced level golden tower pooling is a 128 multiplied by 16 multiplied by 11 feature matrix, the middle feature matrix is 16 128 multiplied by 1 multiplied by 11 three-dimensional matrices, and 16 128-dimensional feature vectors are output through global maximum pooling and global average pooling.
The output of the reduced horizontal pyramid pooling structure is connected with the input of a reduced full-connection network, the reduced full-connection network comprises 16 independent full-connection neural network layers, each layer of input is 128-dimensional vectors, and the output is 128-dimensional vectors.
Compared with the current gait recognition model based on the standard convolution layer structure, the invention designs a compact lightweight gait recognition model (called a student model) by adopting low-cost depth separation convolution, and structurally reduces the parameter quantity of the model.
Example 2
The present embodiment describes a gait recognition model compression method based on the local-global joint knowledge distillation algorithm, which is based on the gait recognition model compression system based on the local-global joint knowledge distillation algorithm in the above embodiment 1.
In this embodiment 2, the two models in the above embodiment 1 are trained to achieve the purpose of gait recognition.
As shown in fig. 7, the gait recognition model compression method based on the local-global joint knowledge distillation algorithm comprises the following steps:
step 1, extracting gait contour sequences in the gait video by using a background subtraction method (the conventional method), and uniformly cutting into image sets with the same size, such as image sets with 64 multiplied by 64 pixels.
The image sets form a data set X, and the data set X is divided into a training set X train And test set X test
With the deep convolutional neural network as a basic structure, two gait recognition models are constructed and respectively recorded as a large-capacity teacher model M t And a light student model M s The model structure is already described in the above embodiment 1, and will not be described here again.
Step 2, training set X train Training teacher model M t The method comprises the steps of carrying out a first treatment on the surface of the The learning rate is set to 0.0001, the iteration number is 80000, the optimizer adopts an Adam optimizer, and 16 sequences of 8 objects are input once per iteration (128 sequences are total, 30 frames of images are randomly selected for each sequence, and the images are scaled to 64×44 pixels).
The loss function adopts a triplet loss function L as shown in a formula (1) tri
Wherein N is tri+ Representing two sample sets of all Euclidean distances other than 0 in a training sample subsetTotal number of sample pairs formed; the training sample subset is the secondary training set X when each training train A set formed by a plurality of sample images selected randomly;
n represents the number of fully connected neural network layers of the teacher network, and t represents the serial number of the fully connected neural network layers of the teacher network;
p represents the number of pedestrians contained in each training sample subset;
i and p respectively represent the serial numbers of the pedestrian samples to be trained in each training sample subset;
k represents the number of video sequences for each pedestrian in each training sample subset;
a. j and k respectively represent the sequence numbers of the pedestrian video sequences in each training sample subset;
m represents the boundary threshold of the loss function;
representing a j-th sample to be trained in the i-th training sample subset;
representing the sum +.>Any sample with the same identity of the represented pedestrian;
representing the sum +.>Any sample with different identities of the represented pedestrians;
sign | I 2 Representing the 2-norm of the matrix;
[] + representing ReLU operation, the calculation method is as follows: [ x ]] + =max {0, x }, max is the maximum valueAnd (3) operating.
By reducing the loss value, the points of the same object are mutually close, the points of different objects are mutually far away, the boundary threshold value M of the loss function is 0.2, and the training aim is to enable the teacher model M t The recognition performance of (a) is best.
Step 3, training set X train Input to a trained teacher model M t And untrained complete student model M s Respectively obtain the same data set in the teacher model M t Multi-dimensional feature matrix of convolution network output of (a)In student model M s Multi-dimensional characteristic matrix of the output of the reduced convolution network ∈>In teacher model M t Multi-dimensional feature matrix of full-connection network outputIn student model M s Multi-dimensional characteristic matrix of the output of the simplified fully-connected network>
Multi-dimensional feature matrixAnd multidimensional feature matrix->Is b x s x c x h x w; multidimensional feature matrix->And multidimensional feature matrix->Is b x n x d; wherein b represents each training sampleThe number of samples in this subset; s represents the number of frames; c represents the number of the output feature matrixes of the convolution layer, h represents the height of the output feature matrixes of the convolution network, w represents the width of the output feature matrixes of the convolution network, and d represents the dimension of the feature matrixes output by the full-connection network.
Step 4, making the difference measurement function L c_dis Calculating a multi-dimensional feature matrixAnd multidimensional feature matrix->Differences between; wherein the difference metric function L c_dis The calculation formula of (2) is as follows:
in the difference measurement function L c_dis Representing the partial distillation loss, the sign is 2 F Representing the F-norm of the matrix.
When the similarity matrix difference of the output characteristics is calculated, the L2 regularization method is further used for normalizing the difference of the characteristic matrices, so that the student model can be guided to learn more effective knowledge from the teacher model.
In this way, the teacher model M t And student model M s The number of channels of the convolved output features may not be consistent, that is, a model of greater or lesser capacity may be designed for knowledge distillation.
Step 5, respectively calculating the multi-dimensional feature matrix by using the formula (3)And multidimensional feature matrix->The difference between each sample of (a) is used as a result of +.>And->A representation;
wherein,representing the distance between all the samples of the same class in the feature matrix output by the teacher model; />Representing the distance between all different types of samples in the feature matrix output by the teacher model; />Representing the distance between all the samples of the same class in the feature matrix output by the student model; />And representing the distance between all different types of samples in the feature matrix output by the student model. The same class of samples refers to picture samples with the same pedestrian identity tag, and the different class of samples refers to picture samples with different pedestrian identity tags.
When the teacher model is trained, the jth sample to be trained with the same category in the ith training sample subset is represented;
when the teacher model is trained, the kth sample to be trained with the same category in the kth training sample subset is represented; />Representing the jth sample to be trained with different categories in the ith training sample subset when training the teacher model;
representing a kth sample to be trained with different categories in a p-th training sample subset when training a teacher model; />Representing the jth sample to be trained with the same category in the ith training sample subset when training the student model;
representing the kth sample to be trained with the same category in the kth training sample subset when training the student model; />Representing the jth sample to be trained with different categories in the ith training sample subset when training the student model;
representing the kth sample to be trained with different categories in the p-th training sample subset when training the student model.
Step 6, using the triplet loss function L shown in the formula (1) tri Respectively calculateAnd->Triple loss->And->The specific calculation formulas are shown in formulas (4) and (5);
/>
wherein m represents the boundary threshold of the loss function, and is set to 0.2;
Representing the total number of sample pairs consisting of two samples with all Euclidean distances other than 0 in one training sample subset in the teacher model;
wherein,representing the total number of sample pairs consisting of two samples with all Euclidean distances other than 0 in one training sample subset in the student model;
calculating the triplet loss by using the Smooth L1 loss function in the formula (6)And->Total distillation loss L of (2) f_dis
Step 7, partial distillation loss L c_dis Total distillation loss L f_dis Loss of tripletAnd->Integrating and calculating to obtain total loss L total The specific calculation formula is as follows:
wherein alpha is a distillation loss weight.
Step 8, setting a student model M s The iteration times are 30000, the optimizer is an Adam optimizer, the learning rate of the former 10000 times is set to be 0.005, the learning rate of the latter 20000 times is set to be 0.001, the optimizer is an Adam optimizer, and the loss value L is reduced total Transferring the knowledge of the teacher model to the student model M s And the recognition performance of the student model is improved.
Step 9. Test set X test The pedestrian video sequence is input to the student network M s And (3) performing recognition to obtain a recognition result.
The compression method of the gait recognition model has the advantages that the combined knowledge distillation algorithm is utilized to carry out local and whole knowledge distillation on the high-capacity gait recognition model (called a teacher model), the student model can be guided to learn more knowledge from the teacher model, and the original recognition effect is maintained as much as possible while the model capacity is reduced.
The method adopts the lightweight model compression and combined knowledge distillation technology, so that the gait recognition accuracy of the model can be effectively ensured while the scale of the model parameters is reduced, the operation cost is reduced, the training times and the reasoning time are reduced, the efficiency of the model is improved, and the method is more suitable for actual scenes with high real-time requirements and large data volume.
Experiments prove that compared with a deep neural network model in the prior art, the gait recognition model compression method disclosed by the invention has the advantages that the model parameter quantity is reduced by 9 times, the calculated quantity is reduced by 19 times, the performance of the model in the public data set CASIA-B is reduced by 2.2%, the actual reasoning time is reduced, and the problem of model efficiency is effectively solved.
The foregoing description is, of course, merely illustrative of preferred embodiments of the present invention, and it should be understood that the present invention is not limited to the above-described embodiments, but is intended to cover all modifications, equivalents and alternatives falling within the spirit and scope of the present invention as defined by the appended claims.

Claims (2)

1. A gait recognition model compression system based on a local-global combined knowledge distillation algorithm is characterized in that,
Comprises a teacher model Mt and a student model M s The method comprises the steps of carrying out a first treatment on the surface of the Wherein:
the teacher model Mt consists of a convolution network, a collection pooling structure, a horizontal pyramid pooling structure and a full-connection network;
the convolution network consists of a backbone network and a plurality of layers of global channels;
the backbone network consists of a first convolution module, a second convolution module and a third convolution module;
the first convolution module is composed of three layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 5×5 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the third layer is a pooling layer, and adopts a maximum pooling layer with pooling core of 2 multiplied by 2 and step length of 2;
the second convolution module consists of three layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the third layer is a pooling layer, and the largest pooling layer with pooling core of 2 multiplied by 2 and step length of 2 is used;
the third convolution module consists of two layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used;
the multi-layer global channel consists of a fourth convolution module and a fifth convolution module;
the fourth convolution module consists of three layers, wherein: the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the third layer is a pooling layer, and a pooling core of 2 multiplied by 2 is adopted;
The fifth convolution module consists of two layers, wherein:
the first layer is a standard convolution layer, and a convolution kernel of 3×3 is used; the second layer is a standard convolution layer, and a convolution kernel of 3×3 is used;
four aggregation pooling structures are arranged in the teacher model Mt, two horizontal pyramid pooling structures are arranged, and one full-connection network is arranged;
defining four collection pooling structures as a first collection pooling structure, a second collection pooling structure and a third collection pooling structure and a fourth collection pooling structure respectively; defining two horizontal pyramid pooling structures as a first horizontal pyramid pooling structure and a second horizontal pyramid pooling structure respectively;
the fully-connected network comprises a first fully-connected sub-network and a second fully-connected sub-network;
the output of the first convolution module is connected with the input of the second convolution module;
the output of the first convolution module is also connected with the input of the fourth convolution module through a first collection pooling structure;
the output of the second convolution module is connected with the input of the third convolution module;
the output of the second convolution module is connected with the input of the second aggregation pooling structure, and the output of the second aggregation pooling structure is connected with the input of the fifth convolution module after corresponding position addition is carried out on the output of the fourth convolution module;
the output of the third convolution module is connected with the input of the third aggregation pooling structure, and the output of the third aggregation pooling structure is connected with the input of the first horizontal pyramid pooling structure after corresponding position addition is carried out on the output of the fifth convolution module;
The output of the third convolution module is also connected with the input of the second horizontal pyramid pooling structure through the fourth aggregate pooling structure;
the output of the first horizontal pyramid pooling structure is connected with the input of the first fully-connected subnetwork;
the output of the second horizontal pyramid pooling structure is connected with the input of the second fully-connected subnetwork;
the outputs of the first fully-connected subnetwork and the second fully-connected subnetwork are used as a teacher model M t An output of (2);
the first and second horizontal pyramid pooling structures have five scales;
the first full-connection sub-network and the second full-connection sub-network respectively comprise 31 independent full-connection neural network layers;
student model M s The method comprises the steps of simplifying a convolution network, assembling a pooling structure, simplifying a horizontal pyramid pooling structure and simplifying a full-connection network, wherein:
the simplified convolution network consists of a sixth convolution module, a seventh convolution module and an eighth convolution module;
the sixth convolution module consists of two layers, wherein: the first layer is a standard convolution layer, and a convolution kernel of 5×5 is used; the second layer is a pooling layer, and the largest pooling layer with pooling core of 2 multiplied by 2 and step length of 2 is used;
the seventh convolution module consists of three layers, wherein:
the first layer is a depth separation convolution layer, and a convolution kernel of 5×5 is used; the second layer is a point convolution layer, and a convolution kernel of 1 multiplied by 1 is used; the third layer is a pooling layer, and adopts a maximum pooling layer with pooling core of 2 multiplied by 2 and step length of 2;
The eighth convolution module consists of two layers, wherein:
the first layer uses a 3 x 3 convolution kernel for the depth-separated convolution layer; the second layer is a point convolution layer, and a convolution kernel of 1 multiplied by 1 is used;
the sixth convolution module, the seventh convolution module and the eighth convolution module are connected in sequence;
definition of student model M s The medium aggregate pooling structure is a fifth aggregate pooling structure;
the output of the eighth convolution module is connected with the input of the simplified horizontal pyramid pooling structure through the fifth aggregate pooling structure;
the output of the simplified horizontal pyramid pooling structure is connected with the input of the simplified fully-connected network;
the simplified horizontal pyramid pooling structure has only one scale; the simplified fully-connected network comprises 16 independent fully-connected neural network layers; output of simplified fully-connected network as student model M s Is provided.
2. A gait recognition model compression method based on a local-global joint knowledge distillation algorithm, the gait recognition model compression system based on the local-global joint knowledge distillation algorithm described in claim 1; it is characterized in that the method comprises the steps of,
the gait recognition model compression method comprises the following steps:
step 1, extracting gait contour sequences in a gait video by using a background subtraction method, uniformly cutting the gait contour sequences into picture sets with the same size to form a data set X, and dividing the data set X into a training set X train And test set X test
Step 2, training set X train Training teacher model M t Setting a learning rate and iteration times, wherein an optimizer adopts an Adam optimizer, and a loss function adopts a triplet loss function L shown in a formula (1) tri
Wherein N is tri+ Representing the total number of sample pairs consisting of two samples of a training sample subset for which all Euclidean distances are not 0; the training sample subset is the secondary training set X when each training train A set formed by a plurality of sample images selected randomly;
n represents the number of fully connected neural network layers of the teacher network, and t represents the serial number of the fully connected neural network layers of the teacher network;
p represents the number of pedestrians contained in each training sample subset;
i and p respectively represent the serial numbers of the pedestrian samples to be trained in each training sample subset;
k represents the number of video sequences for each pedestrian in each training sample subset;
a. j and k respectively represent the sequence numbers of the pedestrian video sequences in each training sample subset;
m represents the boundary threshold of the loss function;
representing a j-th sample to be trained in the i-th training sample subset;
representing the sum +.>Any sample with the same identity of the represented pedestrian;
representing the sum +. >Any sample with different identities of the represented pedestrians;
sign | I 2 Representing the 2-norm of the matrix;
[] + representing ReLU operation, the calculation method is as follows: [ x ]] + =max {0, x }, max being the maximum value operation;
step 3, training set X train Input to a trained teacher model M t And untrained student model M s Respectively obtain the same data set in the teacher model M t Multi-dimensional feature matrix of convolution network output of (a)In student model M s Multi-dimensional characteristic matrix of the output of the reduced convolution network ∈>In teacher model M t Is a multi-dimensional feature matrix of the full-connection network output>In student model M s Multi-dimensional characteristic matrix of the output of the simplified fully-connected network>
Multi-dimensional feature matrixAnd multidimensional feature matrix->Is b x s x c x h x w; multidimensional feature matrix->And multidimensional feature matrix->Is b x n x d; wherein b represents the number of samples in each training sample subset; s represents the number of frames; c represents the number of the output feature matrixes of the convolution layer, h represents the height of the output feature matrixes of the convolution network, w represents the width of the output feature matrixes of the convolution network, and d represents the dimension of the feature matrixes output by the full-connection network;
step 4, making the difference measurement function L c_dis Calculating a multi-dimensional feature matrixAnd multidimensional feature matrix->Differences between; wherein the difference metric function L c_dis The calculation formula of (2) is as follows:
in the difference measurement function L c_dis Representing the loss of partial distillation, signRepresenting the F-norm of the matrix;
step 5, respectively calculating the multi-dimensional feature matrix by using the formula (3)And multidimensional feature matrix->The difference between each sample of (a) is used as a result of +.>A representation;
wherein,representing the distance between all the samples of the same class in the feature matrix output by the teacher model; />Representing the distance between all different types of samples in the feature matrix output by the teacher model; />Representing studentsThe distances among all the samples of the same category in the feature matrix output by the model; />Representing the distance between all different types of samples in the feature matrix output by the student model;
when the teacher model is trained, the jth sample to be trained with the same category in the ith training sample subset is represented;
when the teacher model is trained, the kth sample to be trained with the same category in the kth training sample subset is represented;
representing the jth sample to be trained with different categories in the ith training sample subset when training the teacher model;
representing a kth sample to be trained with different categories in a p-th training sample subset when training a teacher model;
Representing the jth sample to be trained with the same category in the ith training sample subset when training the student model;
representing the kth sample to be trained with the same category in the kth training sample subset when training the student model;
representing the jth sample to be trained with different categories in the ith training sample subset when training the student model;
representing the kth sample to be trained with different categories in the p-th training sample subset when training the student model;
step 6, using the triplet loss function L shown in the formula (1) tri Respectively calculateAnd->Triple loss->And->The specific calculation formulas are shown in formulas (4) and (5);
wherein,representing two samples of a training sample subset of the teacher model with all Euclidean distances other than 0Total number of sample pairs composed;
wherein,representing the total number of sample pairs consisting of two samples with all Euclidean distances other than 0 in one training sample subset in the student model;
calculating the triplet loss by using the Smooth L1 loss function in the formula (6)And->Total distillation loss L of (2) f_dis
Step 7, partial distillation loss L c_dis Total distillation loss L f_dis Loss of tripletAnd->Integrating and calculating to obtain total loss L total The specific calculation formula is as follows:
Wherein alpha is a distillation loss weight;
step 8, setting a student model M s The iteration times are counted, the optimizer is an Adam optimizer, and the loss value L is reduced total Transferring the knowledge of the teacher model to the student model;
step 9. Test set X test The pedestrian video sequence is input to the student network M s And (3) performing recognition to obtain a recognition result.
CN202110824459.3A 2021-07-21 2021-07-21 Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm Active CN113505719B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110824459.3A CN113505719B (en) 2021-07-21 2021-07-21 Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110824459.3A CN113505719B (en) 2021-07-21 2021-07-21 Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm

Publications (2)

Publication Number Publication Date
CN113505719A CN113505719A (en) 2021-10-15
CN113505719B true CN113505719B (en) 2023-11-24

Family

ID=78014088

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110824459.3A Active CN113505719B (en) 2021-07-21 2021-07-21 Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm

Country Status (1)

Country Link
CN (1) CN113505719B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116246349B (en) * 2023-05-06 2023-08-15 山东科技大学 Single-source domain generalization gait recognition method based on progressive subdomain mining
CN116824640B (en) * 2023-08-28 2023-12-01 江南大学 Leg identification method, system, medium and equipment based on MT and three-dimensional residual error network

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108596039A (en) * 2018-03-29 2018-09-28 南京邮电大学 A kind of bimodal emotion recognition method and system based on 3D convolutional neural networks
CN108764462A (en) * 2018-05-29 2018-11-06 成都视观天下科技有限公司 A kind of convolutional neural networks optimization method of knowledge based distillation
CN109034219A (en) * 2018-07-12 2018-12-18 上海商汤智能科技有限公司 Multi-tag class prediction method and device, electronic equipment and the storage medium of image
CN110097084A (en) * 2019-04-03 2019-08-06 浙江大学 Pass through the knowledge fusion method of projection feature training multitask student network
CN112560631A (en) * 2020-12-09 2021-03-26 昆明理工大学 Knowledge distillation-based pedestrian re-identification method
CN112784964A (en) * 2021-01-27 2021-05-11 西安电子科技大学 Image classification method based on bridging knowledge distillation convolution neural network

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108596039A (en) * 2018-03-29 2018-09-28 南京邮电大学 A kind of bimodal emotion recognition method and system based on 3D convolutional neural networks
CN108764462A (en) * 2018-05-29 2018-11-06 成都视观天下科技有限公司 A kind of convolutional neural networks optimization method of knowledge based distillation
CN109034219A (en) * 2018-07-12 2018-12-18 上海商汤智能科技有限公司 Multi-tag class prediction method and device, electronic equipment and the storage medium of image
CN110097084A (en) * 2019-04-03 2019-08-06 浙江大学 Pass through the knowledge fusion method of projection feature training multitask student network
CN112560631A (en) * 2020-12-09 2021-03-26 昆明理工大学 Knowledge distillation-based pedestrian re-identification method
CN112784964A (en) * 2021-01-27 2021-05-11 西安电子科技大学 Image classification method based on bridging knowledge distillation convolution neural network

Also Published As

Publication number Publication date
CN113505719A (en) 2021-10-15

Similar Documents

Publication Publication Date Title
CN111563508B (en) Semantic segmentation method based on spatial information fusion
CN111259786B (en) Pedestrian re-identification method based on synchronous enhancement of appearance and motion information of video
CN106778604B (en) Pedestrian re-identification method based on matching convolutional neural network
CN113239784B (en) Pedestrian re-identification system and method based on space sequence feature learning
CN111814661B (en) Human body behavior recognition method based on residual error-circulating neural network
CN112308158A (en) Multi-source field self-adaptive model and method based on partial feature alignment
CN110414432A (en) Training method, object identifying method and the corresponding device of Object identifying model
CN108288035A (en) The human motion recognition method of multichannel image Fusion Features based on deep learning
CN111639564B (en) Video pedestrian re-identification method based on multi-attention heterogeneous network
CN105678231A (en) Pedestrian image detection method based on sparse coding and neural network
CN113505719B (en) Gait recognition model compression system and method based on local-integral combined knowledge distillation algorithm
CN107301382B (en) Behavior identification method based on deep nonnegative matrix factorization under time dependence constraint
WO2022227292A1 (en) Action recognition method
CN104077742B (en) Human face sketch synthetic method and system based on Gabor characteristic
CN112396036B (en) Method for re-identifying blocked pedestrians by combining space transformation network and multi-scale feature extraction
CN113627266A (en) Video pedestrian re-identification method based on Transformer space-time modeling
CN114330529A (en) Real-time pedestrian shielding detection method based on improved YOLOv4
CN109165698A (en) A kind of image classification recognition methods and its storage medium towards wisdom traffic
CN112766378A (en) Cross-domain small sample image classification model method focusing on fine-grained identification
CN116704611A (en) Cross-visual-angle gait recognition method based on motion feature mixing and fine-granularity multi-stage feature extraction
CN109886160B (en) Face recognition method under non-limited condition
CN113269099B (en) Vehicle re-identification method under heterogeneous unmanned system based on graph matching
CN113505856B (en) Non-supervision self-adaptive classification method for hyperspectral images
CN112966672B (en) Gesture recognition method under complex background
CN116543269B (en) Cross-domain small sample fine granularity image recognition method based on self-supervision and model thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant