CN113128456B - Pedestrian re-identification method based on combined picture generation - Google Patents

Pedestrian re-identification method based on combined picture generation Download PDF

Info

Publication number
CN113128456B
CN113128456B CN202110485010.9A CN202110485010A CN113128456B CN 113128456 B CN113128456 B CN 113128456B CN 202110485010 A CN202110485010 A CN 202110485010A CN 113128456 B CN113128456 B CN 113128456B
Authority
CN
China
Prior art keywords
pedestrian
picture
model
characteristic information
encoder
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110485010.9A
Other languages
Chinese (zh)
Other versions
CN113128456A (en
Inventor
苏迪
张�成
王少博
邱语聃
冀瑞静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202110485010.9A priority Critical patent/CN113128456B/en
Publication of CN113128456A publication Critical patent/CN113128456A/en
Application granted granted Critical
Publication of CN113128456B publication Critical patent/CN113128456B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2155Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a pedestrian re-identification method generated by combining pictures. The pedestrian re-recognition method is based on the confrontation generation network GAN to re-recognize the pedestrian picture, and in the training process of the confrontation generation network, a similar unsupervised learning method is used, a teacher model is adopted to predict the identity (characteristic information) of the pedestrian picture, a feature encoder of the GAN is guided, the information of a generated sample is utilized more fully, the quality of the model is improved, and a student model with higher precision is obtained. In addition, the invention also utilizes the thought of triple loss, expands the data, shortens the difference in the classes, increases the difference between the classes and further improves the identification precision.

Description

Pedestrian re-identification method based on joint picture generation
Technical Field
The invention relates to the technical field of computer vision, in particular to a pedestrian re-identification method generated by combining pictures.
Background
Pedestrian re-identification refers to that a pedestrian picture to be retrieved is known, other pictures of the pedestrian are retrieved from a non-overlapping camera shooting picture database, and the pedestrian re-identification is generally regarded as a metric learning problem. The difficulty is that pictures shot by different cameras have serious intra-class difference on information such as pedestrian appearance, posture, background and the like. Meanwhile, pedestrians can have certain inter-class similarity under a specific visual angle. With the development of deep convolutional networks in recent years, the pedestrian re-identification is greatly progressed by representing a pedestrian picture more robustly and with stronger identification capability, and the representation is even better than the human identification on some data sets.
To reduce the effect of intra-class differences, some methods perform component segmentation on the pedestrian picture, which in turn embeds the component features directly into the pedestrian feature representation through component matching or. Although the robust metric learning strategy solves the problem of insufficient data to some extent, the introduction of additional pedestrian pictures can directly increase the accuracy of the model. With the development of generative confrontation networks, generative models are widely used in augmenting data, and some methods augment the original data set by generating new samples. However, the method for generating the sample has the problems of less distinguishing information of the generated sample, low identity reality degree of the sample, separation of the generated model and the re-recognition model and the like.
In the prior art, information such as Pose or semantics is usually introduced to generate a sample, for example, a method for extracting identity information from a pedestrian picture and generating the pedestrian picture with the corresponding Pose by using GAN in combination with the human Pose is disclosed in FD-GAN (position-guided Feature distinguishing GAN for Robust Person Re-identification) published in NeuroIPS of 2018 computer vision conference; joint characterization and generation Learning for Person Re-identification published in 2019 on computer vision conference CVPR discloses a Joint Learning frame in which a pedestrian identity feature encoder is embedded into a pedestrian Re-identification model, and meanwhile, a pre-trained pedestrian Re-identification model is used as a teacher model to distribute pseudo labels to generated samples. However, in practical applications, pre-trained models with higher recognition accuracy do not necessarily exist.
Disclosure of Invention
In view of this, the invention provides a pedestrian re-identification method combining picture generation, which generates unlabeled samples by using a coder-decoder structure based on a generated countermeasure network (GAN), introduces a teacher model to perform identity prediction on the generated samples, does not need to introduce additional labeling information, a pre-training model and the like, can improve identification precision, and improves the applicability of a pedestrian re-identification algorithm in different scenes.
The pedestrian re-identification method combining the picture generation adopts a generation countermeasure network to complete the picture generation and re-identification, wherein the generation countermeasure network comprises a characteristic encoder, a structure encoder, a generator and a discriminator; the system comprises a pedestrian image input module, a teacher model and a control module, wherein the teacher model is used for identifying characteristic information of an input pedestrian image; taking the feature encoder as a student model, wherein the structure of a teacher model is consistent with that of the feature encoder; the parameter updating mode of the teacher model is as follows: the parameter of the teacher model at the current moment is the weighted sum of the parameter of the student model at the current moment and the parameter of the teacher model at the previous moment, and the initial value of the parameter of the teacher model is the initial value of the parameter of the characteristic encoder;
generating a loss function of the countermeasure network, wherein the loss function also comprises consistency loss of the characteristic information identified by the teacher model at the current time and the characteristic information identified by the characteristic encoder at the current time;
and aiming at the pedestrian picture to be identified, the characteristic encoder in the generation countermeasure network acquires the characteristic information of the pedestrian picture, and compares the characteristic information with the sample set data to complete the identification of the pedestrian.
Preferably, the parameter of the teacher model at the current moment is an exponential weighted average sum of the parameter of the teacher model at the previous moment and the parameter of the feature encoder at the current moment.
Preferably, generating the penalty function for the counterpoise network further includes a triple penalty, wherein the triple penalty L trip Comprises the following steps:
Figure BDA0003049970940000031
wherein,
Figure BDA0003049970940000032
to utilize the pedestrian picture x i Characteristic information f of i Pedestrian picture x with different identities j Structural code s of j Reconstructing the generated picture; e f A feature encoder; d is a distance; m is a set value.
Preferably, after the generation of the confrontation network training is completed, the characteristic information of the pedestrian picture to be identified is directly obtained by the teacher model, and the characteristic information is compared with the sample concentrated data to complete the identification of the pedestrian.
Has the advantages that:
(1) The pedestrian re-recognition method based on the generated countermeasure network GAN is used for re-recognizing the pedestrian picture, a similar unsupervised learning method is used in the training process of generating the countermeasure network, a teacher model is used for predicting the identity (characteristic information) of the pedestrian picture, a feature encoder of the GAN is guided, the information of a generated sample is utilized more fully, the quality of the model is improved, and a student model with higher precision is obtained.
(2) The teacher model updates the model parameters by adopting a weighted average method, and the model parameters do not need to be updated through iterative training, so that the training time is saved, and the complexity of the whole model is reduced.
(3) The invention utilizes the thought of triple loss, expands the data, shortens the intra-class difference, increases the inter-class difference and further improves the identification precision.
Drawings
FIG. 1 is a schematic flow diagram of the process of the present invention;
FIG. 2 is a schematic view of the overall model structure of the present invention;
fig. 3 is a schematic structural diagram of an average teacher model of the present invention.
Detailed Description
The invention is described in detail below by way of example with reference to the accompanying drawings.
The invention provides a pedestrian re-identification method generated by combining pictures, a flow chart of which is shown in figure 1, and the method specifically comprises the following steps:
step 1, reading data from the sample data set, and acquiring a pedestrian picture x in the data set i (x i E.x, i =1,2 …, N is the total number of pictures in the sample data set) and its labeled identity information y k (y k E.g. Y, K =1,2 …, K, K being the number of samplesTotal number of identity categories in the dataset).
Step 2, constructing a generation countermeasure network (GAN):
the generation countermeasure network (GAN) comprises an encoder E, a generator G and a discriminator D, wherein the encoder comprises a structure encoder E s Sum-feature encoder E f (ii) a The overall structure of the resulting countermeasure network is shown in fig. 2.
Pedestrian picture x with identity mark to be read i Input to a constituent encoder E s Sum-feature encoder E f Respectively acquiring the structural information s of the pedestrian pictures i And characteristic information f i The structure information comprises information such as picture background and pedestrian posture, the characteristic information comprises information related to identity recognition, and then a generator G is used for reconstructing the picture; the picture reconstruction of the generator G comprises self reconstruction and cross reconstruction:
(1) Self-reconstruction:
pedestrian picture x i Self-reconstruction is achieved by two methods:
(1-1) Picture of pedestrian x i Characteristic information f of i And structural information s i Reconstructing to generate characteristic information f i Structural information of s i Reconstructed picture of
Figure BDA0003049970940000041
As shown in the second row of the picture in fig. 2; the constraint condition of the reconstruction is that the identity of the reconstructed picture is consistent with that of the original pedestrian picture and the reconstructed picture is consistent on the pixel, so that the loss can be solved by a method of consistent on the pixel, namely the reconstruction uses l 1 The method calculates the loss:
Figure BDA0003049970940000042
wherein G represents a generator, E represents a desired calculation, "| | | | calculation 1 "denotes a 1 A loss calculation method.
(1-2) utilization with Picture x i Having pedestrians of the same identity but with other different structural informationPicture x of k Drawing x i Structural information s of i And picture x k Characteristic information f of k Generating characteristic information f through G reconstruction k Structural information of s i Picture of
Figure BDA0003049970940000051
As illustrated in the third row of the picture in fig. 2; similarly, the constraint condition of the reconstruction is that the generated picture and the original pedestrian picture keep consistent identity and consistent pixel, so the loss can be solved by a consistent pixel method, namely the reconstruction uses one 1 The method calculates the loss:
Figure BDA0003049970940000052
the above two reconstruction methods serve as main constraints on the generator G, which facilitates aggregation of feature information of the same identity and reduction of differences in features within classes.
Meanwhile, the invention introduces identity loss to ensure that the characteristic information of different identities keeps separation:
Figure BDA0003049970940000053
wherein p (y) i |x i ) Is based on the feature information to the picture x i Identity prediction and its true value y i The consistency comparison. The log in the formula is based on e, as follows.
(2) Cross reconstruction:
the cross reconstruction is to mix two pedestrian pictures with different identities to generate one picture, wherein the pedestrian picture x i Providing a feature encoder E f Coded characteristic information f i Pedestrian picture x with different identities j Providing a structural encoder E s Coded structure information s j Reconstructing the generated picture
Figure BDA0003049970940000054
I.e. at x j Structural information s of j Upper hold x i Characteristic information f of i As illustrated in the first row of pictures in fig. 2. The loss cannot be solved by a method of consistency on picture pixels due to no real value, but the characteristic information and the structure information of the generated picture are reconstructed, and the constraint condition is that the characteristic information and the structure information are consistent with the original input information, so that the generated picture and the pedestrian picture from which the characteristic information comes have partial consistency on the identity characteristic, and the reconstructed picture is combined with the generated picture>
Figure BDA0003049970940000057
Respectively performing loss calculation on the characteristic information and the structural information:
Figure BDA0003049970940000055
/>
Figure BDA0003049970940000056
likewise, identity loss is also used to maintain separation of its characteristic information on identity:
Figure BDA0003049970940000061
wherein
Figure BDA0003049970940000063
Based on the characteristic information, the picture is judged>
Figure BDA0003049970940000062
Identity prediction of (2) and its true value y i The consistency comparison.
And finally, judging the authenticity of the generated picture by adopting a discriminator D.
Generating pictures using GAN inevitably introduces a penalty to ensure the authenticity of the generated pictures, said penalty being:
L adv =E[log D(x i )+log(1-D(G(f i ,s j )))] (7)
for cross reconstruction, because the identity information of the generated cross identity picture is difficult to define, the invention not only assumes the identity source is consistent with the identity source of the feature information, but also predicts the identity by using a teacher model and guides the recognition of a feature encoder in the GAN.
Taking a characteristic encoder in the GAN as a student model, and constructing a teacher model, wherein the structure of the teacher model is consistent with that of the characteristic encoder, and the initial parameters of the teacher model are the initial parameters of the characteristic encoder; in the training process of the GAN, the parameters of the teacher model are updated along with the student models, but the parameters are not optimally updated through loss functions like the student models, and the teacher model is updated to be the weighted sum of the parameters of the teacher model at the last moment and the parameters of the student models at the current moment.
In this embodiment, the teacher model parameter θ at the current time t ' use parameter θ of student model at present moment t And teacher model parameter θ 'of last time' t-1 I.e.:
θ′ t =αθ′ t-1 +(1-α)θ t (8)
where α is a smoothing coefficient.
The average weight teacher model can replace a pre-training re-recognition model to label the identity of the generated picture, meanwhile, the teacher model is not updated through a loss function, the complexity of the whole model is reduced, and the structural schematic of the average weight teacher model is shown in fig. 3.
The invention considers that the teacher model has higher pedestrian weight recognition precision than the student model in the early stage of model training. To reconcile student model predictions with teacher models, use l 2 Losses constitute a consistent loss function for both (student and teacher models) predictions:
Figure BDA00030499709400000713
wherein,
Figure BDA0003049970940000071
respectively, the teacher model and the student model predict the identity probability distribution.
In addition, in the cross-reconstruction process, the embodiment also applies to the picture x i 、x j
Figure BDA0003049970940000072
Feature information uses triple loss. Due to the generated picture->
Figure BDA0003049970940000073
Is expected to be in x j Under the structure of (1) hold x i Is inevitably related to x j Characteristically similar, the picture ≥ is generated for enhancement>
Figure BDA00030499709400000712
And identity discrimination, introducing characteristic separation loss, and adopting a similar triple mode to enable the distance between the generated picture and the characteristic information of the two pedestrian identity pictures to be larger than m so as to form a characteristic separation loss function.
Specifically, the three pictures are taken as a triplet, and the feature encoder is taken as a mapping function, so that the loss of the feature triplet can be obtained:
Figure BDA0003049970940000074
/>
wherein,
Figure BDA0003049970940000075
represents->
Figure BDA0003049970940000076
x i The distance of the characteristic information->
Figure BDA0003049970940000077
Represents->
Figure BDA0003049970940000078
x j The distance of the feature information, m, is a feature distance constant.
The total loss function is obtained by weighted addition of the loss functions (equations (1) to (7), and equations (9) to (10)):
Figure BDA0003049970940000079
wherein,
Figure BDA00030499709400000710
is a loss function of self-reconstruction>
Figure BDA00030499709400000711
Is a loss function of the reconstruction of the characteristic information and the structural information in the cross reconstruction, lambda i 、λ id 、λ cons Are the weighting coefficients that control the importance of the loss function.
Optimizing feature encoder E using sample set data f Encoder of structure E s A generator and a discriminator D, finally obtaining a well-trained generated countermeasure network.
And aiming at the pedestrian picture to be identified, identifying the characteristic information of the pedestrian picture by a characteristic encoder, and then comparing the characteristic information with the sample set data to finish the identification of the pedestrian.
To demonstrate the effectiveness of the method of the invention, comparative experiments were performed on the Market1501 and Duke MTMC-reiD datasets, with the experimental results shown in Table 1 being the results of the underlying network, the use of average teacher losses only, the use of triple losses only, and both. Experiments show that the method has the advantages of enabling the accuracy of the pedestrian re-identification model to be higher.
TABLE 1
Figure BDA0003049970940000081
The invention also discloses a published prescription with better effectThe results of comparison are shown in table 2, the upper half is the supervised method without using the generated samples, the lower half is the method using the generated samples, and we test the feature encoder E separately f And teacher model E t . Experiments have shown that the method of the invention is very challenging and more accurate therein.
TABLE 2
Figure BDA0003049970940000082
Figure BDA0003049970940000091
In summary, the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (2)

1. A pedestrian re-identification method combining picture generation adopts a generation countermeasure network to complete picture generation and re-identification, wherein the generation countermeasure network comprises a feature encoder, a structure encoder, a generator and a discriminator; the system is characterized by further comprising a teacher model, wherein the teacher model is used for identifying characteristic information of an input pedestrian picture; the feature encoder is used as a student model, and the structure of a teacher model is consistent with that of the feature encoder; the parameter updating mode of the teacher model is as follows: the parameter of the teacher model at the current moment is the weighted sum of the parameter of the student model at the current moment and the parameter of the teacher model at the previous moment, and the initial value of the parameter of the teacher model is the initial value of the parameter of the feature encoder;
generating a loss function of the countermeasure network, wherein the loss function also comprises consistency loss of the characteristic information identified by the teacher model at the current moment and the characteristic information identified by the characteristic encoder at the current moment;
and aiming at the pedestrian picture to be identified, the characteristic encoder in the generation countermeasure network acquires the characteristic information of the pedestrian picture, and compares the characteristic information with the sample set data to complete the identification of the pedestrian.
2. The pedestrian re-identification method with joint picture generation as claimed in claim 1, wherein generating the loss function of the countermeasure network further comprises a triplet loss, wherein the triplet loss L trip Comprises the following steps:
Figure FDA0003930086150000011
wherein,
Figure FDA0003930086150000012
to utilize the pedestrian picture x i Characteristic information f of i Pedestrian picture x with different identities j Structural code s of j Reconstructing the generated picture; e f A feature encoder; d is a distance; m is a set value. />
CN202110485010.9A 2021-04-30 2021-04-30 Pedestrian re-identification method based on combined picture generation Active CN113128456B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110485010.9A CN113128456B (en) 2021-04-30 2021-04-30 Pedestrian re-identification method based on combined picture generation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110485010.9A CN113128456B (en) 2021-04-30 2021-04-30 Pedestrian re-identification method based on combined picture generation

Publications (2)

Publication Number Publication Date
CN113128456A CN113128456A (en) 2021-07-16
CN113128456B true CN113128456B (en) 2023-04-07

Family

ID=76780750

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110485010.9A Active CN113128456B (en) 2021-04-30 2021-04-30 Pedestrian re-identification method based on combined picture generation

Country Status (1)

Country Link
CN (1) CN113128456B (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109934117A (en) * 2019-02-18 2019-06-25 北京联合大学 Based on the pedestrian's weight recognition detection method for generating confrontation network

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110008842A (en) * 2019-03-09 2019-07-12 同济大学 A kind of pedestrian's recognition methods again for more losing Fusion Model based on depth
CN110427813B (en) * 2019-06-24 2023-06-09 中国矿业大学 Pedestrian re-recognition method of twin generation type countermeasure network based on gesture guidance pedestrian image generation
KR102225022B1 (en) * 2019-08-27 2021-03-08 연세대학교 산학협력단 Person re-identification apparatus and method

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109934117A (en) * 2019-02-18 2019-06-25 北京联合大学 Based on the pedestrian's weight recognition detection method for generating confrontation network

Also Published As

Publication number Publication date
CN113128456A (en) 2021-07-16

Similar Documents

Publication Publication Date Title
CN109815826B (en) Method and device for generating face attribute model
CN112800903B (en) Dynamic expression recognition method and system based on space-time diagram convolutional neural network
AU2014368997B2 (en) System and method for identifying faces in unconstrained media
CN108230291B (en) Object recognition system training method, object recognition method, device and electronic equipment
CN110717411A (en) Pedestrian re-identification method based on deep layer feature fusion
WO2022052530A1 (en) Method and apparatus for training face correction model, electronic device, and storage medium
CN113705290A (en) Image processing method, image processing device, computer equipment and storage medium
CN111931908B (en) Face image automatic generation method based on face contour
CN114220154A (en) Micro-expression feature extraction and identification method based on deep learning
CN117611932B (en) Image classification method and system based on double pseudo tag refinement and sample re-weighting
CN113822790B (en) Image processing method, device, equipment and computer readable storage medium
WO2022166840A1 (en) Face attribute editing model training method, face attribute editing method and device
CN112380374B (en) Zero sample image classification method based on semantic expansion
CN114387641A (en) False video detection method and system based on multi-scale convolutional network and ViT
CN116311483B (en) Micro-expression recognition method based on local facial area reconstruction and memory contrast learning
CN112364791A (en) Pedestrian re-identification method and system based on generation of confrontation network
CN111126155B (en) Pedestrian re-identification method for generating countermeasure network based on semantic constraint
CN112784783A (en) Pedestrian re-identification method based on virtual sample
CN117351550A (en) Grid self-attention facial expression recognition method based on supervised contrast learning
CN114360051A (en) Fine-grained behavior identification method based on progressive hierarchical weighted attention network
CN114240811A (en) Method for generating new image based on multiple images
CN113096133A (en) Method for constructing semantic segmentation network based on attention mechanism
CN116935438A (en) Pedestrian image re-recognition method based on autonomous evolution of model structure
CN111461061A (en) Pedestrian re-identification method based on camera style adaptation
CN113128456B (en) Pedestrian re-identification method based on combined picture generation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant