CN110084108A - Pedestrian re-identification system and method based on GAN neural network - Google Patents

Pedestrian re-identification system and method based on GAN neural network Download PDF

Info

Publication number
CN110084108A
CN110084108A CN201910208515.3A CN201910208515A CN110084108A CN 110084108 A CN110084108 A CN 110084108A CN 201910208515 A CN201910208515 A CN 201910208515A CN 110084108 A CN110084108 A CN 110084108A
Authority
CN
China
Prior art keywords
image
pedestrian
network
feature
loss
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910208515.3A
Other languages
Chinese (zh)
Inventor
高明柯
杜欣军
周燕琼
房家骥
王熠
邵培南
白利娟
夏定江
于楠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
CETC 32 Research Institute
Original Assignee
CETC 32 Research Institute
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by CETC 32 Research Institute filed Critical CETC 32 Research Institute
Priority to CN201910208515.3A priority Critical patent/CN110084108A/en
Publication of CN110084108A publication Critical patent/CN110084108A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands

Abstract

The invention relates to a pedestrian re-identification system and method based on a GAN neural network in the technical field of pedestrian re-identification, which are divided into two stages of image reconstruction and enhancement and pedestrian re-identification. In the image reconstruction and enhancement stage, firstly, a generation network in a GAN neural network is utilized to generate a high-resolution image from a low-resolution image; and then judging whether the generated image or the original image is acquired through a judgment network. If the image is generated, the image is generated again until the network cannot be distinguished, and the image reconstruction is completed; then, enhancing the image by adopting a Retinex algorithm; and in the pedestrian re-identification stage, firstly, extracting color features by using an HSV histogram on the basis of image reconstruction and enhancement, extracting texture features by using SILTP, extracting LOMO features of an image, finally, reducing the dimension of a space by using an XQDA method, and carrying out pedestrian re-identification by using distance measurement. The method has the advantages of high accuracy, good robustness, good generalization capability and the like, and is suitable for video pedestrian monitoring, video pedestrian information retrieval and other applications.

Description

Pedestrian's weight identifying system and method based on GAN neural network
Technical field
The present invention relates to pedestrian's weight identification technology fields, and in particular, to a kind of pedestrian based on GAN neural network knows again Other system and method.
Background technique
Pedestrian identifies that (Person re-identification) is also referred to as pedestrian and identifies again again, is to utilize computer vision skill Art judges the technology that whether there is specific pedestrian in image or video sequence.Pedestrian identifies that mainly there are two committed steps again: Firstly, feature extraction, the i.e. feature of extraction target pedestrian and candidate pedestrian.Then, metric learning, i.e. calculating target pedestrian and time The characteristic similarity for selecting pedestrian judges whether candidate pedestrian is the target to be looked for.Originally researcher introduces some image procossings, mould Formula identifies the maturation method in field, stresses to have studied the available feature of pedestrian, simple classification algorithm.Since two thousand and ten, pedestrian The training library of weight identification technology tends to large-scale, and deep learning frame is widely used and method is trained and reasoning.
Domestic patent CN108520226A proposes a kind of decompose based on body and the pedestrian side of identification again of conspicuousness detection Method.Pedestrian image is resolved into the semantic area with deep decomposition network first, pedestrian is separated from cluttered environment. Next, pedestrian image is divided into fritter, according to the result and salient region of background deduction, effective picture region is automatically selected Domain.LOMO feature is extracted from selected picture region and from whole image finally, extracting PHOG, HSV histogram and SIFT feature. Domestic patent CN105718882A proposes a kind of resolution ratio self-adaptive feature extraction and pedestrian's recognition methods again for merging.This method Biological characteristic and macroscopic features are merged, identifies pedestrian, the difference of Enhanced feature with the method that face characteristic and macroscopic features merge Property, while the performance according to feature on different images scale, different pedestrian's features will be compared on different scale;Together Shi Caiyong Filtering system is first screened with the fusion feature that color characteristic and contour feature obtain, then with face characteristic pair The selection result is supplemented, finally the texture feature extraction on the pedestrian of screening, very big stable color provincial characteristics and weighting face Color characteristic, and the global characteristics and local feature that are extracted using adaptive weighted method fusion obtain fusion feature.
Foreign patent US9396412B2 proposes a kind of recognition methods again of the pedestrian based on semantic color designation.This method is adopted Character image is described with semantic color designation, in combination with traditional color characteristic, calculates the probability point of 11 kinds of basic colors Cloth is as image descriptor.This method is given from rgb value to the mapping of the probability distribution of 11 kinds of colors, constructs semantic histogram conduct Image descriptor.In order to which preferably by semantic color designation and other widely used features, (including color histogram, texture are straight Side's figure and covariance matrix) it combines, this method will be defined as partial descriptor based on apparent affinity model, and pass through phase It measures to form linear combination like property.Different from other conventional methods, it is used as special using the long original vector with antipode Levy pond.Secondly, learning the weight of each similarity measurement by RankBoost algorithm.It is related in foreign patent US10108850B1 And a kind of recognition methods again of the pedestrian based on AlexNet neural network.This method extracts the various of pedestrian in different pictures first Physical trait, such as trunk, foot etc., then by the pre-training of convolutional neural networks and the training of AlexNet, finally by feature Than realizing that pedestrian identifies again.
Summary of the invention
In view of the drawbacks of the prior art, the object of the present invention is to provide a kind of pedestrians based on GAN neural network to identify again System and method.The present invention establishes the neural network structure based on GAN, by image reconstruction and enhancing, solves difference and takes the photograph Under camera and different perspectives, the problem of image definition, and the sensitivity of image under different illumination is reduced by Retinex algorithm Degree;Realize that pedestrian identifies again by LOMO feature combination XQDA method simultaneously, with accuracy is high, robustness is good, generalization ability The advantages that good, is suitable for the practical applications such as video pedestrian monitoring, video pedestrian's information retrieval.
Pedestrian's weight identifying system that the present invention relates to a kind of based on GAN neural network, including image reconstruction and enhancing module, Pedestrian's weight identification module;
Described image, which is rebuild, to be built up with enhancing module for generating network, differentiation network and loss function calculating and counterweight Image enhanced;
Pedestrian's weight identification module is used to rebuild described image the enhancing picture obtained with enhancing module and carries out pedestrian It identifies again.
Preferably, the generation network includes multiple residual blocks, includes two 3 × 3 convolutional layers, volume in each residual block Connection batch standardization layer after lamination chooses PReLU as activation primitive, reconnects two sub-pix convolutional layers and be used to increase feature Size.
Preferably, the differentiation network includes 8 convolutional layers, and as the network number of plies is deepened, Characteristic Number is continuously increased, special Sign size constantly reduces, and chooses LeakyReLU as activation primitive, activates letter eventually by two full articulamentums and sigmoid Number obtains the probability for being predicted as original image.
Preferably, what the differentiation network judgement obtained is to generate image or original image, if image is generated, then again It generates, until differentiating that network cannot be distinguished, then completes image reconstruction.
Preferably, the loss function calculating includes being weighted meter using different weights to content loss and confrontation loss It calculates, the content loss be the Euclidean distance generated between image and original image, and the confrontation loses to generate and allow arbiter Indistinguishable data distribution.
Preferably, the content loss includes the Minimum Mean Square Error Mseloss of pixel space and with 19 layers of Vgg net of pre-training The Euclidean distance Vggloss between image and original image feature is sought survival into based on the ReLU active coating of network.
Preferably, described image, which is rebuild, carries out image enhancement using Retinex algorithm algorithm with enhancing module.
Preferably, the Retinex algorithm algorithm is first according to R, G of pixel, B component by the cromogram of input As being decomposed into three width images, the intensity of the different reflected light of scene medium wavelength is represented;Calculate separately long wave, medium wave and short-wave band Opposite relationship between light and dark between interior pixel, and then determine the color of each pixel, finally by the color line in Retinex chrominance space Property be mapped to rgb space, obtain enhancing image.
Preferably, pedestrian's weight identification module extracts color with HSV histogram on the basis of image reconstruction and enhancing first Feature with SILTP texture feature extraction, then extracts image LOMO feature, carries out dimensionality reduction to space finally by XQDA method, and Pedestrian is carried out using distance metric to identify again.
A kind of recognition methods again of the pedestrian based on GAN neural network, which comprises the steps of:
Step 1, building generates network:
Generating network portion includes multiple residual blocks, includes two 3 × 3 convolutional layers in each residual block, after convolutional layer Connection batch standardization layer chooses PReLU as activation primitive, reconnects two sub-pix convolutional layers for increased feature sizes;
Step 2, building differentiates network:
Differentiate that network portion includes 8 convolutional layers, as the network number of plies is deepened, Characteristic Number is continuously increased, characteristic size Constantly reduce, chooses LeakyReLU as activation primitive, obtained eventually by two full articulamentums and sigmoid activation primitive It is predicted as the probability of original image;
Step 3, loss function calculates:
Loss includes two parts: the content loss of super-resolution image (SR)Weighted sum generator (Gen) confrontation Loss
Wherein X indicates one group of high-definition picture and low-resolution image.
Content loss indicates to generate the Euclidean distance between image and original image, including Mse loss and Vgg loss;
The Minimum Mean Square Error of Mse loss expression pixel space:
Wherein, x, y indicate image coordinate point;R indicates decimation factor;W indicates low resolution (LR) image ILRWidth;H Indicate the height of low-resolution image;RW and rH respectively indicates r times of low-resolution image ILRWidth and height;IHRIt indicates High-definition picture, i.e. original image;GθG(ILR)X, yThe high resolution graphics of networking of making a living complexing;GθGIndicate that generator, θ G indicate The weight and biasing of L layer depth network;
Vgg loss: based on the ReLU active coating of 19 layers of Vgg network of pre-training, image and original image feature are sought survival into Between Euclidean distance, extract the feature map of a certain layer on trained Vgg, image current layer will be generated Feature map corresponding to feature map and original image is compared:
Wherein, WI, jAnd HI, jThe size of each characteristic pattern in VGG network is described;I, j indicate that i-th maximizes pooling Jth time convolution after layer;φ corresponds to the characteristic pattern that certain convolutional layer exports after activation primitive among VGG network, φI, j (IHR)X, yIndicate the characteristic pattern of high-definition picture, φI, j(GθG(ILR))X, yIndicate the characteristic pattern of generation network composograph, The loss function can reflect the error on higher perception level, and mean square error loss item can only reflect between the pixel of low level Error, therefore VGG loss item also known as perceive loss item;
Confrontation lossFor generating the data distribution for making arbiter indistinguishable:
Wherein, DθDFor arbiter, θ D is the weight of arbiter, the probability depending on candidate samples from data distribution;DθD (GθG(ILR)) what is indicated is that arbiter will generate the probability that image prediction is original image;N is sample size;
Step 4, image enhancement is carried out using Retinex algorithm:
The color image of input is decomposed by three width images according to R, G of pixel, B component first, represents scene medium wavelength The intensity of different reflected lights;It calculates separately long wave, the opposite relationship between light and dark in medium wave and short-wave band between pixel, and then determines The color of each pixel obtains enhancing image finally, the color in Retinex chrominance space is linearly mapped to rgb space;
Step 5, feature extraction:
On the basis of step 1-4 carries out image reconstruction and enhancing, color characteristic is extracted with HSV histogram, is mentioned with SILTP Take textural characteristics;
Step 6, LOMO extracts feature:
The down-sampled of 2 × 2average pooling twice is carried out to original image, LOMO is all made of to three images and is mentioned Feature is taken, then by the merging features of three images at a feature vector, is finally used for the especially big value in feature vector Log transformation is inhibited, renormalization later to unit-sized;
Step 7, dimensionality reduction and distance metric are carried out to space using XQDA method:
Enabling Δ=xi-xj indicates the feature difference between 2 samples, and P is Gaussian Profile, and P (Δ | Ω I) it is similar sample Between discrepancy delta meet the Gaussian Profile of difference Ω I in class, the discrepancy delta of P (Δ | Ω E) between foreign peoples's sample meets between class The Gaussian Profile of difference Ω E, and mean value is all 0:
Wherein, T is matrix transposition, and d (xi-xj) is distance function, ∑EAnd ∑IIt is similar sample respectively to collection and dissmilarity Covariance matrix of the sample to collection sample;F (Δ) indicates the distance between 2 samples, if more than 0, then it represents that very in maximum probability It is not similar;
XQDA method carries out feature principal component analysis and similarity-based learning simultaneously, by learning mapping matrix W ∈ Rd×r(r < d), by primitive character xi, xi∈RdIt is mapped to multidimensional subspace, Feature Dimension Reduction is realized, improves the accuracy of characteristic matching, Middle matrix W is made of the corresponding feature vector of preceding r maximum eigenvalue of principal component analysis, and d indicates dimension, therefore, formula (6) Defined in distance function change conversion are as follows:
F (Δ)=(x-z)TW(∑′I -1-∑′E -1)WT(x-z) (7)
Wherein, T be matrix transposition, ∑ 'I -1=WTIW, indicate dissimilar sample to the covariance matrix of collection sample, ∑′E -1=WTEW indicates similar sample to the covariance matrix of collection;X is the sample point at a visual angle, and z is another visual angle Sample point;
Step 8: output recognition result.
Compared with prior art, the present invention have it is following the utility model has the advantages that
1, the present invention establishes the neural network structure based on GAN, by image reconstruction and enhancing, solves different camera shootings Under machine and different perspectives, the problem of image definition, and the sensitivity of image under different illumination is reduced by Retinex algorithm Degree.
2, the present invention by LOMO feature combination XQDA method realize pedestrian identify again, have accuracy height, robustness it is good, The advantages that generalization ability is good is suitable for the practical applications such as video pedestrian monitoring, video pedestrian's information retrieval.
Detailed description of the invention
Upon reading the detailed description of non-limiting embodiments with reference to the following drawings, other feature of the invention, Objects and advantages will become more apparent upon:
Fig. 1 is the recognition methods flow diagram again of the pedestrian based on GAN neural network.
Specific embodiment
The present invention is described in detail combined with specific embodiments below.Following embodiment will be helpful to the technology of this field Personnel further understand the present invention, but the invention is not limited in any way.It should be pointed out that the ordinary skill of this field For personnel, without departing from the inventive concept of the premise, various modifications and improvements can be made.These belong to the present invention Protection scope.
Embodiment
In view of the above demand, the present invention relates to a kind of recognition methods again of the pedestrian based on GAN neural network, are divided into image weight It builds and identifies two stages again with enhancing and pedestrian.Image reconstruction and enhancing stage, first with the generation net in GAN neural network The image of low resolution is generated high-definition picture by network.It then, is to generate image by differentiate network judgement acquisition, still Original image.If generating image, then regenerate, until differentiating that network cannot be distinguished, then completes image reconstruction;It uses again Retinex algorithm enhances image.Pedestrian's weight cognitive phase is straight with HSV on the basis of image reconstruction and enhancing first Side's figure extracts color characteristic, with SILTP texture feature extraction, then image LOMO feature is extracted, finally by XQDA method to sky Between carry out dimensionality reduction, and carry out pedestrian using distance metric and identify again.
As shown in Figure 1, being divided into image reconstruction the present invention relates to a kind of recognition methods again of the pedestrian based on GAN neural network Two stages are identified again with enhancing stage and pedestrian.
The step of image reconstruction is with the enhancing stage is as follows:
Step 1: building generates network
Generating network portion includes multiple residual blocks, includes two 3 × 3 convolutional layers in each residual block, after convolutional layer Connection batch standardization layer chooses PReLU as activation primitive, reconnects two sub-pix convolutional layers for increased feature sizes.
Step 2: building differentiates network
Differentiate that network portion includes 8 convolutional layers, as the network number of plies is deepened, Characteristic Number is continuously increased, characteristic size Constantly reduce, chooses LeakyReLU as activation primitive, obtained eventually by two full articulamentums and sigmoid activation primitive It is predicted as the probability of original image.
Step 3: loss function calculates
Loss function calculates:
Loss includes two parts: the content loss of super-resolution image (SR)Weighted sum generator (Gen) confrontation Loss
Wherein X indicates one group of high-definition picture and low-resolution image.
Content loss indicates to generate the Euclidean distance between image and original image, including Mse loss and Vgg loss;
The Minimum Mean Square Error of Mse loss expression pixel space:
Wherein, x, y indicate image coordinate point;R indicates decimation factor;W indicates low resolution (LR) image ILRWidth;H Indicate the height of low-resolution image;RW and rH respectively indicates r times of low-resolution image ILRWidth and height;IHRIt indicates High-definition picture, i.e. original image;GθG(ILR)X, yThe high resolution graphics of networking of making a living complexing;GθGIndicate that generator, θ G indicate The weight and biasing of L layer depth network;
Vgg loss: based on the ReLU active coating of 19 layers of Vgg network of pre-training, image and original image feature are sought survival into Between Euclidean distance, extract the feature map of a certain layer on trained Vgg, image current layer will be generated Feature map corresponding to feature map and original image is compared:
Wherein, WI, jAnd HI, jThe size of each characteristic pattern in VGG network is described;I, j indicate that i-th maximizes pooling Jth time convolution after layer;φ corresponds to the characteristic pattern that certain convolutional layer exports after activation primitive among VGG network, φI, j (IHR)X, yIndicate the characteristic pattern of high-definition picture, φI, j(GθG(ILR))X, yIndicate the characteristic pattern of generation network composograph, it should Loss function can reflect the error on higher perception level, and mean square error loss item can only reflect between the pixel of low level Error, therefore VGG loss item also known as perceives loss item;
Confrontation lossFor generating the data distribution for making arbiter indistinguishable:
Wherein, DθDFor arbiter, θ D is the weight of arbiter, the probability depending on candidate samples from data distribution;DθD (GθG(ILR)) what is indicated is that arbiter will generate the probability that image prediction is original image;N is sample size;
Step 4: carrying out image enhancement using Retinex algorithm
The color image of input is decomposed by three width images according to R, G of pixel, B component first, represents scene medium wavelength The intensity of different reflected lights;It calculates separately long wave, the opposite relationship between light and dark in medium wave and short-wave band between pixel, and then determines The color of each pixel obtains enhancing image finally, the color in Retinex chrominance space is linearly mapped to rgb space.
Pedestrian's weight cognitive phase step:
Step 1: feature extraction
Color characteristic is extracted with HSV histogram on the basis of image reconstruction and enhancing, with SILTP texture feature extraction. In view of multi-scale information, the down-sampled of 2 × 2average pooling twice is carried out to original image, three images are adopted Feature is extracted with LOMO, then by the merging features of three images at a feature vector, finally, for the spy in feature vector Big value is inhibited using log transformation, renormalization later to unit-sized;
Step 2: dimensionality reduction and distance metric are carried out to space using XQDA method:
Enabling Δ=xi-xj indicates the feature difference between 2 samples, and P is Gaussian Profile, and P (Δ | Ω I) it is similar sample Between discrepancy delta meet the Gaussian Profile of difference Ω I in class, the discrepancy delta of P (Δ | Ω E) between foreign peoples's sample meets between class The Gaussian Profile of difference Ω E, and mean value is all 0:
Wherein, T is matrix transposition, and d (xi-xj) is distance function, ∑EAnd ∑IIt is similar sample respectively to collection and dissmilarity Covariance matrix of the sample to collection sample;F (Δ) indicates the distance between 2 samples, if more than 0, then it represents that very in maximum probability It is not similar;
XQDA method carries out feature principal component analysis and similarity-based learning simultaneously, by learning mapping matrix W ∈ Rd×r(r < d), by primitive character xi, xj∈RdIt is mapped to multidimensional subspace, Feature Dimension Reduction is realized, improves the accuracy of characteristic matching, Middle matrix W is made of the corresponding feature vector of preceding r maximum eigenvalue of principal component analysis, and d indicates dimension, therefore, formula (6) Defined in distance function change conversion are as follows:
F (Δ)=(x-z)TW(∑′I -1-∑′E -1)WT(x-z) (7)
Wherein, T be matrix transposition, ∑ 'I -1=WTIW, indicate dissimilar sample to the covariance matrix of collection sample, ∑′E -1=WTEW indicates similar sample to the covariance matrix of collection;X is the sample point at a visual angle, and z is another visual angle Sample point;
Step 3: output recognition result.
In conclusion the present invention establishes the neural network structure based on GAN, by image reconstruction and enhancing, solve Under different cameras and different perspectives, the problem of image definition, and image under different illumination is reduced by Retinex algorithm Susceptibility;It realizes that pedestrian identifies again by LOMO feature combination XQDA method simultaneously, has accuracy height, robustness good, general The advantages that change ability is good is suitable for the practical applications such as video pedestrian monitoring, video pedestrian's information retrieval.
Specific embodiments of the present invention are described above.It is to be appreciated that the invention is not limited to above-mentioned Particular implementation, those skilled in the art can make a variety of changes or modify within the scope of the claims, this not shadow Ring substantive content of the invention.In the absence of conflict, the feature in embodiments herein and embodiment can any phase Mutually combination.

Claims (10)

1. it is a kind of based on GAN neural network pedestrian weight identifying system, which is characterized in that including image reconstruction and enhancing module, Pedestrian's weight identification module;
Described image, which is rebuild, can generate network with enhancing module, differentiate network and the figure that loss function calculates and counterweight is built up As being enhanced;
Pedestrian's weight identification module can rebuild the enhancing picture progress pedestrian obtained with enhancing module to described image to be known again Not.
2. pedestrian's weight identifying system according to claim 1 based on GAN neural network, characterized in that the generation net Network includes multiple residual blocks, includes two 3 × 3 convolutional layers in each residual block, connection batch standardization layer, chooses after convolutional layer PReLU reconnects two sub-pix convolutional layers as activation primitive for increased feature sizes.
3. pedestrian's weight identifying system according to claim 1 based on GAN neural network, characterized in that the differentiation net Network includes 8 convolutional layers, and as the network number of plies is deepened, Characteristic Number is continuously increased, and characteristic size constantly reduces, and is chosen LeakyReLU obtains being predicted as original image as activation primitive eventually by two full articulamentums and sigmoid activation primitive Probability.
4. pedestrian's weight identifying system according to claim 3 based on GAN neural network, characterized in that the differentiation net What network judgement obtained is to generate image or original image, if image is generated, is then regenerated, until differentiating that network can not area Point, then complete image reconstruction.
5. pedestrian's weight identifying system according to claim 1 based on GAN neural network, characterized in that the loss letter It includes that content loss and confrontation loss are weighted using different weights that number, which calculates, and the content loss is to generate image Euclidean distance between original image, the confrontation loss are used to generate the data distribution for making arbiter indistinguishable.
6. pedestrian's weight identifying system according to claim 5 based on GAN neural network, characterized in that the content damage It loses the Minimum Mean Square Error Mseloss including pixel space and is sought survival based on the ReLU active coating of 19 layers of Vgg network of pre-training At the Euclidean distance Vggloss between image and original image feature.
7. pedestrian's weight identifying system according to claim 1 based on GAN neural network, characterized in that described image weight It builds and carries out image enhancement using Retinex algorithm algorithm with enhancing module.
8. pedestrian's weight identifying system according to claim 7 based on GAN neural network, characterized in that the Retinex The color image of input is decomposed into three width images according to R, G of pixel, B component first by algorithm algorithm, is represented in scene The intensity of the different reflected light of wavelength;Calculate separately long wave, the opposite relationship between light and dark in medium wave and short-wave band between pixel, in turn It determines the color of each pixel, the color in Retinex chrominance space is linearly finally mapped to rgb space, obtain enhancing figure Picture.
9. pedestrian's weight identifying system according to claim 1 based on GAN neural network again, characterized in that pedestrian identifies Module extracts color characteristic with HSV histogram on the basis of image reconstruction and enhancing first, with SILTP texture feature extraction, Image LOMO feature is extracted again, dimensionality reduction is carried out to space finally by XQDA method, and carry out pedestrian using distance metric and know again Not.
10. a kind of recognition methods again of the pedestrian based on GAN neural network, which comprises the steps of:
Step 1, building generates network:
Generating network portion includes multiple residual blocks, includes two 3 × 3 convolutional layers in each residual block, connects after convolutional layer Standardization layer is criticized, PReLU is chosen as activation primitive, reconnects two sub-pix convolutional layers for increased feature sizes;
Step 2, building differentiates network:
Differentiate that network portion includes 8 convolutional layers, as the network number of plies is deepened, Characteristic Number is continuously increased, and characteristic size is continuous Reduce, chooses LeakyReLU as activation primitive, predicted eventually by two full articulamentums and sigmoid activation primitive For the probability of original image;
Step 3, loss function calculates:
Loss includes two parts: the content loss of super-resolution image (SR)Weighted sum generator (Gen) confrontation loss
Wherein X indicates one group of high-definition picture and low-resolution image.
Content loss indicates to generate the Euclidean distance between image and original image, including Mse loss and Vgg loss;
The Minimum Mean Square Error of Mse loss expression pixel space:
Wherein, x, y indicate image coordinate point;R indicates decimation factor;W indicates low resolution (LR) image ILRWidth;H is indicated The height of low-resolution image;RW and rH respectively indicates r times of low-resolution image ILRWidth and height;IHRIndicate high score Resolution image, i.e. original image;GθG(ILR)X, yThe high resolution graphics of networking of making a living complexing;GθGIndicate that generator, θ G indicate L layers The weight and biasing of depth network;
Vgg loss: it based on the ReLU active coating of 19 layers of Vgg network of pre-training, seeks survival between image and original image feature Euclidean distance, extract the feature map of a certain layer on trained Vgg, image current layer will be generated Feature map corresponding to feature map and original image is compared:
Wherein, WI, jAnd HI, jThe size of each characteristic pattern in VGG network is described;After i, j indicate that i-th maximizes pooling layers Jth time convolution;φ corresponds to the characteristic pattern that certain convolutional layer exports after activation primitive among VGG network, φI, j(IHR)X, yTable Show the characteristic pattern of high-definition picture, φI, j(GθG(ILR))X, yIndicate the characteristic pattern of generation network composograph, the loss function Can error on the higher perception level of reflection, and mean square error loss item can only reflect the error between the pixel of low level, because This VGG loses item and also known as perceives loss item;
Confrontation lossFor generating the data distribution for making arbiter indistinguishable:
Wherein, DθDFor arbiter, θ D is the weight of arbiter, the probability depending on candidate samples from data distribution;DθD(GθG (ILR)) what is indicated is that arbiter will generate the probability that image prediction is original image;N is sample size;
Step 4, image enhancement is carried out using Retinex algorithm:
The color image of input is decomposed by three width images according to R, G of pixel, B component first, it is different to represent scene medium wavelength Reflected light intensity;It calculates separately long wave, the opposite relationship between light and dark in medium wave and short-wave band between pixel, and then determines each The color of pixel obtains enhancing image finally, the color in Retinex chrominance space is linearly mapped to rgb space;
Step 5, feature extraction:
On the basis of step 1-4 carries out image reconstruction and enhancing, color characteristic is extracted with HSV histogram, extracts line with SILTP Manage feature;
Step 6, LOMO extracts feature:
The down-sampled of 2 × 2 average pooling twice is carried out to original image, to three images be all made of LOMO extract it is special Sign is finally become for the especially big value in feature vector using log then by the merging features of three images at a feature vector Swap-in row inhibits, renormalization later to unit-sized;
Step 7, dimensionality reduction and distance metric are carried out to space using XQDA method:
Enabling Δ=xi-xj indicates the feature difference between 2 samples, and P is Gaussian Profile, and P (Δ | Ω I) between similar sample Discrepancy delta meet the Gaussian Profile of difference Ω I in class, the discrepancy delta of P (Δ | Ω E) between foreign peoples's sample meets class inherited The Gaussian Profile of Ω E, and mean value is all 0:
Wherein, T is matrix transposition, and d (xi-xj) is distance function, ∑EAnd ∑IIt is similar sample respectively to collection and dissimilar sample To the covariance matrix of collection sample;F (Δ) indicates the distance between 2 samples, if more than 0, then it represents that be not in maximum probability very much It is similar;
XQDA method carries out feature principal component analysis and similarity-based learning simultaneously, by learning mapping matrix W ∈ Rd×r(r < D), by primitive character xi, xj∈RdIt is mapped to multidimensional subspace, Feature Dimension Reduction is realized, improves the accuracy of characteristic matching, wherein Matrix W is made of the corresponding feature vector of preceding r maximum eigenvalue of principal component analysis, and d indicates dimension, therefore, in formula (6) The distance function of definition changes conversion are as follows:
F (Δ)=(x-z)TW(∑′I -1-∑′E -1)WT(x-z) (7)
Wherein, T be matrix transposition, ∑ 'I -1=WTIW, indicate dissimilar sample to the covariance matrix of collection sample, ∑ 'E -1= WTEW indicates similar sample to the covariance matrix of collection;X is the sample point at a visual angle, and z is the sample point at another visual angle;
Step 8: output recognition result.
CN201910208515.3A 2019-03-19 2019-03-19 Pedestrian re-identification system and method based on GAN neural network Pending CN110084108A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910208515.3A CN110084108A (en) 2019-03-19 2019-03-19 Pedestrian re-identification system and method based on GAN neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910208515.3A CN110084108A (en) 2019-03-19 2019-03-19 Pedestrian re-identification system and method based on GAN neural network

Publications (1)

Publication Number Publication Date
CN110084108A true CN110084108A (en) 2019-08-02

Family

ID=67413310

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910208515.3A Pending CN110084108A (en) 2019-03-19 2019-03-19 Pedestrian re-identification system and method based on GAN neural network

Country Status (1)

Country Link
CN (1) CN110084108A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110688966A (en) * 2019-09-30 2020-01-14 华东师范大学 Semantic-guided pedestrian re-identification method
CN110728238A (en) * 2019-10-12 2020-01-24 安徽工程大学 Personnel re-detection method of fusion type neural network
CN111178178A (en) * 2019-12-16 2020-05-19 汇纳科技股份有限公司 Multi-scale pedestrian re-identification method, system, medium and terminal combined with region distribution
CN111275694A (en) * 2020-02-06 2020-06-12 电子科技大学 Attention mechanism guided progressive division human body analytic model and method
CN111340700A (en) * 2020-02-21 2020-06-26 北京中科虹霸科技有限公司 Model generation method, resolution improvement method, image identification method and device
CN111402143A (en) * 2020-06-03 2020-07-10 腾讯科技(深圳)有限公司 Image processing method, device, equipment and computer readable storage medium
CN111460891A (en) * 2020-03-01 2020-07-28 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Automatic driving-oriented vehicle-road cooperative pedestrian re-identification method and system
CN111476353A (en) * 2020-04-07 2020-07-31 中国科学院重庆绿色智能技术研究院 Super-resolution method of GAN image introducing significance
CN111539382A (en) * 2020-05-22 2020-08-14 支付宝(杭州)信息技术有限公司 Image recognition model privacy risk assessment method and device and electronic equipment
CN111582126A (en) * 2020-04-30 2020-08-25 浙江工商大学 Pedestrian re-identification method based on multi-scale pedestrian contour segmentation fusion
CN112016402A (en) * 2020-08-04 2020-12-01 杰创智能科技股份有限公司 Unsupervised learning-based pedestrian re-identification field self-adaption method and device
CN112200106A (en) * 2020-10-16 2021-01-08 中国计量大学 Cross-camera pedestrian re-identification and tracking method
CN112348766A (en) * 2020-11-06 2021-02-09 天津大学 Progressive feature stream depth fusion network for surveillance video enhancement
CN112446361A (en) * 2020-12-16 2021-03-05 上海芯翌智能科技有限公司 Method and equipment for cleaning training data
CN112507617A (en) * 2020-12-03 2021-03-16 青岛海纳云科技控股有限公司 Training method of SRFlow super-resolution model and face recognition method
CN112906614A (en) * 2021-03-08 2021-06-04 中南大学 Pedestrian re-identification method and device based on attention guidance and storage medium
CN113128441A (en) * 2021-04-28 2021-07-16 安徽大学 System and method for identifying vehicle weight by embedding structure of attribute and state guidance
CN113344798A (en) * 2020-02-18 2021-09-03 济南浪潮高新科技投资发展有限公司 Retinex-based dark image enhancement method
CN113536995A (en) * 2021-06-30 2021-10-22 河南大学 Pedestrian re-identification method based on feature mapping space and sample judgment
CN112016402B (en) * 2020-08-04 2024-05-17 杰创智能科技股份有限公司 Self-adaptive method and device for pedestrian re-recognition field based on unsupervised learning

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108520199A (en) * 2018-03-04 2018-09-11 天津大学 Based on radar image and the human action opener recognition methods for generating confrontation model

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108520199A (en) * 2018-03-04 2018-09-11 天津大学 Based on radar image and the human action opener recognition methods for generating confrontation model

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
LEDIG C等: "Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network", 《ARXIV:1609.04802V5 [CS.CV] 25 MAY 2017》 *
SHENGCAI LIAO等: "Person Re-identification by Local Maximal Occurrence Representation and Metric Learning", 《2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR)》 *
李学明: "基于Retinex理论的图像增强算法", 《万方学术期刊数据库》 *
金翠等: "基于随机擦除行人对齐网络的行人重识别方法", 《山东大学学报(工学版)》 *

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110688966B (en) * 2019-09-30 2024-01-09 华东师范大学 Semantic guidance pedestrian re-recognition method
CN110688966A (en) * 2019-09-30 2020-01-14 华东师范大学 Semantic-guided pedestrian re-identification method
CN110728238A (en) * 2019-10-12 2020-01-24 安徽工程大学 Personnel re-detection method of fusion type neural network
CN111178178A (en) * 2019-12-16 2020-05-19 汇纳科技股份有限公司 Multi-scale pedestrian re-identification method, system, medium and terminal combined with region distribution
CN111178178B (en) * 2019-12-16 2023-10-10 汇纳科技股份有限公司 Multi-scale pedestrian re-identification method, system, medium and terminal combined with region distribution
CN111275694B (en) * 2020-02-06 2020-10-23 电子科技大学 Attention mechanism guided progressive human body division analysis system and method
CN111275694A (en) * 2020-02-06 2020-06-12 电子科技大学 Attention mechanism guided progressive division human body analytic model and method
CN113344798A (en) * 2020-02-18 2021-09-03 济南浪潮高新科技投资发展有限公司 Retinex-based dark image enhancement method
CN111340700A (en) * 2020-02-21 2020-06-26 北京中科虹霸科技有限公司 Model generation method, resolution improvement method, image identification method and device
CN111340700B (en) * 2020-02-21 2023-04-25 北京中科虹霸科技有限公司 Model generation method, resolution improvement method, image recognition method and device
CN111460891A (en) * 2020-03-01 2020-07-28 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) Automatic driving-oriented vehicle-road cooperative pedestrian re-identification method and system
CN111476353B (en) * 2020-04-07 2022-07-15 中国科学院重庆绿色智能技术研究院 Super-resolution method of GAN image introducing significance
CN111476353A (en) * 2020-04-07 2020-07-31 中国科学院重庆绿色智能技术研究院 Super-resolution method of GAN image introducing significance
CN111582126B (en) * 2020-04-30 2024-02-27 浙江工商大学 Pedestrian re-recognition method based on multi-scale pedestrian contour segmentation fusion
CN111582126A (en) * 2020-04-30 2020-08-25 浙江工商大学 Pedestrian re-identification method based on multi-scale pedestrian contour segmentation fusion
CN111539382A (en) * 2020-05-22 2020-08-14 支付宝(杭州)信息技术有限公司 Image recognition model privacy risk assessment method and device and electronic equipment
CN111402143A (en) * 2020-06-03 2020-07-10 腾讯科技(深圳)有限公司 Image processing method, device, equipment and computer readable storage medium
CN111402143B (en) * 2020-06-03 2020-09-04 腾讯科技(深圳)有限公司 Image processing method, device, equipment and computer readable storage medium
CN112016402B (en) * 2020-08-04 2024-05-17 杰创智能科技股份有限公司 Self-adaptive method and device for pedestrian re-recognition field based on unsupervised learning
CN112016402A (en) * 2020-08-04 2020-12-01 杰创智能科技股份有限公司 Unsupervised learning-based pedestrian re-identification field self-adaption method and device
CN112200106A (en) * 2020-10-16 2021-01-08 中国计量大学 Cross-camera pedestrian re-identification and tracking method
CN112348766B (en) * 2020-11-06 2023-04-18 天津大学 Progressive feature stream depth fusion network for surveillance video enhancement
CN112348766A (en) * 2020-11-06 2021-02-09 天津大学 Progressive feature stream depth fusion network for surveillance video enhancement
CN112507617A (en) * 2020-12-03 2021-03-16 青岛海纳云科技控股有限公司 Training method of SRFlow super-resolution model and face recognition method
CN112446361A (en) * 2020-12-16 2021-03-05 上海芯翌智能科技有限公司 Method and equipment for cleaning training data
CN112906614A (en) * 2021-03-08 2021-06-04 中南大学 Pedestrian re-identification method and device based on attention guidance and storage medium
CN113128441B (en) * 2021-04-28 2022-10-14 安徽大学 System and method for identifying vehicle weight by embedding structure of attribute and state guidance
CN113128441A (en) * 2021-04-28 2021-07-16 安徽大学 System and method for identifying vehicle weight by embedding structure of attribute and state guidance
CN113536995B (en) * 2021-06-30 2022-11-18 河南大学 Pedestrian re-identification method based on feature mapping space and sample judgment
CN113536995A (en) * 2021-06-30 2021-10-22 河南大学 Pedestrian re-identification method based on feature mapping space and sample judgment

Similar Documents

Publication Publication Date Title
CN110084108A (en) Pedestrian re-identification system and method based on GAN neural network
Jiang et al. Edge-enhanced GAN for remote sensing image superresolution
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN112507997B (en) Face super-resolution system based on multi-scale convolution and receptive field feature fusion
CN108573276B (en) Change detection method based on high-resolution remote sensing image
Ahmad Deep image retrieval using artificial neural network interpolation and indexing based on similarity measurement
Xiao et al. Making of night vision: Object detection under low-illumination
CN106599854B (en) Automatic facial expression recognition method based on multi-feature fusion
CN111625608B (en) Method and system for generating electronic map according to remote sensing image based on GAN model
CN112560831B (en) Pedestrian attribute identification method based on multi-scale space correction
CN107679250A (en) A kind of multitask layered image search method based on depth own coding convolutional neural networks
CN109558806A (en) The detection method and system of high score Remote Sensing Imagery Change
CN110929593A (en) Real-time significance pedestrian detection method based on detail distinguishing and distinguishing
CN109635726B (en) Landslide identification method based on combination of symmetric deep network and multi-scale pooling
CN109543632A (en) A kind of deep layer network pedestrian detection method based on the guidance of shallow-layer Fusion Features
CN111652273B (en) Deep learning-based RGB-D image classification method
CN115423734B (en) Infrared and visible light image fusion method based on multi-scale attention mechanism
Liang et al. An improved DualGAN for near-infrared image colorization
CN113870160B (en) Point cloud data processing method based on transformer neural network
Cho et al. Semantic segmentation with low light images by modified CycleGAN-based image enhancement
CN115359372A (en) Unmanned aerial vehicle video moving object detection method based on optical flow network
CN114724155A (en) Scene text detection method, system and equipment based on deep convolutional neural network
CN111428758A (en) Improved remote sensing image scene classification method based on unsupervised characterization learning
CN113052200A (en) Sonar image target detection method based on yolov3 network
Guo et al. Dual graph U-Nets for hyperspectral image classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190802

RJ01 Rejection of invention patent application after publication