CN112801020A - Pedestrian re-identification method and system based on background graying - Google Patents
Pedestrian re-identification method and system based on background graying Download PDFInfo
- Publication number
- CN112801020A CN112801020A CN202110174952.5A CN202110174952A CN112801020A CN 112801020 A CN112801020 A CN 112801020A CN 202110174952 A CN202110174952 A CN 202110174952A CN 112801020 A CN112801020 A CN 112801020A
- Authority
- CN
- China
- Prior art keywords
- bgg
- background
- stream
- pedestrian
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 238000010586 diagram Methods 0.000 claims abstract description 18
- 238000000605 extraction Methods 0.000 claims abstract description 10
- 238000012545 processing Methods 0.000 claims abstract description 8
- 206010039509 Scab Diseases 0.000 claims abstract description 7
- 230000006870 function Effects 0.000 claims description 17
- 230000004913 activation Effects 0.000 claims description 12
- 238000004590 computer program Methods 0.000 claims description 12
- 238000013507 mapping Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 6
- 230000004044 response Effects 0.000 claims description 6
- 238000012360 testing method Methods 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 3
- 230000007246 mechanism Effects 0.000 claims description 3
- 238000011176 pooling Methods 0.000 claims description 3
- 230000009467 reduction Effects 0.000 claims description 3
- 230000009286 beneficial effect Effects 0.000 abstract description 3
- 230000003313 weakening effect Effects 0.000 abstract description 2
- 238000001514 detection method Methods 0.000 description 4
- 230000011218 segmentation Effects 0.000 description 4
- 210000000746 body region Anatomy 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000036544 posture Effects 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/103—Static body considered as a whole, e.g. static pedestrian or occupant recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/52—Surveillance or monitoring of activities, e.g. for recognising suspicious objects
- G06V20/53—Recognition of crowd images, e.g. recognition of crowd congestion
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Human Computer Interaction (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention relates to a pedestrian re-identification method and a system based on background graying, wherein the method comprises the following steps: s1, carrying out background graying processing on the original image to obtain a processed BGg image; s2, performing feature extraction on BGg images by using a two-way network based on ResNet50, wherein one way is background gray level Stream BGg-Stream, and the other way is global Stream G-Stream; s3, interacting the two networks through cascade connection; s4, connecting the obtained feature diagrams of the two networks at each stage with an SCAB module to serve as input feature diagrams of next layer convolution of BGg-stream, so that the two networks are connected and the two features are combined; and S5, updating the network parameters by adopting a triple loss function, calculating the similarity, and finally outputting a permutation sequence. The method and the system are beneficial to weakening background interference and improving the accuracy of pedestrian re-identification.
Description
Technical Field
The invention belongs to the field of computer vision, and particularly relates to a pedestrian re-identification method and system based on background graying.
Background
Human re-identification (ReID) is a well studied problem in computer vision, by retrieving images of specific pedestrians from a large number of images taken by different cameras. The method is the most basic visual identification problem in video monitoring and has wide application prospect. The conventional methods proposed by ReID mostly use low-order features of color histograms and texture histograms, and use metric learning to find a distance function that minimizes the distance between images from the same class and maximizes the distance between different classes. ReID has been studied in academia for many years, and has not achieved a significant breakthrough until the last few years with the development of deep learning. However, the problem that cannot be ignored is that the depth convolution characteristic is a high-dimensional characteristic, and is easily influenced by factors such as human postures, object shielding, different illumination intensities, background clutter and the like.
Many methods have been proposed over the past few years to obtain more robust features. But these methods tend to take the entire image directly as input. Such global information contains not only pedestrian features but also clutter features of the background. Currently, effective methods for alleviating the influence of background clutter are mainly divided into two types: 1) methods based on body region detection, such as pose and keypoint estimation, extract body information in images using methods of region detection. 2) A method based on human body segmentation. Recent existing image segmentation methods including FCN, mask R-CNN, jppnet, etc. have been able to obtain excellent effects in terms of background truncation.
But in a practical unconstrained scenario, pedestrian re-identification remains a very challenging task. How to extract the discriminative and robust features of the background clutter invariance is a core problem, because the non-pedestrian part in the background can cause great interference to the features of the foreground information.
Most of the existing methods for solving background interference are based on body region detection or using segmentation to filter out the background, but they still have one of the following limitations. First, additional detection models and segmentation models need to be pre-trained, as well as additional data acquisition efforts. Secondly, potential data set deviation between the attitude estimation and the reid can cause partitioning errors and destroy the original complete body type characteristics of pedestrians. And thirdly, although the background is partially cut off, the background still exists around the pedestrians and participates in model training with the weight equivalent to the pedestrian area. By doing so, these methods do not really address the background clutter problem, which does not present a fundamental solution. Fourthly, the defects brought by the strong and hard segmentation not only destroy the original structure and smoothness of the image, but also completely abandon all background information. Some context information may sometimes be useful context information, and ignoring all of it may ignore some clues about the task of pedestrian re-identification.
Disclosure of Invention
The invention aims to provide a pedestrian re-identification method and system based on background graying, which are beneficial to weakening background interference and improving the accuracy of pedestrian re-identification.
In order to achieve the purpose, the invention adopts the technical scheme that: a pedestrian re-identification method based on background graying comprises the following steps: s1, carrying out background gray processing on the original image, and keeping the foreground information of the image unchanged to obtain a processed BGg image, namely a background gray image;
s2, based on a ResNet50 two-way network, wherein one way is background gray level Stream BGg-Stream, the background gray level image obtained in the step S1 is subjected to feature extraction, and the other way is global Stream G-Stream, and the original image is subjected to feature extraction;
s3, interacting the two networks obtained in the step S2 through cascade connection;
s4, connecting the two networks and combining the two characteristics by using a feature diagram obtained by the two networks at each stage as an input feature diagram of the next layer of convolution of BGg-stream through a space-channel attention mechanism module, namely an SCAB module;
and S5, after feature extraction, updating network parameters by adopting a triple loss function, then carrying out similarity calculation, and finally outputting a permutation sequence.
Further, the input of the BGg-stream is BGg images, and the background gray-scale formula of the images is as follows:
BGg(i,j)=0.299×R(i,j)+0.587×G(i,j)+0.114×B(i,j)
where BGg (i, j) is the pixel value of BGg image at row i and column j, and R, G, B is the three channels of the RGB image.
Further, the SCAB module comprises a channel attention module and a space attention module.
Further, the channel attention module is implemented as follows: given input feature mapping Fi∈RC*H*WFirstly, the spatial information of the feature map is aggregated by using average pooling operation to generate a spatial context descriptor Fi∈RC*1*1Compressing and converting the spatial information into a channel; then the activation size of the concealment is set to Fi∈RC/r*1*1To reduce the parameter overhead, wherein R is the reduction rate; thus, the channel attention module is represented as:
Fii=σ(θ(R(ζ(δ(Fi)))))
wherein FiThe method comprises the steps that the characteristics of BGg-Stream of each stage are represented, sigma is a sigmoid activation function, and R is a ReLU activation function; δ is the average pool number; θ and ζ are represented as two distinct tie layers;representing element-wise multiplication.
Further, the space attention module is implemented as follows: mapping F to input featuresi∈RC*H*WWhereinC is the total number of channels, H × W is the size of the feature map, then the space attention module is represented as:
Fiii=σ(C(Fi))
wherein sigma is sigmoid activation function, and the output of the space attention module is Fiii∈R1*H*W。
Further, a local feature response maximization strategy, namely an LRM strategy, is applied in the testing phase; in the test process, the feature map is divided into n regions with proper number according to the level, and the feature with the maximum response is extracted from each part of features to be used as the feature of the part.
The invention also provides a pedestrian re-identification system based on background graying, which comprises a memory, a processor and a computer program which is stored on the memory and can run on the processor, wherein when the computer program is run by the processor, the steps of the method are realized.
Compared with the prior art, the invention has the following beneficial effects:
1. the present invention does not require additional training and data sets.
2. The invention can keep the integrity and effectiveness of the original body type information of the pedestrian.
3. The invention can accurately position the human body area and process all backgrounds including the whole body range of the pedestrian.
4. The invention can not be interfered by strong color in the background and can leave useful information.
5. The method and the device enable the model to be concentrated in the learning of the foreground information in the learning process, and further weaken background interference.
Drawings
FIG. 1 is a flow chart of a method implementation of an embodiment of the present invention.
Fig. 2 is a schematic structural diagram of a channel attention module in the embodiment of the present invention.
Fig. 3 is a schematic structural diagram of a space attention module according to an embodiment of the present invention.
Detailed Description
The invention is further explained below with reference to the drawings and the embodiments.
It should be noted that the following detailed description is exemplary and is intended to provide further explanation of the disclosure. Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this application belongs.
It is noted that the terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of example embodiments according to the present application. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, and it should be understood that when the terms "comprises" and/or "comprising" are used in this specification, they specify the presence of stated features, steps, operations, devices, components, and/or combinations thereof, unless the context clearly indicates otherwise.
As shown in fig. 1, the present embodiment provides a pedestrian re-identification method based on background graying, which includes the following steps:
s1, performing background graying processing on the original image by using mask, while the foreground information of the image remains unchanged, to obtain a processed BGg image, i.e., a background grayed image (an image in which the foreground is an RGB image and the background is a grayscale image).
S2, based on a ResNet50 two-way network, wherein one way is background gray level Stream BGg-Stream, the background gray level image obtained in the step S1 is subjected to feature extraction, and the other way is global Stream G-Stream, and the original image is subjected to feature extraction;
s3, interacting the two networks obtained in the step S2 through cascade connection;
s4, connecting the two networks and combining the two characteristics by using a feature diagram obtained by the two networks at each stage as an input feature diagram of the next layer of convolution of BGg-stream through a space-channel attention mechanism module, namely an SCAB module;
and S5, after feature extraction, updating network parameters by adopting a triple loss function, then carrying out similarity calculation, and finally outputting a permutation sequence.
In this embodiment, the input of BGg-stream is BGg image, and the background gray-scale formula of the image is:
BGg(i,j)=0.299×R(i,j)+0.587×G(i,j)+0.114×B(i,j)
where BGg (i, j) is the pixel value of BGg image at row i and column j, and R, G, B is the three channels of the RGB image.
The SCAB module comprises a channel attention module and a space attention module.
In this embodiment, the complete structure of the channel attention module is shown in fig. 2. The channel attention module is realized as follows: given input feature mapping Fi∈RC*H*WFirstly, the spatial information of the feature map is aggregated by using average pooling operation to generate a spatial context descriptor Fi∈RC*1*1Compressing and converting the spatial information into a channel; then the activation size of the concealment is set to Fi∈RC/r*1*1To reduce the parameter overhead, wherein R is the reduction rate; thus, the channel attention module is represented as:
Fii=σ(θ(R(ζ(δ(Fi)))))
wherein FiThe method comprises the steps that the characteristics of BGg-Stream of each stage are represented, sigma is a sigmoid activation function, and R is a ReLU activation function; δ is the average pool number; θ and ζ are represented as two distinct tie layers;representing element-wise multiplication.
In this embodiment, the complete structure of the space attention module is shown in fig. 3. The space attention module is realized as follows: mapping F to input featuresi∈RC*H*WWhere C is the total number of channels and H × W is the size of the feature map, the spatial attention module is expressed as:
Fiii=σ(C(Fi))
wherein sigma is sigmoid activation function, and the output of the space attention module is Fiii∈R1*H*W。
In this embodiment, a local feature response maximization strategy, i.e. an LRM strategy, is applied in the testing phase; in the test process, the feature map is horizontally divided into n regions of an appropriate number, where n is 8 in this embodiment, and the feature with the largest response is extracted for each partial feature as the feature of the partial feature.
The embodiment also provides a pedestrian re-identification system based on background graying, which comprises a memory, a processor and a computer program stored on the memory and capable of running on the processor, wherein when the computer program is run by the processor, the steps of the method are realized.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
The foregoing is directed to preferred embodiments of the present invention, other and further embodiments of the invention may be devised without departing from the basic scope thereof, and the scope thereof is determined by the claims that follow. However, any simple modification, equivalent change and modification of the above embodiments according to the technical essence of the present invention are within the protection scope of the technical solution of the present invention.
Claims (7)
1. A pedestrian re-identification method based on background graying is characterized by comprising the following steps:
s1, carrying out background gray processing on the original image, and keeping the foreground information of the image unchanged to obtain a processed BGg image, namely a background gray image;
s2, based on a ResNet50 two-way network, wherein one way is background gray level Stream BGg-Stream, the background gray level image obtained in the step S1 is subjected to feature extraction, and the other way is global Stream G-Stream, and the original image is subjected to feature extraction;
s3, interacting the two networks obtained in the step S2 through cascade connection;
s4, connecting the two networks and combining the two characteristics by using a feature diagram obtained by the two networks at each stage as an input feature diagram of the next layer of convolution of BGg-stream through a space-channel attention mechanism module, namely an SCAB module;
and S5, after feature extraction, updating network parameters by adopting a triple loss function, then carrying out similarity calculation, and finally outputting a permutation sequence.
2. The pedestrian re-identification method based on background graying according to claim 1, wherein the input of BGg-stream is BGg image, and the background graying formula of the image is as follows:
BGg(i,j)=0.299×R(i,j)+0.587×G(i,j)+0.114×B(i,j)
where BGg (i, j) is the pixel value of BGg image at row i and column j, and R, G, B is the three channels of the RGB image.
3. The pedestrian re-identification method based on background graying according to claim 1, wherein the SCAB module comprises a channel attention module and a space attention module.
4. The pedestrian re-identification method based on background graying according to claim 3, wherein the channel attention module is implemented as follows: given input feature mapping Fi∈Rc*H*WFirstly, the spatial information of the feature map is aggregated by using average pooling operation to generate a spatial context descriptor Fi∈RC*1*1Compressing and converting the spatial information into a channel; then the activation size of the concealment is set to Fi∈RC/r*1*1To reduce the parameter overhead, wherein R is the reduction rate; thus, the channel attention module is represented as:
Fii=σ(θ(R(ζ(δ(Fi)))))
wherein FiThe method comprises the steps that the characteristics of BGg-Stream of each stage are represented, sigma is a sigmoid activation function, and R is a ReLU activation function; δ is the average pool number; θ and ζ are represented as two distinct tie layers;representing element-wise multiplication.
5. The pedestrian re-identification method based on background graying according to claim 3, wherein the spatial attention module is implemented as follows: mapping F to input featuresi∈RC*H*WWhere C is the total number of channels and H × W is the size of the feature map, the spatial attention module is expressed as:
Fiii=σ(C(Fi))
wherein sigma is sigmoid activation function, and the output of the space attention module is Fiii∈R1*H*W。
6. The pedestrian re-identification method based on background graying according to claim 1, wherein a local feature response maximization strategy (LRM strategy) is applied in a test phase; in the test process, the feature map is divided into n regions with proper number according to the level, and the feature with the maximum response is extracted from each part of features to be used as the feature of the part.
7. A pedestrian re-identification system based on background graying, characterized by comprising a memory, a processor and a computer program stored on the memory and capable of being run on the processor, which when run by the processor implements the method steps according to any one of claims 1 to 6.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110174952.5A CN112801020B (en) | 2021-02-09 | 2021-02-09 | Pedestrian re-identification method and system based on background graying |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110174952.5A CN112801020B (en) | 2021-02-09 | 2021-02-09 | Pedestrian re-identification method and system based on background graying |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112801020A true CN112801020A (en) | 2021-05-14 |
CN112801020B CN112801020B (en) | 2022-10-14 |
Family
ID=75814893
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110174952.5A Active CN112801020B (en) | 2021-02-09 | 2021-02-09 | Pedestrian re-identification method and system based on background graying |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112801020B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113627241A (en) * | 2021-06-29 | 2021-11-09 | 厦门市美亚柏科信息股份有限公司 | Background suppression method and system for re-identification of pedestrians |
CN114820687A (en) * | 2022-05-31 | 2022-07-29 | 厦门市美亚柏科信息股份有限公司 | Closed-loop end-to-end pedestrian re-identification background suppression method and system |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150304634A1 (en) * | 2011-08-04 | 2015-10-22 | John George Karvounis | Mapping and tracking system |
CN108960114A (en) * | 2018-06-27 | 2018-12-07 | 腾讯科技(深圳)有限公司 | Human body recognition method and device, computer readable storage medium and electronic equipment |
CN110232330A (en) * | 2019-05-23 | 2019-09-13 | 复钧智能科技(苏州)有限公司 | A kind of recognition methods again of the pedestrian based on video detection |
US20190313056A1 (en) * | 2018-04-10 | 2019-10-10 | Facebook, Inc. | Automated cinematic decisions based on descriptive models |
EP3570207A1 (en) * | 2018-05-15 | 2019-11-20 | IDEMIA Identity & Security Germany AG | Video cookies |
CN111291633A (en) * | 2020-01-17 | 2020-06-16 | 复旦大学 | Real-time pedestrian re-identification method and device |
CN111462183A (en) * | 2020-03-31 | 2020-07-28 | 山东大学 | Behavior identification method and system based on attention mechanism double-current network |
CN111860309A (en) * | 2020-07-20 | 2020-10-30 | 汪秀英 | Face recognition method and system |
-
2021
- 2021-02-09 CN CN202110174952.5A patent/CN112801020B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150304634A1 (en) * | 2011-08-04 | 2015-10-22 | John George Karvounis | Mapping and tracking system |
US20200404245A1 (en) * | 2011-08-04 | 2020-12-24 | Trx Systems, Inc. | Mapping and tracking system with features in three-dimensional space |
US20190313056A1 (en) * | 2018-04-10 | 2019-10-10 | Facebook, Inc. | Automated cinematic decisions based on descriptive models |
EP3570207A1 (en) * | 2018-05-15 | 2019-11-20 | IDEMIA Identity & Security Germany AG | Video cookies |
CN108960114A (en) * | 2018-06-27 | 2018-12-07 | 腾讯科技(深圳)有限公司 | Human body recognition method and device, computer readable storage medium and electronic equipment |
CN110232330A (en) * | 2019-05-23 | 2019-09-13 | 复钧智能科技(苏州)有限公司 | A kind of recognition methods again of the pedestrian based on video detection |
CN111291633A (en) * | 2020-01-17 | 2020-06-16 | 复旦大学 | Real-time pedestrian re-identification method and device |
CN111462183A (en) * | 2020-03-31 | 2020-07-28 | 山东大学 | Behavior identification method and system based on attention mechanism double-current network |
CN111860309A (en) * | 2020-07-20 | 2020-10-30 | 汪秀英 | Face recognition method and system |
Non-Patent Citations (2)
Title |
---|
李艳凤,张斌,孙嘉,陈后金,朱锦雷: "基于多池化融合与背景消除网络的跨数据集行人再识别方法", 《通信学报》 * |
马志艳等: "基于视觉的茶作物行间行走路径规划研究", 《农机化研究》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113627241A (en) * | 2021-06-29 | 2021-11-09 | 厦门市美亚柏科信息股份有限公司 | Background suppression method and system for re-identification of pedestrians |
CN114820687A (en) * | 2022-05-31 | 2022-07-29 | 厦门市美亚柏科信息股份有限公司 | Closed-loop end-to-end pedestrian re-identification background suppression method and system |
Also Published As
Publication number | Publication date |
---|---|
CN112801020B (en) | 2022-10-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10719940B2 (en) | Target tracking method and device oriented to airborne-based monitoring scenarios | |
CN109558832B (en) | Human body posture detection method, device, equipment and storage medium | |
CN104268594B (en) | A kind of video accident detection method and device | |
CN110334762B (en) | Feature matching method based on quad tree combined with ORB and SIFT | |
US8620026B2 (en) | Video-based detection of multiple object types under varying poses | |
CN108108751B (en) | Scene recognition method based on convolution multi-feature and deep random forest | |
CN111696110B (en) | Scene segmentation method and system | |
CN108256562A (en) | Well-marked target detection method and system based on Weakly supervised space-time cascade neural network | |
CN109886159B (en) | Face detection method under non-limited condition | |
CN112801020B (en) | Pedestrian re-identification method and system based on background graying | |
CN111738344A (en) | Rapid target detection method based on multi-scale fusion | |
CN109685045A (en) | A kind of Moving Targets Based on Video Streams tracking and system | |
CN111723687A (en) | Human body action recognition method and device based on neural network | |
CN111914698A (en) | Method and system for segmenting human body in image, electronic device and storage medium | |
CN113066089B (en) | Real-time image semantic segmentation method based on attention guide mechanism | |
CN110852327A (en) | Image processing method, image processing device, electronic equipment and storage medium | |
CN115082966B (en) | Pedestrian re-recognition model training method, pedestrian re-recognition method, device and equipment | |
CN114266894A (en) | Image segmentation method and device, electronic equipment and storage medium | |
CN109523570A (en) | Beginning parameter transform model method and device | |
CN111951298B (en) | Target tracking method integrating time sequence information | |
CN117710841A (en) | Small target detection method and device for aerial image of unmanned aerial vehicle | |
CN117437691A (en) | Real-time multi-person abnormal behavior identification method and system based on lightweight network | |
Özyurt et al. | A new method for classification of images using convolutional neural network based on Dwt-Svd perceptual hash function | |
CN113628206B (en) | License plate detection method, device and medium | |
CN114820755A (en) | Depth map estimation method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |