CN107766890B - Improved method for discriminant graph block learning in fine-grained identification - Google Patents

Improved method for discriminant graph block learning in fine-grained identification Download PDF

Info

Publication number
CN107766890B
CN107766890B CN201711040828.XA CN201711040828A CN107766890B CN 107766890 B CN107766890 B CN 107766890B CN 201711040828 A CN201711040828 A CN 201711040828A CN 107766890 B CN107766890 B CN 107766890B
Authority
CN
China
Prior art keywords
image
discriminant
vector
multiplied
detector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711040828.XA
Other languages
Chinese (zh)
Other versions
CN107766890A (en
Inventor
冀中
赵可心
张锁平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201711040828.XA priority Critical patent/CN107766890B/en
Publication of CN107766890A publication Critical patent/CN107766890A/en
Application granted granted Critical
Publication of CN107766890B publication Critical patent/CN107766890B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

An improved method for discriminant graph block learning in fine-grained recognition comprises the following steps: extracting image blocks with distinguishing properties in an original image, comprising the following steps: obtaining a feature map from the original image through a convolution pooling layer in a convolution neural network, and regarding a vector of each space fixed position in the feature map as a detector corresponding to a corresponding position image block in the original image; supposing that a detector with the highest response in a discriminant area in an original image is learned, carrying out convolution operation on the detector and a feature map to obtain a new response map; selecting the position with the maximum value in the new response image to obtain an image block with distinguishing property; learning features of tiles having discriminative properties and for classification, comprising: obtaining a local saliency map according to an image block with distinguishing properties; the local saliency map is encoded using a spatially weighted fisher vector. The method learns the discriminant characteristics more suitable for a fine-grained identification task, and reduces the interference of background information in a discriminant image block so as to improve the classification precision.

Description

Improved method for discriminant graph block learning in fine-grained identification
Technical Field
The invention relates to discriminative graph block learning in fine-grained identification. In particular, to an improved method for discriminant tile learning in fine-grained identification by spatially weighting an image descriptor to obtain spatially weighted fisher vectors according to a response map.
Background
In recent years, fine-grained recognition attracts more and more attention in the field of target recognition, which is to recognize subclasses of a certain large class of targets, such as flowers, birds, dogs, automobiles and the like, which generally have the same structure, so that how to learn features with distinguishing properties in images becomes a main task of fine-grained recognition.
In past research, the field of fine-grained identification mainly includes two tasks: local localization and characterization. In addition to image category labels, fine-grained datasets usually provide additional labeling of target bounding boxes and local parts, and many previous tasks more or less rely on these additional labeling, but fine-grained classification usually requires expert-level knowledge, and ordinary people have difficulty in completing the task, which makes manual labeling expensive. In recent years, more research is focused on methods without any additional labeled information, the method related to the invention only needs image category labels, and does not need local labeling, and local features with discriminability are learned by a weak supervision method.
For the feature description of images, CNN features have made a breakthrough on many benchmarks. The traditional method is to encode local information and then fuse the local information into an integral feature representation, the CNN features are different from the CNN features, the global learning can be directly realized, a feature extractor does not need to be designed manually, and the current fine-grained identification method is based on the CNN and is used for learning the fine and unique features of the image through other algorithms.
Disclosure of Invention
The invention aims to solve the technical problem of providing an improved method for learning discriminant image blocks in fine-grained identification, which can accurately learn detailed features and abandon disordered background information in small image blocks, thereby improving classification precision and avoiding the need of global feature assistance.
The technical scheme adopted by the invention is as follows: an improved method for learning discriminant graph blocks in fine-grained recognition comprises the following steps:
1) extracting image blocks with distinguishing properties in an original image, comprising the following steps:
(1) obtaining a characteristic image of C multiplied by H multiplied by W size from an original image through a convolution pooling layer in a convolution neural network, wherein C is the number of pipelines, H is the height, W is the width, and a C multiplied by 1 vector of each space fixed position in the characteristic image is regarded as a detector corresponding to a corresponding position image block in the original image;
(2) supposing that a detector with the highest response in a discriminant area in an original image is learned, carrying out convolution operation on the detector with the size of C multiplied by 1 and a feature map with the size of C multiplied by H multiplied by W to obtain a new response map with the size of H multiplied by W;
(3) selecting the position with the maximum value in the new H multiplied by W response image to obtain a 1 multiplied by 1 image block with distinguishing property;
2) learning features of tiles having discriminative properties and for classification, comprising:
(1) obtaining a local saliency map according to an image block with distinguishing properties;
(2) the local saliency map is encoded using a spatially weighted fisher vector.
Step 1) step (2) the step of learning the detector with the highest response in a discriminant region in the original image is as follows:
(1) assuming that the number of discriminative pattern block detectors of each type of image is n, the images share M types, and the number of required detectors is nM;
(2) respectively performing convolution operation on nM Cx 1 x 1 detectors and Cx H x W feature maps to obtain new feature maps, and performing global maximum pooling on the new feature maps to obtain a nM-dimension feature vector;
(3) averaging each class of eigenvectors in the nM dimensional eigenvectors to obtain an M dimensional vector;
(4) introducing the M-dimensional vector into a Softmax loss function, training a Cx 1 x 1 detector by using a back propagation algorithm, and obtaining a detector with the highest response in a discriminant region in an original image after the training is finished;
step 2) the step (1) of obtaining the local saliency map is as follows: a local saliency map Q is computed from a saliency map S derived from the original image, as follows:
Figure BDA0001452217710000021
where p is the pixel of the discriminative tile, i is the detection site, and when the ith detection site contains pixel p, then Di(p) 1, otherwise Di(p) 0, s (p) the saliency map of the whole image, q (p) the local saliency map, Z a normalized constant such that maxq (p) 1.
Step 2) the step (2) comprises the following steps:
suppose vector I is (z)1,…,zN) Is a series of D-dimension characteristic vectors extracted from an image, and the Fisher vector of a picture I encodes phi (I) to (u)1,v1,…,uk,vk) Is the mean square error ukSum covariance vkAccumulation of (u)k,vkWritten as follows:
Figure BDA0001452217710000022
Figure BDA0001452217710000023
where j 1, D represents the vector dimension, and o (μ) is definedkkkK1, …, K) are parameters of a gaussian mixture model, qikIs each vector z of mode k in the hybrid modeliWherein i is 1, …, N.
For each vector ziIntroducing a spatial weighting term Q (p)i),ujkAnd vjkThe weighted result is expressed as:
Figure BDA0001452217710000024
wherein, Q (p)i) Is a local saliency map, uijk、vijkRespectively, formula (2) (3), important features can be learned by introducing spatial weights.
The improved method for learning the discriminant image block in the fine-grained recognition combines the CNN characteristic and the Fisher vector to learn the discriminant characteristic more suitable for a fine-grained recognition task, and reduces the interference of background information in the discriminant image block so as to improve the classification precision. The method mainly aims at the improvement of the existing discriminant graphic block learning, and for the discriminant region detected by the learned graphic block detector, except the discriminant feature of the target object, redundant background information often exists, so that the method utilizes the local saliency map and the Fisher vector coding to improve the limit, and fully utilizes the characteristics of the discriminant region to be effectively used for the classification task. The advantages are mainly reflected in that:
1) the novelty is as follows: the most efficient and popular feature representation method at present is CNN, but the present invention combines fisher vector coding with CNN features for specific problems. Because the most important thing in the fine-grained identification problem is the learning of the discriminant features, and the backgrounds in the data sets are usually very similar, the method introduces the Fisher vector in a novel way, and can effectively reduce the interference of the background information in the discriminant image blocks.
2) Effectiveness: compared with the original method, the local salient map and Fisher vector coding-based method designed by the invention can effectively learn local features, the traditional CNN usually needs a rectangle with a fixed size as input, which contains invalid information of the background, and the method can effectively reduce the interference of background noise, so that the learned local features are more discriminative, and the classification precision is improved.
3) The practicability is as follows: the method is simple and feasible, is an end-to-end network, and can be effectively used for fine-grained identification.
Drawings
FIG. 1 is a flow chart of an improved method of discriminative tile learning in fine-grained recognition of the present invention.
Detailed Description
1. An improved method for learning discriminant graph blocks in fine-grained recognition is characterized by comprising the following steps:
1) extracting image blocks with distinguishing properties in an original image, comprising the following steps:
(1) obtaining a characteristic image of C multiplied by H multiplied by W size from an original image through a convolution pooling layer in a convolution neural network, wherein C is the number of pipelines, H is the height, W is the width, and a C multiplied by 1 vector of each space fixed position in the characteristic image is regarded as a detector corresponding to a corresponding position image block in the original image;
(2) supposing that a detector with the highest response in a discriminant area in an original image is learned, carrying out convolution operation on the detector with the size of C multiplied by 1 and a feature map with the size of C multiplied by H multiplied by W to obtain a new response map with the size of H multiplied by W;
(3) selecting the position with the maximum value in the new H multiplied by W response image to obtain a 1 multiplied by 1 image block with distinguishing property;
2) learning features of tiles having discriminative properties and for classification, comprising:
(1) obtaining a local saliency map according to an image block with distinguishing properties;
(2) encoding the local saliency map using spatially weighted fisher vectors;
2. the improved method for discriminative patch learning in fine granularity recognition as claimed in claim 1, wherein the step 1) step (2) of learning the detector with the highest response in a discriminative region in the original image comprises the steps of:
(1) assuming that the number of discriminative pattern block detectors of each type of image is n, the images share M types, and the number of required detectors is nM;
(2) respectively performing convolution operation on nM Cx 1 x 1 detectors and Cx H x W feature maps to obtain new feature maps, and performing global maximum pooling on the new feature maps to obtain a nM-dimension feature vector;
(3) averaging each class of eigenvectors in the nM dimensional eigenvectors to obtain an M dimensional vector;
(4) introducing the M-dimensional vector into a Softmax loss function, training a Cx 1 x 1 detector by using a back propagation algorithm, and obtaining a detector with the highest response in a discriminant region in an original image after the training is finished;
3. the improved method for discriminative graph block learning in fine grain identification as claimed in claim 1, wherein the step 2) the step (1) is to obtain the local saliency map as follows: a local saliency map Q is computed from a saliency map S derived from the original image, as follows:
Figure BDA0001452217710000041
where p is the pixel of the discriminative tile, i is the detection site, and when the ith detection site contains pixel p, then Di(p) 1, otherwise Di(p) 0, s (p) the saliency map of the whole image, q (p) the local saliency map, Z a normalized constant such that maxq (p) 1.
4. The improved method for discriminative tile learning in fine granularity recognition as claimed in claim 1, wherein the step 2) and the step (2) comprise:
suppose vector I is (z)1,…,zN) Is a series of D-dimension characteristic vectors extracted from an image, and the Fisher vector of a picture I encodes phi (I) to (u)1,v1,…,uk,vk) Is the mean square error ukSum covariance vkAccumulation of (u)k,vkWritten as follows:
Figure BDA0001452217710000042
Figure BDA0001452217710000043
where j 1, D represents the vector dimension, and o (μ) is definedkkkK1, …, K) are parameters of a gaussian mixture model, qikIs each vector z of mode k in the hybrid modeliWherein i is 1, …, N.
For each vector ziIntroducing a spatial weighting term Q (p)i),ujkAnd vjkThe weighted result is expressed as:
Figure BDA0001452217710000044
wherein, Q (p)i) Is a local saliency map, uijk、vijkRespectively, formula (2) (3), important features can be learned by introducing spatial weights.
Specific examples are given below in connection with fig. 1:
FIG. 1 depicts a flow diagram of the architecture of the present invention. The structure of the invention mainly comprises three parts, as shown in figure 1. The method is based on a VGG-16 model, and the model has 16 layers. The implementation process is divided into two stages: a training phase and a testing phase.
In the training stage, the parameters of the detector are mainly learned, and the process is shown as (i) and (ii) in fig. 1.
(1) First, the input image passes through a pre-trained convolutional neural network VGG-16, conv4-3 outputs a feature map with the size of 512 × 28 × 28, and therefore, the size of each detector is 512 × 1 × 1. Setting the number of detectors of each class to 10, there are 2000 detectors for the CUB200-2011 dataset;
(2) convolving each detector with the obtained 512 × 28 × 28 feature map to obtain a response map with the size of 28 × 28;
(3) after the response map is subjected to global maximum pooling, a 2000-dimensional feature vector is obtained,
(4) the feature vectors of each class in the 2000-dimensional vectors are averaged to obtain a 200-dimensional vector, the vector after the average value pooling is transmitted into a Softmax loss function to be trained through a back propagation algorithm, and then a detector capable of extracting discriminant image blocks for each class can be obtained.
And (3) a testing stage, as shown in (r) and (c) of FIG. 1. For the trained detector, the steps (1) to (3) in the training stage are repeated, so that a response graph with the size of 1 × 1 can be obtained, and the part with the distinguishing property of each picture can be identified. And then calculating a local saliency map of the image, wherein the local saliency map is obtained by two parts of the local map and the image saliency map, and a discriminant image block extracted from the original image is multiplied by the global saliency map, so as to obtain the local saliency map by the formula (1). The local saliency map is used to indicate the possibility that the pixel belongs to the foreground, which can effectively reduce the interference of the background. Designing weights for the Fisher vectors of the images according to the local saliency maps to obtain space-weighted Fisher vectors, and learning important features in a fine-grained identification task by introducing the weights so as to finally realize classification of the fine-grained images.

Claims (1)

1. An improved method for learning discriminant graph blocks in fine-grained recognition is characterized by comprising the following steps:
1) extracting image blocks with distinguishing properties in an original image, comprising the following steps:
(1) obtaining a characteristic diagram of C multiplied by H multiplied by W size from an original image through a convolution pooling layer in a convolution neural network, wherein C is the number of pipelines, H is the height, and W is the width;
(2) supposing that a detector with the highest response in a discriminant area in an original image is learned, carrying out convolution operation on the detector with the size of C multiplied by 1 and a feature map with the size of C multiplied by H multiplied by W to obtain a new response map with the size of H multiplied by W; the step of learning the detector with the highest response in a discriminant region in the original image is as follows:
(2.1) assuming that the number of discriminant pattern block detectors of each type of image is n, the images share M types, and the number of required detectors is nM;
(2.2) performing convolution operation on nM Cx 1 x 1 detectors and Cx H x W feature maps to obtain new feature maps, and performing global maximum pooling on the new feature maps to obtain an nM-dimension feature vector;
(2.3) averaging each type of feature vector in the feature vectors of nM dimension to obtain a vector of M dimension;
(2.4) transmitting the M-dimensional vector into a Softmax loss function, training a Cx 1 x 1 detector by using a back propagation algorithm, and obtaining a detector with the highest response in a discriminant area in an original image after the training is finished;
(3) selecting the position with the maximum value in the new H multiplied by W response image to obtain a 1 multiplied by 1 image block with distinguishing property;
2) learning features of tiles having discriminative properties and for classification, comprising:
(1) obtaining a local saliency map according to an image block with distinguishing properties; the local saliency map is obtained by: a local saliency map Q is computed from a saliency map S derived from the original image, as follows:
Figure FDA0003096889330000011
where p is the pixel of the discriminant block, m represents the mth detection position, m is 1, …, Α, and when the mth detection position contains pixel p, then D ism(p) 1, otherwise Dm(p) 0, s (p) is the saliency map of the whole image, q (p) is the local saliency map, Z is a normalized constant such that maxq (p) 1;
(2) encoding the local saliency map using spatially weighted fisher vectors; the method comprises the following steps:
suppose vector I is (z)1,…,zN) Is a series of D-dimension characteristic vectors extracted from an image, and the Fisher vector of a picture I encodes phi (I) to (u)1,v1,…,uk,vk) Is the mean square error ukSum covariance vkAccumulation of (u)k,vkWritten as follows:
Figure FDA0003096889330000012
Figure FDA0003096889330000013
where j 1, D represents the vector dimension, and o (μ) is definedkkkK1, …, K) are parameters of a gaussian mixture model, qikIs each vector z of mode k in the hybrid modeliWherein i ═ 1, …, N;
for each vector ziIntroducing a spatial weighting term Q (p)i),ujkAnd vjkThe weighted result is expressed as:
Figure FDA0003096889330000021
wherein, Q (p)i) Is a local saliency map, uijk、vijkRespectively, formula (2) (3), important features can be learned by introducing spatial weights.
CN201711040828.XA 2017-10-31 2017-10-31 Improved method for discriminant graph block learning in fine-grained identification Active CN107766890B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711040828.XA CN107766890B (en) 2017-10-31 2017-10-31 Improved method for discriminant graph block learning in fine-grained identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711040828.XA CN107766890B (en) 2017-10-31 2017-10-31 Improved method for discriminant graph block learning in fine-grained identification

Publications (2)

Publication Number Publication Date
CN107766890A CN107766890A (en) 2018-03-06
CN107766890B true CN107766890B (en) 2021-09-14

Family

ID=61271840

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711040828.XA Active CN107766890B (en) 2017-10-31 2017-10-31 Improved method for discriminant graph block learning in fine-grained identification

Country Status (1)

Country Link
CN (1) CN107766890B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102102161B1 (en) * 2018-05-18 2020-04-20 오드컨셉 주식회사 Method, apparatus and computer program for extracting representative feature of object in image
CN110019915B (en) * 2018-07-25 2022-04-12 北京京东尚科信息技术有限公司 Method and device for detecting picture and computer readable storage medium
CN109409384A (en) * 2018-09-30 2019-03-01 内蒙古科技大学 Image-recognizing method, device, medium and equipment based on fine granularity image
CN109815973A (en) * 2018-12-07 2019-05-28 天津大学 A kind of deep learning method suitable for the identification of fish fine granularity
CN109948628B (en) * 2019-03-15 2023-01-03 中山大学 Target detection method based on discriminant region mining
CN110197202A (en) * 2019-04-30 2019-09-03 杰创智能科技股份有限公司 A kind of local feature fine granularity algorithm of target detection
CN110309858B (en) * 2019-06-05 2022-07-01 大连理工大学 Fine-grained image classification method based on discriminant learning
CN110363233B (en) * 2019-06-28 2021-05-28 西安交通大学 Fine-grained image recognition method and system of convolutional neural network based on block detector and feature fusion
CN110796183A (en) * 2019-10-17 2020-02-14 大连理工大学 Weak supervision fine-grained image classification algorithm based on relevance-guided discriminant learning
CN111062438B (en) * 2019-12-17 2023-06-16 大连理工大学 Image propagation weak supervision fine granularity image classification algorithm based on correlation learning
CN112927221B (en) * 2020-12-09 2022-03-29 广州市玄武无线科技股份有限公司 Image fine-grained feature-based reproduction detection method and system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104573744A (en) * 2015-01-19 2015-04-29 上海交通大学 Fine granularity classification recognition method and object part location and feature extraction method thereof
WO2016168235A1 (en) * 2015-04-17 2016-10-20 Nec Laboratories America, Inc. Fine-grained image classification by exploring bipartite-graph labels
CN106778705A (en) * 2017-02-04 2017-05-31 中国科学院自动化研究所 A kind of pedestrian's individuality dividing method and device
CN106778804A (en) * 2016-11-18 2017-05-31 天津大学 The zero sample image sorting technique based on category attribute transfer learning
CN106778807A (en) * 2016-11-22 2017-05-31 天津大学 The fine granularity image classification method of dictionary pair is relied on based on public dictionary pair and class
CN106951872A (en) * 2017-03-24 2017-07-14 江苏大学 A kind of recognition methods again of the pedestrian based on unsupervised depth model and hierarchy attributes

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104573744A (en) * 2015-01-19 2015-04-29 上海交通大学 Fine granularity classification recognition method and object part location and feature extraction method thereof
WO2016168235A1 (en) * 2015-04-17 2016-10-20 Nec Laboratories America, Inc. Fine-grained image classification by exploring bipartite-graph labels
CN106778804A (en) * 2016-11-18 2017-05-31 天津大学 The zero sample image sorting technique based on category attribute transfer learning
CN106778807A (en) * 2016-11-22 2017-05-31 天津大学 The fine granularity image classification method of dictionary pair is relied on based on public dictionary pair and class
CN106778705A (en) * 2017-02-04 2017-05-31 中国科学院自动化研究所 A kind of pedestrian's individuality dividing method and device
CN106951872A (en) * 2017-03-24 2017-07-14 江苏大学 A kind of recognition methods again of the pedestrian based on unsupervised depth model and hierarchy attributes

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Fully Convolutional Attention Localization Networks: Efficient Attention Localization for Fine-Grained Recognition;Xiao Liu 等;《arXiv》;20160404;第1-10页 *
Picking Deep Filter Responses for Fine-grained Image Recognition;Xiaopeng Zhang 等;《2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)》;20161212;论文第1-5节 *
自上而下注意图分割的细粒度图像分类;冯语姗 等;《中国图象图形学报》;20160930;第1147-1154页 *

Also Published As

Publication number Publication date
CN107766890A (en) 2018-03-06

Similar Documents

Publication Publication Date Title
CN107766890B (en) Improved method for discriminant graph block learning in fine-grained identification
CN110443143B (en) Multi-branch convolutional neural network fused remote sensing image scene classification method
Cao et al. Landmark recognition with sparse representation classification and extreme learning machine
CN104915636B (en) Remote sensing image road recognition methods based on multistage frame significant characteristics
CN104820718B (en) Image classification and search method based on geographic location feature Yu overall Vision feature
CN110796168A (en) Improved YOLOv 3-based vehicle detection method
CN110717411A (en) Pedestrian re-identification method based on deep layer feature fusion
CN111583263A (en) Point cloud segmentation method based on joint dynamic graph convolution
CN111079847B (en) Remote sensing image automatic labeling method based on deep learning
CN103080979B (en) From the system and method for photo synthesis portrait sketch
US20230118864A1 (en) Lifted semantic graph embedding for omnidirectional place recognition
CN105184298A (en) Image classification method through fast and locality-constrained low-rank coding process
CN105574548A (en) Hyperspectral data dimensionality-reduction method based on sparse and low-rank representation graph
CN104616319B (en) Multiple features selection method for tracking target based on support vector machines
CN103927511A (en) Image identification method based on difference feature description
US20210319242A1 (en) Dense and Discriminative Neural Network Architectures for Improved Object Detection and Instance Segmentation
CN106874862A (en) People counting method based on submodule technology and semi-supervised learning
CN111860297A (en) SLAM loop detection method applied to indoor fixed space
CN115690152A (en) Target tracking method based on attention mechanism
CN112232297A (en) Remote sensing image scene classification method based on depth joint convolution activation
CN115565080A (en) Multi-feature fusion SAR image sea ice classification method based on self-attention mechanism
CN116935196A (en) Novel brain inspired target detection method for blocked target in complex environment
Schenkel et al. Domain adaptation for semantic segmentation using convolutional neural networks
CN105718858A (en) Pedestrian recognition method based on positive-negative generalized max-pooling
Rahimi et al. Uav sensor fusion with latent-dynamic conditional random fields in coronal plane estimation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant