CN112132004A - A fine-grained image recognition method based on multi-view feature fusion - Google Patents

A fine-grained image recognition method based on multi-view feature fusion Download PDF

Info

Publication number
CN112132004A
CN112132004A CN202010992253.7A CN202010992253A CN112132004A CN 112132004 A CN112132004 A CN 112132004A CN 202010992253 A CN202010992253 A CN 202010992253A CN 112132004 A CN112132004 A CN 112132004A
Authority
CN
China
Prior art keywords
feature
loss function
bilinear
fine
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010992253.7A
Other languages
Chinese (zh)
Other versions
CN112132004B (en
Inventor
黄伟锋
张甜
常东良
马占宇
柳斐
王丹
刘念
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South To North Water Transfer Middle Route Information Technology Co ltd
Beijing University of Posts and Telecommunications
Original Assignee
South To North Water Transfer Middle Route Information Technology Co ltd
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South To North Water Transfer Middle Route Information Technology Co ltd, Beijing University of Posts and Telecommunications filed Critical South To North Water Transfer Middle Route Information Technology Co ltd
Priority to CN202010992253.7A priority Critical patent/CN112132004B/en
Publication of CN112132004A publication Critical patent/CN112132004A/en
Application granted granted Critical
Publication of CN112132004B publication Critical patent/CN112132004B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Multimedia (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

一种基于多视角特征融合的细粒度图像识别方法,涉及图像处理技术领域,解决现有细粒度图像识别方法忽略了图像的细节信息以及存在图像之间视觉差异的适应能力较差,且引入的损失函数复杂,增加了模型的参数量等问题,包括双线性特征提取步骤,抑制分支学习步骤,同类对比模块学习步骤,中心损失计算步骤以及模型优化损失函数计算步骤;本发明引入抑制分支,通过抑制图像中最显著的区域,迫使网络寻找易混淆类别间微妙的判别性特征。引入同类对比学习模块,将同类样本的特征向量进行融合,增加同一类别下不同图像的交互信息。还引入中心损失函数,最小化特征与对应类中心之间的距离,使学到的特征更具判别性。提升了细粒度图像识别的准确率。

Figure 202010992253

A fine-grained image recognition method based on multi-view feature fusion, relates to the technical field of image processing, and solves the problem that existing fine-grained image recognition methods ignore the detailed information of images and have poor adaptability to visual differences between images, and the introduced The loss function is complex, which increases the number of parameters of the model, including the bilinear feature extraction step, the suppression branch learning step, the similar comparison module learning step, the center loss calculation step and the model optimization loss function calculation step; the invention introduces the suppression branch, By suppressing the most salient regions in the image, the network is forced to look for subtle discriminative features between confusing classes. A similar comparison learning module is introduced to fuse the feature vectors of similar samples to increase the interactive information of different images under the same category. A center loss function is also introduced to minimize the distance between features and corresponding class centers, making the learned features more discriminative. Improves the accuracy of fine-grained image recognition.

Figure 202010992253

Description

一种基于多视角特征融合的细粒度图像识别方法A fine-grained image recognition method based on multi-view feature fusion

技术领域technical field

本发明涉及图像处理技术领域,具体涉及一种基于多视角特征融合的细粒度图像识别方法。The invention relates to the technical field of image processing, in particular to a fine-grained image recognition method based on multi-view feature fusion.

背景技术Background technique

细粒度图像分类是在区分出基本类别的基础上,进行更精细的子类划分,例如鸟类或犬类等。因此该问题需要捕获细微的类间差异,充分挖掘图像的判别性特征。Fine-grained image classification is based on distinguishing basic categories, and then performs more fine-grained sub-categories, such as birds or dogs. Therefore, this problem needs to capture subtle inter-class differences and fully exploit the discriminative features of images.

细粒度物体在现实生活中广泛存在,与之相对应的细粒度图像识别是计算机视觉识别中的一个重要研究课题。当前细粒度图像识别主要存在以下三个方面的挑战:(1)由于姿势、背景和拍摄角度的不同,同一类别外表看来可能具有较大的差异。(2)不同类别由于属于同一父类,它们之间的差异仅存在于一些细微的区域中,例如鸟的喙和尾巴等。(3)对细粒度图像的收集、标注费时费力。如图1所示。Fine-grained objects exist widely in real life, and the corresponding fine-grained image recognition is an important research topic in computer vision recognition. The current fine-grained image recognition mainly has the following three challenges: (1) Due to the different poses, backgrounds and shooting angles, the appearance of the same category may appear to be quite different. (2) Since different categories belong to the same parent category, the differences between them only exist in some subtle areas, such as the beak and tail of birds. (3) The collection and labeling of fine-grained images is time-consuming and labor-intensive. As shown in Figure 1.

现有的方法主要通过以下三个方面达到识别的目的:(1)基于定位-分类网络进行细粒度图像识别。(2)通过开发用于细粒度识别的强大深度模型来直接学习更具判别力的表征。(3)结合图像的全局特征与局部特征实现图像的细粒度分类。The existing methods mainly achieve the purpose of recognition through the following three aspects: (1) Fine-grained image recognition based on localization-classification network. (2) Directly learn more discriminative representations by developing powerful deep models for fine-grained recognition. (3) Combine the global features and local features of images to achieve fine-grained classification of images.

现有技术1,双线性池化细粒度图像分类(Bilinear pooling),通过预训练的孪生卷积神经网络(convolutional neural networks)提取特征,并在特征的各个通道层面进行双线性池化,得到特征的高阶表示,从而增强特征的判别能力。该方法得益于通过一种新的池化方式,实现了细粒度图像识别准确率的提升。Prior art 1, bilinear pooling fine-grained image classification (Bilinear pooling), extracts features through pre-trained twin convolutional neural networks (convolutional neural networks), and performs bilinear pooling at each channel level of the features, A high-order representation of the feature is obtained, thereby enhancing the discriminative ability of the feature. This method benefits from a new pooling method, which improves the accuracy of fine-grained image recognition.

该方法提出了一种新的双线性池化方式,但在细粒度图像类别间关系、模型参数量、细节区域的数量等方面没有针对细粒度图像识别进行有效的设计。即没有考虑到细粒度图像蕴含多种细节信息,类间差异小,类内差异大等因素的影响。This method proposes a new bilinear pooling method, but there is no effective design for fine-grained image recognition in terms of the relationship between fine-grained image categories, the amount of model parameters, and the number of detail regions. That is, it does not take into account the influence of factors such as fine-grained images containing a variety of detailed information, small differences between classes, and large differences within classes.

现有技术2,多注意多分类约束网络(Multi-Attention Multi-ClassConstraint),该方法通过一次压缩-多次扩展(one-squeeze multi-excitation)模块提取输入图像的多个注意(attention)区域,然后引入度量学习(metric learning),采用triplet loss和softmax loss来训练网络,将同类特征的相同注意拉近,将不同注意或不同类的特征推开,从而加强了部件之间的关系,实现了细粒度图像识别准确率的提升。Prior art 2, Multi-Attention Multi-Class Constraint, this method extracts multiple attention regions of the input image through a one-squeeze multi-excitation module, Then metric learning is introduced, triplet loss and softmax loss are used to train the network, the same attention of the same features is drawn closer, and the features of different attention or different classes are pushed apart, thereby strengthening the relationship between components and realizing Improved accuracy of fine-grained image recognition.

该方法主要利用度量学习来改善特征空间中的样本分布,因此,它对于挖掘一对图像之间视觉差异的适应能力较差。而且引入的损失函数较为复杂,需要构建大量样本对,极大地增加了模型的参数量。This method mainly utilizes metric learning to improve the sample distribution in the feature space, so it is less adaptable for mining visual differences between a pair of images. Moreover, the introduced loss function is relatively complex, and a large number of sample pairs need to be constructed, which greatly increases the number of parameters of the model.

发明内容SUMMARY OF THE INVENTION

本发明为解决现有细粒度图像识别方法忽略了图像的细节信息以及存在图像之间视觉差异的适应能力较差,且引入的损失函数复杂,增加了模型的参数量等问题,提供一种基于多视角特征融合的细粒度图像识别方法。In order to solve the problems that the existing fine-grained image recognition method ignores the detailed information of the image and the visual difference between the images has poor adaptability, the introduced loss function is complex, and the parameter quantity of the model is increased, the present invention provides a method based on A fine-grained image recognition method based on multi-view feature fusion.

一种基于多视角特征融合的细粒度图像识别方法,该方法由以下步骤实现:A fine-grained image recognition method based on multi-view feature fusion, the method is realized by the following steps:

步骤一、双线性特征提取;Step 1. Bilinear feature extraction;

将原始图像输入至双线性特征提取网络,将不同卷积层输出的特征图进行融合,获得双线性特征向量;所述特征提取网络采用在数据集ImageNet预训练的网络结构;The original image is input into the bilinear feature extraction network, and the feature maps output by different convolutional layers are fused to obtain bilinear feature vectors; the feature extraction network adopts the network structure pre-trained in the dataset ImageNet;

步骤二、抑制分支学习,具体过程为:Step 2: Suppress branch learning, the specific process is:

步骤二一、根据步骤一所述特征提取网络不同卷积层输出的特征图的大小及阈值生成注意图;Step 21: Generate an attention map according to the size and threshold of the feature maps output by different convolutional layers of the feature extraction network described in step 1;

步骤二二、根据步骤二一所述的注意图生成抑制掩码,将所述抑制掩码覆盖到所述原始图像上生成局部区域被掩蔽的抑制图像;Step 22, generating a suppression mask according to the attention map described in step 21, and overlaying the suppression mask on the original image to generate a suppressed image in which the local area is masked;

步骤二三、将步骤二所述的抑制图像采用步骤一进行双线性特征提取,获得双线性特征向量,将所述双线性特征向量输入至全连接层,获得预测的类别概率值,并对该预测的类别概率值计算多分类交叉熵;Step 2 and 3: Use step 1 to perform bilinear feature extraction on the suppressed image described in step 2, obtain bilinear feature vector, input the bilinear feature vector to the fully connected layer, and obtain the predicted class probability value, And calculate the multi-class cross entropy for the predicted class probability value;

步骤三、同类对比模块学习;Step 3. Similar comparison module learning;

步骤三一、随机选择与原始图像同一类别下其他N张图像作为正样本图像;Step 31: Randomly select other N images in the same category as the original image as positive sample images;

步骤三二、将目标图像与步骤三一所述的正样本图像送入步骤一所述的特征提取网络进行双线性特征向量融合,获得融合特征综合了同类别下多张图像的双线性特征向量;Step 32: The target image and the positive sample image described in step 31 are sent to the feature extraction network described in step 1 for bilinear feature vector fusion, and the fusion feature is obtained. Feature vector;

步骤三三、将步骤三二获得的同类别下多张图像的双线性特征向量求平均,获得一个融合的特征向量,将融合后的特征向量输入全连接层,获得预测的概率,对得到的同类别的预测概率计算多分类交叉熵;Step 33: Average the bilinear eigenvectors of multiple images of the same category obtained in step 32 to obtain a fused feature vector, and input the fused feature vector into the fully connected layer to obtain the predicted probability. Calculate the multi-class cross entropy of the predicted probability of the same category;

步骤四、中心损失函数LC计算;Step 4: Calculate the central loss function LC ;

令vi为第i个样本的双线性特征,ci为样本i对应类别所有样本的平均特征,即类中心,N为当前批量的样本个数,则中心损失函数LC的公式如下:Let v i be the bilinear feature of the ith sample, c i be the average feature of all samples in the corresponding category of sample i, that is, the class center, and N be the number of samples in the current batch, then the formula of the center loss function L C is as follows:

Figure BDA0002691299670000021
Figure BDA0002691299670000021

步骤五、模型优化损失函数计算;Step 5. Model optimization loss function calculation;

将原始图像的双线性特征向量的交叉熵损失函数,抑制图像的双线性特征向量的交叉熵损失函数,融合特征的交叉熵损失函数以及中心损失函数加权求和,获得模型优化的损失函数。The cross-entropy loss function of the bilinear feature vector of the original image, the cross-entropy loss function of the bilinear feature vector of the suppressed image, the cross-entropy loss function of the fusion feature and the central loss function are weighted and summed to obtain the loss function of the model optimization. .

本发明的有益效果:本发明综合考虑了细粒度图像类内差异大,类间差异小,背景噪声影响大等因素,引入一个抑制分支,通过抑制图像中最显著的区域,从而迫使网络寻找易混淆类别间微妙的判别性特征。还引入了一个同类对比学习模块,将同类样本的特征向量进行融合,从而增加了同一类别下不同图像的交互信息。同时,引入了中心损失函数,最小化特征与对应类中心之间的距离,使学到的特征更具判别性。Beneficial effects of the present invention: The present invention comprehensively considers factors such as large intra-class differences, small inter-class differences, and large background noise in fine-grained images, and introduces a suppression branch, which suppresses the most significant area in the image, thereby forcing the network to find easy-to-use Confuse subtle discriminative features between classes. A class comparison learning module is also introduced to fuse the feature vectors of samples of the same class, thereby increasing the interactive information of different images under the same class. At the same time, a center loss function is introduced to minimize the distance between the feature and the corresponding class center, making the learned features more discriminative.

综合以上几点,本发明在判断过程综合利用了全局特征和局部特征,在多个细粒度图像分类任务上取得了明显的性能提升,相比现有的方法更具鲁棒性,且易于实际部署。提升了细粒度图像识别的准确率。Based on the above points, the present invention comprehensively utilizes global features and local features in the judgment process, and achieves obvious performance improvement in multiple fine-grained image classification tasks, which is more robust and easy to practice compared with existing methods. deploy. Improves the accuracy of fine-grained image recognition.

附图说明Description of drawings

图1为中图1a、图1b、图1c和图1d均为现有的4组细粒度图像示意图;Fig. 1 is a schematic diagram of four existing groups of fine-grained images in Fig. 1a, Fig. 1b, Fig. 1c and Fig. 1d;

图2为本发明所述的一种基于多视角特征融合的细粒度图像识别方法中双线性特征提取的示意图;2 is a schematic diagram of bilinear feature extraction in a fine-grained image recognition method based on multi-view feature fusion according to the present invention;

图3为本发明所述的一种基于多视角特征融合的细粒度图像识别方法中同类对比学习的示意图;FIG. 3 is a schematic diagram of similar comparison learning in a fine-grained image recognition method based on multi-view feature fusion according to the present invention;

图4为本发明所述的一种基于多视角特征融合的细粒度图像识别方法中模型优化损失函数计算的示意图;4 is a schematic diagram of model optimization loss function calculation in a fine-grained image recognition method based on multi-view feature fusion according to the present invention;

图5为本发明所述的一种基于多视角特征融合的细粒度图像识别方法获得的特征可视化效果图。FIG. 5 is a feature visualization effect diagram obtained by a fine-grained image recognition method based on multi-view feature fusion according to the present invention.

具体实施方式Detailed ways

结合图2至图5说明本实施方式,一种基于多视角特征融合的细粒度图像识别方法,该方法由以下步骤实现:The present embodiment, a fine-grained image recognition method based on multi-view feature fusion, is described with reference to FIGS. 2 to 5 , and the method is implemented by the following steps:

步骤一、双线性特征提取步骤:采用在ImageNet上的预训练的ResNet-50网络结构,输入大小固定的原始图像,将不同卷积层输出的特征图进行融合,得到双线性特征向量。Step 1. Bilinear feature extraction step: Using the pre-trained ResNet-50 network structure on ImageNet, input the original image with a fixed size, and fuse the feature maps output by different convolutional layers to obtain bilinear feature vectors.

结合图2,在特征提取步骤中,使用在数据集ImageNet预训练的网络作为特征提取的基础网络,常用的图像分类网络如VGGNet、GoogLeNet、ResNet等,可对其进行微调(fine-tune)使模型适应特定的任务。具体来说,将原始图像输入至特征提取网络,得到最后两个卷积层输出的特征图(feature map),分别记为

Figure BDA0002691299670000041
其中D1、D2表示两种特征的通道数,H和W分别表示特征图的高度和宽度。为了解决融合后特征维数过高的问题,同时保证生成的特征向量中所包含的特征信息足够多,只随机抽取F2中n个通道的特征图与F1进行融合。记特征图F1与F2中各个位置沿通道的特征向量为
Figure BDA0002691299670000042
Figure BDA0002691299670000043
两个特征向量相乘后可得到双线性矩阵
Figure BDA0002691299670000044
将特征图中各个位置对应的双线性矩阵进行相加,并将矩阵展为一个向量,即为双线性向量
Figure BDA0002691299670000045
其中D=D1×D2。该双线性向量提供了比线性模型更强的特征表示。Combined with Figure 2, in the feature extraction step, the network pre-trained in the dataset ImageNet is used as the basic network for feature extraction. Commonly used image classification networks such as VGGNet, GoogLeNet, ResNet, etc., can be fine-tuned (fine-tune). Models are adapted to specific tasks. Specifically, the original image is input to the feature extraction network, and the feature maps output by the last two convolutional layers are obtained, which are recorded as
Figure BDA0002691299670000041
Among them, D1 and D2 represent the number of channels of the two features, and H and W represent the height and width of the feature map, respectively. In order to solve the problem that the feature dimension is too high after fusion, and at the same time ensure that the generated feature vector contains enough feature information, only the feature maps of n channels in F2 are randomly extracted and fused with F1. The feature vectors of each position along the channel in the feature maps F 1 and F 2 are written as
Figure BDA0002691299670000042
Figure BDA0002691299670000043
The bilinear matrix can be obtained by multiplying the two eigenvectors
Figure BDA0002691299670000044
Add the bilinear matrices corresponding to each position in the feature map, and expand the matrix into a vector, which is a bilinear vector
Figure BDA0002691299670000045
where D=D1×D2. This bilinear vector provides a stronger feature representation than the linear model.

步骤二、抑制分支学习步骤:Step 2. Suppression branch learning steps:

A、注意图生成步骤:根据特征图的大小及阈值生成注意图。A. Attention map generation step: generate an attention map according to the size and threshold of the feature map.

B、图像抑制步骤:根据注意图生成抑制掩码,覆盖到原始图像上生成局部区域被掩蔽的抑制图像。B. Image suppression step: generate a suppression mask according to the attention map, and overlay it on the original image to generate a suppressed image with masked local areas.

C、多分类交叉熵计算步骤:抑制图像经由步骤一得到双线性特征向量,再输入至全连接层得到预测概率值,对得到的类别预测值计算多分类交叉熵。C. Multi-classification cross-entropy calculation step: The suppression image obtains a bilinear feature vector through step 1, and then inputs it to the fully connected layer to obtain the predicted probability value, and calculates the multi-classification cross-entropy for the obtained class prediction value.

在抑制分支学习步骤中,包含以下三个方面:In the suppression branch learning step, the following three aspects are included:

步骤A对特征提取网络中卷积层输出的特征图

Figure BDA0002691299670000046
的各个通道求平均值pd,按该平均值进行排序,选择top-5的值计算熵:Step A: The feature map output by the convolutional layer in the feature extraction network
Figure BDA0002691299670000046
Calculate the average value of each channel p d , sort by this average value, and select the top-5 value to calculate the entropy:

通过比较熵和阈值δ的大小来构建注意图A:Construct the attention map A by comparing the magnitude of entropy and threshold δ:

Figure BDA0002691299670000048
Figure BDA0002691299670000048

步骤B将注意图放大至原始图像大小,计算其平均值m,以m*θ作为阈值,将注意图中大于阈值的元素设置为0,将其他元素设置为1,从而获得抑制掩码M:Step B enlarges the attention map to the original image size, calculates its average value m, uses m*θ as the threshold, sets the elements in the attention map larger than the threshold value to 0, and sets other elements to 1, thereby obtaining the suppression mask M:

Figure BDA0002691299670000049
Figure BDA0002691299670000049

计算所述注意图的平均值m,设定范围在0~1之间的阈值θ,Calculate the average value m of the attention map, and set a threshold θ ranging from 0 to 1,

步骤C将抑制掩码覆盖到原始图像,从而得到局部区域被掩蔽的抑制图像:Step C covers the suppression mask to the original image, so as to obtain the suppressed image with the masked local area:

Is(x,y)=I(x,y)*M(x,y)Is (x,y)=I( x ,y)*M(x,y)

式中,I(x,y)为原始图像中I中(x,y)位置的值。In the formula, I(x,y) is the value of the (x,y) position in I in the original image.

由于抑制了图像最显著的区域,从而分散了注意力,迫使神经网络从其他区域学到判别性信息。还可以降低网络对训练样本的依赖,防止过拟合,进一步提高了模型的鲁棒性。Distraction by suppressing the most salient regions of the image forces the neural network to learn discriminative information from other regions. It can also reduce the network's dependence on training samples, prevent overfitting, and further improve the robustness of the model.

步骤三、同类对比模块学习步骤:Step 3. Learning steps for similar comparison modules:

A、图像采样步骤:随机选择同一类别下其他N张图像作为正样本。A. Image sampling step: randomly select other N images in the same category as positive samples.

B、特征融合步骤:将目标图像和随机采样的正样本图像经步骤一得到的双线性特征向量进行融合,得到的融合特征综合了同一类别下多张图像的特征信息。B. Feature fusion step: fuse the target image and the randomly sampled positive sample image through the bilinear feature vector obtained in step 1, and the obtained fusion feature integrates the feature information of multiple images under the same category.

C、融合特征损失函数计算:将融合后的特征向量直接输入全连接层得到预测概率,对得到的类别预测值计算多分类交叉熵。C. Calculation of fusion feature loss function: The fused feature vector is directly input into the fully connected layer to obtain the prediction probability, and the multi-class cross entropy is calculated for the obtained category prediction value.

结合图3,步骤A随机选择与输入图像属于同一类别的N张图像,都送入步骤一的双线性特征提取网络。Referring to Figure 3, step A randomly selects N images belonging to the same category as the input image, and sends them to the bilinear feature extraction network in step 1.

步骤B对步骤A输出的多张同类图像的双线性特征向量求平均,得到一个融合的特征向量:Step B averages the bilinear feature vectors of multiple images of the same type output in step A to obtain a fused feature vector:

Figure BDA0002691299670000051
Figure BDA0002691299670000051

式中,j为特征向量的位置,V(j)为特征向量在第j个位置的值,T为选取的正样本数量;Vr(j)为第r个正样本在在第j个位置的值;In the formula, j is the position of the feature vector, V(j) is the value of the feature vector at the jth position, T is the number of positive samples selected; V r (j) is the rth positive sample at the jth position. the value of;

步骤四、中心损失计算步骤;Step 4: Calculate the center loss;

A、类中心生成步骤:在训练过程中,不断更新网络学到的各个类别中心的特征向量。A. Class center generation step: During the training process, the feature vector of each class center learned by the network is continuously updated.

B、中心损失计算步骤:以每个输入图像得到的双线性特征向量与类中心向量之前的距离作为中心损失,在训练过程中不断优化。B. Center loss calculation step: The distance between the bilinear feature vector obtained from each input image and the class center vector is used as the center loss, which is continuously optimized during the training process.

本实施方式中,对每个类别都计算一个特征向量作为对应类别的类中心,这个特征向量随着训练的进行会不断的更新。通过惩罚每个样本的双线性特征向量和对应类别样本中心的偏移,使得同一类别的样本尽量聚合在一起,省去了复杂的样本对构造过程。令vi为第i个样本的双线性特征,ci为样本i对应类别所有样本的平均特征,即类中心,N为当前批量的样本个数,公式如下:In this embodiment, a feature vector is calculated for each category as the class center of the corresponding category, and this feature vector will be continuously updated as the training progresses. By penalizing the offset of the bilinear feature vector of each sample and the center of the corresponding class of samples, the samples of the same class are aggregated together as much as possible, eliminating the complex sample pair construction process. Let vi be the bilinear feature of the ith sample, ci be the average feature of all samples of the corresponding category of sample i , namely the class center, N is the number of samples in the current batch, the formula is as follows:

Figure BDA0002691299670000061
Figure BDA0002691299670000061

步骤五、模型优化损失函数计算步骤:Step 5. Model optimization loss function calculation steps:

将原始图像的双线性特征的交叉熵损失函数,抑制图像的双线性特征的交叉熵损失函数,融合特征的交叉熵损失函数以及中心损失函数加权求和,即可得到模型优化的损失函数。The weighted summation of the cross-entropy loss function of the bilinear feature of the original image, the cross-entropy loss function of the suppressed bilinear feature of the image, the cross-entropy loss function of the fusion feature, and the central loss function can obtain the optimized loss function of the model .

结合图4,记原始图像的双线性特征向量的交叉熵损失为LCE1,抑制图像的双线性特征向量的交叉熵损失为LCE2,融合特征的交叉熵损失为LCE3,中心损失为LC,将这些损失函数加权求和,即可得到模型优化的损失函数L:Referring to Figure 4, the cross-entropy loss of the bilinear feature vector of the original image is L CE1 , the cross-entropy loss of the bi-linear feature vector of the suppressed image is L CE2 , the cross-entropy loss of the fusion feature is L CE3 , and the center loss is L C , the weighted summation of these loss functions, the loss function L optimized by the model can be obtained:

Figure BDA0002691299670000062
Figure BDA0002691299670000062

式中,λ为中心损失函数的权重。where λ is the weight of the central loss function.

结合图5说明本实施方式,图5中,第一行为数据集中随机选取的原始图像,第二行为原始图像输入的全局分支获得的类激活图,第三行为抑制分支获得的类激活图。可以看出,在全局分支中,网络学习到了图像最显著的区域,如鸟的喙、汽车前照灯等,在抑制分支中,网络学习到了有利于细粒度分类的微妙的特征,如鸟的躯干、车轮等。多视角结合,使得网络模型的判断依据更为全面,既能显著性区域,又能微妙地捕捉到细粒度特征。This embodiment is described with reference to FIG. 5 . In FIG. 5 , the first row is a randomly selected original image in the data set, the second row is a class activation map obtained by the global branch input from the original image, and the third row is a class activation map obtained by the suppression branch. It can be seen that in the global branch, the network learns the most salient regions of the image, such as bird's beak, car headlights, etc., and in the suppression branch, the network learns subtle features that are conducive to fine-grained classification, such as bird's beak, car headlights, etc. Torso, wheels, etc. The combination of multiple perspectives makes the judgment basis of the network model more comprehensive, which can capture both salient regions and fine-grained features subtly.

本实施方式所述的细粒度图像识别方法,引入了一种新的数据增强方式,通过注意图引导来抑制图像中的部件区域,从而分散注意力,使网络学习到更多互补区域特征信息。引入了一个同类对比模块,融合了来自同一类别多张图像的特征信息,使同一类别图像的表示在嵌入空间尽可能接近,从而提高了分类性能。The fine-grained image recognition method described in this embodiment introduces a new data enhancement method, which suppresses component regions in the image through attention map guidance, thereby distracting attention and enabling the network to learn more complementary region feature information. A homogenous comparison module is introduced, which fuses feature information from multiple images of the same category to make the representations of images of the same category as close as possible in the embedding space, thereby improving the classification performance.

以上所述实施例的各技术特征可以进行任意的组合,为使描述简洁,未对上述实施例中的各个技术特征所有可能的组合都进行描述,然而,只要这些技术特征的组合不存在矛盾,都应当认为是本说明书记载的范围。The technical features of the above-described embodiments can be combined arbitrarily. For the sake of brevity, all possible combinations of the technical features in the above-described embodiments are not described. However, as long as there is no contradiction between the combinations of these technical features, All should be regarded as the scope described in this specification.

以上所述实施例仅表达了本发明的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些都属于本发明的保护范围。因此,本发明专利的保护范围应以所附权利要求为准。The above-mentioned embodiments only represent several embodiments of the present invention, and the descriptions thereof are specific and detailed, but should not be construed as a limitation on the scope of the invention patent. It should be pointed out that for those skilled in the art, without departing from the concept of the present invention, several modifications and improvements can be made, which all belong to the protection scope of the present invention. Therefore, the protection scope of the patent of the present invention shall be subject to the appended claims.

Claims (4)

1.一种基于多视角特征融合的细粒度图像识别方法,其特征是:该方法由以下步骤实现:1. a fine-grained image recognition method based on multi-view feature fusion, is characterized in that: the method is realized by the following steps: 步骤一、双线性特征提取;Step 1. Bilinear feature extraction; 将原始图像输入至双线性特征提取网络,将不同卷积层输出的特征图进行融合,获得双线性特征向量;所述特征提取网络采用在数据集ImageNet预训练的网络结构;The original image is input into the bilinear feature extraction network, and the feature maps output by different convolutional layers are fused to obtain bilinear feature vectors; the feature extraction network adopts the network structure pre-trained in the dataset ImageNet; 步骤二、抑制分支学习,具体过程为:Step 2: Suppress branch learning, the specific process is: 步骤二一、根据步骤一所述特征提取网络不同卷积层输出的特征图的大小及阈值生成注意图;Step 21: Generate an attention map according to the size and threshold of the feature maps output by different convolutional layers of the feature extraction network described in step 1; 步骤二二、根据步骤二一所述的注意图生成抑制掩码,将所述抑制掩码覆盖到所述原始图像上生成局部区域被掩蔽的抑制图像;Step 22, generating a suppression mask according to the attention map described in step 21, and overlaying the suppression mask on the original image to generate a suppressed image in which the local area is masked; 步骤二三、将步骤二所述的抑制图像采用步骤一进行双线性特征提取,获得双线性特征向量,将所述双线性特征向量输入至全连接层,获得预测的类别概率值,并对该预测的类别概率值计算多分类交叉熵;Step 2 and 3: Use step 1 to perform bilinear feature extraction on the suppressed image described in step 2, obtain bilinear feature vector, input the bilinear feature vector to the fully connected layer, and obtain the predicted class probability value, And calculate the multi-class cross entropy for the predicted class probability value; 步骤三、同类对比模块学习;Step 3. Similar comparison module learning; 步骤三一、随机选择与原始图像同一类别下其他N张图像作为正样本图像;Step 31: Randomly select other N images in the same category as the original image as positive sample images; 步骤三二、将目标图像与步骤三一所述的正样本图像送入步骤一所述的特征提取网络进行双线性特征向量融合,获得融合特征综合了同类别下多张图像的双线性特征向量;Step 32: The target image and the positive sample image described in step 31 are sent to the feature extraction network described in step 1 for bilinear feature vector fusion, and the fusion feature is obtained. Feature vector; 步骤三三、将步骤三二获得的同类别下多张图像的双线性特征向量求平均,获得一个融合的特征向量,将融合后的特征向量输入全连接层,获得预测的概率,对得到的同类别的预测概率计算多分类交叉熵;Step 33: Average the bilinear eigenvectors of multiple images of the same category obtained in step 32 to obtain a fused feature vector, and input the fused feature vector into the fully connected layer to obtain the predicted probability. Calculate the multi-class cross entropy of the predicted probability of the same category; 步骤四、中心损失函数LC计算;Step 4: Calculate the central loss function LC ; 令vi为第i个样本的双线性特征,ci为样本i对应类别所有样本的平均特征,即类中心,N为当前批量的样本个数,则中心损失函数LC的公式如下:Let v i be the bilinear feature of the ith sample, c i be the average feature of all samples in the corresponding category of sample i, that is, the class center, and N be the number of samples in the current batch, then the formula of the center loss function L C is as follows:
Figure FDA0002691299660000011
Figure FDA0002691299660000011
步骤五、模型优化损失函数计算;Step 5. Model optimization loss function calculation; 将原始图像的双线性特征向量的交叉熵损失函数,抑制图像的双线性特征向量的交叉熵损失函数,融合特征的交叉熵损失函数以及中心损失函数加权求和,获得模型优化的损失函数。The cross-entropy loss function of the bilinear feature vector of the original image, the cross-entropy loss function of the bilinear feature vector of the suppressed image, the cross-entropy loss function of the fusion feature, and the central loss function are weighted and summed to obtain the loss function of the model optimization. .
2.根据权利要求1所述的一种基于多视角特征融合的细粒度图像识别方法,其特征在于:步骤二一中,生成注意图的具体过程为:2. a kind of fine-grained image recognition method based on multi-view feature fusion according to claim 1, is characterized in that: in step 21, the specific process of generating attention map is: 对特征提取网络中最后一个卷积层输出的特征图
Figure FDA0002691299660000025
的各个通道求平均值pd,式中,D为特征的通道数,H和W分别为特征图的高度和宽度;按该平均值进行排序,获得熵E的公式为:
Feature map output from the last convolutional layer in the feature extraction network
Figure FDA0002691299660000025
Calculate the average value p d of each channel of , where D is the number of channels of the feature, H and W are the height and width of the feature map, respectively; according to the average value, the formula for obtaining the entropy E is:
Figure FDA0002691299660000021
Figure FDA0002691299660000021
通过比较熵和阈值δ的大小构建注意图A:Construct an attention map A by comparing the magnitude of entropy and threshold δ:
Figure FDA0002691299660000022
Figure FDA0002691299660000022
式中,Fk为按通道排序后的各个通道对应的二维特征图;In the formula, F k is the two-dimensional feature map corresponding to each channel sorted by channel; 步骤二二中,生成抑制掩码的具体过程为:In step 22, the specific process of generating the suppression mask is as follows: 将步骤二一中的注意图放大至原始图像大小,计算所述注意图的平均值m,设定范围在0~1之间的阈值θ,以m*θ作为阈值,将注意图中大于阈值m*θ的元素设置为0,将其他元素设置为1,获得抑制掩码M:Enlarge the attention map in step 21 to the original image size, calculate the average value m of the attention map, set a threshold θ between 0 and 1, take m*θ as the threshold, and set the attention map larger than the threshold Elements of m*θ are set to 0, other elements are set to 1, and the suppression mask M is obtained:
Figure FDA0002691299660000023
Figure FDA0002691299660000023
式中,A(x,y)为注意图A中(x,y)位置的值;In the formula, A(x, y) is the value of the (x, y) position in the attention map A; 将抑制掩码覆盖到原始图像,获得局部区域被掩蔽的抑制图像Is(x,y);Covering the suppression mask to the original image to obtain the suppressed image Is ( x , y) where the local area is masked; Is(x,y)=I(x,y)*M(x,y)Is (x,y)=I( x ,y)*M(x,y) 式中,I(x,y)为原始图像中I中(x,y)位置的值。In the formula, I(x,y) is the value of the (x,y) position in I in the original image.
3.根据权利要求1所述的一种基于多视角特征融合的细粒度图像识别方法,其特征在于:步骤三三中,对同类别下多张图像的双线性特征向量求平均,获得一个融合的特征向量,用下式表示为:3. a kind of fine-grained image recognition method based on multi-view feature fusion according to claim 1 is characterized in that: in step 33, the bilinear feature vectors of multiple images under the same category are averaged to obtain a The fused feature vector is expressed as:
Figure FDA0002691299660000024
Figure FDA0002691299660000024
式中,j为特征向量的位置,V(j)为特征向量在第j个位置的值,T为选取的正样本数量;Vr(j)为第r个正样本在在第j个位置的值;In the formula, j is the position of the feature vector, V(j) is the value of the feature vector at the jth position, T is the number of positive samples selected; V r (j) is the rth positive sample at the jth position. the value of;
4.根据权利要求1所述的一种基于多视角特征融合的细粒度图像识别方法,其特征在于:步骤五中,原始图像的双线性特征向量的交叉熵损失函数为LCE1,抑制图像的双线性特征向量的交叉熵损失函数为LCE2,融合特征的交叉熵损失函数为LCE3,中心损失函数为LC,将其进行加权求和,获得模型优化的损失函数L,最终实现细粒度图像识别;用下式表示为:4. a kind of fine-grained image recognition method based on multi-view feature fusion according to claim 1, is characterized in that: in step 5, the cross entropy loss function of the bilinear feature vector of the original image is L CE1 , suppressing the image The cross-entropy loss function of the bilinear feature vector is L CE2 , the cross-entropy loss function of the fusion feature is L CE3 , and the central loss function is L C , which are weighted and summed to obtain the model-optimized loss function L, and finally realize Fine-grained image recognition; expressed as:
Figure FDA0002691299660000031
Figure FDA0002691299660000031
式中,λ为中心损失函数的权重。In the formula, λ is the weight of the central loss function.
CN202010992253.7A 2020-09-21 2020-09-21 Fine granularity image recognition method based on multi-view feature fusion Active CN112132004B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010992253.7A CN112132004B (en) 2020-09-21 2020-09-21 Fine granularity image recognition method based on multi-view feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010992253.7A CN112132004B (en) 2020-09-21 2020-09-21 Fine granularity image recognition method based on multi-view feature fusion

Publications (2)

Publication Number Publication Date
CN112132004A true CN112132004A (en) 2020-12-25
CN112132004B CN112132004B (en) 2024-06-25

Family

ID=73841694

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010992253.7A Active CN112132004B (en) 2020-09-21 2020-09-21 Fine granularity image recognition method based on multi-view feature fusion

Country Status (1)

Country Link
CN (1) CN112132004B (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112712066A (en) * 2021-01-19 2021-04-27 腾讯科技(深圳)有限公司 Image recognition method and device, computer equipment and storage medium
CN112733912A (en) * 2020-12-31 2021-04-30 华侨大学 Fine-grained image recognition method based on multi-grained countermeasure loss
CN112766378A (en) * 2021-01-19 2021-05-07 北京工商大学 Cross-domain small sample image classification model method focusing on fine-grained identification
CN112800927A (en) * 2021-01-25 2021-05-14 北京工业大学 AM-Softmax loss-based butterfly image fine granularity identification method
CN112990270A (en) * 2021-02-10 2021-06-18 华东师范大学 Automatic fusion method of traditional feature and depth feature
CN113065443A (en) * 2021-03-25 2021-07-02 携程计算机技术(上海)有限公司 Training method, recognition method, system, device and medium of image recognition model
CN113255793A (en) * 2021-06-01 2021-08-13 之江实验室 Fine-grained ship identification method based on contrast learning
CN113449613A (en) * 2021-06-15 2021-09-28 北京华创智芯科技有限公司 Multitask long-tail distribution image recognition method, multitask long-tail distribution image recognition system, electronic device and medium
CN113569814A (en) * 2021-09-06 2021-10-29 海南智晶科技有限公司 Unsupervised pedestrian re-identification method based on feature consistency
CN113642571A (en) * 2021-07-12 2021-11-12 中国海洋大学 Fine-grained image identification method based on saliency attention mechanism
CN113705489A (en) * 2021-08-31 2021-11-26 中国电子科技集团公司第二十八研究所 Remote sensing image fine-grained airplane identification method based on priori regional knowledge guidance
CN114119979A (en) * 2021-12-06 2022-03-01 西安电子科技大学 A Fine-Grained Image Classification Method Based on Segmentation Mask and Self-Attention Neural Network
CN114332562A (en) * 2021-12-29 2022-04-12 深圳市商汤科技有限公司 Neural network model training method, neural network model training device, neural network model detecting device, neural network model training device, neural network training device, and medium
CN114676777A (en) * 2022-03-25 2022-06-28 中国科学院软件研究所 A self-supervised learning method for fine-grained image classification based on Siamese network
CN115424086A (en) * 2022-07-26 2022-12-02 北京邮电大学 Multi-view fine-granularity identification method and device, electronic equipment and medium
CN117725483A (en) * 2023-09-26 2024-03-19 电子科技大学 Supervised signal classification method based on neural network

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109685115A (en) * 2018-11-30 2019-04-26 西北大学 A kind of the fine granularity conceptual model and learning method of bilinearity Fusion Features
CN110135502A (en) * 2019-05-17 2019-08-16 东南大学 A Fine-grained Image Recognition Method Based on Reinforcement Learning Strategy
CN110210550A (en) * 2019-05-28 2019-09-06 东南大学 Image fine granularity recognition methods based on integrated study strategy
CN110222636A (en) * 2019-05-31 2019-09-10 中国民航大学 The pedestrian's attribute recognition approach inhibited based on background
CN110807465A (en) * 2019-11-05 2020-02-18 北京邮电大学 A fine-grained image recognition method based on channel loss function
CN111523534A (en) * 2020-03-31 2020-08-11 华东师范大学 A method of image description

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109685115A (en) * 2018-11-30 2019-04-26 西北大学 A kind of the fine granularity conceptual model and learning method of bilinearity Fusion Features
CN110135502A (en) * 2019-05-17 2019-08-16 东南大学 A Fine-grained Image Recognition Method Based on Reinforcement Learning Strategy
CN110210550A (en) * 2019-05-28 2019-09-06 东南大学 Image fine granularity recognition methods based on integrated study strategy
CN110222636A (en) * 2019-05-31 2019-09-10 中国民航大学 The pedestrian's attribute recognition approach inhibited based on background
CN110807465A (en) * 2019-11-05 2020-02-18 北京邮电大学 A fine-grained image recognition method based on channel loss function
CN111523534A (en) * 2020-03-31 2020-08-11 华东师范大学 A method of image description

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
黄伟锋;张甜;常东良;闫冬;王嘉希;王丹;马占宇;: "基于多视角融合的细粒度图像分类方法", 信号处理, no. 09 *

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112733912A (en) * 2020-12-31 2021-04-30 华侨大学 Fine-grained image recognition method based on multi-grained countermeasure loss
CN112733912B (en) * 2020-12-31 2023-06-09 华侨大学 Fine-grained image recognition method based on multi-grained adversarial loss
CN112766378A (en) * 2021-01-19 2021-05-07 北京工商大学 Cross-domain small sample image classification model method focusing on fine-grained identification
CN112712066A (en) * 2021-01-19 2021-04-27 腾讯科技(深圳)有限公司 Image recognition method and device, computer equipment and storage medium
CN112766378B (en) * 2021-01-19 2023-07-21 北京工商大学 A cross-domain small-sample image classification model method focusing on fine-grained recognition
CN112800927A (en) * 2021-01-25 2021-05-14 北京工业大学 AM-Softmax loss-based butterfly image fine granularity identification method
CN112800927B (en) * 2021-01-25 2024-03-29 北京工业大学 Butterfly image fine-granularity identification method based on AM-Softmax loss
CN112990270B (en) * 2021-02-10 2023-04-07 华东师范大学 Automatic fusion method of traditional feature and depth feature
CN112990270A (en) * 2021-02-10 2021-06-18 华东师范大学 Automatic fusion method of traditional feature and depth feature
CN113065443A (en) * 2021-03-25 2021-07-02 携程计算机技术(上海)有限公司 Training method, recognition method, system, device and medium of image recognition model
CN113255793A (en) * 2021-06-01 2021-08-13 之江实验室 Fine-grained ship identification method based on contrast learning
CN113255793B (en) * 2021-06-01 2021-11-30 之江实验室 Fine-grained ship identification method based on contrast learning
CN113449613A (en) * 2021-06-15 2021-09-28 北京华创智芯科技有限公司 Multitask long-tail distribution image recognition method, multitask long-tail distribution image recognition system, electronic device and medium
CN113449613B (en) * 2021-06-15 2024-02-27 北京华创智芯科技有限公司 Multi-task long tail distribution image recognition method, system, electronic equipment and medium
CN113642571A (en) * 2021-07-12 2021-11-12 中国海洋大学 Fine-grained image identification method based on saliency attention mechanism
CN113642571B (en) * 2021-07-12 2023-10-10 中国海洋大学 Fine granularity image recognition method based on salient attention mechanism
CN113705489A (en) * 2021-08-31 2021-11-26 中国电子科技集团公司第二十八研究所 Remote sensing image fine-grained airplane identification method based on priori regional knowledge guidance
CN113705489B (en) * 2021-08-31 2024-06-07 中国电子科技集团公司第二十八研究所 Remote sensing image fine-granularity airplane identification method based on priori regional knowledge guidance
CN113569814A (en) * 2021-09-06 2021-10-29 海南智晶科技有限公司 Unsupervised pedestrian re-identification method based on feature consistency
CN114119979A (en) * 2021-12-06 2022-03-01 西安电子科技大学 A Fine-Grained Image Classification Method Based on Segmentation Mask and Self-Attention Neural Network
CN114332562A (en) * 2021-12-29 2022-04-12 深圳市商汤科技有限公司 Neural network model training method, neural network model training device, neural network model detecting device, neural network model training device, neural network training device, and medium
CN114676777A (en) * 2022-03-25 2022-06-28 中国科学院软件研究所 A self-supervised learning method for fine-grained image classification based on Siamese network
CN115424086A (en) * 2022-07-26 2022-12-02 北京邮电大学 Multi-view fine-granularity identification method and device, electronic equipment and medium
CN117725483A (en) * 2023-09-26 2024-03-19 电子科技大学 Supervised signal classification method based on neural network
CN117725483B (en) * 2023-09-26 2025-02-07 电子科技大学 A supervised signal classification method based on neural network

Also Published As

Publication number Publication date
CN112132004B (en) 2024-06-25

Similar Documents

Publication Publication Date Title
CN112132004A (en) A fine-grained image recognition method based on multi-view feature fusion
Yang et al. A survey of DNN methods for blind image quality assessment
Eigen et al. Nonparametric image parsing using adaptive neighbor sets
CN113408492A (en) Pedestrian re-identification method based on global-local feature dynamic alignment
CN105787501B (en) Power transmission line corridor region automatically selects the vegetation classification method of feature
CN113011329A (en) Pyramid network based on multi-scale features and dense crowd counting method
CN111178208A (en) Pedestrian detection method, device and medium based on deep learning
CN113688894A (en) Fine-grained image classification method fusing multi-grained features
Sun et al. Encoding spectral and spatial context information for hyperspectral image classification
CN110503113B (en) Image saliency target detection method based on low-rank matrix recovery
CN107169417A (en) Strengthened based on multinuclear and the RGBD images of conspicuousness fusion cooperate with conspicuousness detection method
CN109165698A (en) A kind of image classification recognition methods and its storage medium towards wisdom traffic
Zhang et al. Boosting transferability of physical attack against detectors by redistributing separable attention
CN110276363A (en) A small bird target detection method based on density map estimation
Wang et al. Study on the method of transmission line foreign body detection based on deep learning
CN114495170A (en) A method and system for pedestrian re-identification based on local suppression of self-attention
CN113591928A (en) Vehicle weight identification method and system based on multi-view and convolution attention module
CN113128308A (en) Pedestrian detection method, device, equipment and medium in port scene
CN109447014A (en) A kind of online behavioral value method of video based on binary channels convolutional neural networks
Han et al. The edge-preservation multi-classifier relearning framework for the classification of high-resolution remotely sensed imagery
Tang et al. Deep saliency quality assessment network with joint metric
Cao et al. A new region proposal network for far-infrared pedestrian detection
CN111695436B (en) A high spatial resolution remote sensing image scene classification method based on object enhancement
CN116434010A (en) Multi-view pedestrian attribute identification method
CN114694019B (en) Building migration extraction method in remote sensing images based on anomaly detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant