CN114782779B - Small sample image feature learning method and device based on feature distribution migration - Google Patents
Small sample image feature learning method and device based on feature distribution migration Download PDFInfo
- Publication number
- CN114782779B CN114782779B CN202210487387.2A CN202210487387A CN114782779B CN 114782779 B CN114782779 B CN 114782779B CN 202210487387 A CN202210487387 A CN 202210487387A CN 114782779 B CN114782779 B CN 114782779B
- Authority
- CN
- China
- Prior art keywords
- class
- distribution
- sample
- feature
- samples
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 238000013508 migration Methods 0.000 title claims abstract description 22
- 230000005012 migration Effects 0.000 title claims abstract description 22
- 238000012937 correction Methods 0.000 claims abstract description 13
- 230000006870 function Effects 0.000 claims description 18
- 238000011176 pooling Methods 0.000 claims description 12
- 238000012549 training Methods 0.000 claims description 10
- 238000012360 testing method Methods 0.000 claims description 9
- 239000013598 vector Substances 0.000 claims description 9
- 238000010586 diagram Methods 0.000 claims description 7
- 230000004913 activation Effects 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 6
- 239000011159 matrix material Substances 0.000 claims description 6
- 238000004422 calculation algorithm Methods 0.000 claims description 4
- 238000000605 extraction Methods 0.000 claims description 4
- 238000005259 measurement Methods 0.000 claims description 4
- 238000007781 pre-processing Methods 0.000 claims description 4
- 238000010606 normalization Methods 0.000 claims description 3
- 238000012545 processing Methods 0.000 claims description 3
- 230000003044 adaptive effect Effects 0.000 claims 2
- 230000000694 effects Effects 0.000 abstract description 3
- 238000005516 engineering process Methods 0.000 description 3
- 101001121408 Homo sapiens L-amino-acid oxidase Proteins 0.000 description 2
- 102100026388 L-amino-acid oxidase Human genes 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 101000827703 Homo sapiens Polyphosphoinositide phosphatase Proteins 0.000 description 1
- 102100023591 Polyphosphoinositide phosphatase Human genes 0.000 description 1
- 101100012902 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) FIG2 gene Proteins 0.000 description 1
- 101100233916 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) KAR5 gene Proteins 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000007635 classification algorithm Methods 0.000 description 1
- 238000013145 classification model Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/243—Classification techniques relating to the number of classes
- G06F18/2431—Multiple classes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Software Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Medical Informatics (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于特征分布迁移的小样本图像特征学习方法及装置,在前期利用基类的数据结合梯度下降的方法,对嵌入模块以及分布学习模块的参数进行优化,后期进行分布矫正时,并不需要额外的参数设置;另外,通常假设特征表示中的每个维度都遵循高斯分布,这样高斯分布的均值和方差可以在类似的类别之间传递,减少偏差,以便这些类别的统计数据在足够的样本数下得到更好的估计,再利用分布矫正模型,对样本的分布进行矫正,从而更为精准的对新类样本进行分类。同时可以与任何分类器和特征提取器配对,无需额外的参数,解决了小样本图像分类中存在的原型偏差问题,改善了图像的分类效果,具有很高的实用价值。
The invention discloses a small-sample image feature learning method and device based on feature distribution migration. In the early stage, the data of the base class is combined with the method of gradient descent to optimize the parameters of the embedding module and the distribution learning module, and when the distribution is corrected in the later stage , does not require additional parameter settings; in addition, it is usually assumed that each dimension in the feature representation follows a Gaussian distribution, so that the mean and variance of the Gaussian distribution can be transferred between similar categories, reducing bias, so that the statistics of these categories A better estimate is obtained with a sufficient number of samples, and then the distribution correction model is used to correct the distribution of the samples, so as to classify the new class samples more accurately. At the same time, it can be paired with any classifier and feature extractor without additional parameters. It solves the problem of prototype deviation in small sample image classification, improves the classification effect of images, and has high practical value.
Description
技术领域Technical Field
本发明涉及图像分类技术领域,尤其涉及一种基于特征分布迁移的小样本图像特征学习方法及装置。The present invention relates to the technical field of image classification, and in particular to a small sample image feature learning method and device based on feature distribution migration.
背景技术Background Art
近年来,随着计算机技术的发展,人们浏览的信息日益丰富,每天都有大量图片被上传到网络,由于数量巨大,人工已经无法对此进行分类。在很多大样本图像分类任务上,机器的识别性能已经超越人类。然而,当样本量比较少时,机器的识别水平仍与人类存在较大差距。因此,研究高效可靠的图片分类算法有很迫切的社会需求。In recent years, with the development of computer technology, people browse more and more information. A large number of pictures are uploaded to the Internet every day. Due to the huge number, it is no longer possible to classify them manually. In many large-sample image classification tasks, the recognition performance of machines has surpassed that of humans. However, when the sample size is relatively small, the recognition level of machines is still far behind that of humans. Therefore, there is an urgent social need to study efficient and reliable image classification algorithms.
小样本分类(Few-shot Classification)属于小样本学习(Few-shot Learning)范畴,往往包含类别空间不相交的两类数据,即基类数据和新类数据。小样本分类旨在利用基类数据学习的知识和新类数据的少量标记样本(支持样本)来学习分类规则,准确预测新类任务中未标记样本(查询样本)的类别。Few-shot classification belongs to the category of few-shot learning, which often contains two types of data in disjoint category space, namely base class data and new class data. Few-shot classification aims to use the knowledge learned from base class data and a small number of labeled samples (support samples) of new class data to learn classification rules and accurately predict the category of unlabeled samples (query samples) in new class tasks.
针对现有的技术缺点而言,首先,对标记样本极少的小样本分类任务来说,现有的深度学习技术并不适用。因而,如何基于基类数据和标记样本极少的新类数据,来学习高辨识度的特征表示,是一个值得探索的问题。其次,对分布原型的偏差而言,由于标记样本极少,常常使得学习的原型偏差(bias)较大。因此,如何通过减少原型偏差来提高小样本图像分类性能,也是一项具有挑战的任务。最后,在基类数据特征的判别性以及基类数据特征在新类数据上的可迁移性存在着误差,易导致分类的不准确性。Regarding the shortcomings of existing technologies, first of all, existing deep learning technologies are not applicable to small sample classification tasks with very few labeled samples. Therefore, how to learn highly recognizable feature representations based on base class data and new class data with very few labeled samples is a problem worth exploring. Secondly, as for the deviation of distribution prototypes, due to the very few labeled samples, the bias of the learned prototypes is often large. Therefore, how to improve the performance of small sample image classification by reducing prototype bias is also a challenging task. Finally, there are errors in the discriminability of base class data features and the transferability of base class data features to new class data, which can easily lead to inaccurate classification.
发明内容Summary of the invention
本发明针对上述小样本图像分类中的原型偏差问题,提出一种基于特征分布迁移的小样本图像特征学习方法及装置,主要是通常假设特征表示中的每个维度都遵循高斯分布,这样高斯分布的均值和方差可以在类似的类别之间传递,结合比较样本间或者样本与分布原型间的距离来判断类别,这些类别的统计数据在足够的样本数下得到更好的估计,改善了图像的分类效果。In response to the prototype deviation problem in the above-mentioned small sample image classification, the present invention proposes a small sample image feature learning method and device based on feature distribution migration. It is mainly assumed that each dimension in the feature representation follows a Gaussian distribution, so that the mean and variance of the Gaussian distribution can be transferred between similar categories. The category is judged by comparing the distance between samples or between the sample and the distribution prototype. The statistical data of these categories can be better estimated with a sufficient number of samples, thereby improving the classification effect of the image.
为了实现上述目的,本发明提供如下技术方案:In order to achieve the above object, the present invention provides the following technical solutions:
一方面,本发明提供了一种基于特征分布迁移的小样本图像特征学习方法,包括以下步骤:On the one hand, the present invention provides a small sample image feature learning method based on feature distribution migration, comprising the following steps:
S1、对数据进行预处理,其中数据包括训练集和测试集;S1. Preprocess the data, where the data includes a training set and a test set;
S2,利用基类数据预训练嵌入模块fθ,得到良好的特征空间;S2, use the base class data to pre-train the embedding module f θ to obtain a good feature space;
S3,将Dtrain输入到嵌入模块fθ,得到样本特征图,将其输入到分布学习模块gφ中,最小化损失函数,优化分布学习模块gφ;S3, input D train into the embedding module f θ to obtain the sample feature map, input it into the distribution learning module g φ , minimize the loss function, and optimize the distribution learning module g φ ;
S4,将新类数据分为支持集和查询集将支持集经过嵌入模块fθ和分布学习模块gφ计算每类的分布原型和 S4, divide the new class data into support sets and queryset Will support set After the embedding module f θ and the distribution learning module g φ, the distribution prototype of each class is calculated and
S5,计算基类数据中各类的类别概率,选取最大的前n个类别,将n个类别的分布与当前类别的分布合并,得到矫正后每类的分布原型和 S5, calculate the category probability of each category in the base category data, select the largest first n categories, merge the distribution of n categories with the distribution of the current category, and obtain the distribution prototype of each category after correction and
S6,计算新类查询样本的预测概率。S6, calculate the predicted probability of the new class query sample.
进一步地,步骤S1的预处理方法为:Furthermore, the preprocessing method of step S1 is:
S11,将数据分为 和两部分,且这两部分的类别空间互斥,将Dtrain用以在训练过程中调整参数,Dtest作为新类数据测评模型性能;S11, the data Divided into and The class spaces of the two parts are mutually exclusive. D train is used to adjust parameters during training, and D test is used as new class data to evaluate model performance.
S12,对于C-way K-shot分类任务,从Dtrain中随机选出C个类别,每个类别中随机选出M个样本,其中K个样本作为支持样本Si,其余M-K个样本作为查询样本Qi,Si和Qi构成一个任务Ti;同样地,对于Dtest有任务 S12, for the C-way K-shot classification task, randomly select C categories from D train , randomly select M samples from each category, K samples as support samples Si , and the remaining MK samples as query samples Qi . Si and Qi constitute a task Ti ; similarly, for D test, there is a task
进一步地,步骤S2中使用包含四个卷积块的嵌入模块fθ对图像提取特征,其中含有卷积层、池化层和非线性激活函数,每个卷积块使用窗口大小为3*3的卷积核,一个批量归一化,一个RELU非线性层,一个2×2最大池化层,裁剪了最后两个块的最大池化层。Furthermore, in step S2, an embedding module f θ containing four convolutional blocks is used to extract features from the image, which contains convolutional layers, pooling layers and non-linear activation functions. Each convolutional block uses a convolution kernel with a window size of 3*3, a batch normalization, a RELU non-linear layer, a 2×2 maximum pooling layer, and the maximum pooling layers of the last two blocks are cropped.
进一步地,步骤S3中分布学习模块gφ由两个全连接层组成,用以提取图像特征的分布表示,得到类中每个样本的均值与方差。Furthermore, the distribution learning module g φ in step S3 is composed of two fully connected layers to extract the distribution representation of image features and obtain the mean and variance of each sample in the class.
进一步地,步骤S3中最小化损失函数使用的是梯度下降算法,不断地调整权重ω和偏差b,使得损失函数的值变得越来越小。Furthermore, in step S3, the gradient descent algorithm is used to minimize the loss function, and the weight ω and the bias b are continuously adjusted so that the value of the loss function becomes smaller and smaller.
进一步地,步骤S3具体包括:Furthermore, step S3 specifically includes:
S31、将基类中的Dtrain输入嵌入模块fθ中,依次经过卷积层、池化层和激活函数,得每个类的样本特征图;S31, embed the D train input in the base class into the module f θ , and pass through the convolution layer, pooling layer and activation function in sequence to obtain the sample feature map of each class;
S32,计算出每个类样本特征图的均值μc和方差σc,与预训练样本特征的空间分布相比较,对嵌入模块fθ的参数加以调整,根据公式(1)和(2)计算每个类的均值μc和方差σc:S32, calculate the mean μ c and variance σ c of the feature map of each class sample, compare with the spatial distribution of the pre-training sample features, adjust the parameters of the embedding module f θ , and calculate the mean μ c and variance σ c of each class according to formulas (1) and (2):
式中xi表示为基类中C的第i个样本的特征向量,nc表示为C类中的样本总数;Where xi represents the feature vector of the i-th sample in the base class C, and nc represents the total number of samples in class C;
S33,将各类样本特征图输入到分布学习模块gφ,得到每个样本均值和方差利用高斯分布公式(3)计算每个样本xi类别概率:S33, input each type of sample feature map into the distribution learning module g φ to obtain the mean of each sample and variance The Gaussian distribution formula (3) is used to calculate the category probability of each sample xi :
式中∑c表示为C类特征的协方差矩阵,其计算公式如公式(4)所示:Where ∑ c represents the covariance matrix of C-type features, and its calculation formula is shown in formula (4):
S34,利用交叉熵公式最小化损失函数,优化分布学习模块gφ参数,公式如(5)所示:S34, using the cross entropy formula to minimize the loss function and optimize the distribution learning module gφ parameters, the formula is shown in (5):
式中y表示为一组带有标签的特征向量。Where y is represented as a set of labeled feature vectors.
进一步地,步骤4具体如下:Furthermore, step 4 is as follows:
S41,每个任务由支持集和查询集组成;S41, each task Supported by and queryset composition;
S42,将支持集输入嵌入模块fθ,得到每个类样本特征图的均值μc和方差σc;S42, will support the collection Input the embedding module f θ to obtain the mean μ c and variance σ c of the feature map of each class sample;
S43,将各类样本特征图输入到分布学习模块gφ,得到每个样本均值和方差 S43, input each type of sample feature map into the distribution learning module g φ to obtain the mean of each sample and variance
S44,根据每个样本的均值和方差利用公式(6)和(7)计算中每个类的分布原型和 S44, based on the mean of each sample and variance Using formulas (6) and (7) The distribution prototype of each class in and
式(6)中,Sc表示为支持集中的第C类,xi表示为支持集中的第C类中的样本,表示为样本xi的均值,μc表示为第C类的均值即第C个类的分布,式(6)整体表示为求第C类的加权调和平均数,以表示在模型中不同类的分布原型的位置,用以收紧类内关系和满足识别差距;In formula (6), Sc represents the support set The Cth class in the The samples in the Cth class in It is represented as the mean of sample xi , μc is represented as the mean of the Cth class, i.e., the distribution of the Cth class. The overall expression of formula (6) is to find the weighted harmonic mean of the Cth class, which is used to represent the position of the distribution prototypes of different classes in the model, in order to tighten the intra-class relationship and meet the recognition gap;
式(7)的目的是求类别C的方差,用以在足够的类别信息下消除单个数据的类无关表示,减少整体类别信息的幅度变化。The purpose of formula (7) is to find the variance of category C, so as to eliminate the class-independent representation of a single data under sufficient category information and reduce the amplitude variation of the overall category information.
进一步地,步骤S5具体为:Furthermore, step S5 is specifically as follows:
S51,计算基类样本数据中各个类的类别概率,公式如下:S51, calculate the class probability of each class in the base class sample data, the formula is as follows:
式中,表示为基类样本数据中类别C的均值与方差服从高斯分布,表示为支持集中的第C类的均值,Sd表示为将支持集第C类的分布作为输入,与基类样本数据中的第C类的分布相比较的距离集;In the formula, It is expressed as the mean and variance of category C in the base class sample data obeying Gaussian distribution, Support set The mean of the Cth class in the support set S d is represented by The distribution of the Cth class is used as input, and the distance set compared with the distribution of the Cth class in the base class sample data;
S52,选取最大的前n个类别,将n个类别的分布与当前类别的分布合并,公式如下:S52, select the largest first n categories, and merge the distribution of n categories with the distribution of the current category. The formula is as follows:
式中,topn(·)表示为一个从输入距离集Sd中选择顶部元素的操作符,SN用以存储关于特征向量最近的n个最近的基类样本数据;In the formula, topn(·) represents an operator that selects the top elements from the input distance set Sd , and S N is used to store the n nearest base class sample data closest to the feature vector;
S53,将合并后的类输入公式(6)和(7),得到矫正后的每个类的分布原型和 S53, input the merged classes into formulas (6) and (7) to obtain the corrected distribution prototype of each class and
式(6)中,Sc表示为支持集中的第C类,xi表示为支持集中的第C类中的样本,表示为样本xi的均值,μc表示为第C类的均值即第C个类的分布,式(6)整体表示为求第C类的加权调和平均数,以表示在模型中不同类的分布原型的位置,用以收紧类内关系和满足识别差距;In formula (6), Sc represents the support set The Cth class in the The samples in the Cth class in It is represented as the mean of sample xi , μc is represented as the mean of the Cth class, i.e., the distribution of the Cth class. The overall expression of formula (6) is to find the weighted harmonic mean of the Cth class, which is used to represent the position of the distribution prototypes of different classes in the model, in order to tighten the intra-class relationship and meet the recognition gap;
式(7)的目的是求类别C的方差,用以在足够的类别信息下消除单个数据的类无关表示,减少整体类别信息的幅度变化。The purpose of formula (7) is to find the variance of category C, so as to eliminate the class-independent representation of a single data under sufficient category information and reduce the amplitude variation of the overall category information.
进一步地,步骤S6具体为:Further, step S6 is specifically as follows:
S61,将新类数据的查询集的样本信息输入嵌入模块fθ,得到每个类样本特征图的均值μc和方差σc;S61, the query set of new class data The sample information is input into the embedding module f θ to obtain the mean μ c and variance σ c of the sample feature map of each class;
S62,将各类样本特征图输入到分布学习模块gφ,得到每个样本均值和方差 S62, input each type of sample feature map into the distribution learning module g φ to obtain the mean of each sample and variance
S63,将每个样本的均值和方差输入公式(3),计算出新类查询样本的预测概率,将其输入到度量模块,输出对应的类别标签:S63, the mean of each sample and variance Enter formula (3) to calculate the predicted probability of the new class query sample, input it into the measurement module, and output the corresponding category label:
式中∑c表示为C类特征的协方差矩阵,其计算公式如公式(4)所示:Where ∑ c represents the covariance matrix of C-type features, and its calculation formula is shown in formula (4):
另一方面,本发明还提供了一种面向小样本图像的任务自适应度量学习装置,用以实现上述的任一项方法,包括以下模块:On the other hand, the present invention also provides a task-adaptive metric learning device for small sample images, which is used to implement any of the above methods, and includes the following modules:
嵌入模块,用于对图像样本进行特征提取处理,构造特征空间,其中,所述图像样本包括基类样本、新类支持样本和查询样本;An embedding module, used for performing feature extraction processing on image samples and constructing a feature space, wherein the image samples include base class samples, new class support samples and query samples;
分布学习模块,用于提取图像特征的分布表示,得到类中每个样本的均值与方差;The distribution learning module is used to extract the distribution representation of image features and obtain the mean and variance of each sample in the class;
分布矫正模块,目的在于用基类样本的分布对新类样本进行分布矫正,构建图像分布矫正模型;The distribution correction module aims to use the distribution of base class samples to correct the distribution of new class samples and build an image distribution correction model;
度量模块,用于利用优化后的基类样本的分布对新类查询集样本进行分类,获取类别标签。The metric module is used to classify the new class query set samples using the optimized distribution of the base class samples and obtain the category labels.
与现有技术相比,本发明的有益效果为:Compared with the prior art, the present invention has the following beneficial effects:
本发明建立了一种基于特征分布迁移的小样本图像特征学习方法及装置,可以与任何分类器和特征提取器配对,无需额外的参数,解决了小样本图像分类中存在的原型偏差问题,改善了图像的分类效果,具有很高的实用价值。The present invention establishes a small sample image feature learning method and device based on feature distribution migration, which can be paired with any classifier and feature extractor without the need for additional parameters. It solves the prototype bias problem in small sample image classification, improves the image classification effect, and has high practical value.
本发明的装置在前期利用基类的数据结合梯度下降的方法,对嵌入模块以及分布学习模块的参数进行优化,后期进行分布矫正时,并不需要额外的参数设置;另外,通常假设特征表示中的每个维度都遵循高斯分布,这样高斯分布的均值和方差可以在类似的类别之间传递,减少偏差,以便这些类别的统计数据在足够的样本数下得到更好的估计,再利用分布矫正模型,对样本的分布进行矫正,从而更为精准的对新类样本进行分类。The device of the present invention uses the data of the base class in combination with the gradient descent method to optimize the parameters of the embedding module and the distribution learning module in the early stage, and no additional parameter setting is required when the distribution correction is performed in the later stage; in addition, it is usually assumed that each dimension in the feature representation follows a Gaussian distribution, so that the mean and variance of the Gaussian distribution can be transferred between similar categories to reduce deviations, so that the statistical data of these categories can be better estimated with a sufficient number of samples, and then the distribution correction model is used to correct the distribution of the samples, so as to more accurately classify new class samples.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明中记载的一些实施例,对于本领域普通技术人员来讲,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings required for use in the embodiments will be briefly introduced below. Obviously, the drawings described below are only some embodiments recorded in the present invention. For ordinary technicians in this field, other drawings can also be obtained based on these drawings.
图1为本发明实施例提供的基于特征分布迁移的小样本图像特征学习方法流程图。FIG1 is a flow chart of a small sample image feature learning method based on feature distribution migration provided in an embodiment of the present invention.
图2为本发明实施例提供的基于特征分布迁移的小样本图像特征学习模型的迁移学习和分布迁移的特征学习网络结构图。FIG2 is a diagram showing a feature learning network structure of transfer learning and distribution migration of a small sample image feature learning model based on feature distribution migration provided in an embodiment of the present invention.
图3为本发明实施例提供的分布矫正模块的流程图。FIG3 is a flow chart of a distribution correction module provided in an embodiment of the present invention.
图4为本发明实施例提供的基于特征分布迁移的小样本图像特征学习装置功能模块示意图。FIG4 is a schematic diagram of functional modules of a small sample image feature learning device based on feature distribution migration provided in an embodiment of the present invention.
具体实施方式DETAILED DESCRIPTION
下面结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本发明一部分实施例,而不是全部的实施例。本发明中的实施例,本领域技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The following is a clear and complete description of the technical solutions in the embodiments of the present invention in conjunction with the drawings in the embodiments of the present invention. Obviously, the described embodiments are part of the embodiments of the present invention, but not all of the embodiments. The embodiments of the present invention and all other embodiments obtained by those skilled in the art without creative work are within the scope of protection of the present invention.
根据本文公开的一个方面,提供了一种基于特征分布迁移的小样本图像特征学习方法,如图1所示,包括以下阶段步骤:According to one aspect disclosed herein, a small sample image feature learning method based on feature distribution migration is provided, as shown in FIG1 , including the following phase steps:
S1、对数据进行预处理,其中数据包括训练集和测试集;S1. Preprocess the data, where the data includes a training set and a test set;
具体地,步骤S1的预处理方法包括:Specifically, the preprocessing method of step S1 includes:
S11,将数据分为 S11, the data Divided into
和两部分,且这两部分的类别空间互斥,将Dtrain用以在训练过程中调整参数,Dtest作为新类数据测评模型性能; and The two parts have mutually exclusive category spaces. D train is used to adjust parameters during training, and D test is used as new category data to evaluate model performance.
S12,对于C-way K-shot分类任务,从Dtrain中随机选出C个类别,每个类别中随机选出M个样本,其中K个样本作为支持样本Si,其余M-K个样本作为查询样本Qi,Si和Qi构成一个任务Ti;同样地,对于Dtest有任务 S12, for the C-way K-shot classification task, randomly select C categories from D train , randomly select M samples from each category, K samples as support samples Si , and the remaining MK samples as query samples Qi . Si and Qi constitute a task Ti ; similarly, for D test, there is a task
S2,利用基类数据预训练嵌入模块fθ,得到良好的特征空间;S2, use the base class data to pre-train the embedding module f θ to obtain a good feature space;
进一步地,步骤S2中使用包含四个卷积块的嵌入模块fθ对图像提取特征,其中含有卷积层、池化层和非线性激活函数,每个卷积块使用窗口大小为3*3的卷积核,一个批量归一化,一个RELU非线性层,一个2×2最大池化层,裁剪了最后两个块的最大池化层。例如,对于84×84×3RGB图像,每个块使用一个带有64个滤波器的3x3的卷积核。Furthermore, in step S2, an embedding module f θ containing four convolutional blocks is used to extract features from the image, which contains convolutional layers, pooling layers and non-linear activation functions. Each convolutional block uses a convolution kernel with a window size of 3*3, a batch normalization, a RELU non-linear layer, a 2×2 maximum pooling layer, and the maximum pooling layers of the last two blocks are cropped. For example, for an 84×84×3 RGB image, each block uses a 3x3 convolution kernel with 64 filters.
S3,将Dtrain输入到嵌入模块fθ,得到样本特征图,将其输入到分布学习模块gφ中,最小化损失函数,优化分布学习模块gφ;S3, input D train into the embedding module f θ to obtain the sample feature map, input it into the distribution learning module g φ , minimize the loss function, and optimize the distribution learning module g φ ;
其中,分布学习模块gφ由两个全连接层组成,用以提取图像特征的分布表示,得到类中每个样本的均值与方差。Among them, the distribution learning module g φ consists of two fully connected layers, which is used to extract the distribution representation of image features and obtain the mean and variance of each sample in the class.
最小化损失函数使用的是梯度下降算法,不断地调整权重ω和偏差b,使得损失函数的值变得越来越小。也可以替换为随机梯度下降或者批量梯度下降等。The gradient descent algorithm is used to minimize the loss function, which continuously adjusts the weight ω and the deviation b to make the value of the loss function smaller and smaller. It can also be replaced by stochastic gradient descent or batch gradient descent.
具体地,步骤S3包括:Specifically, step S3 includes:
S31、将基类中的Dtrain输入嵌入模块fθ中,依次经过卷积层、池化层和激活函数,得每个类的样本特征图;S31, embed the D train input in the base class into the module f θ , and pass through the convolution layer, pooling layer and activation function in sequence to obtain the sample feature map of each class;
S32,计算出每个类样本特征图的均值μc和方差σc,与预训练样本特征的空间分布相比较,对嵌入模块fθ的参数加以调整,根据公式(1)和(2)计算每个类的均值μc和方差σc:S32, calculate the mean μ c and variance σ c of the feature map of each class sample, compare with the spatial distribution of the pre-training sample features, adjust the parameters of the embedding module f θ , and calculate the mean μ c and variance σ c of each class according to formulas (1) and (2):
式中xi表示为基类中C的第i个样本的特征向量,nc表示为C类中的样本总数;Where xi represents the feature vector of the i-th sample in the base class C, and nc represents the total number of samples in class C;
S33,将各类样本特征图输入到分布学习模块gφ,得到每个样本均值和方差利用高斯分布公式(3)计算每个样本xi类别概率:S33, input each type of sample feature map into the distribution learning module g φ to obtain the mean of each sample and variance The Gaussian distribution formula (3) is used to calculate the category probability of each sample xi :
式中∑c表示为C类特征的协方差矩阵,其计算公式如公式(4)所示:Where ∑ c represents the covariance matrix of C-type features, and its calculation formula is shown in formula (4):
S34,利用交叉熵公式最小化损失函数,优化分布学习模块gφ参数,公式如(5)所示:S34, using the cross entropy formula to minimize the loss function and optimize the distribution learning module gφ parameters, the formula is shown in (5):
式中y表示为一组带有标签的特征向量。Where y is represented as a set of labeled feature vectors.
S4,将支持集经过嵌入模块fθ和分布学习模块gφ计算每类的分布原型和 S4, will support the collection After the embedding module f θ and the distribution learning module g φ, the distribution prototype of each class is calculated and
具体地,步骤4包括:Specifically, step 4 includes:
S41,将新类数据分为支持集和查询集每个任务由支持集和查询集组成;S41, divide the new class data into support sets and queryset Each task Supported by and queryset composition;
S42,将支持集输入嵌入模块fθ,得到每个类样本特征图的均值μc和方差σc;S42, will support the collection Input the embedding module f θ to obtain the mean μ c and variance σ c of the feature map of each class sample;
S43,将各类样本特征图输入到分布学习模块gφ,得到每个样本均值和方差 S43, input each type of sample feature map into the distribution learning module g φ to obtain the mean of each sample and variance
S44,根据每个样本的均值和方差利用公式(6)和(7)计算中每个类的分布原型和 S44, based on the mean of each sample and variance Using formulas (6) and (7) The distribution prototype of each class in and
式(6)中,Sc表示为支持集中的第C类,xi表示为支持集中的第C类中的样本,表示为样本xi的均值,μc表示为第C类的均值即第C个类的分布,式(6)整体表示为求第C类的加权调和平均数,以表示在模型中不同类的分布原型的位置,用以收紧类内关系和满足识别差距;In formula (6), Sc represents the support set The Cth class in the The samples in the Cth class in It is represented as the mean of sample xi , μc is represented as the mean of the Cth class, i.e., the distribution of the Cth class. The overall expression of formula (6) is to find the weighted harmonic mean of the Cth class, which is used to represent the position of the distribution prototypes of different classes in the model, in order to tighten the intra-class relationship and meet the recognition gap;
式(7)的目的是求类别C的方差,用以在足够的类别信息下消除单个数据的类无关表示,减少整体类别信息的幅度变化。The purpose of formula (7) is to find the variance of category C, so as to eliminate the class-independent representation of a single data under sufficient category information and reduce the amplitude variation of the overall category information.
S5,计算基类数据中各类的类别概率,选取最大的前n个类别,将n个类别的分布与当前类别的分布合并,得到矫正后每类的分布原型和 S5, calculate the category probability of each category in the base category data, select the largest first n categories, merge the distribution of n categories with the distribution of the current category, and obtain the distribution prototype of each category after correction and
具体地,步骤S5包括:Specifically, step S5 includes:
S51,计算基类样本数据中各个类的类别概率,公式如下:S51, calculate the class probability of each class in the base class sample data, the formula is as follows:
式中,表示为基类样本数据中类别C的均值与方差服从高斯分布,表示为支持集中的第C类的均值,Sd表示为将支持集第C类的分布作为输入,与基类样本数据中的第C类的分布相比较的距离集;In the formula, It is expressed as the mean and variance of category C in the base class sample data obeying Gaussian distribution, Support set The mean of the Cth class in the support set S d is represented by The distribution of the Cth class is used as input, and the distance set compared with the distribution of the Cth class in the base class sample data;
S52,选取最大的前n个类别,将n个类别的分布与当前类别的分布合并,公式如下:S52, select the largest first n categories, and merge the distribution of n categories with the distribution of the current category. The formula is as follows:
式中,topn(·)表示为一个从输入距离集Sd中选择顶部元素的操作符,SN用以存储关于特征向量最近的n个最近的基类样本数据;In the formula, topn(·) represents an operator that selects the top elements from the input distance set Sd , and S N is used to store the n nearest base class sample data closest to the feature vector;
S53,将合并后的类输入公式(6)和(7),得到矫正后的每个类的分布原型和 S53, input the merged classes into formulas (6) and (7) to obtain the corrected distribution prototype of each class and
式(6)中,Sc表示为支持集中的第C类,xi表示为支持集中的第C类中的样本,表示为样本xi的均值,μc表示为第C类的均值即第C个类的分布,式(6)整体表示为求第C类的加权调和平均数,以表示在模型中不同类的分布原型的位置,用以收紧类内关系和满足识别差距;In formula (6), Sc represents the support set The Cth class in the The samples in the Cth class in It is represented as the mean of sample xi , μc is represented as the mean of the Cth class, i.e., the distribution of the Cth class. The overall expression of formula (6) is to find the weighted harmonic mean of the Cth class, which is used to represent the position of the distribution prototypes of different classes in the model, in order to tighten the intra-class relationship and meet the recognition gap;
式(7)的目的是求类别C的方差,用以在足够的类别信息下消除单个数据的类无关表示,减少整体类别信息的幅度变化。The purpose of formula (7) is to find the variance of category C, so as to eliminate the class-independent representation of a single data under sufficient category information and reduce the amplitude variation of the overall category information.
S6,计算新类查询样本的预测概率,输出类别标签。S6, calculates the predicted probability of the new class query sample and outputs the category label.
具体地,步骤S6包括:Specifically, step S6 includes:
S61,将新类数据的查询集的样本信息输入嵌入模块fθ,得到每个类样本特征图的均值μc和方差σc;S61, the query set of new class data The sample information is input into the embedding module f θ to obtain the mean μ c and variance σ c of the sample feature map of each class;
S62,将各类样本特征图输入到分布学习模块gφ,得到每个样本均值和方差 S62, input each type of sample feature map into the distribution learning module g φ to obtain the mean of each sample and variance
S63,将每个样本的均值和方差输入公式(3),计算出新类查询样本的预测概率,将其输入到度量模块,输出对应的类别标签:S63, the mean of each sample and variance Enter formula (3) to calculate the predicted probability of the new class query sample, input it into the measurement module, and output the corresponding category label:
式中∑c表示为C类特征的协方差矩阵,其计算公式如公式(4)所示:Where ∑ c represents the covariance matrix of C-type features, and its calculation formula is shown in formula (4):
另一方面,本发明还提供了一种面向小样本图像的任务自适应度量学习装置,用以实现上述的任一项方法,包括以下模块:On the other hand, the present invention also provides a task-adaptive metric learning device for small sample images, which is used to implement any of the above methods, and includes the following modules:
嵌入模块,用于对图像样本进行特征提取处理,构造特征空间,其中,所述图像样本包括基类样本、新类支持样本和查询样本;An embedding module, used for performing feature extraction processing on image samples and constructing a feature space, wherein the image samples include base class samples, new class support samples and query samples;
分布学习模块,用于提取图像特征的分布表示,得到类中每个样本的均值与方差;The distribution learning module is used to extract the distribution representation of image features and obtain the mean and variance of each sample in the class;
分布矫正模块,目的在于用基类样本的分布对新类样本进行分布矫正,构建图像分布矫正模型;The distribution correction module aims to use the distribution of base class samples to correct the distribution of new class samples and build an image distribution correction model;
度量模块,用于利用优化后的基类样本的分布对新类查询集样本进行分类,获取类别标签。The metric module is used to classify the new class query set samples using the optimized distribution of the base class samples and obtain the category labels.
本发明基于特征分布迁移的小样本图像特征学习方法,建立在现成的预训练特征抽取器和分类模型之上,可以与任何分类器和特征提取器配对,无需额外的参数;基于特征分布迁移的小样本图像特征学习方法装置中,采用分布学习模块,通常假设特征表示中的每个维度都遵循高斯分布,这样高斯分布的均值和方差可以在类似的类别之间传递,通过计算每个样本的均值和方差,计算出对应类别的加权调和平均数,用以表示模型中不同类别的分布原型的位置,比较与每个类别代表的距离,学会了对样本进行分类。目的在于通过融合与查询样本相连样本特征及它们之间距离的相似性,更新查询样本并分类。The present invention is based on a small sample image feature learning method of feature distribution migration, which is built on the existing pre-trained feature extractor and classification model, and can be paired with any classifier and feature extractor without additional parameters; in the small sample image feature learning method device based on feature distribution migration, a distribution learning module is adopted, and it is usually assumed that each dimension in the feature representation follows a Gaussian distribution, so that the mean and variance of the Gaussian distribution can be transferred between similar categories, and the weighted harmonic mean of the corresponding category is calculated by calculating the mean and variance of each sample, which is used to represent the position of the distribution prototype of different categories in the model, and the distance from each category representative is compared, so that the samples are learned to be classified. The purpose is to update and classify the query sample by fusing the similarity of the sample features connected to the query sample and the distance between them.
以上结合附图对所提出的基于特征分布迁移的小样本图像特征学习方法及模型的具体实施方式进行了阐述。通过以上实施方式的描述,所属领域的技术人员可以清楚的了解该方法以及装置的实施。The specific implementation of the proposed small sample image feature learning method and model based on feature distribution migration is described above in conjunction with the accompanying drawings. Through the description of the above implementation, those skilled in the art can clearly understand the implementation of the method and device.
需要说明的是,在附图和说明书正文中,未描述的实现方式,均为所属技术领域中普通技术人员所知的形式,未进行详细说明。此外,上述对各元件和方法的定义并不仅限于实例中提到的各种具体结构、形状或方式,本领域普通技术人员可对其进行简单地更改或替换。It should be noted that the implementation methods not described in the drawings and the main body of the specification are all forms known to ordinary technicians in the relevant technical field and are not described in detail. In addition, the above definitions of various elements and methods are not limited to the various specific structures, shapes or methods mentioned in the examples, and ordinary technicians in the field can simply change or replace them.
此外,除非特别描述或必须依序发生地步骤,上述步骤地顺序并无限制于以上所列,且可根据所需设计而变化或重新安排。并且上述实例可基于设计及可靠度地考虑,彼此混合搭配使用或与其他实例混合搭配使用,即不同实施中的技术特征可以自由组合形成更多地实施例子。在此提供的算法和显示不与任何特定计算机、虚拟系统或者其他设备固有相关。各种通用系统也可以与基于在此地启示一起使用。根据上面的描述,构造这类系统所要求的结构是显而易见的。此外,本文公开的也不针对任何特定的编程语言。但是应当了解,可以利用各种编程语言实现在此描述的本文公开的内容,并且上面对特定语言所做的描述是为了披露本文公开的最佳实施方式。In addition, unless the steps are specifically described or must occur in sequence, the order of the above steps is not limited to the above listed, and can be changed or rearranged according to the desired design. And the above examples can be mixed and matched with each other or with other examples based on design and reliability considerations, that is, the technical features in different implementations can be freely combined to form more implementation examples. The algorithms and displays provided herein are not inherently related to any specific computer, virtual system or other device. Various general systems can also be used together with the revelations based on this. According to the above description, the structure required to construct such a system is obvious. In addition, what is disclosed herein is not directed to any specific programming language. However, it should be understood that various programming languages can be used to implement the content disclosed herein described herein, and the above description of specific languages is to disclose the best implementation method disclosed herein.
类似的,应当理解,为了使本文尽量精简并且帮助理解各个公开方面中的一个或多个,在上面对本文公开的示例性实施例的描述中,本文公开的各个特征有时被一起分组到单个实施例、图、或者对其的描述中。然而,并不应将该公开的方法解释成反映如下示意图:即要求所保护的本文公开的要求比在每个权力要求中所明确记载的特征具有更多的特征。更确切地说,如下面的权力要求书所反映的那样,公开方面在于少于前面公开的单个实施例的所有特征。因此,遵循具体实施方式的权利要求书由此明确地并入该具体实施方式,其中每个权利要求本身都作为本公开的单独实施例子。Similarly, it should be understood that in order to make this document as concise as possible and to aid in understanding one or more of the various disclosed aspects, in the above description of the exemplary embodiments disclosed herein, the various features disclosed herein are sometimes grouped together into a single embodiment, figure, or description thereof. However, the disclosed method should not be interpreted as reflecting the following schematic diagram: the requirements disclosed herein that are claimed to be protected have more features than the features explicitly recorded in each claim. More specifically, as reflected in the claims below, the disclosed aspects are less than all the features of the single embodiment disclosed above. Therefore, the claims that follow the specific embodiment are hereby expressly incorporated into the specific embodiment, wherein each claim itself serves as a separate embodiment of the present disclosure.
以上所述实施例,仅为本申请的具体实施方式,用以说明本申请的技术方案,而非对其限制,本申请的保护范围并不局限于此,尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:任何熟悉本技术领域的技术人员在本申请揭露的技术范围内,其依然可以对前述实施例所记载的技术方案进行修改或可轻易想到变化,或者对其中部分技术特殊进行等同替换;而这些修改、变化或者替换,并不使相应技术方案的本质脱离本申请实施例技术方案的精神和范围。都应涵盖在本申请的保护范围之内。因此,本申请的保护范围应所述以权利要求的保护范围为准。The above-described embodiments are only specific implementation methods of the present application, which are used to illustrate the technical solutions of the present application, rather than to limit them. The protection scope of the present application is not limited thereto. Although the present application is described in detail with reference to the aforementioned embodiments, ordinary technicians in the field should understand that any technician familiar with the technical field can still modify the technical solutions recorded in the aforementioned embodiments within the technical scope disclosed in the present application, or can easily think of changes, or make equivalent replacements for some of the technical specialties therein; and these modifications, changes or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the embodiments of the present application. They should all be included in the protection scope of the present application. Therefore, the protection scope of the present application should be based on the protection scope of the claims.
Claims (9)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210487387.2A CN114782779B (en) | 2022-05-06 | 2022-05-06 | Small sample image feature learning method and device based on feature distribution migration |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210487387.2A CN114782779B (en) | 2022-05-06 | 2022-05-06 | Small sample image feature learning method and device based on feature distribution migration |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114782779A CN114782779A (en) | 2022-07-22 |
CN114782779B true CN114782779B (en) | 2023-06-02 |
Family
ID=82436001
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210487387.2A Active CN114782779B (en) | 2022-05-06 | 2022-05-06 | Small sample image feature learning method and device based on feature distribution migration |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114782779B (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101750611A (en) * | 2009-12-02 | 2010-06-23 | 哈尔滨工程大学 | Underwater robot object detection device and detection method |
CN105469111A (en) * | 2015-11-19 | 2016-04-06 | 浙江大学 | Small sample set object classification method on basis of improved MFA and transfer learning |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021253226A1 (en) * | 2020-06-16 | 2021-12-23 | Huawei Technologies Co., Ltd. | Learning proxy mixtures for few-shot classification |
CN111858991A (en) * | 2020-08-06 | 2020-10-30 | 南京大学 | A Few-Sample Learning Algorithm Based on Covariance Metrics |
CN113610151B (en) * | 2021-08-05 | 2022-05-03 | 哈尔滨理工大学 | Small sample image classification system based on prototype network and self-encoder |
CN114328921B (en) * | 2021-12-27 | 2024-04-09 | 湖南大学 | Small sample entity relation extraction method based on distribution calibration |
CN114333027B (en) * | 2021-12-31 | 2024-05-14 | 之江实验室 | Cross-domain novel facial expression recognition method based on combined and alternate learning frames |
CN114387474A (en) * | 2022-01-12 | 2022-04-22 | 南通大学 | Small sample image classification method based on Gaussian prototype classifier |
CN114387473A (en) * | 2022-01-12 | 2022-04-22 | 南通大学 | Small sample image classification method based on base class sample characteristic synthesis |
-
2022
- 2022-05-06 CN CN202210487387.2A patent/CN114782779B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101750611A (en) * | 2009-12-02 | 2010-06-23 | 哈尔滨工程大学 | Underwater robot object detection device and detection method |
CN105469111A (en) * | 2015-11-19 | 2016-04-06 | 浙江大学 | Small sample set object classification method on basis of improved MFA and transfer learning |
Also Published As
Publication number | Publication date |
---|---|
CN114782779A (en) | 2022-07-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Thai et al. | Image classification using support vector machine and artificial neural network | |
CN117597703B (en) | Multi-scale converter for image analysis | |
CN110879982B (en) | A crowd counting system and method | |
JP7028322B2 (en) | Information processing equipment, information processing methods and information processing programs | |
CN115690522B (en) | An object detection method based on multi-pooling fusion channel attention and its application | |
WO2024066111A1 (en) | Image processing model training method and apparatus, image processing method and apparatus, and device and medium | |
CN109840560A (en) | Based on the image classification method for incorporating cluster in capsule network | |
CN110826534A (en) | Face key point detection method and system based on local principal component analysis | |
WO2020183267A1 (en) | Image search method and image search system | |
CN114492634B (en) | Fine granularity equipment picture classification and identification method and system | |
WO2011096010A1 (en) | Pattern recognition device | |
CN114782779B (en) | Small sample image feature learning method and device based on feature distribution migration | |
CN110211122A (en) | A kind of detection image processing method and processing device | |
CN111242839A (en) | Image scaling and cutting method based on scale grade | |
CN118351594A (en) | Motion gesture recognition method based on machine learning | |
CN118430054A (en) | A face recognition method and system based on AI intelligence | |
CN117934970A (en) | Image classifier optimization method based on local features and position attention | |
CN109583406B (en) | Facial Expression Recognition Method Based on Feature Attention Mechanism | |
CN114723998B (en) | Small sample image classification method and device based on large-boundary Bayesian prototype learning | |
CN116704610A (en) | Behavior classification method, apparatus, computer device, storage medium, and program product | |
CN117557857A (en) | Detection network light weight method combining progressive guided distillation and structural reconstruction | |
CN116503398A (en) | Insulator pollution flashover detection method and device, electronic equipment and storage medium | |
Lee et al. | Ensemble Algorithm of Convolution Neural Networks for Enhancing Facial Expression Recognition | |
CN112699898B (en) | A method for image orientation recognition based on multi-layer feature fusion | |
CN114596433A (en) | Insulator identification method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |