CN111611972A - Recognition method of crop leaf species based on multi-view and multi-task ensemble learning - Google Patents

Recognition method of crop leaf species based on multi-view and multi-task ensemble learning Download PDF

Info

Publication number
CN111611972A
CN111611972A CN202010485899.6A CN202010485899A CN111611972A CN 111611972 A CN111611972 A CN 111611972A CN 202010485899 A CN202010485899 A CN 202010485899A CN 111611972 A CN111611972 A CN 111611972A
Authority
CN
China
Prior art keywords
learning
model
view
task
models
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010485899.6A
Other languages
Chinese (zh)
Other versions
CN111611972B (en
Inventor
田青
梅承
孙灏铖
张恒
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN202010485899.6A priority Critical patent/CN111611972B/en
Publication of CN111611972A publication Critical patent/CN111611972A/en
Application granted granted Critical
Publication of CN111611972B publication Critical patent/CN111611972B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

本发明涉及一种基于多视图多任务集成学习的作物叶片种类识别方法,该方法选取叶片图像作为原始数据集,并进行特征提取,得到若干视图下的数据集;利用CNN模型作为基学习器,对若干视图下的数据集与原始数据集分别进行单独的集成学习;然后固定所有基学习器的参数,并去除掉基学习器中全连接分类器的最后一层,然后将所有模型的输出拼接起来,并添加新的分类器,对若干视图进行联合特征选择,使得其验证集准确率达到期望,得到多个视图下的模型;再利用多任务学习,对叶片种类进行识别。本发明强化了模型的准确度和泛化能力,整体上解决了传统深度学习模型训练数据不足,模型简单地堆叠深度导致的泛化能力弱的问题。The invention relates to a method for identifying crop leaf species based on multi-view and multi-task integrated learning. The method selects a leaf image as an original data set, and performs feature extraction to obtain data sets under several views; using a CNN model as a basic learner, Perform separate ensemble learning on the datasets under several views and the original datasets; then fix the parameters of all basic learners, remove the last layer of the fully connected classifier in the basic learners, and then splicing the outputs of all models Then, a new classifier is added, and joint feature selection is performed on several views, so that the accuracy of the validation set reaches the expectation, and the model under multiple views is obtained; and then the multi-task learning is used to identify the leaf species. The invention strengthens the accuracy and generalization ability of the model, and solves the problem of insufficient generalization ability caused by insufficient training data of the traditional deep learning model and simple stacking of the model.

Description

基于多视图多任务集成学习的作物叶片种类识别方法Recognition method of crop leaf species based on multi-view and multi-task ensemble learning

技术领域technical field

本发明属于人工智能领域,在传统深度学习模型的基础上提出一种提高识别作物叶片及其存在病害效果的改进方法。The invention belongs to the field of artificial intelligence, and provides an improved method for improving the effect of identifying crop leaves and their existing diseases on the basis of a traditional deep learning model.

背景技术Background technique

当下粮食安全问题日益严峻。有许多因素正威胁着粮食安全,其中植物病害对全球范围的粮食安全构成严重威胁。以往对农作物病害的识别大多采用人工方式,但人工识别存在诸多不足。随着精准农业的兴起,运用信息技术辅助农业生产为农作物病害的识别提供了新思路,图像处理技术就是其中之一,其对农作物病害识别具有传统方法所不具备的各种优点,即实时性强、速度快、误判率低,甚至还可以及时提供防治病害传播的必要方法。The current food security problem is becoming more and more serious. Many factors are threatening food security, among which plant diseases pose a serious threat to food security on a global scale. In the past, the identification of crop diseases was mostly manual, but there are many shortcomings in manual identification. With the rise of precision agriculture, the use of information technology to assist agricultural production provides new ideas for the identification of crop diseases. Image processing technology is one of them. It has various advantages for crop disease identification that traditional methods do not have, that is, real-time performance. It is strong, fast, and has a low misjudgment rate, and it can even provide the necessary methods to prevent and control the spread of diseases in time.

目前通过图像识别农作物病害的难点主要在于图像分割、特征提取与分类识别。At present, the difficulty of identifying crop diseases through images mainly lies in image segmentation, feature extraction and classification recognition.

解决这些难点的主要方法有阈值分割法、边缘检测法、数学形态学法、支持向量机法与模糊聚类法等。尽管这些方法已经取得了很好的分类效果,但这些方法采用了传统的机器学习方法,通过使低级视觉特征与多种算法的结合来识别病害。这就导致它们也有一些局限性。The main methods to solve these difficulties include threshold segmentation method, edge detection method, mathematical morphology method, support vector machine method and fuzzy clustering method. Although these methods have achieved good classification results, these methods employ traditional machine learning methods to identify diseases by combining low-level visual features with multiple algorithms. This makes them also have some limitations.

其中阈值分割法的特点是简单,执行效率高,但阈值的选取,作物病虫害区域的颜色、纹理等特征往往与非病害区域有着较大的差别。而边缘检测法的分割效率依赖于边缘检测算子,鲁棒性较差。数学形态学法的缺点则在于由各种几何基元的并集、交集和差集构成的目标与人类对形状的感觉有一定的差异。糊聚类法收敛速度慢、必须先确定分类数等局限件,支持向量机法的性能又过于依赖核函数和对样本的训练速度。此外以上方法提取特征时间点过于单一,大多数的特征提取都是在农作物病虫害症状十分明显时才进行,严重影响了实时性,不能做到早识别、早防治。而且对噪声和初始化数据的敏感这又导致分割精度产生的影响在具有复杂生长环境的农作物病害图像分割中尤为突出,例如当图像背景复杂或叶片呈粉状时,识别工作将很困难。同时,它们大多依赖手工制作的特征,无法解决语义问题间隙。Among them, the threshold segmentation method is characterized by simplicity and high execution efficiency, but the selection of the threshold, the color, texture and other characteristics of crop disease and insect pest areas are often quite different from non-disease areas. The segmentation efficiency of the edge detection method depends on the edge detection operator, and the robustness is poor. The disadvantage of mathematical morphology is that the object formed by the union, intersection and difference of various geometric primitives is different from the human sense of shape. The fuzzy clustering method has a slow convergence speed and must first determine the number of classifications and other limitations. The performance of the support vector machine method depends too much on the kernel function and the training speed of the samples. In addition, the above methods extract feature time points too single, and most feature extraction is performed when the symptoms of crop diseases and insect pests are very obvious, which seriously affects the real-time performance, and cannot achieve early identification and early control. Moreover, it is sensitive to noise and initialization data, which in turn leads to the impact of segmentation accuracy, which is particularly prominent in the segmentation of crop disease images with complex growth environments. For example, when the image background is complex or the leaves are powdery, the recognition work will be difficult. At the same time, they mostly rely on hand-crafted features and cannot resolve the semantic gap.

与它们相比CNN作为一个深度学习模型可以从数据中自动发现越来越高层次的特征,并且在许多不同领域都取得了显著的成功。尤其是在在图像识别领域,CNN在学习数据充足时有稳定的表现。对于一般的大规模图像分类问题,卷积神经网络可用于构建阶层分类器(hierarchical classifier),也可以在精细分类识别(fine-grained recognition)中用于提取图像的判别特征以供其它分类器进行学习。对于后者,特征提取可以人为地将图像的不同部分分别输入卷积神经网络,也可以由卷积神经网络通过无监督学习自行提取。这些资料均表明CNN在图像识别领域取得了巨大的成功,因此,近年来,有许多研究者利用CNN方法进行植物病害诊断。Compared to them, CNN as a deep learning model can automatically discover increasingly higher-level features from data, and has achieved remarkable success in many different fields. Especially in the field of image recognition, CNN has a stable performance when the learning data is sufficient. For general large-scale image classification problems, convolutional neural networks can be used to construct hierarchical classifiers, and can also be used to extract discriminative features of images in fine-grained recognition for other classifiers. study. For the latter, feature extraction can artificially input different parts of the image into the convolutional neural network, or it can be extracted by the convolutional neural network itself through unsupervised learning. These data all show that CNN has achieved great success in the field of image recognition. Therefore, in recent years, many researchers have used CNN methods for plant disease diagnosis.

目前,虽然基于深度学习的图像识别方法在准确率上已经强于很多传统算法,但大量的作物识别的模型依旧面对模型泛化能力一般的问题,其主要原因有二:一是数据集规模的限制。作物病害图片的人工获取和人工标注标签耗时耗力,导致可供模型训练的数据较少,传统解决方法是使用数据增强扩充数据集,但提高模型泛化能力的程度十分有限。二是大量的作物识别的模型只是简单的使用卷积神经网络与全连接层的简单堆叠,且都只是在单视图的思路下实现,然而,物体的不同视图描述了物体的不同特性,这一缺陷导致模型本身泛化能力不强。At present, although the image recognition method based on deep learning has better accuracy than many traditional algorithms, a large number of crop recognition models still face the problem of generalization ability of the model. There are two main reasons: First, the scale of the data set limits. The manual acquisition and manual labeling of crop disease pictures is time-consuming and labor-intensive, resulting in less data for model training. The traditional solution is to use data augmentation to expand the data set, but the degree of improving the generalization ability of the model is very limited. Second, a large number of crop recognition models simply use a simple stack of convolutional neural networks and fully-connected layers, and are only implemented under the idea of a single view. However, different views of an object describe different characteristics of the object. This The defect leads to the poor generalization ability of the model itself.

发明内容SUMMARY OF THE INVENTION

本发明为了解决现有技术中存在的问题,提供一种模型精度高且泛化能力强的基于多视图多任务集成学习的作物叶片种类识别方法。In order to solve the problems existing in the prior art, the present invention provides a method for identifying crop leaf species based on multi-view and multi-task integrated learning with high model accuracy and strong generalization ability.

为了达到上述目的,本发明提出的技术方案为:一种基于多视图多任务集成学习的作物叶片种类识别方法,In order to achieve the above purpose, the technical solution proposed by the present invention is: a method for identifying crop leaf species based on multi-view and multi-task integrated learning,

首先,选取叶片图像作为原始数据集,并对原始数据集进行特征提取,得到若干单一视图下的数据集;First, the leaf image is selected as the original data set, and feature extraction is performed on the original data set to obtain several data sets under a single view;

然后,利用CNN模型作为基学习器,对若干单一视图下的数据集与原始数据集分别进行单独的集成学习;Then, using the CNN model as the base learner, separate ensemble learning is performed on several single-view datasets and original datasets;

单独集成学习完成后固定所有基学习器的参数,并去除掉所有基学习器中全连接分类器的最后一层,然后将所有CNN模型的输出拼接起来,并添加新的全连接分类器,对若干视图进行联合特征选择,使得其验证集准确率达到期望值,得到多个视图下的模型,完成多视图集成学习;After the individual ensemble learning is completed, the parameters of all base learners are fixed, and the last layer of the fully connected classifier in all base learners is removed, and then the outputs of all CNN models are spliced together, and a new fully connected classifier is added. Joint feature selection is performed on several views, so that the accuracy of the validation set reaches the expected value, and models under multiple views are obtained to complete multi-view ensemble learning;

再利用多任务学习,共享不同任务已学到的特征表示,对叶片种类进行识别。Then, multi-task learning is used to share the feature representations learned by different tasks to identify leaf species.

对上述技术方案的进一步设计为:所述基学习器的选取方法为:选取深度学习图像识别模型组成模型族,然后给模型族中的每个模型编号,再在这些模型中随机选择一定数量的模型来作为一个单一视图下数据集的所有基学习器;对于原始数据集,使用模型族中所有模型作为基学习器。The further design of the above-mentioned technical scheme is: the selection method of the basic learner is: select a deep learning image recognition model to form a model family, then number each model in the model family, and then randomly select a certain number of these models. The model is used as all base learners of the dataset under a single view; for the original dataset, all models in the model family are used as base learners.

所述深度学习图像识别模型包括GoogleNet,VGG,Resnet等。The deep learning image recognition models include GoogleNet, VGG, Resnet and so on.

所述多视图集成学习中的损失函数为:The loss function in the multi-view ensemble learning is:

Figure BDA0002519221270000021
Figure BDA0002519221270000021

本发明的技术方案与现有技术相比,具有的有益效果为:Compared with the prior art, the technical solution of the present invention has the following beneficial effects:

本发明采用了CNN集成学习,将多种成熟的CNN模型作为基学习器。基于此,利用多视图学习,针对作物叶片的边缘,灰度,纹理,质地和原始图片五个视图进行训练模型。最后又利用多任务学习,共享不同任务已学到的特征表示。强化了模型的准确度和泛化能力,整体上解决了传统深度学习模型训练数据不足,模型简单地堆叠深度导致的泛化能力弱的问题。The invention adopts CNN integrated learning, and uses a variety of mature CNN models as basic learners. Based on this, the multi-view learning is used to train the model for the five views of the edge, grayscale, texture, texture and original picture of the crop leaf. Finally, multi-task learning is used to share the feature representations learned by different tasks. The accuracy and generalization ability of the model are strengthened, and the problem of insufficient generalization ability caused by the lack of training data of traditional deep learning models and the simple stacking depth of the model is generally solved.

本专利采用了多任务学习,通过在所有任务之间共享隐藏层,同时保留特定任务的输出层来实现隐层参数的硬共享,使多个任务在并行训练中共享不同任务已学到的特征表示,降低了过拟合的风险。This patent adopts multi-task learning, which realizes the hard sharing of hidden layer parameters by sharing the hidden layer among all tasks while retaining the output layer of a specific task, so that multiple tasks can share the learned features of different tasks in parallel training. representation, reducing the risk of overfitting.

附图说明Description of drawings

图1为本发明的模型设计图;Fig. 1 is the model design drawing of the present invention;

图2为本发明的模型训练流程图;Fig. 2 is the model training flow chart of the present invention;

图3为VGG模型结构图;Figure 3 is a structural diagram of the VGG model;

图4为按灰度提取后的图片的集成学习模型;Fig. 4 is the integrated learning model of the picture after the grayscale extraction;

图5为按纹理提取的图片;Figure 5 is a picture extracted by texture;

图6为多视图学习结构图;Figure 6 is a multi-view learning structure diagram;

图7为多任务学习分类器。Figure 7 is a multi-task learning classifier.

具体实施方式Detailed ways

下面结合附图以及具体实施例对本发明进行详细说明。The present invention will be described in detail below with reference to the accompanying drawings and specific embodiments.

本实施例的基于多视图多任务集成学习的作物叶片种类识别方法的模型设计如图1所示,步骤一、取叶片图像作为原始数据集,并对原始数据集进行特征提取,得到若干视图下的数据集;利用CNN模型作为基学习器,对若干视图下的数据集与原始数据集分别进行单独的集成学习;The model design of the crop leaf species identification method based on multi-view and multi-task ensemble learning in this embodiment is shown in Figure 1. Step 1: Take the leaf image as the original data set, and perform feature extraction on the original data set to obtain several views. The data set; using the CNN model as the base learner, separate ensemble learning is performed on the data set under several views and the original data set;

原始数据集图片通过特定的卷积核可以得到不同视图下的图片,例如灰度,质地,边缘,纹理等(图5为展示了按纹理提取的新图片)。这些卷积核需要进行设计,如果我们想要得到图片纹理,我们可以设置卷积核的参数为(-1,0,1;-2,0,2;-1,0,1)。卷积核的参数可以根据具体任务的要求来设计。对原始数据集经过提取后,便可以得到若干不同视图下的数据集。The original dataset images can be obtained under different views through a specific convolution kernel, such as grayscale, texture, edge, texture, etc. (Figure 5 shows a new image extracted by texture). These convolution kernels need to be designed. If we want to get the image texture, we can set the parameters of the convolution kernel to (-1, 0, 1; -2, 0, 2; -1, 0, 1). The parameters of the convolution kernel can be designed according to the requirements of specific tasks. After the original dataset is extracted, several datasets under different views can be obtained.

这些不同视图的图片连同原始图片在不同的视图中进行单独的训练。视图定义如公式(1)所示;These pictures of different views are trained separately in different views along with the original pictures. The view definition is shown in formula (1);

Figure BDA0002519221270000031
Figure BDA0002519221270000031

其中,视图的总数目为1+|view|,v0表示没有通过特征提取的视图,即原始图片,而vi(1≤i≤|view|)表示通过第i(1≤i≤|view|)种特征提取方式提取后的视图。Among them, the total number of views is 1+|view|, v 0 represents the view that has not passed the feature extraction, that is, the original picture, and v i (1≤i≤|view|) means the i (1≤i≤|view) |) views extracted by feature extraction methods.

本实施例中模型训练过程如图2所示,集成学习的过程如下:The model training process in this embodiment is shown in Figure 2, and the process of ensemble learning is as follows:

首先,我们定义一个模型族,模型族定义如公式(2)所示,其中模型的总数为||。First, we define a model family, which is defined as formula (2), where the total number of models is ||.

Figure BDA0002519221270000041
Figure BDA0002519221270000041

这里我们会用一些成熟的深度学习图像识别模型,如GoogleNet,VGG,Resnet等组成模型族。例如VGG深度卷积神经网络,它探索了卷积神经网络的深度和其性能之间的关系,通过反复的堆叠3*3的小型卷积核和2*2的最大池化层,成功的构建了16~19层深的卷积神经网络。目前为止,VGGNet依然被用来提取图像的特征。VGGNet的网络结构如图3所示。VGGNet包含很多级别的网络,深度从11层到19层不等,比较常用的是VGGNet-16和VGGNet-19。VGGNet把网络分成了5段,每段都把多个3*3的卷积网络串联在一起,每段卷积后面接一个最大池化层,最后面是3个全连接层和一个softmax层。Here we will use some mature deep learning image recognition models, such as GoogleNet, VGG, Resnet, etc. to form a model family. For example, the VGG deep convolutional neural network, which explores the relationship between the depth of the convolutional neural network and its performance, is successfully constructed by repeatedly stacking 3*3 small convolution kernels and 2*2 max pooling layers Convolutional neural network with 16-19 layers deep. So far, VGGNet is still used to extract image features. The network structure of VGGNet is shown in Figure 3. VGGNet contains many levels of networks, ranging from 11 to 19 layers in depth. The more commonly used ones are VGGNet-16 and VGGNet-19. VGGNet divides the network into 5 segments, each segment connects multiple 3*3 convolutional networks together, each segment of convolution is followed by a maximum pooling layer, and the last is 3 fully connected layers and a softmax layer.

然后,给模型族中的每个模型编号,再在这些模型中随机选择一定数量的模型来作为一个视图下的所有基学习器。例如我们选取五个模型构成模型族,分别编号为1、2、3、4、5。例如,在灰度视图下,我们从模型族中选择三个模型作为灰度图像下的所有基学习器,假设随机选择为1、2、5,那么第1、2、5个模型将作为灰度图像的所有模型。在原始的视图下,我们将使用所有模型,而不进行随机选择。Then, each model in the model family is numbered, and a certain number of models are randomly selected among these models to serve as all base learners under a view. For example, we select five models to form a model family, numbered 1, 2, 3, 4, and 5 respectively. For example, in the grayscale view, we select three models from the model family as all the base learners under the grayscale image, assuming that 1, 2, and 5 are randomly selected, then the 1st, 2nd, and 5th models will be used as grayscale. All models of degree images. In the original view, we will use all models without random selection.

对于模型族中的单个模型来说,需要根据我们自己的数据集的类别和规模来设计新的全连接分类器。如在keras框架下,我们可以根据具体的数据集的规模来设置多个全连接层,前几层全连接层的激活函数可以设置为“relu”,这几层的作用是用于特征的学习;最后一层设为“softmax”,这一层的作用是用于计算图片属于每一个类别的概率;然后我们需要冻结单个模型的部分卷积基。一般在深度学习中,将一个或三个卷积层和一个池化层作为一组,然后若干组构成卷积基。这里的冻结操作通常是冻结除了最后一组外的所有卷积层和池化层。在keras框架下,我们可以将卷积层(池化层)的trainable参数设置为False来冻结卷积层(池化层)。接下来,根据具体的数据集(不同视图对应的图片)训练模型,使得验证集的准确率达到期望的效果。接着解冻卷积基的最后几组,通常可以解冻最后第三组和第二组。我们可以将卷积层(池化层)的trainable参数设置为True来解冻卷积层(池化层)。最后联合训练模型,使得验证集地准确率达到期望的效果。在训练结束后我们将冻结整个基学习器(单个模型)。For a single model in a family of models, a new fully connected classifier needs to be designed according to the class and size of our own dataset. For example, under the keras framework, we can set up multiple fully connected layers according to the scale of the specific data set. The activation function of the first several fully connected layers can be set to "relu", and these layers are used for feature learning. ; The last layer is set to "softmax", the role of this layer is to calculate the probability that the image belongs to each category; then we need to freeze part of the convolution base of a single model. Generally in deep learning, one or three convolutional layers and a pooling layer are used as a group, and then several groups constitute a convolutional base. The freezing operation here is usually freezing all convolutional and pooling layers except the last group. Under the keras framework, we can freeze the convolutional layer (pooling layer) by setting the trainable parameter of the convolutional layer (pooling layer) to False. Next, train the model according to the specific data set (images corresponding to different views), so that the accuracy of the validation set can achieve the desired effect. Then unfreeze the last few groups of the convolution base, usually the last third and second groups can be unfrozen. We can unfreeze the convolutional layer (pooling layer) by setting the trainable parameter of the convolutional layer (pooling layer) to True. Finally, jointly train the model so that the accuracy of the validation set achieves the desired effect. We freeze the entire base learner (single model) after training.

在单个视图下,我们去除掉所有的基学习器中全连接分类器的最后一层,然后将所有的模型的输出(去除了最后一层后,输出的是特征,而不是类别)拼接起来,并添加新的分类器,然后将多个基学习器进行集成学习,使得其验证集准确率达到期望的效果,这样就可以得到单个视图下的集成学习模型,如图4所示。Under a single view, we remove the last layer of the fully connected classifier in all base learners, and then splicing the outputs of all models (after removing the last layer, the output is the feature, not the category), And add a new classifier, and then perform ensemble learning on multiple base learners, so that the accuracy of the validation set can achieve the desired effect, so that the ensemble learning model under a single view can be obtained, as shown in Figure 4.

步骤二、单个视图中,集成学习的效果达到预期后我们会固定单个视图下的所有基学习器的参数。之后去除掉所有的单个视图中全连接分类器的最后一层,然后将所有的模型的输出拼接起来,并添加新的分类器,然后将多个视图进行联合特征选择,使得其验证集准确率达到期望的效果,这样就可以得到多个视图下的模型,完成多视图的集成学习。部分结构如图6所示。Step 2. In a single view, after the effect of ensemble learning is as expected, we will fix the parameters of all basic learners in a single view. Then remove the last layer of the fully connected classifier in all single views, then splicing the outputs of all models, adding a new classifier, and then performing joint feature selection on multiple views to make the validation set accuracy rate To achieve the desired effect, the model under multiple views can be obtained, and the integrated learning of multiple views can be completed. Part of the structure is shown in Figure 6.

多视图集成学习中涉及的损失函数如下:The loss functions involved in multi-view ensemble learning are as follows:

第i(0≤i≤|view|)个视图下的数据集我们定义为

Figure BDA0002519221270000057
其中
Figure BDA0002519221270000058
Figure BDA0002519221270000056
由于图片的维度不确定,所以第j个样本xij的维度中的height与width不是确定的数(之后会对其进行预处理,使得基学习器的所有输入样本维度都相同,但允许不同的基学习器的输入样本有所差异。),而标签的维度都相同,都为类别总数K。We define the dataset under the i (0≤i≤|view|) view as
Figure BDA0002519221270000057
in
Figure BDA0002519221270000058
Figure BDA0002519221270000056
Since the dimension of the picture is uncertain, the height and width in the dimension of the jth sample x ij are not definite numbers (it will be preprocessed later, so that all input samples of the basic learner have the same dimension, but allow different The input samples of the base learner are different.), and the dimensions of the labels are the same, which are the total number of categories K.

假设我们在第i个视图vi下,使用第t个基学习器mt,使用softmax进行多分类,损失函数使用多分类交叉熵,我们可以得到视图vi下的第j个样本在基学习器mt下属于第k类的概率为公式(3)Suppose we are in the i -th view vi, use the t-th base learner m t , use softmax for multi-classification, and the loss function uses multi-class cross-entropy, we can get the j-th sample under the view v i in the base learning The probability of belonging to the kth class under the device m t is formula (3)

Figure BDA0002519221270000051
Figure BDA0002519221270000051

因此第j个样本的损失函数为So the loss function for the jth sample is

Figure BDA0002519221270000052
Figure BDA0002519221270000052

那么我们可以得到在第i个视图vi下在第t个基学习器mt下的损失函数为Then we can get the loss function under the t-th base learner m t under the i-th view v i as

Figure BDA0002519221270000053
Figure BDA0002519221270000053

由于我们在视图v0下使用所有的基学习器,而在第i(1≤i≤|view|)个视图下集成学习的模型是随机选择的,随机选择的总数为p(p<|model|)。我们可以得到视图v0下选择的模型为mi,m2,…,m|modej|,第i(1≤i≤|view|)个视图下选择的模型为

Figure BDA0002519221270000059
我们在每个视图的每个模型下使用正则化约束
Figure BDA00025192212700000510
则我们可以得到多视图下的损失函数为Since we use all base learners under view v 0 , and the model learned by the ensemble under the i (1≤i≤|view|)th view is randomly selected, the total number of random selections is p(p<|model |). We can get that the model selected under the view v 0 is m i ,m 2 ,…,m |modej| , and the model selected under the i-th (1≤i≤|view|) view is
Figure BDA0002519221270000059
We use regularization constraints under each model per view
Figure BDA00025192212700000510
Then we can get the loss function under multi-view as

Figure BDA0002519221270000054
Figure BDA0002519221270000054

Figure BDA0002519221270000055
Figure BDA0002519221270000055

总体的损失函数为:The overall loss function is:

Figure BDA0002519221270000061
Figure BDA0002519221270000061

步骤三、利用多任务学习,共享不同任务已学到的特征表示,对叶片种类进行识别。Step 3: Use multi-task learning to share the feature representations learned by different tasks to identify leaf species.

多任务学习的核心思想是多个任务并行训练并共享不同任务已学到的特征表示,这样可以充分利用模型学习到的特征,也节省资源。如选定三个任务,分别为农作物的种类,病害的种类和病害的严重程度。在我们的网络中,我们将原始的全连接分类器的最后一层去除掉,然后在倒数第二层后添加图7所示的分类器,其中第一个任务用于农作物的种类的分类,在keras中可将最后一层的激活函数设为“softmax”,第二个任务用于病害的分类,在keras中可将最后一层的激活函数设为“softmax”,第三个用于预测病害的严重程度,在keras中可将最后一层的激活函数设为“mse”。然后将多视图学习后的模型参数固定,之后进行训练,使得其验证集准确率达到期望的效果,这样就可以实现多任务学习。多任务学习是一种归纳迁移方法,充分利用隐含在多个相关任务训练信号中的特定领域信息。在后向传播过程中,多任务学习允许共享隐层中专用于某个任务的特征被其他任务使用;多任务学习将可以学习到可适用于几个不同任务的特征,这样的特征在单任务学习网络中往往不容易学到。The core idea of multi-task learning is that multiple tasks are trained in parallel and share the feature representations learned by different tasks, which can make full use of the features learned by the model and save resources. If three tasks are selected, they are the type of crops, the type of disease and the severity of the disease. In our network, we remove the last layer of the original fully-connected classifier, and then add the classifier shown in Figure 7 after the penultimate layer, where the first task is to classify the types of crops, In keras, the activation function of the last layer can be set to "softmax", and the second task is used for disease classification. In keras, the activation function of the last layer can be set to "softmax", and the third task is used for prediction The severity of the disease. In keras, the activation function of the last layer can be set to "mse". Then, the model parameters after multi-view learning are fixed, and then the training is performed to make the accuracy of the validation set achieve the desired effect, so that multi-task learning can be realized. Multi-task learning is an inductive transfer method that exploits domain-specific information implicit in training signals for multiple related tasks. In the process of backward propagation, multi-task learning allows the features in the shared hidden layer that are dedicated to a certain task to be used by other tasks; multi-task learning will learn features that are applicable to several different tasks, and such features can be used in a single task. Learning in the network is often not easy to learn.

本发明的技术方案不局限于上述各实施例,凡采用等同替换方式得到的技术方案均落在本发明要求保护的范围内。The technical solutions of the present invention are not limited to the above-mentioned embodiments, and all technical solutions obtained by adopting equivalent replacement methods fall within the protection scope of the present invention.

Claims (4)

1. A crop leaf type identification method based on multi-view and multi-task ensemble learning is characterized in that:
selecting a leaf image as an original data set, and performing feature extraction on the original data set to obtain a plurality of data sets under a single view;
the CNN model is used as a base learner, and independent integrated learning is respectively carried out on a data set and an original data set under a plurality of single views;
after the independent integrated learning is finished, parameters of all the base learners are fixed, the last layer of all the fully connected classifiers in all the base learners is removed, then the outputs of all the CNN models are spliced, new fully connected classifiers are added, and the combined feature selection is carried out on a plurality of views, so that the accuracy of a verification set reaches an expected value, the models under a plurality of views are obtained, and the multi-view integrated learning is finished;
and (3) utilizing multi-task learning to share the learned feature representations of different tasks and identify the blade types.
2. The crop leaf type identification method based on multi-view and multi-task ensemble learning as claimed in claim 1, wherein the selection method of the base learner is as follows: selecting a deep learning image recognition model to form a model family, numbering each model in the model family, and randomly selecting a certain number of models from the models to serve as all base learners of a data set under a single view; for the raw data set, all models in the model family are used as basis learners.
3. The method for identifying the crop leaf types based on the multi-view and multi-task ensemble learning as claimed in claim 2, wherein: the deep learning image recognition model includes GoogleNet, VGG, Resnet, and the like.
4. The method for identifying the crop leaf types based on the multi-view and multi-task ensemble learning as claimed in claim 2, wherein the loss function in the multi-view ensemble learning is as follows:
Figure FDA0002519221260000011
CN202010485899.6A 2020-06-01 2020-06-01 Crop leaf type identification method based on multi-view multi-task integrated learning Active CN111611972B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010485899.6A CN111611972B (en) 2020-06-01 2020-06-01 Crop leaf type identification method based on multi-view multi-task integrated learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010485899.6A CN111611972B (en) 2020-06-01 2020-06-01 Crop leaf type identification method based on multi-view multi-task integrated learning

Publications (2)

Publication Number Publication Date
CN111611972A true CN111611972A (en) 2020-09-01
CN111611972B CN111611972B (en) 2024-01-05

Family

ID=72201702

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010485899.6A Active CN111611972B (en) 2020-06-01 2020-06-01 Crop leaf type identification method based on multi-view multi-task integrated learning

Country Status (1)

Country Link
CN (1) CN111611972B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112418219A (en) * 2020-11-24 2021-02-26 广东工业大学 Color and shape recognition method and related device for garment fabric pieces
CN112712106A (en) * 2020-12-07 2021-04-27 西安交通大学 Mechanical equipment health state identification method based on multi-view confrontation self-encoder
CN113191391A (en) * 2021-04-07 2021-07-30 浙江省交通运输科学研究院 Road disease classification method aiming at three-dimensional ground penetrating radar map
WO2023109319A1 (en) * 2021-12-14 2023-06-22 Ping An Technology (Shenzhen) Co., Ltd. Systems and methods for crop disease diagnosis
CN116523136A (en) * 2023-05-05 2023-08-01 中国自然资源航空物探遥感中心 Mineral resource space intelligent prediction method and device based on multi-model integrated learning

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845549A (en) * 2017-01-22 2017-06-13 珠海习悦信息技术有限公司 A kind of method and device of the scene based on multi-task learning and target identification
CN109508650A (en) * 2018-10-23 2019-03-22 浙江农林大学 A kind of wood recognition method based on transfer learning

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106845549A (en) * 2017-01-22 2017-06-13 珠海习悦信息技术有限公司 A kind of method and device of the scene based on multi-task learning and target identification
CN109508650A (en) * 2018-10-23 2019-03-22 浙江农林大学 A kind of wood recognition method based on transfer learning

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
HOUJEUNG HAN ET AL.: "MULTI-VIEW VISUAL SPEECH RECOGNITION BASED ON MULTI TASK LEARNING", 《IEEE》 *
何雪梅: "多视图聚类算法综述", 《软件导刊》 *
许景辉等: "基于迁移学习的卷积神经网络玉米病害图像识别", 《农业机械学报》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112418219A (en) * 2020-11-24 2021-02-26 广东工业大学 Color and shape recognition method and related device for garment fabric pieces
CN112712106A (en) * 2020-12-07 2021-04-27 西安交通大学 Mechanical equipment health state identification method based on multi-view confrontation self-encoder
CN112712106B (en) * 2020-12-07 2022-12-09 西安交通大学 Mechanical equipment health state identification method based on multi-view confrontation self-encoder
CN113191391A (en) * 2021-04-07 2021-07-30 浙江省交通运输科学研究院 Road disease classification method aiming at three-dimensional ground penetrating radar map
WO2023109319A1 (en) * 2021-12-14 2023-06-22 Ping An Technology (Shenzhen) Co., Ltd. Systems and methods for crop disease diagnosis
CN116523136A (en) * 2023-05-05 2023-08-01 中国自然资源航空物探遥感中心 Mineral resource space intelligent prediction method and device based on multi-model integrated learning

Also Published As

Publication number Publication date
CN111611972B (en) 2024-01-05

Similar Documents

Publication Publication Date Title
CN108717568B (en) A Method of Image Feature Extraction and Training Based on 3D Convolutional Neural Network
US11195051B2 (en) Method for person re-identification based on deep model with multi-loss fusion training strategy
CN111611972A (en) Recognition method of crop leaf species based on multi-view and multi-task ensemble learning
CN107292298B (en) Ox face recognition method based on convolutional neural networks and sorter model
CN113378632A (en) Unsupervised domain pedestrian re-identification algorithm based on pseudo label optimization
CN113469236A (en) Deep clustering image recognition system and method for self-label learning
CN108830326A (en) A kind of automatic division method and device of MRI image
CN108389211A (en) Based on the image partition method for improving whale Optimization of Fuzzy cluster
CN106778687A (en) Method for viewing points detecting based on local evaluation and global optimization
CN113032613B (en) A 3D model retrieval method based on interactive attention convolutional neural network
CN111488917A (en) Garbage image fine-grained classification method based on incremental learning
CN109063719A (en) A kind of image classification method of co-ordinative construction similitude and category information
CN111127423B (en) Rice pest and disease identification method based on CNN-BP neural network algorithm
CN110097060A (en) A kind of opener recognition methods towards trunk image
CN103279944A (en) Image division method based on biogeography optimization
CN114581451A (en) Scattering map neural network-based brain magnetic resonance image segmentation method
Chen et al. Deep convolutional network for citrus leaf diseases recognition
CN110766082A (en) Plant leaf disease and insect pest degree classification method based on transfer learning
CN109102019A (en) Image classification method based on HP-Net convolutional neural networks
Hu et al. Learning salient features for flower classification using convolutional neural network
Yu Research progress of crop disease image recognition based on wireless network communication and deep learning
Shukla et al. Plant disease detection and localization using GRADCAM
CN109934292A (en) A cost-sensitivity-assisted learning method for unbalanced polarimetric SAR terrain classification
CN115511838A (en) Plant disease high-precision identification method based on group intelligent optimization
CN112308002B (en) A single-stage deep learning network-based method for seabed biometrics identification and detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant