CN108564166A - Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link - Google Patents
Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link Download PDFInfo
- Publication number
- CN108564166A CN108564166A CN201810238288.4A CN201810238288A CN108564166A CN 108564166 A CN108564166 A CN 108564166A CN 201810238288 A CN201810238288 A CN 201810238288A CN 108564166 A CN108564166 A CN 108564166A
- Authority
- CN
- China
- Prior art keywords
- image
- layer
- network
- training
- size
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了基于带对称跨层连接的卷积神经网络半监督特征学习方法,包含如下步骤:生成无类标受损图像数据集;构建跨层连接卷积神经网络;预训练图像恢复神经网络;提取网络参数构建分类网络;训练分类网络。本发明利用对无类标图像数据的恢复任务,预训练神经网络,从而提高对有类标图像的分类效果,实现半监督特征学习。此外,通过在传统卷积神经网络自动编码器中加入对称跨层连接,使得网络更易优化,并增强网络中层特征抽象能力,使得无监督图像恢复任务得到的网络权重更易迁移于有监督学习任务。本发明实现了高效、准确的基于卷积神经网络的半监督学习方法,因此具有较高的实用价值。
The invention discloses a semi-supervised feature learning method based on a convolutional neural network with symmetrical cross-layer connections, comprising the following steps: generating a damaged image data set without class labels; constructing a cross-layer connection convolutional neural network; pre-training the image restoration neural network ; Extract network parameters to build a classification network; train the classification network. The invention utilizes the recovery task of image data without class labels to pre-train neural networks, thereby improving the classification effect of images with class labels and realizing semi-supervised feature learning. In addition, by adding symmetric cross-layer connections to the traditional convolutional neural network autoencoder, the network is easier to optimize, and the feature abstraction ability of the middle layer of the network is enhanced, so that the network weight obtained from the unsupervised image restoration task can be transferred to the supervised learning task more easily. The invention realizes an efficient and accurate semi-supervised learning method based on convolutional neural network, and thus has high practical value.
Description
技术领域technical field
本发明涉及图像半监督特征学习,尤其涉及基于带对称跨层连接的卷积神经网络(Convolutional Neural Network,CNN)半监督特征学习方法。The invention relates to image semi-supervised feature learning, in particular to a semi-supervised feature learning method based on a convolutional neural network (Convolutional Neural Network, CNN) with symmetric cross-layer connections.
背景技术Background technique
随着信息技术的不断飞速发展,各个领域每天都在以惊人的速度产生各种类型的图像数据。在大量的图像数据获取、传播过程中,如何更好地理解图像语义信息,并借此完成人类才可完成的任务,是现今人工智能与模式识别领域的重要挑战。人们迫切地希望计算机能够帮助人类更好地获取并利用海量图像数据。With the continuous rapid development of information technology, various fields are generating various types of image data at an astonishing speed every day. In the process of acquiring and disseminating a large amount of image data, how to better understand the semantic information of the image and use it to complete the tasks that humans can complete is an important challenge in the field of artificial intelligence and pattern recognition today. People urgently hope that computers can help humans better acquire and utilize massive image data.
互联网中的图像数据往往都是以没有类标的形式存在的,仅有少量结构化数据或用于科研的图像数据具有类标。因此如何使用大量无类标数据,辅助少量有类标数据的理解与学习,成为人工智能领域亟待解决的问题。图像的半监督特征学习作为利用无类标数据的重要方法,一直受到工业界和学术界的广泛关注,并经常作为各种图像相关国际学术会议的重要主题,是人工智能和模式识别领域一个非常重要的研究课题。其基础思想是利用无类标图像中提取出的结构信息,利用一定技术手段,将无类标信息与有类标数据特征相关联,从而辅助有类标图像的理解与学习。Image data on the Internet often exists in a form without class labels, and only a small amount of structured data or image data used for scientific research has class labels. Therefore, how to use a large amount of unlabeled data to assist the understanding and learning of a small amount of labeled data has become an urgent problem in the field of artificial intelligence. As an important method of using unlabeled data, the semi-supervised feature learning of images has been widely concerned by industry and academia, and is often used as an important topic of various image-related international academic conferences. It is a very important topic in the field of artificial intelligence and pattern recognition. important research topic. The basic idea is to use the structural information extracted from the unlabeled image and use certain technical means to associate the unlabeled information with the features of the labeled data, thereby assisting the understanding and learning of the labeled image.
近几年,基于深度神经网络,特别是深度卷积神经网络的方法广泛应用于许多计算机视觉和模式识别任务,在许多高层图像理解任务,如图像分类,图像分割等问题上取得令人瞩目的效果。但其仍有一些缺点使其在应用中收到限制,其中很重要的一点即是需要大量有类标图像数据。在有类标数据量有限的情况下,深度学习方法的表现往往不尽如人意。如何将半监督特征学习思想应用于深度学习领域,已成为当前研究热点,对积极推进社会信息化进程起到重要作用。在创造了无可替代的社会价值的同时,该领域仍有许多关键技术问题尚未解决,仍有许多功能实现需要进一步完善,因此,如何利用深度卷积神经网络,更有效地在半监督情况下对图像进行理解,以更灵活地实现计算机视觉的研究,具有深远的意义。In recent years, methods based on deep neural networks, especially deep convolutional neural networks, have been widely used in many computer vision and pattern recognition tasks, and have achieved remarkable results in many high-level image understanding tasks, such as image classification, image segmentation, etc. Effect. However, there are still some shortcomings that restrict its application, one of which is that it requires a large amount of class-labeled image data. In the case of limited amount of labeled data, the performance of deep learning methods is often unsatisfactory. How to apply the idea of semi-supervised feature learning to the field of deep learning has become a current research hotspot and plays an important role in actively promoting the process of social informatization. While creating irreplaceable social value, there are still many key technical issues in this field that have not yet been resolved, and there are still many functional realizations that need to be further improved. Therefore, how to use deep convolutional neural networks more effectively in semi-supervised It is of far-reaching significance to understand the image to realize the research of computer vision more flexibly.
发明内容Contents of the invention
发明目的:本发明所要解决的技术问题是针对现有技术的不足,提供一种基于带跨层连接的卷积神经网络的半监督特征学习方法,通过在海量无类标数据中对卷积神经网络进行预训练,最终提高在有类标数据中的表现。Purpose of the invention: the technical problem to be solved by the present invention is to provide a kind of semi-supervised feature learning method based on the convolutional neural network with cross-layer connection for the deficiencies in the prior art. The network is pre-trained to eventually improve performance on labeled data.
为了解决上述技术问题,本发明公开了一种基于带跨层连接的卷积神经网络(Convolutional Neural Network,CNN)的半监督特征学习方法,包含如下步骤:In order to solve the above-mentioned technical problems, the present invention discloses a semi-supervised feature learning method based on a convolutional neural network (Convolutional Neural Network, CNN) with cross-layer connections, comprising the following steps:
步骤1,生成无类标和有类标数据集:采集有类标和无类标图像数据,对每张图像做随机裁剪和归一化处理,得到有类标图像集合X0和无类标图像集合Y,根据图像分辨率大小,对集合Y中图像进行不同的破坏,得到破坏后的无类标图像集X1,设Z为有类标图像类标向量,Z={z1,z2,…,zn},zi表示第i张图像类标,i取值为1~n,则(X1,Y)组成用于非监督预训练的无类标训练数据集合,(X0,Z)作为有监督训练的有类标训练数据集合;Step 1, generate unlabeled and labeled data sets: collect labeled and unlabeled image data, perform random cropping and normalization processing on each image, and obtain the labeled image set X 0 and unlabeled image set X 0 The image set Y, according to the image resolution, destroys the images in the set Y differently, and obtains the destroyed image set X 1 without class labels. Let Z be the class label vector of images with class labels, Z={z 1 ,z 2 ,...,z n }, z i represents the i-th image class label, and the value of i is 1~n, then (X 1 ,Y) constitutes a non-label training data set for unsupervised pre-training, (X 0 , Z) as a set of labeled training data for supervised training;
步骤2,构建预训练图像恢复网络:根据输入图像大小构建图像恢复网络,设网络总深度为D层,D为偶数,其中前D/2层为卷积层,后D/2层为反卷积层,卷积核大小取为3x3,步长为1或2,根据网络深度和图像大小决定步长变化率。输入为步骤1中破坏后的图像集X1中的图像,输出为网络恢复后图像;Step 2, build a pre-trained image recovery network: build an image recovery network according to the size of the input image, set the total depth of the network to be D layers, D is an even number, where the first D/2 layer is a convolutional layer, and the last D/2 layer is a deconvolution layer In the product layer, the size of the convolution kernel is 3x3, and the step size is 1 or 2. The rate of change of the step size is determined according to the depth of the network and the size of the image. The input is the image in the image set X 1 destroyed in step 1, and the output is the image after the network restores;
步骤3,训练图像恢复网络:使用ADAM(Kingma,Diederik P.,and Jimmy Ba."Adam:A method for stochastic optimization."arXiv preprint arXiv:1412.6980(2014).)优化算法,采用步骤1得到的训练集(X1,Y)对步骤2所构建网络进行训练,以集合X1中受破坏图像作为输入,并以集合Y中对应的无损图像作为网络监督信息,训练后记录图像恢复网络前D/2层每一层权重W和偏置b;Step 3, training image restoration network: use ADAM (Kingma, Diederik P., and Jimmy Ba."Adam:A method for stochastic optimization."arXiv preprint arXiv:1412.6980(2014).) Optimization algorithm, using the training obtained in step 1 Set (X 1 , Y) to train the network constructed in step 2, take the damaged image in the set X 1 as input, and use the corresponding lossless image in the set Y as the network supervision information, record the image after training and restore the network before D/ 2 layers of weight W and bias b for each layer;
步骤4,构建有监督分类网络,以步骤2构建的图像恢复网络为模板,根据输入图像大小构建D/2层网络,均为卷积层,步长变化与步骤2中构建网络一致。并加入Max-pooling层和Softmax层,同时对卷积层参数使用步骤3中训练好网络对应权重W和偏置b进行初始化;Step 4, build a supervised classification network, use the image restoration network built in step 2 as a template, and build a D/2 layer network according to the size of the input image, all of which are convolutional layers, and the step size change is consistent with the network built in step 2. And add the Max-pooling layer and Softmax layer, and initialize the parameters of the convolutional layer using the corresponding weight W and bias b of the network trained in step 3;
步骤5,训练分类网络,将步骤4中构建并初始化的分类网络使用ADAM优化算法,在有类标图像数据上进行训练,直到算法收敛。Step 5, train the classification network, use the ADAM optimization algorithm to train the classification network constructed and initialized in step 4 on the image data with class labels until the algorithm converges.
步骤1具体包括如下步骤:Step 1 specifically includes the following steps:
步骤1-1,采集有类标和无类标图像数据,对每幅图像进行裁剪,采用随机裁剪的方式,得到大小相同的图像块,其中图像块大小取决于原始图像大小和模型大小,对于小于50*50的低分辨率图像(如CIFAR-10数据集),裁剪大小为29*29,对于大于225*225高分辨率的自然图像(如PASCAL VOC数据集),裁剪大小为225*225,若分辨率在二者之间,则先进行缩放到相近分辨率,再进行裁剪。将裁剪后所有图像集合记为X’;Step 1-1, collecting image data with and without class labels, cutting each image, and using random cutting to obtain image blocks of the same size, where the size of the image block depends on the size of the original image and the size of the model, for For low-resolution images smaller than 50*50 (such as the CIFAR-10 dataset), the crop size is 29*29, and for natural images with high resolution larger than 225*225 (such as the PASCAL VOC dataset), the crop size is 225*225 , if the resolution is between the two, first scale to a similar resolution and then crop. Record all image collections after cropping as X';
步骤1-2,将裁剪后的图像块进行归一化和中心化处理,首先计算裁剪后图像数据集合X’上每个像素的均值和标准差,设X’上的所有图像均值图像为标准差为std,对于一张特定图像x,对其进行归一化和中心化处理如下:Step 1-2, normalize and centralize the cropped image blocks, first calculate the mean and standard deviation of each pixel on the cropped image data set X', set the mean image of all images on X' as The standard deviation is std. For a specific image x, it is normalized and centered as follows:
x′为图像x处理后的图像;处理后图像中,有类标图像集合记为X0,无类标图像集合记为Y。x' is the processed image of image x; in the processed image, the set of images with class labels is marked as X 0 , and the set of images without class labels is marked as Y.
步骤1-3,对于有类标图像,将其处理后图像集合X0和对应类标向量Z组成有类标训练数据(X0,Z),Z={z1,z2,…,zn},zi表示第i张图像类标。Step 1-3, for the labeled image, the processed image set X 0 and the corresponding labeled vector Z form the labeled training data (X 0 , Z), Z={z 1 ,z 2 ,…,z n }, z i represents the class label of the i-th image.
步骤1-4,对于无类标图像集合Y中的图像,进行破坏,加高斯噪声或将图像中像素值置为0,若裁剪后为低分辨率图像(分辨率小于50*50),则采取加高斯噪声方法,若裁剪后为高分辨率图像(分辨率大于等于50*50),则采取像素点置为0方法,置0的像素点为随机选取10个相邻的8*8区域,得到破坏后的无类标图像集X1,其与无类标图像集合Y组成无类标训练数据集合(X1,Y)。Steps 1-4, destroy the images in the unlabeled image set Y, add Gaussian noise or set the pixel value in the image to 0, if it is a low-resolution image after cropping (resolution less than 50*50), then The method of adding Gaussian noise is adopted. If the cropped image is a high-resolution image (resolution greater than or equal to 50*50), the pixel point is set to 0, and the pixel point set to 0 is randomly selected from 10 adjacent 8*8 areas. , to obtain the destroyed unlabeled image set X 1 , which forms the unlabeled training data set (X 1 , Y) with the unlabeled image set Y.
步骤2具体包括如下步骤:Step 2 specifically includes the following steps:
步骤2-1,设图像恢复网络总深度为D层,D为偶数,其中前D/2层使用卷积层,后D/2层使用反卷积层,卷积核大小为3x3,步长为1或2,每隔k层步长为1第k+1层步长为2,0<k<D/2-1,重复n次。k和n大小由网络深度和图像块大小决定:对于29*29低分辨率图像,k=4,n=3,对于225*225高分辨率图像,k=2,n=5,在每层卷积层和反卷积层之后添加BatchNormalization层和ReLU(Rectified Linear Unit)非线性层,网络输入为步骤1中生成的破坏后的图像集合X1中图像,输出为网络恢复后图像。网络参数包括卷积层和反卷积层的权重W和偏置b,以及BatchNormalization层权重γ和偏置β。Step 2-1, set the total depth of the image restoration network to be D layers, D is an even number, where the first D/2 layer uses a convolution layer, and the last D/2 layer uses a deconvolution layer, the convolution kernel size is 3x3, and the step size is is 1 or 2, the step size of every k layer is 1, the step size of the k+1 layer is 2, 0<k<D/2-1, repeat n times. The k and n sizes are determined by the network depth and image block size: for 29*29 low-resolution images, k=4, n=3, for 225*225 high-resolution images, k=2, n=5, in each layer A BatchNormalization layer and a ReLU (Rectified Linear Unit) nonlinear layer are added after the convolutional layer and the deconvolutional layer. The input of the network is the image in the destroyed image set X 1 generated in step 1, and the output is the restored image of the network. The network parameters include the weight W and bias b of the convolutional layer and the deconvolution layer, and the weight γ and bias β of the BatchNormalization layer.
步骤2-2,每隔两层,在卷积层和反卷积层之间加入对称跨层连接:设COi表示第i层卷积层输出,DIi代表第i层反卷积层输入,DOi代表第i层反卷积层输出,则跨层连接表示为:Step 2-2, add symmetrical cross-layer connections between the convolutional layer and the deconvolution layer every two layers: let CO i represent the output of the i-th convolutional layer, and DI i represent the input of the i-th deconvolutional layer , DO i represents the output of the i-th deconvolution layer, then the cross-layer connection is expressed as:
DID-i+1=DOD-i+COi,DI D-i+1 = DO Di + CO i ,
则DID+1为网络输出,CO0为网络输入,第一次跨层连接从输入层开始连接到输出层,网络最终层输入与无损图像集合Y中对应图像计算欧氏距离作为之后网络训练的损失函数 Then DI D+1 is the network output, CO 0 is the network input, the first cross-layer connection starts from the input layer and connects to the output layer, the final layer input of the network and the corresponding image in the lossless image set Y calculate the Euclidean distance as the subsequent network training loss function
其中Xi为图像集X1中第i张图像,Yi为图像集Y中第i张图像,为神经网络代表的函数,N为训练图像数量,θ为网络所有可训练参数,包括卷积层和反卷积层的权重W和偏置b,以及BatchNormalization层权重γ和偏置β。Where X i is the i-th image in the image set X 1 , Y i is the i-th image in the image set Y, is the function represented by the neural network, N is the number of training images, θ is all trainable parameters of the network, including the weight W and bias b of the convolutional layer and the deconvolution layer, and the weight γ and bias β of the BatchNormalization layer.
步骤3具体包括如下步骤:Step 3 specifically includes the following steps:
步骤3-1,利用ADAM优化算法进行梯度反向传播训练神经网络,学习率设置为1e-4,训练持续n1轮(一般为20轮),在第n2轮(一般为第8轮)和第n3轮(一般为第16轮)结束后,学习率分别设置为1e-5和1e-6;;Step 3-1, use the ADAM optimization algorithm to train the neural network with gradient backpropagation, the learning rate is set to 1e-4, and the training lasts for n 1 rounds (generally 20 rounds), and at n 2 rounds (generally the 8th round) and after the n 3rd round (generally the 16th round), the learning rates are set to 1e-5 and 1e-6 respectively;
步骤3-2,为说明带跨层连接的梯度反向传播步骤,设图像恢复网络深度为7层,以步骤2-2中方式添加跨层连接,设X0为网络输入,Xi为第i层卷积层输出,跨层链接具体将X1连接到第5层输入,将X0连接到第7层输入。此时在前向计算时,得到图像恢复网络输出X7为:Step 3-2, to illustrate the step of gradient backpropagation with cross-layer connections, set the depth of the image restoration network to 7 layers, add cross-layer connections in the manner in step 2-2, let X 0 be the network input, and Xi be the first The output of the i-layer convolutional layer, the cross-layer link specifically connects X 1 to the input of layer 5, and connects X 0 to the input of layer 7. At this time, in the forward calculation, the image restoration network output X 7 is obtained as:
X7=f7(X0,X6);X 7 = f 7 (X 0 , X 6 );
步骤3-3,将X7进一步展开表示为:Step 3-3, further expand X 7 as:
X7=f7(X0,X6)X 7 = f 7 (X 0 , X 6 )
=f7(X0,f6(X5))= f 7 (X 0 , f 6 (X 5 ))
=f7(X0,f6(f5(X1,X4)))= f 7 (X 0 , f 6 (f 5 (X 1 , X 4 )))
=f7(X0,f6(f5(X1,f4(X3))))= f 7 (X 0 , f 6 (f 5 (X 1 , f 4 (X 3 ))))
=f7(X0,f6(f5(X1,f4(fk(X2)))))= f 7 (X 0 , f 6 (f 5 (X 1 , f 4 (f k (X 2 )))))
步骤3-4,在梯度反向传播时,网络中的第i层直接从其顶层获得梯度来更新该层的参数θi,在本方法中,θi具体包括卷积/反卷积层权重Wi和偏置bi,以及BatchNormalization层权重γi和偏置βi以该网络第一层为例,为更新第一层参数θ1需要计算损失函数ζ关于θ1的偏导数:Step 3-4, during gradient backpropagation, the i-th layer in the network directly obtains the gradient from its top layer to update the parameters θ i of this layer. In this method, θ i specifically includes the convolution/deconvolution layer weights W i and bias b i , and BatchNormalization layer weight γ i and bias β i take the first layer of the network as an example, in order to update the first layer parameter θ 1 , it is necessary to calculate the partial derivative of the loss function ζ with respect to θ 1 :
步骤3-5,在得到每一层对应顶层的偏导数之后,使用ADAM算法对应的更新规则更新每一层参数,训练在步骤1得到的无类标训练数据集合(X1,Y)上进行,以X1中受破坏图像作为输入,以Y中对应清晰图像作为监督信息以步骤3-4中方法更新参数,直到使用了所有训练数据训练20轮之后停止。Step 3-5, after obtaining the partial derivative of each layer corresponding to the top layer, use the update rule corresponding to the ADAM algorithm to update the parameters of each layer, and the training is performed on the unlabeled training data set (X 1 , Y) obtained in step 1 , take the damaged image in X 1 as input, and use the corresponding clear image in Y as supervision information to update the parameters in steps 3-4 until all training data are used for 20 rounds of training.
步骤4具体包括如下步骤:Step 4 specifically includes the following steps:
构建有监督分类网络,先提取步骤3中训练好图像恢复网络卷积层参数W、b、γ和β,根据输入图像大小构建D/2层网络,每层均为卷积层,步长变化与步骤2中构建网络一致,其中在最后一层卷积层后加入最大值池化层(Max-pooling)层,在其后根据分类任务类标数量N,N为有监督训练的有类标训练数据集合(X0,Z)中类标向量Z的最大可能取值,加入N类的Softmax层。使用提取的参数W、b、γ和β,采用直接赋值的方式,将有监督分类网络对应参数初始化。To build a supervised classification network, first extract the parameters W, b, γ, and β of the convolutional layer of the image restoration network trained in step 3, and construct a D/2 layer network according to the size of the input image, each layer is a convolutional layer, and the step size changes Consistent with the construction of the network in step 2, in which the maximum pooling layer (Max-pooling) layer is added after the last convolutional layer, and then according to the number of class labels for classification tasks N, N is the class label for supervised training The maximum possible value of the class label vector Z in the training data set (X 0 , Z) is added to the Softmax layer of N classes. Using the extracted parameters W, b, γ and β, the corresponding parameters of the supervised classification network are initialized by direct assignment.
步骤5具体包括如下步骤:Step 5 specifically includes the following steps:
训练分类网络,将步骤4中构建并初始化的分类网络使用ADAM优化算法在有类标训练数据集合(X0,Z)上进行训练,初始学习率设置为1e-4,训练持续n4轮(一般为200轮),在第n5、n6和n7轮(一般n5为80、n6为120、n7为160)结束后,将当前学习率乘以0.2得到新学习率,n4轮后直到网络收敛。To train the classification network, use the ADAM optimization algorithm to train the classification network constructed and initialized in step 4 on the training data set (X 0 , Z) with class labels. The initial learning rate is set to 1e-4, and the training lasts for n 4 rounds ( Generally 200 rounds), after the n 5 , n 6 and n 7 rounds (generally n 5 is 80, n 6 is 120, n 7 is 160), multiply the current learning rate by 0.2 to get the new learning rate, n After 4 rounds until the network converges.
本发明针对图像半监督特征学习的深度卷积神经网络方法,本发明具有如下特征:1)本发明在使用深度神经网络进行预训练时时,加入了跨层连接,使得网络可以更快收敛,同时使得网络中层特征可以保留更多图像抽象信息;2)本发明方法不同于以往针对特定数据类型的半监督特征学习方法,可以应用于几乎所有图像数据,具有普适性。The present invention is aimed at the deep convolutional neural network method of image semi-supervised feature learning. The present invention has the following characteristics: 1) when the present invention uses the deep neural network for pre-training, a cross-layer connection is added, so that the network can converge faster, and at the same time The middle-level features of the network can retain more image abstract information; 2) The method of the present invention is different from the previous semi-supervised feature learning method for specific data types, and can be applied to almost all image data and has universal applicability.
有益效果:本发明充分考虑了卷积神经网络逐层连接在非监督特征学习,加入跨层连接保证网络可以从无类标数据中提取出足够有用的抽象信息,从而更好地辅助有类标数据的分类,提升图像分类的准确率。Beneficial effects: the present invention fully considers the layer-by-layer connection of the convolutional neural network in unsupervised feature learning, adding cross-layer connections to ensure that the network can extract enough useful abstract information from data without class labels, thereby better assisting class-labeled data. Data classification improves the accuracy of image classification.
附图说明Description of drawings
下面结合附图和具体实施方式对本发明做更进一步的具体说明,本发明的上述或其他方面的优点将会变得更加清楚。The present invention will be further described below in conjunction with the accompanying drawings and specific embodiments, and the advantages of the above and other aspects of the present invention will become clearer.
图1为本发明流程图。Fig. 1 is the flow chart of the present invention.
图2为网络结构示意图。Figure 2 is a schematic diagram of the network structure.
图3a为实施例中的一副原图。Fig. 3a is a pair of original pictures in the embodiment.
图3b为图3a加噪声后图像。Figure 3b is the image after adding noise in Figure 3a.
图3c为图3a的恢复图像。Figure 3c is the restored image of Figure 3a.
图4a为实施例中的一副原图。Fig. 4a is a pair of original pictures in the embodiment.
图4b为实施例中的一副原图。Fig. 4b is a pair of original pictures in the embodiment.
图4c为图4a对应的特征图。Figure 4c is a feature map corresponding to Figure 4a.
图4d为图4b对应的特征图。Figure 4d is the feature map corresponding to Figure 4b.
具体实施方式Detailed ways
下面结合附图及实施例对本发明做进一步说明。The present invention will be further described below in conjunction with the accompanying drawings and embodiments.
如图1所示,本发明公开了一种基于带跨层连接的卷积神经网络的图像修复方法,包含如下步骤:As shown in Figure 1, the present invention discloses an image repair method based on a convolutional neural network with cross-layer connections, comprising the following steps:
步骤1,生成无类标和有类标数据集:采集有类标和无类标图像数据,对每张图像做随机裁剪和归一化处理,得到有类标图像集X0和无类标清晰图像集合Y,根据图像分辨率大小,对Y中图像进行不同的破坏,得到破坏后无类标的图像集X1,设Z为有类标图像类标向量,Z={z1,z2,…,zn},zi表示第i张图像类标,则(X1,Y)组成用于非监督预训练的神经网络的训练集,(X0,Z)作为有监督训练的训练集;Step 1, generate unlabeled and labeled data sets: collect labeled and unlabeled image data, perform random cropping and normalization on each image, and obtain labeled image set X 0 and unlabeled image data The clear image set Y, according to the image resolution, destroys the images in Y differently, and obtains the image set X 1 without class labels after destruction. Let Z be the class label vector of images with class labels, Z={z 1 ,z 2 ,...,z n }, z i represents the class label of the i-th image, then (X 1 ,Y) constitutes the training set of the neural network for unsupervised pre-training, and (X 0 ,Z) is used as the training set for supervised training set;
步骤2,构建预训练图像恢复网络:根据输入图像大小构建,设网络总深度为D层,D为偶数,其中前D/2层为卷积层,后D/2层为反卷积层,卷积核大小取为3x3,步长为1或2,根据网络深度和图像大小决定步长变化率。输入为步骤1中破坏图像集X1中图像,输出为网络恢复后图像;Step 2, build a pre-trained image restoration network: build according to the size of the input image, set the total depth of the network to be D layers, D is an even number, where the first D/2 layer is a convolutional layer, and the last D/2 layer is a deconvolutional layer, The size of the convolution kernel is 3x3, the step size is 1 or 2, and the step size change rate is determined according to the network depth and image size. The input is the image in the damaged image set X 1 in step 1, and the output is the image after the network restoration;
步骤3,训练图像恢复网络:使用ADAM(adaptive moment estimation,自适应矩估计)优化算法,采用步骤1得到的训练集(X1,Y)对步骤2所构建网络进行训练,以集合X1中受破坏图像作为输入,并以集合Y中对应的无损图像作为网络监督信息,训练后记录图像恢复网络前D/2层每一层权重W和偏置b;Step 3, training the image restoration network: using the ADAM (adaptive moment estimation, adaptive moment estimation) optimization algorithm, using the training set (X 1 , Y) obtained in step 1 to train the network constructed in step 2, to set X 1 The damaged image is used as input, and the corresponding lossless image in the set Y is used as the network supervision information, and the weight W and bias b of each layer of D/2 layers before the image restoration network are recorded after training;
步骤4,构建有监督分类网络,以步骤2构建的图像恢复网络为模板,根据输入图像大小构建D/2层网络,均为卷积层,步长变化与步骤2中构建网络一致。并加入Max-pooling层和Softmax层,同时对卷积层参数使用步骤3中训练好网络对应权重W和偏置b进行初始化;Step 4, build a supervised classification network, use the image restoration network built in step 2 as a template, and build a D/2 layer network according to the size of the input image, all of which are convolutional layers, and the step size change is consistent with the network built in step 2. And add the Max-pooling layer and Softmax layer, and initialize the parameters of the convolutional layer using the corresponding weight W and bias b of the network trained in step 3;
步骤5,训练分类网络,将步骤4中构建并初始化的分类网络使用ADAM优化算法,在有类标图像数据上进行训练,直到算法收敛。Step 5, train the classification network, use the ADAM optimization algorithm to train the classification network constructed and initialized in step 4 on the image data with class labels until the algorithm converges.
步骤1具体包括如下步骤:Step 1 specifically includes the following steps:
本步骤描述数据预处理过程,采集有类标和无类标图像数据,对每幅图像进行裁剪,采用随机裁剪的方式,得到大小相同的图像块,其中图像块大小取决于原始图像大小和模型大小,对于小于50*50的低分辨率图像(如CIFAR-10数据集),裁剪大小为29*29,对于大于225*225高分辨率的自然图像(如PASCAL VOC数据集),裁剪大小为225*225,若分辨率在二者之间,则先进行缩放到相近分辨率,再进行裁剪。将裁剪后所有图像集合记为X’;将裁剪后的图像块进行归一化和中心化处理,首先计算裁剪后图像数据集合X’上每个像素的均值和标准差,设X’上的所有图像均值图像为标准差为std,对于一张特定图像x,对其进行归一化和中心化处理如下:This step describes the data preprocessing process, collecting image data with and without class labels, cutting each image, and using random cutting to obtain image blocks of the same size, where the size of the image block depends on the size of the original image and the model Size, for low-resolution images smaller than 50*50 (such as the CIFAR-10 data set), the cropping size is 29*29, and for natural images with high resolution larger than 225*225 (such as the PASCAL VOC data set), the cropping size is 225*225, if the resolution is between the two, first scale to a similar resolution and then crop. Record all image sets after cropping as X'; normalize and centralize the cropped image blocks, first calculate the mean and standard deviation of each pixel on the cropped image data set X', let X' be All images mean image is The standard deviation is std. For a specific image x, it is normalized and centered as follows:
x′为图像x处理后的图像;处理后图像中,有类标图像集合记为X0,无类标图像集合记为Y。x' is the processed image of image x; in the processed image, the set of images with class labels is marked as X 0 , and the set of images without class labels is marked as Y.
对于有类标图像,将其处理后图像集合X0和对应类标向量Z组成有类标训练数据(X0,Z),Z={z1,z2,…,zn},zi表示第i张图像类标。For a labeled image, the processed image set X 0 and the corresponding labeled vector Z form the labeled training data (X 0 , Z), Z={z 1 ,z 2 ,…,z n }, z i Indicates the class label of the i-th image.
对于无类标清晰图像集合Y中的图像,进行破坏,加高斯噪声或将图像中像素值置为0,若裁剪后为低分辨率图像(分辨率小于50*50),则采取加高斯噪声方法,若裁剪后为高分辨率图像(分辨率大于等于50*50),则采取像素点置为0方法,置0的像素点为随机选取10个相邻的8*8区域。得到破坏后的无类标图像集X1,其与清晰图像集Y组成无类标训练数据(X1,Y)。For the image in the unclassified image set Y, destroy it, add Gaussian noise or set the pixel value in the image to 0, if it is a low-resolution image after cropping (resolution less than 50*50), then add Gaussian noise method, if the cropped image is a high-resolution image (resolution greater than or equal to 50*50), the pixel point is set to 0, and the pixel point set to 0 is randomly selected from 10 adjacent 8*8 areas. The destroyed unlabeled image set X 1 is obtained, which forms the unlabeled training data (X 1 , Y) with the clear image set Y.
步骤2具体包括如下步骤:Step 2 specifically includes the following steps:
本步骤描述对预训练神经网络模型的构建过程,设网络总深度为D层,D为偶数,其中前D/2层使用卷积层,后D/2层使用反卷积层,卷积核大小为3x3,步长为1或2,每隔k层步长为1第k+1层步长为2,0<k<D/2-1,重复n次。根据网络深度和图像块大小调整k和n大小。在每层卷积层和反卷积层之后添加BatchNormalization层和ReLU(Rectified Linear Unit)非线性层。网络输入为步骤1中生成的破坏后的图像,输出为网络恢复后图像;每隔两层,在卷积层和反卷积层之间加入对称跨层连接。具体地,设COi表示第i层卷积层输出,DIi代表第i层反卷积层输入,DOi代表第i层反卷积层输出,则跨层连接可表示为:This step describes the construction process of the pre-trained neural network model. The total depth of the network is set to D layers, and D is an even number. The first D/2 layer uses a convolution layer, and the last D/2 layer uses a deconvolution layer, and the convolution kernel The size is 3x3, the step size is 1 or 2, the step size of every k layer is 1, the step size of the k+1 layer is 2, 0<k<D/2-1, repeat n times. Adjust k and n sizes according to network depth and image patch size. Add BatchNormalization layer and ReLU (Rectified Linear Unit) nonlinear layer after each convolution layer and deconvolution layer. The input of the network is the damaged image generated in step 1, and the output is the restored image of the network; every two layers, a symmetric cross-layer connection is added between the convolutional layer and the deconvolutional layer. Specifically, let CO i represent the output of the i-th convolutional layer, DI i represent the input of the i-th deconvolution layer, and DO i represent the output of the i-th deconvolution layer, then the cross-layer connection can be expressed as:
DID-i+1=DOD-i+COi DI D-i+1 =DO Di +CO i
特别的,DID+1为网络输出,CO0为网络输入,即第一次跨层连接从输入层开始连接到输出层。网络最终层输入与原图像数据集Y中对应图像计算欧氏距离作为损失函数:In particular, DI D+1 is the network output, and CO 0 is the network input, that is, the first cross-layer connection is connected from the input layer to the output layer. Calculate the Euclidean distance between the input of the final layer of the network and the corresponding image in the original image data set Y as the loss function:
其中Xi为无类标受损图像集X1中第i张图像,Yi为无类标清晰图像集Y中第i张图像,为神经网络代表的函数,N为训练图像数量,θ为网络所有可训练参数,包括卷积层和反卷积层的权重W和偏置b,以及BatchNormalizatiion层权重γ和偏置β。Where X i is the i-th image in the unlabeled damaged image set X 1 , and Y i is the i-th image in the unlabeled clear image set Y, is the function represented by the neural network, N is the number of training images, θ is all trainable parameters of the network, including the weight W and bias b of the convolutional layer and the deconvolution layer, and the weight γ and bias β of the BatchNormalizatiion layer.
图2是网络结构简单示意图,左边图中,Corrupted data是网络输入数据,restored data是网络输出数据,conv1,conv2以及c3…c6为卷积层,d3…d6,,deconv1,deconv2为反卷积层。右边图描述一个跨层连接的细节,图中conv为卷积层,deconv为反卷积层,ReLU和BatchNorm分别代表ReLU层和BatchNormalization层。Figure 2 is a simple schematic diagram of the network structure. In the left figure, Corrupted data is the network input data, restored data is the network output data, conv1, conv2 and c3...c6 are convolution layers, d3...d6,, deconv1, deconv2 are deconvolution Floor. The figure on the right describes the details of a cross-layer connection. In the figure, conv is the convolution layer, deconv is the deconvolution layer, and ReLU and BatchNorm represent the ReLU layer and the BatchNormalization layer, respectively.
步骤3具体包括如下步骤:Step 3 specifically includes the following steps:
本步骤描述对预训练神经网络模型的训练过程,利用ADAM优化算法进行梯度反向传播训练神经网络,学习率设置为1e-4。训练持续20轮,在第8轮和第16轮结束后,学习率分别设置为1e-5和1e-6。This step describes the training process of the pre-trained neural network model. The ADAM optimization algorithm is used to perform gradient backpropagation to train the neural network, and the learning rate is set to 1e-4. The training lasts for 20 epochs, and after the 8th and 16th epochs, the learning rate is set to 1e-5 and 1e-6, respectively.
为说明带跨层连接的梯度反向传播步骤,设该网络深度为7层,以步骤2-2中方式添加跨层连接,设X0为网络输入,Xi为第i层卷积层输出,跨层连接具体将X1连接到第5层输入,将X0连接到第7层输入。此时在前向计算时,得到图像恢复网络输出X7为:To illustrate the step of gradient backpropagation with cross-layer connections, let the network depth be 7 layers, add cross-layer connections in the way in step 2-2, let X 0 be the network input, and X i be the i-th convolutional layer output , the cross-layer connection specifically connects X 1 to the layer 5 input and X 0 to the layer 7 input. At this time, in the forward calculation, the image restoration network output X 7 is obtained as:
X7=f7(X0,X6)X 7 = f 7 (X 0 , X 6 )
X7可以进一步表示为: X7 can be further expressed as:
X7=f7(X0,X6)X 7 = f 7 (X 0 , X 6 )
=f7(X0,f6(X5))= f 7 (X 0 , f 6 (X 5 ))
=f7(X0,f6(f5(X1,X4)))= f 7 (X 0 , f 6 (f 5 (X 1 , X 4 )))
=f7(X0,f6(f5(X1,f4(X3))))= f 7 (X 0 , f 6 (f 5 (X 1 , f 4 (X 3 ))))
=f7(X0,f6(f5(X1,f4(fk(X2)))))= f 7 (X 0 , f 6 (f 5 (X 1 , f 4 (f k (X 2 )))))
其中X1和X2表示第1、2个卷积层得到的特征图。Where X 1 and X 2 represent the feature maps obtained by the first and second convolutional layers.
在梯度反向传播时,网络中的第i层直接从其顶层获得梯度来更新该层的参数θi,在本方法中,θi具体包括卷积/反卷积层权重Wi和偏置bi,以及BatchNormalization层权重γi和偏置βi以该网络第一层为例,为更新第一层参数θ1需要计算损失函数ζ关于θ1的偏导数:During gradient backpropagation, the i-th layer in the network directly obtains the gradient from its top layer to update the parameters θ i of this layer. In this method, θ i specifically includes the convolution/deconvolution layer weight W i and bias b i , and BatchNormalization layer weight γ i and bias β i take the first layer of the network as an example, in order to update the first layer parameters θ 1 need to calculate the partial derivative of the loss function ζ with respect to θ 1 :
在得到每一层对应顶层的偏导数之后,使用ADAM算法对应的更新规则更新每一层参数,训练在步骤1得到的无类标数据集合(X1,Y)上进行,以X1中受破坏图像作为输入,以Y中对应清晰图像作为监督信息以步骤3-4中方法更新参数,直到使用了所有训练数据训练20轮之后停止。After obtaining the partial derivative of each layer corresponding to the top layer, use the update rule corresponding to the ADAM algorithm to update the parameters of each layer, and the training is carried out on the unlabeled data set (X 1 , Y) obtained in step 1 . The damaged image is used as input, and the corresponding clear image in Y is used as the supervisory information to update the parameters in steps 3-4 until all training data are used for 20 rounds of training.
步骤4具体包括如下步骤:Step 4 specifically includes the following steps:
本步骤描述有监督分类网络的构建过程,先提取步骤3中训练好图像恢复网络卷积层参数W、b、γ和β,根据输入图像大小构建D/2层网络,每层均为卷积层,步长变化与步骤2中构建网络一致,其中在最后一层卷积层后加入最大值池化层(Max-pooling)层,在其后根据分类任务类标数量N,N为有监督训练数据(X0,Z)中类标向量Z的最大可能取值,加入N类的Softmax层。使用提取的参数W、b、γ和β,采用直接赋值的方式,将有监督分类网络对应参数初始化。This step describes the construction process of the supervised classification network. First, extract the parameters W, b, γ, and β of the convolutional layer of the image restoration network trained in step 3, and construct a D/2 layer network according to the size of the input image, and each layer is a convolution Layer, the step size change is consistent with the network constructed in step 2, where the maximum pooling layer (Max-pooling) layer is added after the last convolutional layer, and then according to the number of classification tasks N, N is supervised The maximum possible value of the class label vector Z in the training data (X 0 , Z) is added to the Softmax layer of N classes. Using the extracted parameters W, b, γ and β, the corresponding parameters of the supervised classification network are initialized by direct assignment.
步骤5具体包括如下步骤:Step 5 specifically includes the following steps:
本步骤描述有监督分类网络的训练过程,将步骤4中构建并初始化的分类网络使用ADAM优化算法在有类标图像数据集(X0,Z)上进行训练,初始学习率设置为1e-4,训练持续n4轮(一般为200轮),在第n5、n6和n7轮(一般n5为80、n6为120、n7为160)结束后,将当前学习率乘以0.2得到新学习率,n4轮后直到网络收敛。This step describes the training process of the supervised classification network. The classification network constructed and initialized in step 4 is trained on the labeled image dataset (X 0 , Z) using the ADAM optimization algorithm, and the initial learning rate is set to 1e-4 , the training lasts for n 4 rounds (generally 200 rounds), after the n 5 , n 6 and n 7 rounds (generally n 5 is 80, n 6 is 120, n 7 is 160), the current learning rate is multiplied by 0.2 to get the new learning rate, n after 4 rounds until the network converges.
实施例1Example 1
本实施例描述CIFAR-10上的半监督特征学习,包括以下部分:This example describes semi-supervised feature learning on CIFAR-10, including the following parts:
1、首先将CIFAR-10数据集中50000张自然图像均匀地分为两部分,一部分包含4000张有类标图像,另一部分包含46000张无类标图像。1. First, the 50,000 natural images in the CIFAR-10 dataset are evenly divided into two parts, one part contains 4000 images with class labels, and the other part contains 46000 images without class labels.
2、对于每张32*32大小的无类标图像,在训练时,随机截取29*29的图像块,在图像上附加均值为0,标准差为30的高斯噪声。对于加噪声后的图像和不加噪声的图像分别进行归一化,形成无类标训练集。2. For each unlabeled image of size 32*32, during training, a 29*29 image block is randomly intercepted, and Gaussian noise with a mean value of 0 and a standard deviation of 30 is added to the image. Normalize the image after adding noise and the image without noise to form a non-labeled training set.
3、构建18层带跨层连接的卷积神经网络,使用ADAM算法在生成的无类标图像上进行训练。网络收敛后保留前9层,利用其网络权重构建对应的分类网络。3. Construct an 18-layer convolutional neural network with cross-layer connections, and use the ADAM algorithm to train on the generated unlabeled images. After the network converges, the first 9 layers are retained, and the corresponding classification network is constructed using their network weights.
4、在另外4000张有类标图像上训练分类网络,使用ADAM算法训练直到收敛,在原始图像的50000张测试集上进行测试,并报告准确率如下表1:4. Train the classification network on another 4,000 images with class labels, use the ADAM algorithm to train until convergence, test on the 50,000 original image test set, and report the accuracy rate as shown in Table 1:
表1Table 1
其中最后一行是该方法准确率,可以看到该方法达到目前很多利用GAN进行的半监督学习准确率且相比不使用预训练(No pre-training行)和不加跨层连接(Pre-training without shortcut行)相同网络,准确率均有较大提升。The last line is the accuracy of the method. It can be seen that the method has reached the accuracy of many semi-supervised learning using GAN and compared with no pre-training (No pre-training line) and no cross-layer connection (Pre-training Without shortcut line) the same network, the accuracy rate has been greatly improved.
实施例2Example 2
本实施例描述利用Imagenet数据集和Pascal VOC 2007数据的大规模半监督特征学习,包括以下部分:This embodiment describes the large-scale semi-supervised feature learning using Imagenet dataset and Pascal VOC 2007 data, including the following parts:
1、首先在Imagenet自然图像数据集上,随机截取225*225的图像块,对于每个图像块将其随机35个8*8图像区域像素置为0,对于置0处理的图像和原始图像分别进行归一化,形成无类标训练集。1. First, randomly intercept 225*225 image blocks on the Imagenet natural image data set, and set 35 random 8*8 image area pixels to 0 for each image block, and set the 0-set image and the original image respectively Perform normalization to form a non-labeled training set.
2、构建32层带跨层连接的卷积神经网络,使用ADAM算法在生成的无类标图像上进行训练。网络收敛后保留前16层,利用其网络权重构建对应的分类网络。2. Construct a 32-layer convolutional neural network with cross-layer connections, and use the ADAM algorithm to train on the generated unlabeled images. After the network converges, the first 16 layers are retained, and the corresponding classification network is constructed using their network weights.
3、对PASCAL VOC 2007自然图像数据,随机截取225*225的图像块,并将图像块以50%概率进行水平翻转并归一化后,得到有类标数据。3. For PASCAL VOC 2007 natural image data, 225*225 image blocks are randomly intercepted, and the image blocks are horizontally flipped and normalized with a probability of 50% to obtain class-labeled data.
4、在生成的有类标数据上训练分类网络,使用ADAM算法训练直到收敛,使用测试集进行测试,并报告准最终确率如下表:4. Train the classification network on the generated class-labeled data, use the ADAM algorithm to train until convergence, use the test set for testing, and report the accuracy rate as follows:
表2Table 2
其中最后一行是本方法准确率,可以看到该方法比目前同类方法准确率高1%左右,且相比不加跨层连接(Ours without shortcut行)相同网络,准确率均有较大提升。图3a~图3c是该实施例预训练网络对图像的恢复效果,图3a为原图,图3b为加噪声后图像,图3c为恢复图像,可以看到预训练网络可以很好地学习到图像细节信息。图4a~图4d是该实施例学习到的特征可视化效果,图4a和图4b为原图,图4c和图4d分别为图4a和图4b对应的特征图,每一张图像中的狗脸部分在特征图中都十分明显,可以说明该方法学习到的特征很好地捕捉到了图像深层语义信息。The last line is the accuracy rate of this method. It can be seen that the accuracy rate of this method is about 1% higher than that of the current similar methods, and compared with the same network without cross-layer connection (Ours without shortcut line), the accuracy rate has been greatly improved. Figures 3a to 3c are the image recovery effects of the pre-trained network in this embodiment. Figure 3a is the original image, Figure 3b is the image after adding noise, and Figure 3c is the restored image. It can be seen that the pre-trained network can learn well Image details. Figures 4a to 4d are the feature visualization effects learned in this embodiment. Figures 4a and 4b are the original images, and Figures 4c and 4d are the feature maps corresponding to Figures 4a and 4b respectively. The dog faces in each image Some of them are very obvious in the feature map, which shows that the features learned by this method capture the deep semantic information of the image well.
本发明提供了基于带对称跨层连接的卷积神经网络半监督特征学习方法,具体实现该技术方案的方法和途径很多,以上所述仅是本发明的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。本实施例中未明确的各组成部分均可用现有技术加以实现。The present invention provides a semi-supervised feature learning method based on a convolutional neural network with symmetric cross-layer connections. There are many methods and approaches to specifically realize the technical solution. The above is only a preferred embodiment of the present invention. It should be pointed out that for this technology Those of ordinary skill in the art can make some improvements and modifications without departing from the principle of the present invention, and these improvements and modifications should also be regarded as the protection scope of the present invention. All components that are not specified in this embodiment can be realized by existing technologies.
Claims (6)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810238288.4A CN108564166A (en) | 2018-03-22 | 2018-03-22 | Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201810238288.4A CN108564166A (en) | 2018-03-22 | 2018-03-22 | Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN108564166A true CN108564166A (en) | 2018-09-21 |
Family
ID=63532070
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201810238288.4A Pending CN108564166A (en) | 2018-03-22 | 2018-03-22 | Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN108564166A (en) |
Cited By (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN110543563A (en) * | 2019-08-20 | 2019-12-06 | 暨南大学 | A hierarchical text classification method and system |
| CN110598629A (en) * | 2019-09-11 | 2019-12-20 | 北京百度网讯科技有限公司 | Search space construction method and device of hyper network and electronic equipment |
| WO2020061972A1 (en) * | 2018-09-27 | 2020-04-02 | 电子科技大学 | Convolutional neural network-based diabetic retinopathy diagnosing technique |
| CN111263226A (en) * | 2020-01-17 | 2020-06-09 | 中国科学技术大学 | Video processing method, video processing device, electronic equipment and medium |
| CN112396136A (en) * | 2019-08-15 | 2021-02-23 | 北京神州泰岳软件股份有限公司 | Handwritten number recognition method and device, electronic equipment and readable storage medium |
| CN114341951A (en) * | 2019-08-20 | 2022-04-12 | 阿卡蒂科有限公司 | Method and system for diagnosing diseases, physiological changes or other internal conditions of crustaceans by non-invasive means |
| CN115147426A (en) * | 2022-09-06 | 2022-10-04 | 北京大学 | Method and system for model training and image segmentation based on semi-supervised learning |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106204449A (en) * | 2016-07-06 | 2016-12-07 | 安徽工业大学 | A kind of single image super resolution ratio reconstruction method based on symmetrical degree of depth network |
| AU2017101166A4 (en) * | 2017-08-25 | 2017-11-02 | Lai, Haodong MR | A Method For Real-Time Image Style Transfer Based On Conditional Generative Adversarial Networks |
| CN107316307A (en) * | 2017-06-27 | 2017-11-03 | 北京工业大学 | A kind of Chinese medicine tongue image automatic segmentation method based on depth convolutional neural networks |
| CN107784308A (en) * | 2017-10-09 | 2018-03-09 | 哈尔滨工业大学 | Conspicuousness object detection method based on the multiple dimensioned full convolutional network of chain type |
-
2018
- 2018-03-22 CN CN201810238288.4A patent/CN108564166A/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106204449A (en) * | 2016-07-06 | 2016-12-07 | 安徽工业大学 | A kind of single image super resolution ratio reconstruction method based on symmetrical degree of depth network |
| CN107316307A (en) * | 2017-06-27 | 2017-11-03 | 北京工业大学 | A kind of Chinese medicine tongue image automatic segmentation method based on depth convolutional neural networks |
| AU2017101166A4 (en) * | 2017-08-25 | 2017-11-02 | Lai, Haodong MR | A Method For Real-Time Image Style Transfer Based On Conditional Generative Adversarial Networks |
| CN107784308A (en) * | 2017-10-09 | 2018-03-09 | 哈尔滨工业大学 | Conspicuousness object detection method based on the multiple dimensioned full convolutional network of chain type |
Non-Patent Citations (8)
Cited By (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2020061972A1 (en) * | 2018-09-27 | 2020-04-02 | 电子科技大学 | Convolutional neural network-based diabetic retinopathy diagnosing technique |
| CN112396136A (en) * | 2019-08-15 | 2021-02-23 | 北京神州泰岳软件股份有限公司 | Handwritten number recognition method and device, electronic equipment and readable storage medium |
| CN110543563A (en) * | 2019-08-20 | 2019-12-06 | 暨南大学 | A hierarchical text classification method and system |
| CN110543563B (en) * | 2019-08-20 | 2022-03-08 | 暨南大学 | Hierarchical text classification method and system |
| CN114341951A (en) * | 2019-08-20 | 2022-04-12 | 阿卡蒂科有限公司 | Method and system for diagnosing diseases, physiological changes or other internal conditions of crustaceans by non-invasive means |
| CN110598629A (en) * | 2019-09-11 | 2019-12-20 | 北京百度网讯科技有限公司 | Search space construction method and device of hyper network and electronic equipment |
| CN110598629B (en) * | 2019-09-11 | 2023-06-06 | 北京百度网讯科技有限公司 | Super-network search space construction method and device and electronic equipment |
| CN111263226A (en) * | 2020-01-17 | 2020-06-09 | 中国科学技术大学 | Video processing method, video processing device, electronic equipment and medium |
| CN111263226B (en) * | 2020-01-17 | 2021-10-22 | 中国科学技术大学 | Video processing method, device, electronic device and medium |
| CN115147426A (en) * | 2022-09-06 | 2022-10-04 | 北京大学 | Method and system for model training and image segmentation based on semi-supervised learning |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN108564166A (en) | Based on the semi-supervised feature learning method of the convolutional neural networks with symmetrical parallel link | |
| CN113378965B (en) | Multi-label image identification method and system based on DCGAN and GCN | |
| CN108133188B (en) | Behavior identification method based on motion history image and convolutional neural network | |
| CN111275092B (en) | An Image Classification Method Based on Unsupervised Domain Adaptation | |
| CN103345656B (en) | A kind of data identification method based on multitask deep neural network and device | |
| CN108875696A (en) | The Off-line Handwritten Chinese Recognition method of convolutional neural networks is separated based on depth | |
| CN107016406A (en) | The pest and disease damage image generating method of network is resisted based on production | |
| CN113065459B (en) | Video instance segmentation method and system based on dynamic condition convolution | |
| CN110213244A (en) | A kind of network inbreak detection method based on space-time characteristic fusion | |
| CN106228177A (en) | Daily life subject image recognition methods based on convolutional neural networks | |
| CN106980641A (en) | The quick picture retrieval system of unsupervised Hash and method based on convolutional neural networks | |
| CN106355248A (en) | Deep convolution neural network training method and device | |
| CN108961272A (en) | It is a kind of to fight the generation method for generating the skin disease image of network based on depth convolution | |
| CN107145893A (en) | An image recognition algorithm and system based on convolutional deep network | |
| CN111368734B (en) | A micro-expression recognition method based on normal expression assistance | |
| CN105678248A (en) | Face key point alignment algorithm based on deep learning | |
| CN108537747A (en) | A kind of image repair method based on the convolutional neural networks with symmetrical parallel link | |
| CN114565880B (en) | Method, system and equipment for detecting counterfeit video based on optical flow tracking | |
| Jiang et al. | Face2nodes: learning facial expression representations with relation-aware dynamic graph convolution networks | |
| CN114463810A (en) | Training method and device for face recognition model | |
| CN119991683B (en) | Industrial large model anomaly detection method integrating image mask and mixed expert | |
| CN108986091A (en) | Casting defect image detecting method based on depth Hash network | |
| CN107463932A (en) | A kind of method that picture feature is extracted using binary system bottleneck neutral net | |
| CN111275076B (en) | Image saliency detection method based on feature selection and feature fusion | |
| Salem et al. | Semantic image inpainting using self-learning encoder-decoder and adversarial loss |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180921 |