WO2020119624A1 - Class-sensitive edge detection method based on deep learning - Google Patents

Class-sensitive edge detection method based on deep learning Download PDF

Info

Publication number
WO2020119624A1
WO2020119624A1 PCT/CN2019/123968 CN2019123968W WO2020119624A1 WO 2020119624 A1 WO2020119624 A1 WO 2020119624A1 CN 2019123968 W CN2019123968 W CN 2019123968W WO 2020119624 A1 WO2020119624 A1 WO 2020119624A1
Authority
WO
WIPO (PCT)
Prior art keywords
deep learning
class
feature
edge
edge detection
Prior art date
Application number
PCT/CN2019/123968
Other languages
French (fr)
Chinese (zh)
Inventor
王磊
徐成俊
程俊
Original Assignee
中国科学院深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中国科学院深圳先进技术研究院 filed Critical 中国科学院深圳先进技术研究院
Publication of WO2020119624A1 publication Critical patent/WO2020119624A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection

Definitions

  • the invention relates to the field of image processing, in particular to the edge detection technology of images.
  • Edge detection is a basic problem in the field of image processing.
  • the research on edge segmentation has always been the basic task of computer vision and the technical foundation of many researches and applications.
  • cell segmentation of medical images, road segmentation in autonomous driving, etc. are all specific applications of edge detection technology.
  • edge detection algorithms Sobel operator, Laplacian operator, and Canny operator. These methods have their own applicable situations, and the efficiency and robustness of obtaining edge images are limited.
  • Some technologies have gradually proposed the use of deep learning to obtain image edges, such as references [1] and [2]. Although these technologies have improved efficiency and robustness, they have been greatly improved.
  • the edge of is relatively thicker than the real edge, and the detection accuracy still needs to be improved.
  • these edge detection methods can only obtain binary edge images, and cannot classify the obtained edges, and further obtain the semantic information of the edges.
  • the present invention provides an edge detection method with high detection accuracy and capable of obtaining category information of edges.
  • a class-sensitive edge detection method based on deep learning using a deep learning neural network model to obtain edge classification detection results of several class targets.
  • the deep learning neural network model is a CNN convolutional neural network model, which includes a feature extraction step 1, an upsampling step 2, and a feature fusion and classification step 3.
  • the feature extraction step 1 includes five stages of feature extraction, S1-S5, to obtain edge feature maps of binary classification with different scales.
  • an adaptive scale transformation label is used to supervise the training.
  • the up-sampling step 2 is up-sampling the edge features extracted by the feature extraction step 1, the up-sampled image size is consistent with the size of the image to be detected; the feature fusion and classification step 3 is based on Based on the result of the upsampling step 2, feature fusion is performed, and the edge categories are classified at the same time.
  • F new ⁇ F 1 ,E,F 2 ,E,...,F n ,E ⁇ ,
  • E represents the set of edge feature images obtained by supervised learning in feature extraction step 1
  • F represents the feature obtained in the sixth stage S6
  • F n means that F has n channels in total
  • n is the number of categories of the target to be identified
  • F new is the new feature obtained after feature fusion
  • the sixth-stage feature S6 is the up-sampling of the fifth-stage feature S5.
  • the loss function of the CNN convolutional neural network model is used as follows: the loss function of the reset weight is used in the supervision of the binary image of the S1-S5 stage; the general category is used in the multi-class edge supervision of the S6 stage Cross loss function; the loss function of the reset weight used in the final feature fusion and classification.
  • the obtained edges of different categories are represented by different colors.
  • the present invention can detect specific target edges and classify the obtained edges at the same time. Compared with the existing deep learning edge detection method, it not only improves the detection accuracy, but also obtains a more detailed image Edge, and almost no post-processing is required; and the function has been expanded to provide higher performance guarantee for other tasks based on edge, such as target segmentation and instance segmentation.
  • Figure 1 is an algorithm block diagram of the present invention.
  • Figure 2 is a comparison chart using different loss functions.
  • the algorithm block diagram of the class-sensitive edge detection method based on deep learning is shown in Figure 1.
  • the feature extraction step 1 includes five stages, namely S1 to S5, which can obtain edge feature maps of binary classification with different scales, that is, S1-edge and S2-edge in FIG. 1.
  • the size of the S1 to S4 process is reduced twice, and the S4 size is as large as the S5 size.
  • the up-sampling step 2 includes the S6 stage, as well as S1-edge-upsample, S2-edge-upsample, and S3-edge-upsample.
  • S6 is the up-sampling of S5 features, the size of the up-sampling is consistent with the image to be detected.
  • S1-edge-upsample, S2-edge-upsample, and S3-edge-upsample are the up-sampling of S1-edge, S2-edge, and S3-edge, respectively, and the size is also consistent with the image to be detected.
  • Feature fusion and classification step 3 is to fuse all the up-sampled features and get the final multi-class target edge detection result.
  • Feature fusion and classification are as follows:
  • F new ⁇ F 1 ,E,F 2 ,E,...,F n ,E ⁇
  • E represents the edge feature image set supervised and learned in the middle stage
  • F is the feature obtained in the sixth stage
  • F n means that there are n channels in total, and n is the number of categories that want to identify the target.
  • F new is a new feature obtained after feature fusion.
  • the new feature F new is obtained by fusion for the final multi-category classification of edges.
  • supervised learning is adopted for CNN convolutional neural network.
  • Supervise training of multi-scale binary edge images in S1-S5 process that is, only perform edge two-class classification; then perform multi-classification of edge pixels at S6, classification output at S6 stage, which is used for preliminary supervised learning of the model ,
  • the number of output channels is consistent with the number of categories to be detected.
  • the training data label because the feature scales obtained in each stage are different during feature extraction, such as S1-edge, S2-edge, and S3-edge in Figure 1, the training data label only provides one, so it is proposed
  • An adaptive scale transformation label is used for supervised training of intermediate processes. When performing stage supervision training, the label will be adjusted to the size of the corresponding stage feature map according to the size of the current feature map, and then the loss function calculation will be performed.
  • the reset weight loss function and the general cross loss entropy function as the loss function of the entire neural network, specifically: the reset weight is used in the supervision of the binary image in the S1-S5 stage The loss function of Classification; the multi-class edge supervision in the Classification of S6 stage uses the general cross loss function; the loss function of the final weight used in the final feature fusion and classification. After many experiments, the results show that the cross-reuse loss function and the general cross loss entropy function can improve the detection accuracy and refine the edges.
  • the present invention uses two data sets for training and testing.
  • One is the SBD (Semantic Boundary Dataset) dataset.
  • the SBD dataset contains 11355 pictures, of which 8498 are used for training and 2857 for verification.
  • the second is the Cityscapes dataset, which contains 5000 images, of which 2975 are used for training, 500 for verification, and 1525 for testing.
  • the input image size is 400x400, and the evaluation index used is F-measure.
  • the calculation method is as follows:
  • the 20 types of verification indicators on the SBD data set are as follows:
  • FIG. 2(c) The effect of the present invention is shown in Figure 2, where (a) and (d) are the original pictures, (b) and (e) are the detection results obtained by using only the loss function of reset weights, (c) and (f) Test results obtained by using the method of the present invention.
  • Fig. 2(c) different categories of areas are marked with different colors, in which chairs and sofas are represented by red and green frames, namely dark and light frames in Fig. 2(c).

Abstract

A class-sensitive edge detection method based on deep learning. A deep learning neural network model is used to obtain edge classification detection results of several class targets. A deep supervision method is used for model training. Labels for adaptive scale transformation are used during a training process. A loss function of a reset weight and a general cross-loss entropy function are used intersectingly. By employing the present method, specific target edges may be detected and the obtained edges may be classified at the same time. Compared to existing deep learning edge detection methods, the present invention not only improves the detection accuracy, but also obtains more detailed image edges while requiring little subsequent reprocessing. In addition, functions have been expanded, which may provide better performance guarantee for other tasks which have edges as the basis, such as target segmentation and instance segmentation.

Description

一种基于深度学习的类别敏感型边缘检测方法A category-sensitive edge detection method based on deep learning 技术领域Technical field
本发明涉及图像处理领域,尤其涉及图像的边缘检测技术。The invention relates to the field of image processing, in particular to the edge detection technology of images.
背景技术Background technique
边缘检测是图像处理领域中的基础问题,对边缘分割的研究一直是计算机视觉的基本任务,也是很多研究和应用的技术基础。例如医学图像的细胞分割,自动驾驶中道路分割等,都是边缘检测技术的具体应用。Edge detection is a basic problem in the field of image processing. The research on edge segmentation has always been the basic task of computer vision and the technical foundation of many researches and applications. For example, cell segmentation of medical images, road segmentation in autonomous driving, etc. are all specific applications of edge detection technology.
传统的边缘检测算法有三种:Sobel算子,Laplacian算子,和Canny算子。这些方法有各自的适用情形,得到边缘图像的效率和鲁棒性都有局限。随着人工智能的快速发展,逐渐有技术提出采用深度学习的方式来得到图像边缘,如参考文献[1]、[2],这些技术在效率和鲁棒性上虽然有较大提升,但得到的边缘相对于真实的边缘较粗,检测精度仍有待提高。而且这些边缘检测方法都只能得到二值边缘图像,不能对得到边缘进行分类,进一步得到边缘的语义信息。There are three traditional edge detection algorithms: Sobel operator, Laplacian operator, and Canny operator. These methods have their own applicable situations, and the efficiency and robustness of obtaining edge images are limited. With the rapid development of artificial intelligence, some technologies have gradually proposed the use of deep learning to obtain image edges, such as references [1] and [2]. Although these technologies have improved efficiency and robustness, they have been greatly improved. The edge of is relatively thicker than the real edge, and the detection accuracy still needs to be improved. Moreover, these edge detection methods can only obtain binary edge images, and cannot classify the obtained edges, and further obtain the semantic information of the edges.
参考文献:references:
[1]S.Xie and Z.Tu.Holistically-nested edge detection.In IJCV.Springer,2017[1]S.Xie and Z.Tu.Holistically-nested edge detection.In IJCV.Springer,2017
[2]G.Bertasius,J.Shi,and L.Torresani.DeepEdge:A multiscale bifurcated deep network for top-down contour detection.In IEEE CVPR,pages 4380–4389,2015[2]G.Bertasius,J.Shi,and L.Torresani.DeepEdge:A multiscale bifurcated deep network for top-down contour detection.In IEEE CVPR,pages 4380–4389,2015
发明内容Summary of the invention
针对上述现有技术的缺陷,本发明提供了一种检测精度高、能得到边缘的类别信息的边缘检测方法。In view of the above-mentioned defects of the prior art, the present invention provides an edge detection method with high detection accuracy and capable of obtaining category information of edges.
一种基于深度学习的类别敏感型边缘检测方法,采用深度学习神经网络模型得到若干类别目标的边缘分类检测结果。A class-sensitive edge detection method based on deep learning, using a deep learning neural network model to obtain edge classification detection results of several class targets.
进一步地,将待检测图像输入所述深度学习神经网络模型,所述深度学习神经网络模型为CNN卷积神经网络模型,其中包括特征提取步骤1、上采样步骤2,特征融合及分类步骤3。Further, the image to be detected is input to the deep learning neural network model. The deep learning neural network model is a CNN convolutional neural network model, which includes a feature extraction step 1, an upsampling step 2, and a feature fusion and classification step 3.
进一步地,所述特征提取步骤1包括5个阶段的特征提取,S1-S5,得到不同尺度的二值分类的边缘特征图。Further, the feature extraction step 1 includes five stages of feature extraction, S1-S5, to obtain edge feature maps of binary classification with different scales.
进一步地,在训练所述CNN卷积神经网络模型时,采用一个自适应尺度变换的标签来监督训练。Further, when training the CNN convolutional neural network model, an adaptive scale transformation label is used to supervise the training.
进一步地,所述上采样步骤2,是对所述特征提取步骤1提取的边缘特征进行上采样,上采样的图像尺寸与待检测图像大小一致;所述特征融合及分类步骤3,是根据所述上采样步骤2的结果,进行特征融合,同时对边缘的类别进行分类。Further, the up-sampling step 2 is up-sampling the edge features extracted by the feature extraction step 1, the up-sampled image size is consistent with the size of the image to be detected; the feature fusion and classification step 3 is based on Based on the result of the upsampling step 2, feature fusion is performed, and the edge categories are classified at the same time.
进一步地,所述特征融合及分类的方式为:Further, the feature fusion and classification methods are:
E={E 1,E 2,E 3}, E={E 1 ,E 2 ,E 3 },
F={F 1,F 2,…,F n}, F={F 1 ,F 2 ,...,F n },
F new={F 1,E,F 2,E,…,F n,E}, F new ={F 1 ,E,F 2 ,E,...,F n ,E},
其中E代表特征提取步骤1中通过监督学习得到的边缘特征图像集合,E i代表第i个阶段得到边缘特征的上采样特征,i=1、2、3,F代表第6阶段S6得到的特征,F n表示F一共有n个通道,n为要识别目标的类别数,F new是特征融合后得到的新特征,其中第6阶段的特征S6是对第5阶段特征S5的上采样。 Where E represents the set of edge feature images obtained by supervised learning in feature extraction step 1, E i represents the up-sampling feature of the edge feature obtained in the i-th stage, i=1, 2, 3, and F represents the feature obtained in the sixth stage S6 , F n means that F has n channels in total, n is the number of categories of the target to be identified, F new is the new feature obtained after feature fusion, and the sixth-stage feature S6 is the up-sampling of the fifth-stage feature S5.
进一步地,所述CNN卷积神经网络模型的损失函数使用方式为:在S1-S5阶段的二值图像的监督中使用重置权重的损失函数;在S6阶段的多类别边缘监督中使用一般的交叉损失函数;在最后特征融合及分类中使用的重置权重的损失函数。Further, the loss function of the CNN convolutional neural network model is used as follows: the loss function of the reset weight is used in the supervision of the binary image of the S1-S5 stage; the general category is used in the multi-class edge supervision of the S6 stage Cross loss function; the loss function of the reset weight used in the final feature fusion and classification.
进一步地,得到的不同类别边缘采用不同颜色表示。Further, the obtained edges of different categories are represented by different colors.
本发明的有益效果为:本发明能够对特定的目标边缘进行检测,同时对得到的边缘进行分类,相较于现有的深度学习边缘检测方法,不仅提高了检测精度,得到更加细化的图像边缘,且几乎不需要后期再处理;而且在功能上进行了扩展,能够对以边缘为基础的其他任务,如目标分割、实例分割等任务提供更高的性能保障。The beneficial effects of the present invention are: the present invention can detect specific target edges and classify the obtained edges at the same time. Compared with the existing deep learning edge detection method, it not only improves the detection accuracy, but also obtains a more detailed image Edge, and almost no post-processing is required; and the function has been expanded to provide higher performance guarantee for other tasks based on edge, such as target segmentation and instance segmentation.
附图说明BRIEF DESCRIPTION
图1为本发明的算法框图。Figure 1 is an algorithm block diagram of the present invention.
图2为使用不同损失函数对比图。Figure 2 is a comparison chart using different loss functions.
具体实施方式detailed description
基于深度学习的类别敏感型边缘检测方法的算法框图如图1所示。首先将待检测图像输入CNN卷积神经网络模型,该模型包括特征提取步骤1,上采样步 骤2,以及特征融合及分类步骤3。特征提取步骤1中包括5个阶段,即S1至S5,可以得到不同尺度的二值分类的边缘特征图,即图1中的S1-edge、S2-edge等。其中S1到S4过程的尺寸每次缩小两倍,S4尺寸和S5尺寸一样大。The algorithm block diagram of the class-sensitive edge detection method based on deep learning is shown in Figure 1. First, input the image to be detected into the CNN convolutional neural network model, which includes feature extraction step 1, up-sampling step 2, and feature fusion and classification step 3. The feature extraction step 1 includes five stages, namely S1 to S5, which can obtain edge feature maps of binary classification with different scales, that is, S1-edge and S2-edge in FIG. 1. The size of the S1 to S4 process is reduced twice, and the S4 size is as large as the S5 size.
上采样步骤2中包括S6阶段,以及得到S1-edge-upsample,S2-edge-upsample,S3-edge-upsample。其中,S6是对S5特征的上采样,上采样的大小与待检测图像一致。S1-edge-upsample,S2-edge-upsample,S3-edge-upsample分别是对S1-edge,S2-edge,S3-edge的上采样,大小也与待检测图像一致。The up-sampling step 2 includes the S6 stage, as well as S1-edge-upsample, S2-edge-upsample, and S3-edge-upsample. Among them, S6 is the up-sampling of S5 features, the size of the up-sampling is consistent with the image to be detected. S1-edge-upsample, S2-edge-upsample, and S3-edge-upsample are the up-sampling of S1-edge, S2-edge, and S3-edge, respectively, and the size is also consistent with the image to be detected.
特征融合及分类步骤3,是把所有上采样得到的特征进行一个融合,并得到最后的多类别目标边缘检测结果。特征融合及分类的方式如下:Feature fusion and classification step 3 is to fuse all the up-sampled features and get the final multi-class target edge detection result. Feature fusion and classification are as follows:
E={E 1,E 2,E 3} E={E 1 ,E 2 ,E 3 }
F={F 1,F 2,…,F n} F={F 1 ,F 2 ,...,F n }
F new={F 1,E,F 2,E,…,F n,E} F new = {F 1 ,E,F 2 ,E,…,F n ,E}
其中E代表中间阶段监督学习得的边缘特征图像集合,E i代表第i个阶段得到的边缘特征的上采样,i=1、2、3。其中F是第6阶段得到的特征,F n表示F一共有n个通道,n是想要识别目标的类别数。F new是特征融合后得到的新特征。最后用融合得到新特征F new进行最后的边缘多类别分类。 Where E represents the edge feature image set supervised and learned in the middle stage, E i represents the up-sampling of the edge feature obtained in the i-th stage, i=1, 2, 3. Where F is the feature obtained in the sixth stage, F n means that there are n channels in total, and n is the number of categories that want to identify the target. F new is a new feature obtained after feature fusion. Finally, the new feature F new is obtained by fusion for the final multi-category classification of edges.
为了更好的得到边缘像素点多分类的效果,对CNN卷积神经网络采用监督学习的方式。在S1-S5过程进行多尺度的二值边缘图像的监督训练,即仅进行边缘的二分类;在S6时再进行边缘像素的多分类,S6阶段输出的classification,其用于模型初步的监督学习,其输出的通道数量与要检测的类别数量一致。在监督学习过程中,由于特征提取时,每个阶段得到的特征尺度不一样,如图1中的S1-edge,S2-edge,S3-edge,而训练数据的标签只提供了一个,因此提出了一种自适应尺度变换的标签用来进行中间过程的监督训练。进行阶段监督训练的时候,会根据当前的特征图的大小,标签自适应调整为相应阶段特征图的尺寸大小,然后再进行损失函数计算。In order to better obtain the effect of multi-classification of edge pixels, supervised learning is adopted for CNN convolutional neural network. Supervise training of multi-scale binary edge images in S1-S5 process, that is, only perform edge two-class classification; then perform multi-classification of edge pixels at S6, classification output at S6 stage, which is used for preliminary supervised learning of the model , The number of output channels is consistent with the number of categories to be detected. In the supervised learning process, because the feature scales obtained in each stage are different during feature extraction, such as S1-edge, S2-edge, and S3-edge in Figure 1, the training data label only provides one, so it is proposed An adaptive scale transformation label is used for supervised training of intermediate processes. When performing stage supervision training, the label will be adjusted to the size of the corresponding stage feature map according to the size of the current feature map, and then the loss function calculation will be performed.
对于深度学习中的损失函数,在已有的使用深度学习方法来做边缘检测的算法中,都只使用了重置权重的损失函数,该损失函数会使预测得到的边缘比较粗,需要进行后期的细化操作。在本发明中,我们交叉使用重置权重的损失函数和一般的交叉损失熵函数作为整个神经网络的损失函数,具体为:在S1-S5阶段的 二值图像的监督中使用的是重置权重的损失函数;在S6阶段的Classification多类别边缘监督使用的是一般的交叉损失函数;最后特征融合及分类中使用的重置权重的损失函数。经过多次实验,结果表明交叉使用重置权重的损失函数和一般的交叉损失熵函数,既能够提升检测精度又能细化边缘。For the loss function in deep learning, in the existing algorithms that use deep learning methods for edge detection, only the loss function of reset weight is used. This loss function will make the predicted edge rougher, and needs to be post-processed. Refinement operation. In the present invention, we alternately use the reset weight loss function and the general cross loss entropy function as the loss function of the entire neural network, specifically: the reset weight is used in the supervision of the binary image in the S1-S5 stage The loss function of Classification; the multi-class edge supervision in the Classification of S6 stage uses the general cross loss function; the loss function of the final weight used in the final feature fusion and classification. After many experiments, the results show that the cross-reuse loss function and the general cross loss entropy function can improve the detection accuracy and refine the edges.
本发明采用两个数据集进行训练和测试。一个是SBD(Semantic Boundary Dataset)数据集,SBD数据集包含11355张图片,其中8498张用来训练,2857张用来验证。第二个是Cityscapes数据集,它包含了5000张图片,其中2975张用来训练,500张用来验证,1525张用来测试。The present invention uses two data sets for training and testing. One is the SBD (Semantic Boundary Dataset) dataset. The SBD dataset contains 11355 pictures, of which 8498 are used for training and 2857 for verification. The second is the Cityscapes dataset, which contains 5000 images, of which 2975 are used for training, 500 for verification, and 1525 for testing.
其中输入的图像尺寸为400x400,使用的评价指标是F-measure,。其计算方式如下:The input image size is 400x400, and the evaluation index used is F-measure. The calculation method is as follows:
Figure PCTCN2019123968-appb-000001
Figure PCTCN2019123968-appb-000001
其中
Figure PCTCN2019123968-appb-000002
TP,FP,FN的含义见下表:
among them
Figure PCTCN2019123968-appb-000002
The meaning of TP, FP, FN is shown in the table below:
Figure PCTCN2019123968-appb-000003
Figure PCTCN2019123968-appb-000003
在SBD数据集上20类的验证指标如下:The 20 types of verification indicators on the SBD data set are as follows:
Figure PCTCN2019123968-appb-000004
Figure PCTCN2019123968-appb-000004
本发明的效果如图2所示,其中(a)、(d)为原始图片,(b)、(e) 为只使用重置权重的损失函数得到的检测结果,(c)、(f)为使用本发明方法得到的测试结果。图2(c)中采用不同颜色标识出不同的类别区域,其中椅子和沙发分别用红色框与绿色框表示,即图2(c)中的深色框、浅色框。The effect of the present invention is shown in Figure 2, where (a) and (d) are the original pictures, (b) and (e) are the detection results obtained by using only the loss function of reset weights, (c) and (f) Test results obtained by using the method of the present invention. In Fig. 2(c), different categories of areas are marked with different colors, in which chairs and sofas are represented by red and green frames, namely dark and light frames in Fig. 2(c).
所述领域的普通技术人员应当理解,以上所述仅为本发明的具体实施例而已,并不用于限制本发明,凡在本发明的精神和原则之内,所做的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。Persons of ordinary skill in the field should understand that the above is only specific embodiments of the present invention and is not intended to limit the present invention. Any modification, equivalent replacement, or replacement made within the spirit and principle of the present invention, Improvements, etc., should be included in the protection scope of the present invention.

Claims (8)

  1. 一种基于深度学习的类别敏感型边缘检测方法,其特征在于,采用深度学习神经网络模型得到若干类别目标的边缘分类检测结果。A class-sensitive edge detection method based on deep learning is characterized in that a deep learning neural network model is used to obtain edge classification detection results of several class targets.
  2. 如权利要求1所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,将待检测图像输入所述深度学习神经网络模型,所述深度学习神经网络模型为CNN卷积神经网络模型,其中包括特征提取步骤(1)、上采样步骤(2),特征融合及分类步骤(3)。The class-sensitive edge detection method based on deep learning according to claim 1, wherein the image to be detected is input to the deep learning neural network model, and the deep learning neural network model is a CNN convolutional neural network model, It includes feature extraction step (1), up-sampling step (2), feature fusion and classification step (3).
  3. 如权利要求2所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,所述特征提取步骤(1)包括5个阶段的特征提取(S1-S5),得到不同尺度的二值分类的边缘特征图。The class-sensitive edge detection method based on deep learning according to claim 2, wherein the feature extraction step (1) includes five stages of feature extraction (S1-S5) to obtain binary classification at different scales Edge feature map.
  4. 如权利要求2或3所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,在训练所述CNN卷积神经网络模型时,采用一个自适应尺度变换的标签来监督训练。The class-sensitive edge detection method based on deep learning according to claim 2 or 3, characterized in that, when training the CNN convolutional neural network model, an adaptive scale transformation label is used to supervise the training.
  5. 如权利要求2所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,所述上采样步骤(2),是对所述特征提取步骤(1)提取的边缘特征进行上采样,上采样的图像尺寸与待检测图像大小一致;所述特征融合及分类步骤(3),是根据所述上采样步骤(2)的结果,进行特征融合,同时对边缘的类别进行分类。The class-sensitive edge detection method based on deep learning according to claim 2, wherein the up-sampling step (2) is to up-sample the edge features extracted by the feature extraction step (1). The size of the sampled image is consistent with the size of the image to be detected; the feature fusion and classification step (3) is to perform feature fusion based on the result of the upsampling step (2) and classify the edge category at the same time.
  6. 如权利要求5所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,所述特征融合及分类的方式为:The class-sensitive edge detection method based on deep learning according to claim 5, wherein the feature fusion and classification methods are:
    E={E 1,E 2,E 3}, E={E 1 ,E 2 ,E 3 },
    F={F 1,F 2,…,F n}, F={F 1 ,F 2 ,...,F n },
    F new={F 1,E,F 2,E,…,F n,E}, F new ={F 1 ,E,F 2 ,E,...,F n ,E},
    其中E代表特征提取步骤(1)中通过监督学习得到的边缘特征图像集合,E i代表第i个阶段得到边缘特征的上采样特征,i=1、2、3,F代表第6阶段(S6)得到的特征,F n表示F一共有n个通道,n为要识别目标的类别数,F new是特征 融合后得到的新特征,其中第6阶段的特征(S6)是对第5阶段特征(S5)的上采样。 Where E represents the edge feature image set obtained by supervised learning in the feature extraction step (1), E i represents the up-sampling feature of the edge feature obtained at the i-th stage, i=1, 2, 3, and F represents the sixth stage (S6 ) The obtained features, F n means that F has n channels in total, n is the number of categories of the target to be identified, F new is the new feature obtained after feature fusion, and the features of stage 6 (S6) are the features of stage 5 (S5) Upsampling.
  7. 如权利要求6所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,所述CNN卷积神经网络模型的损失函数的使用方式为:在特征提取步骤(1)的二值图像的监督学习中使用重置权重的损失函数;在第6阶段(S6)的多类别边缘监督学习中使用一般的交叉损失函数;在特征融合及分类步骤(3)中使用重置权重的损失函数。The class-sensitive edge detection method based on deep learning according to claim 6, wherein the loss function of the CNN convolutional neural network model is used as follows: the binary image in the feature extraction step (1) In supervised learning, the loss function of reset weight is used; in the sixth stage (S6) multi-class edge supervised learning, the general cross loss function is used; in the feature fusion and classification step (3), the loss function of reset weight is used.
  8. 如权利要求1-7任一项所述的基于深度学习的类别敏感型边缘检测方法,其特征在于,得到的不同类别边缘采用不同颜色表示。The method for class-sensitive edge detection based on deep learning according to any one of claims 1-7, wherein the obtained different class edges are represented by different colors.
PCT/CN2019/123968 2018-12-12 2019-12-09 Class-sensitive edge detection method based on deep learning WO2020119624A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811520198.0A CN109741351A (en) 2018-12-12 2018-12-12 A kind of classification responsive type edge detection method based on deep learning
CN201811520198.0 2018-12-12

Publications (1)

Publication Number Publication Date
WO2020119624A1 true WO2020119624A1 (en) 2020-06-18

Family

ID=66359395

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/123968 WO2020119624A1 (en) 2018-12-12 2019-12-09 Class-sensitive edge detection method based on deep learning

Country Status (2)

Country Link
CN (1) CN109741351A (en)
WO (1) WO2020119624A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115588024A (en) * 2022-11-25 2023-01-10 东莞市兆丰精密仪器有限公司 Artificial intelligence-based complex industrial image edge extraction method and device

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109741351A (en) * 2018-12-12 2019-05-10 中国科学院深圳先进技术研究院 A kind of classification responsive type edge detection method based on deep learning
CN110310254B (en) * 2019-05-17 2022-11-29 广东技术师范大学 Automatic room corner image grading method based on deep learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018023734A1 (en) * 2016-08-05 2018-02-08 深圳大学 Significance testing method for 3d image
CN108710919A (en) * 2018-05-25 2018-10-26 东南大学 A kind of crack automation delineation method based on multi-scale feature fusion deep learning
CN109741351A (en) * 2018-12-12 2019-05-10 中国科学院深圳先进技术研究院 A kind of classification responsive type edge detection method based on deep learning

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104952073B (en) * 2015-06-15 2017-12-15 上海交通大学 Scene Incision method based on deep learning
CN105069807B (en) * 2015-08-28 2018-03-23 西安工程大学 A kind of stamped workpieces defect inspection method based on image procossing
CN107610140A (en) * 2017-08-07 2018-01-19 中国科学院自动化研究所 Near edge detection method, device based on depth integration corrective networks

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018023734A1 (en) * 2016-08-05 2018-02-08 深圳大学 Significance testing method for 3d image
CN108710919A (en) * 2018-05-25 2018-10-26 东南大学 A kind of crack automation delineation method based on multi-scale feature fusion deep learning
CN109741351A (en) * 2018-12-12 2019-05-10 中国科学院深圳先进技术研究院 A kind of classification responsive type edge detection method based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YU, ZHIDING ET AL.: "CASENet: Deep Category-Aware Semantic Edge Detection", 2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 31 December 2017 (2017-12-31), XP033249517, DOI: 20200303111926X *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115588024A (en) * 2022-11-25 2023-01-10 东莞市兆丰精密仪器有限公司 Artificial intelligence-based complex industrial image edge extraction method and device

Also Published As

Publication number Publication date
CN109741351A (en) 2019-05-10

Similar Documents

Publication Publication Date Title
CN110543837B (en) Visible light airport airplane detection method based on potential target point
CN110348319B (en) Face anti-counterfeiting method based on face depth information and edge image fusion
US10846566B2 (en) Method and system for multi-scale cell image segmentation using multiple parallel convolutional neural networks
CN110781897B (en) Semantic edge detection method based on deep learning
CN106408030B (en) SAR image classification method based on middle layer semantic attribute and convolutional neural networks
CN104361313B (en) A kind of gesture identification method merged based on Multiple Kernel Learning heterogeneous characteristic
CN105138998B (en) Pedestrian based on the adaptive sub-space learning algorithm in visual angle recognition methods and system again
Zhang et al. Infrared image segmentation for photovoltaic panels based on Res-UNet
WO2020119624A1 (en) Class-sensitive edge detection method based on deep learning
CN109740686A (en) A kind of deep learning image multiple labeling classification method based on pool area and Fusion Features
CN106126585B (en) The unmanned plane image search method combined based on quality grading with perceived hash characteristics
Zang et al. Traffic sign detection based on cascaded convolutional neural networks
CN104036284A (en) Adaboost algorithm based multi-scale pedestrian detection method
CN108734200B (en) Human target visual detection method and device based on BING (building information network) features
CN103106265A (en) Method and system of classifying similar images
CN112926652B (en) Fish fine granularity image recognition method based on deep learning
CN110008899B (en) Method for extracting and classifying candidate targets of visible light remote sensing image
CN103186776A (en) Human detection method based on multiple features and depth information
Intwala et al. Indian sign language converter using convolutional neural networks
CN111026898A (en) Weak supervision image emotion classification and positioning method based on cross space pooling strategy
CN114492634B (en) Fine granularity equipment picture classification and identification method and system
Li et al. HEp-2 specimen classification via deep CNNs and pattern histogram
Khellal et al. Pedestrian classification and detection in far infrared images
CN106203448A (en) A kind of scene classification method based on Nonlinear Scale Space Theory
CN113487610B (en) Herpes image recognition method and device, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19897270

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03.11.2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19897270

Country of ref document: EP

Kind code of ref document: A1