CN116309429A - A Chip Defect Detection Method Based on Deep Learning - Google Patents
A Chip Defect Detection Method Based on Deep Learning Download PDFInfo
- Publication number
- CN116309429A CN116309429A CN202310240534.0A CN202310240534A CN116309429A CN 116309429 A CN116309429 A CN 116309429A CN 202310240534 A CN202310240534 A CN 202310240534A CN 116309429 A CN116309429 A CN 116309429A
- Authority
- CN
- China
- Prior art keywords
- convolution
- feature
- chip
- network
- resnet
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 47
- 230000007547 defect Effects 0.000 title claims abstract description 45
- 238000013135 deep learning Methods 0.000 title claims abstract description 16
- 230000004927 fusion Effects 0.000 claims abstract description 17
- 238000000605 extraction Methods 0.000 claims abstract description 16
- 238000012549 training Methods 0.000 claims abstract description 10
- 230000001629 suppression Effects 0.000 claims abstract description 8
- 238000012360 testing method Methods 0.000 claims abstract description 4
- 238000000034 method Methods 0.000 claims description 17
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims description 12
- 238000013527 convolutional neural network Methods 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 8
- 230000004913 activation Effects 0.000 claims description 6
- 239000011159 matrix material Substances 0.000 claims description 6
- 238000011176 pooling Methods 0.000 claims description 6
- 238000005070 sampling Methods 0.000 claims description 6
- 238000010586 diagram Methods 0.000 claims description 5
- 230000004048 modification Effects 0.000 claims description 4
- 238000012986 modification Methods 0.000 claims description 4
- 230000002776 aggregation Effects 0.000 claims description 3
- 238000004220 aggregation Methods 0.000 claims description 3
- 238000013528 artificial neural network Methods 0.000 claims description 3
- 230000008901 benefit Effects 0.000 claims description 3
- 230000008859 change Effects 0.000 claims description 3
- 230000008034 disappearance Effects 0.000 claims description 3
- 230000006870 function Effects 0.000 claims description 3
- 230000001788 irregular Effects 0.000 claims description 3
- 238000007781 pre-processing Methods 0.000 claims description 3
- 238000013519 translation Methods 0.000 claims description 3
- 238000013139 quantization Methods 0.000 claims 1
- 238000002372 labelling Methods 0.000 abstract 1
- 238000005516 engineering process Methods 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/806—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
- G06T2207/20104—Interactive definition of region of interest [ROI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/30—Computing systems specially adapted for manufacturing
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
本发明公开了一种基于深度学习的芯片缺陷检测方法,包括如下步骤:过高速工业相机采集不同种类的芯片表面缺陷图像;通过图像标注工具对芯片表面图像缺陷进行位置和类型的标注,构建缺陷数据集,将缺陷数据集分为训练数据集D和测试数据集T;基于改进的FasterR‑CNN框架,构建用于芯片缺陷检测的深度学习网络模型;该一种基于深度学习的芯片缺陷检测方法通过采用可变形卷积重构特征提取网络、多尺度特征融合、感兴趣区域对齐(ROI对齐)、软非最大值抑制等措施对FasterR‑CNN算法进行改进,解决了芯片缺陷体积小、形状复杂等问题,提高了缺陷检测的精度。
The invention discloses a chip defect detection method based on deep learning, which includes the following steps: collecting different types of chip surface defect images through a high-speed industrial camera; using an image labeling tool to mark the position and type of chip surface image defects, and constructing defects Data set, the defect data set is divided into training data set D and test data set T; based on the improved FasterR-CNN framework, a deep learning network model for chip defect detection is constructed; this chip defect detection method based on deep learning Improve the Faster R-CNN algorithm by adopting deformable convolution reconstruction feature extraction network, multi-scale feature fusion, region of interest alignment (ROI alignment), soft non-maximum suppression and other measures to solve the problem of small size and complex shape of chip defects and other issues, improving the accuracy of defect detection.
Description
技术领域technical field
本发明涉及人工智能技术领域,具体为一种基于深度学习的芯片缺陷检测方法。The invention relates to the technical field of artificial intelligence, in particular to a chip defect detection method based on deep learning.
背景技术Background technique
芯片在通电之后会产生一个启动指令来传递信0号以及传输数据,也可以让家电智联起来,是高端制造业的核心基础。但是芯片的制造工艺非常复杂,在生产过程中极容易会产生一些不符合预期的结构,造成目标缺陷。近年来,深度学习技术得到迅速发展,基于深度学习在大量数据中的强大学习能力和特征提取能力,因此很多研究者尝试将深度学习技术应用在产品的缺陷检测中,使得缺陷检测的效率得到大幅提高。然而,小型目标检测问题仍是工业产品表面缺陷检测领域的难点之一,因此提高对芯片这种小目标的缺陷检测率成为亟待解决的问题。After the chip is powered on, it will generate a start command to transmit signal 0 and transmit data, and it can also make home appliances intelligently connected, which is the core foundation of high-end manufacturing. However, the chip manufacturing process is very complicated, and it is very easy to produce some unexpected structures during the production process, resulting in target defects. In recent years, deep learning technology has developed rapidly. Based on the powerful learning ability and feature extraction ability of deep learning in a large amount of data, many researchers try to apply deep learning technology to product defect detection, which greatly improves the efficiency of defect detection. improve. However, the problem of small target detection is still one of the difficulties in the field of surface defect detection of industrial products, so improving the defect detection rate of small targets such as chips has become an urgent problem to be solved.
发明内容Contents of the invention
本发明的目的在于针对现有技术的不足之处,提供一种基于深度学习的芯片缺陷检测方法,以解决背景技术中所提出的问题。The purpose of the present invention is to provide a chip defect detection method based on deep learning to solve the problems raised in the background technology, aiming at the deficiencies of the prior art.
为实现上述目的,本发明提供如下技术方案:包括如下步骤:To achieve the above object, the present invention provides the following technical solutions: comprising the steps of:
步骤一:通过高速工业相机采集不同种类的芯片表面缺陷图像;Step 1: Collect different types of chip surface defect images through high-speed industrial cameras;
步骤二:通过图像标注工具对芯片表面图像缺陷进行位置和类型的标注,构建缺陷数据集,将缺陷数据集分为训练数据集D和测试数据集T;Step 2: Use the image annotation tool to mark the position and type of the chip surface image defect, construct a defect data set, and divide the defect data set into a training data set D and a test data set T;
步骤三:基于改进的FasterR-CNN框架,构建用于芯片缺陷检测的深度学习网络模型:Step 3: Based on the improved FasterR-CNN framework, build a deep learning network model for chip defect detection:
1)选取特征提取网络,经典的特征提取网络有VggNet、ResNet、GoogleNet等,其中ResNet网络的优点在于其结构的残差单元,解决了梯度消失问题,使网络的准确率上升,故选取ResNet-50用作特征提取网络。ResNet-50网络中包含49个卷积层、一个全连接层,ResNet-50网络结构可以分成七个部分,第一部分主要负责对输入数据卷积、正则化、激活函数、最大化池的计算。其中第一部分不包含残差块,第二、三、四、五部分结构都包含了残差块。在ResNet-50网络结构中,残差块都有三层卷积,那该网络中共有49个卷积层,加上最后的全连接层总共是50层;1) Select the feature extraction network. Classical feature extraction networks include VggNet, ResNet, GoogleNet, etc. Among them, the advantage of the ResNet network lies in the residual unit of its structure, which solves the problem of gradient disappearance and increases the accuracy of the network. Therefore, ResNet- 50 is used as a feature extraction network. The ResNet-50 network contains 49 convolutional layers and a fully connected layer. The ResNet-50 network structure can be divided into seven parts. The first part is mainly responsible for the calculation of input data convolution, regularization, activation function, and maximization pool. The first part does not contain a residual block, and the second, third, fourth, and fifth parts of the structure all contain a residual block. In the ResNet-50 network structure, the residual block has three layers of convolution, so there are a total of 49 convolution layers in the network, plus the last fully connected layer is a total of 50 layers;
2)可变形卷积重构特征提取网络,可变形卷积是指卷积核在每一个元素上额外增加了一个参数方向参数,这样卷积核就能在训练过程中扩展到很大的范围。由于芯片表面缺陷的形状不同,没有固定的几何结构。因此,针对未知变化能力差、泛化能力弱的ResNet-50,引入可变形卷积的思想,提高神经网络对不规则目标的识别能力。其中可变形卷积的实现过程为:2) Deformable convolution reconstruction feature extraction network, deformable convolution means that the convolution kernel adds an additional parameter direction parameter to each element, so that the convolution kernel can be expanded to a large range during the training process . Due to the different shapes of defects on the chip surface, there is no fixed geometry. Therefore, for ResNet-50, which has poor ability of unknown change and weak generalization ability, the idea of deformable convolution is introduced to improve the recognition ability of neural network for irregular targets. The implementation process of deformable convolution is as follows:
子步骤1:对采集的芯片图像调整输入尺寸以及进行预处理操作。Sub-step 1: adjust the input size and perform preprocessing operations on the collected chip image.
子步骤2:根据输入的图像,利用传统的卷积核提取特征图。Sub-step 2: According to the input image, use the traditional convolution kernel to extract the feature map.
子步骤3:得到可变形卷积的变形的偏移量:把得到的特征图作为输入,对特征图再添加一个卷积层。Sub-step 3: Get the deformed offset of the deformable convolution: take the obtained feature map as input, and add another convolution layer to the feature map.
子步骤4:偏移层是2N(其中2是指x和y两个方向,N是指卷积核大小),由于需要在二维平面上做平移,因此只需改变x值和y值,其中x值和y值分别代表在图像数据像素在x和y方向上的偏移量。Sub-step 4: The offset layer is 2N (where 2 refers to the two directions of x and y, and N refers to the size of the convolution kernel). Since translation is required on the two-dimensional plane, only the x and y values need to be changed. Wherein the x value and the y value represent the offset of the pixel in the image data in the x and y directions, respectively.
其中在训练过程中,用于生成输出特征的卷积核和用于生成偏移量的卷积核是同步学习的。During the training process, the convolution kernel used to generate output features and the convolution kernel used to generate offsets are learned synchronously.
3)多尺度特征融合,由于芯片缺陷检测本身为小目标检测的特点,因此引入特征金字塔FPN来进行特征融合。特征融合即将深层语义信息融合成浅层特征图,利用深层特征丰富语义信息,同时利用浅层特征的特征。将特征金字塔FPN引入到更快的R-CNN,将所有层的特征集成到特征金字塔中,同时将ResNet-50网络中第二、三、四、五层的残差块的特征激活输出作为输入。通过1x1卷积将C2-C5的通道数减少到256,得到M2-M5(以C2和M2为例,C2代表Conv2所对应的一系列残差结构的特征矩阵,M2代表C2通过1×1卷积后的残差结构的特征矩阵)。通过上采样添加相同大小的浅层特征图和深度特征图,得到FPN的输出P2-P5,然后采用3x3卷积。特征图P6是通过对FPN的输出P5上两个最大的池进行缩减采样获得的,输出多尺度融合特征组合。利用RPN中的融合特征生成目标候选帧,通过分类获得检测结果;3) Multi-scale feature fusion. Since the chip defect detection itself is characterized by small target detection, the feature pyramid FPN is introduced for feature fusion. Feature fusion is to fuse deep semantic information into shallow feature maps, use deep features to enrich semantic information, and use shallow features at the same time. Introduce the feature pyramid FPN into the faster R-CNN, integrate the features of all layers into the feature pyramid, and at the same time use the feature activation output of the second, third, fourth, and fifth layers of the residual block in the ResNet-50 network as input . The number of channels of C2-C5 is reduced to 256 by 1x1 convolution, and M2-M5 is obtained (take C2 and M2 as examples, C2 represents the feature matrix of a series of residual structures corresponding to Conv2, and M2 represents C2 through 1×1 volume feature matrix of the residual structure after product). The shallow feature map and deep feature map of the same size are added by upsampling to obtain the output P2-P5 of the FPN, and then 3x3 convolution is used. The feature map P6 is obtained by downsampling the two largest pools on the output P5 of the FPN, outputting a multi-scale fusion feature combination. Use the fusion features in RPN to generate target candidate frames, and obtain detection results through classification;
4)感兴趣区域对齐(ROI对齐)。ROIAlign与ROI池的不同之处不仅在于简单地量化然后池化,还在于使用区域特征聚合方法将其转换为连续操作。感兴趣区域对齐的具体实现步骤如下:4) Region of interest alignment (ROI alignment). ROIAlign differs from ROI pooling not only by simply quantizing and then pooling, but also by using a region feature aggregation method to convert it into a continuous operation. The specific implementation steps of the alignment of the region of interest are as follows:
子步骤1:循环访问所有候选区域,并保持映射的候选区域浮点坐标未量化。Sub-step 1: Loop through all candidate regions and keep the floating-point coordinates of the mapped candidate regions unquantized.
子步骤2:候选区域分为z×z单元格,并且每个单元格也不量化。Sub-step 2: The candidate area is divided into z×z cells, and each cell is not quantized.
子步骤3:确定每个单元中采样点的浮点坐标,使用双线性插值方法计算采样点的浮点坐标,然后可以获得固定维度的ROI输出。Sub-step 3: Determine the floating-point coordinates of the sampling points in each unit, calculate the floating-point coordinates of the sampling points using a bilinear interpolation method, and then obtain a fixed-dimensional ROI output.
5)软非极大值抑制(SoftNMS)。软非最大值抑制是对所有的候选框去做重新的估分,对比较差的框保留但会抑制分数,其最终实现结果是SoftNMS对框的分数做了一个修改,保留了较高的召回率。SoftNMS可用以下公式来表示:5) Soft non-maximum suppression (SoftNMS). Soft non-maximum suppression is to re-evaluate all candidate frames, retain the poorer frames but suppress the scores, and the final realization result is that SoftNMS has made a modification to the frame scores, retaining a higher recall Rate. SoftNMS can be expressed by the following formula:
作为本发明的优选技术方案,所述SoftNMS公式中Si是检测框i的分数,M是最高分数框,所述bi是检测帧的集合,所述Nt是设置的阈值,所述NMS将IOU大于阈值的检测帧的评分设置为0,而软NMS则对IOU大于阈值的检测帧的评分进行衰减,可以缓解目标漏检和误检的问题,所述软NMS的IOU阈值为0.5,最低分数为0.05。As a preferred technical solution of the present invention, Si is the score of the detection frame i in the SoftNMS formula, M is the highest score frame, the bi is the collection of detection frames, and the Nt is the threshold set, and the NMS will set the IOU greater than The score of the detection frame of the threshold is set to 0, while the soft NMS attenuates the score of the detection frame whose IOU is greater than the threshold, which can alleviate the problem of target missed detection and false detection. The IOU threshold of the soft NMS is 0.5, and the minimum score is 0.05.
与现有技术相比,本发明提供了一种基于深度学习的芯片缺陷检测方法,具备以下有益效果:Compared with the prior art, the present invention provides a chip defect detection method based on deep learning, which has the following beneficial effects:
本发明的目的在于针对芯片缺陷检测的特点和需求,以FasterR-CNN为基本框架,由特征提取网络、区域推荐网络、检测网络组成。故而提出一种基于深度学习的芯片缺陷检测算法,该算法通过采用可变形卷积重构特征提取网络、多尺度特征融合、感兴趣区域对齐(ROI对齐)、软非最大值抑制等措施对FasterR-CNN算法进行改进,解决了芯片缺陷体积小、形状复杂等问题,提高了缺陷检测的精度。The purpose of the present invention is to aim at the characteristics and requirements of chip defect detection, take Faster R-CNN as the basic framework, and consist of a feature extraction network, a region recommendation network, and a detection network. Therefore, a chip defect detection algorithm based on deep learning is proposed. The algorithm adopts measures such as deformable convolution reconstruction feature extraction network, multi-scale feature fusion, alignment of regions of interest (ROI alignment), soft non-maximum suppression, etc. -The CNN algorithm is improved to solve the problems of small size and complex shape of chip defects, and improve the accuracy of defect detection.
附图说明Description of drawings
图1为本发明ResNet50网络结构图;Fig. 1 is the network structural diagram of ResNet50 of the present invention;
图2为本发明可变形卷积示意图;Fig. 2 is a schematic diagram of deformable convolution in the present invention;
图3为本发明多尺度特征融合网络结构图。Fig. 3 is a structure diagram of the multi-scale feature fusion network of the present invention.
具体实施方式Detailed ways
下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The following will clearly and completely describe the technical solutions in the embodiments of the present invention with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some, not all, embodiments of the present invention. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.
请参阅图1-3,本实施方案中:包括如下步骤:Please refer to Fig. 1-3, in this embodiment: comprise the following steps:
步骤一:通过高速工业相机采集不同种类的芯片表面缺陷图像;Step 1: Collect different types of chip surface defect images through high-speed industrial cameras;
步骤二:通过图像标注工具对芯片表面图像缺陷进行位置和类型的标注,构建缺陷数据集,将缺陷数据集分为训练数据集D和测试数据集T;Step 2: Use the image annotation tool to mark the position and type of the chip surface image defect, construct a defect data set, and divide the defect data set into a training data set D and a test data set T;
步骤三:基于改进的FasterR-CNN框架,构建用于芯片缺陷检测的深度学习网络模型:Step 3: Based on the improved FasterR-CNN framework, build a deep learning network model for chip defect detection:
1)选取特征提取网络,经典的特征提取网络有VggNet、ResNet、GoogleNet等,其中ResNet网络的优点在于其结构的残差单元,解决了梯度消失问题,使网络的准确率上升,故选取ResNet-50用作特征提取网络。ResNet-50网络中包含49个卷积层、一个全连接层,ResNet-50网络结构可以分成七个部分,第一部分主要负责对输入数据卷积、正则化、激活函数、最大化池的计算。其中第一部分不包含残差块,第二、三、四、五部分结构都包含了残差块。在ResNet-50网络结构中,残差块都有三层卷积,那该网络中共有49个卷积层,加上最后的全连接层总共是50层;1) Select the feature extraction network. Classical feature extraction networks include VggNet, ResNet, GoogleNet, etc. Among them, the advantage of the ResNet network lies in the residual unit of its structure, which solves the problem of gradient disappearance and increases the accuracy of the network. Therefore, ResNet- 50 is used as a feature extraction network. The ResNet-50 network contains 49 convolutional layers and a fully connected layer. The ResNet-50 network structure can be divided into seven parts. The first part is mainly responsible for the calculation of input data convolution, regularization, activation function, and maximization pool. The first part does not contain a residual block, and the second, third, fourth, and fifth parts of the structure all contain a residual block. In the ResNet-50 network structure, the residual block has three layers of convolution, so there are a total of 49 convolution layers in the network, plus the last fully connected layer is a total of 50 layers;
2)可变形卷积重构特征提取网络,可变形卷积是指卷积核在每一个元素上额外增加了一个参数方向参数,这样卷积核就能在训练过程中扩展到很大的范围。由于芯片表面缺陷的形状不同,没有固定的几何结构。因此,针对未知变化能力差、泛化能力弱的ResNet-50,引入可变形卷积的思想,提高神经网络对不规则目标的识别能力。其中可变形卷积的实现过程为:2) Deformable convolution reconstruction feature extraction network, deformable convolution means that the convolution kernel adds an additional parameter direction parameter to each element, so that the convolution kernel can be expanded to a large range during the training process . Due to the different shapes of defects on the chip surface, there is no fixed geometry. Therefore, for ResNet-50, which has poor ability of unknown change and weak generalization ability, the idea of deformable convolution is introduced to improve the recognition ability of neural network for irregular targets. The implementation process of deformable convolution is as follows:
子步骤1:对采集的芯片图像调整输入尺寸以及进行预处理操作。Sub-step 1: adjust the input size and perform preprocessing operations on the collected chip image.
子步骤2:根据输入的图像,利用传统的卷积核提取特征图。Sub-step 2: According to the input image, use the traditional convolution kernel to extract the feature map.
子步骤3:得到可变形卷积的变形的偏移量:把得到的特征图作为输入,对特征图再添加一个卷积层。Sub-step 3: Get the deformed offset of the deformable convolution: take the obtained feature map as input, and add another convolution layer to the feature map.
子步骤4:偏移层是2N(其中2是指x和y两个方向,N是指卷积核大小),由于需要在二维平面上做平移,因此只需改变x值和y值,其中x值和y值分别代表在图像数据像素在x和y方向上的偏移量。Sub-step 4: The offset layer is 2N (where 2 refers to the two directions of x and y, and N refers to the size of the convolution kernel). Since translation is required on the two-dimensional plane, only the x and y values need to be changed. Wherein the x value and the y value represent the offset of the pixel in the image data in the x and y directions, respectively.
其中在训练过程中,用于生成输出特征的卷积核和用于生成偏移量的卷积核是同步学习的。During the training process, the convolution kernel used to generate output features and the convolution kernel used to generate offsets are learned synchronously.
3)多尺度特征融合,由于芯片缺陷检测本身为小目标检测的特点,因此引入特征金字塔FPN来进行特征融合。特征融合即将深层语义信息融合成浅层特征图,利用深层特征丰富语义信息,同时利用浅层特征的特征。将特征金字塔FPN引入到更快的R-CNN,将所有层的特征集成到特征金字塔中,同时将ResNet-50网络中第二、三、四、五层的残差块的特征激活输出作为输入。通过1x1卷积将C2-C5的通道数减少到256,得到M2-M5(以C2和M2为例,C2代表Conv2所对应的一系列残差结构的特征矩阵,M2代表C2通过1×1卷积后的残差结构的特征矩阵)。通过上采样添加相同大小的浅层特征图和深度特征图,得到FPN的输出P2-P5,然后采用3x3卷积。特征图P6是通过对FPN的输出P5上两个最大的池进行缩减采样获得的,输出多尺度融合特征组合。利用RPN中的融合特征生成目标候选帧,通过分类获得检测结果;3) Multi-scale feature fusion. Since the chip defect detection itself is characterized by small target detection, the feature pyramid FPN is introduced for feature fusion. Feature fusion is to fuse deep semantic information into shallow feature maps, use deep features to enrich semantic information, and use shallow features at the same time. Introduce the feature pyramid FPN into the faster R-CNN, integrate the features of all layers into the feature pyramid, and at the same time use the feature activation output of the second, third, fourth, and fifth layers of the residual block in the ResNet-50 network as input . The number of channels of C2-C5 is reduced to 256 by 1x1 convolution, and M2-M5 is obtained (take C2 and M2 as examples, C2 represents the feature matrix of a series of residual structures corresponding to Conv2, and M2 represents C2 through 1×1 volume feature matrix of the residual structure after product). The shallow feature map and deep feature map of the same size are added by upsampling to obtain the output P2-P5 of the FPN, and then 3x3 convolution is used. The feature map P6 is obtained by downsampling the two largest pools on the output P5 of the FPN, outputting a multi-scale fusion feature combination. Use the fusion features in RPN to generate target candidate frames, and obtain detection results through classification;
4)感兴趣区域对齐(ROI对齐)。ROIAlign与ROI池的不同之处不仅在于简单地量化然后池化,还在于使用区域特征聚合方法将其转换为连续操作。感兴趣区域对齐的具体实现步骤如下:4) Region of interest alignment (ROI alignment). ROIAlign differs from ROI pooling not only by simply quantizing and then pooling, but also by using a region feature aggregation method to convert it into a continuous operation. The specific implementation steps of the alignment of the region of interest are as follows:
子步骤1:循环访问所有候选区域,并保持映射的候选区域浮点坐标未量化。Sub-step 1: Loop through all candidate regions and keep the floating-point coordinates of the mapped candidate regions unquantized.
子步骤2:候选区域分为z×z单元格,并且每个单元格也不量化。Sub-step 2: The candidate area is divided into z×z cells, and each cell is not quantized.
子步骤3:确定每个单元中采样点的浮点坐标,使用双线性插值方法计算采样点的浮点坐标,然后可以获得固定维度的ROI输出。Sub-step 3: Determine the floating-point coordinates of the sampling points in each unit, calculate the floating-point coordinates of the sampling points using a bilinear interpolation method, and then obtain a fixed-dimensional ROI output.
5)软非极大值抑制(SoftNMS)。软非最大值抑制是对所有的候选框去做重新的估分,对比较差的框保留但会抑制分数,其最终实现结果是SoftNMS对框的分数做了一个修改,保留了较高的召回率。SoftNMS可用以下公式来表示:5) Soft non-maximum suppression (SoftNMS). Soft non-maximum suppression is to re-evaluate all candidate frames, retain the poorer frames but suppress the scores, and the final realization result is that SoftNMS has made a modification to the frame scores, retaining a higher recall Rate. SoftNMS can be expressed by the following formula:
需要注意的是:SoftNMS公式中Si是检测框i的分数,M是最高分数框,bi是检测帧的集合,Nt是设置的阈值,NMS将IOU大于阈值的检测帧的评分设置为0,而软NMS则对IOU大于阈值的检测帧的评分进行衰减,可以缓解目标漏检和误检的问题,软NMS的IOU阈值为0.5,最低分数为0.05。It should be noted that in the SoftNMS formula, Si is the score of the detection frame i, M is the highest score frame, bi is the set of detection frames, Nt is the set threshold, NMS sets the score of the detection frame with IOU greater than the threshold to 0, and Soft NMS attenuates the scores of detection frames whose IOU is greater than the threshold, which can alleviate the problem of missed and false detection of targets. The IOU threshold of soft NMS is 0.5, and the minimum score is 0.05.
最后应说明的是:以上所述仅为本发明的优选实施例而已,并不用于限制本发明,尽管参照前述实施例对本发明进行了详细的说明,对于本领域的技术人员来说,其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换。凡在本发明的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本发明的保护范围之内。Finally, it should be noted that: the above is only a preferred embodiment of the present invention, and is not intended to limit the present invention. Although the present invention has been described in detail with reference to the foregoing embodiments, for those skilled in the art, it still The technical solutions recorded in the foregoing embodiments may be modified, or some technical features thereof may be equivalently replaced. Any modifications, equivalent replacements, improvements, etc. made within the spirit and principles of the present invention shall be included within the protection scope of the present invention.
Claims (2)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310240534.0A CN116309429A (en) | 2023-03-14 | 2023-03-14 | A Chip Defect Detection Method Based on Deep Learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310240534.0A CN116309429A (en) | 2023-03-14 | 2023-03-14 | A Chip Defect Detection Method Based on Deep Learning |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116309429A true CN116309429A (en) | 2023-06-23 |
Family
ID=86833799
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310240534.0A Pending CN116309429A (en) | 2023-03-14 | 2023-03-14 | A Chip Defect Detection Method Based on Deep Learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116309429A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117237683A (en) * | 2023-11-10 | 2023-12-15 | 闽都创新实验室 | Chip defect intelligent detection system based on improved neural network |
CN118261877A (en) * | 2024-03-29 | 2024-06-28 | 泰山科技学院 | Method for evaluating and maintaining service life of printed circuit electronic component |
CN118840337A (en) * | 2024-07-03 | 2024-10-25 | 江苏省特种设备安全监督检验研究院 | Crane track defect identification method based on convolutional neural network |
-
2023
- 2023-03-14 CN CN202310240534.0A patent/CN116309429A/en active Pending
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117237683A (en) * | 2023-11-10 | 2023-12-15 | 闽都创新实验室 | Chip defect intelligent detection system based on improved neural network |
CN117237683B (en) * | 2023-11-10 | 2024-03-26 | 闽都创新实验室 | Intelligent chip defect detection system based on improved neural network |
CN118261877A (en) * | 2024-03-29 | 2024-06-28 | 泰山科技学院 | Method for evaluating and maintaining service life of printed circuit electronic component |
CN118261877B (en) * | 2024-03-29 | 2024-10-11 | 泰山科技学院 | Method for evaluating and maintaining service life of printed circuit electronic component |
CN118840337A (en) * | 2024-07-03 | 2024-10-25 | 江苏省特种设备安全监督检验研究院 | Crane track defect identification method based on convolutional neural network |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111401384B (en) | Transformer equipment defect image matching method | |
CN110544251B (en) | Dam crack detection method based on multi-migration learning model fusion | |
CN116309429A (en) | A Chip Defect Detection Method Based on Deep Learning | |
CN110689599B (en) | 3D visual saliency prediction method based on non-local enhancement generation countermeasure network | |
CN111368825B (en) | Pointer positioning method based on semantic segmentation | |
CN113870335A (en) | Monocular depth estimation method based on multi-scale feature fusion | |
CN114881871A (en) | Attention-fused single image rain removing method | |
CN112288776B (en) | A Target Tracking Method Based on Multi-Time Step Pyramid Codec | |
CN113850324B (en) | Multispectral target detection method based on Yolov4 | |
CN112884788B (en) | An optic cup and optic disc segmentation method and imaging method based on rich context network | |
CN113361493B (en) | A Facial Expression Recognition Method Robust to Different Image Resolutions | |
CN114612456B (en) | A method for automatic semantic segmentation and recognition of steel billets based on deep learning | |
CN113449691A (en) | Human shape recognition system and method based on non-local attention mechanism | |
CN116310325A (en) | A Semantic Segmentation Method of Large Format Remote Sensing Imagery from PATCH to REGION Architecture | |
Peng et al. | Litchi detection in the field using an improved YOLOv3 model | |
CN114445620A (en) | An Object Segmentation Method to Improve Mask R-CNN | |
CN118230131A (en) | Image recognition and target detection method | |
CN116993737A (en) | Lightweight fracture segmentation method based on convolutional neural network | |
CN116630387A (en) | Monocular Image Depth Estimation Method Based on Attention Mechanism | |
CN115424017B (en) | A method, device and storage medium for segmenting the interior and exterior contours of a building | |
CN115936992A (en) | Garbage image super-resolution method and system of lightweight transform | |
CN113112450B (en) | A method for small target detection in remote sensing images guided by image pyramid | |
CN114926734A (en) | Solid waste detection device and method based on feature aggregation and attention fusion | |
CN116740790B (en) | Face detection method and device based on transducer | |
Zhu et al. | Multi-class segmentation of navel orange surface defects based on improved deeplabv3+ |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |