WO2021249233A1 - Image processing method, target recognition model training method, and target recognition method - Google Patents

Image processing method, target recognition model training method, and target recognition method Download PDF

Info

Publication number
WO2021249233A1
WO2021249233A1 PCT/CN2021/097562 CN2021097562W WO2021249233A1 WO 2021249233 A1 WO2021249233 A1 WO 2021249233A1 CN 2021097562 W CN2021097562 W CN 2021097562W WO 2021249233 A1 WO2021249233 A1 WO 2021249233A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
convolutional layer
convolution
feature
processed
Prior art date
Application number
PCT/CN2021/097562
Other languages
French (fr)
Chinese (zh)
Inventor
刘道学
耿天宝
杨铭
于健
胡伟
肖丽娜
张尧尘
Original Assignee
中铁四局集团有限公司
安徽数智建造研究院有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 中铁四局集团有限公司, 安徽数智建造研究院有限公司 filed Critical 中铁四局集团有限公司
Publication of WO2021249233A1 publication Critical patent/WO2021249233A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components

Definitions

  • the invention belongs to the technical field of target recognition, and more specifically, relates to an image processing method, a target recognition model training method and a target recognition method.
  • the original image collected in this recognition method is the photo of the back of the hand taken by optical imaging, but the target imaging that needs to be detected in the tunnel detection is through the radar detection band signal Then the digital information obtained by analog-to-digital conversion is presented in the form of images. Therefore, the above recognition methods cannot be simply applied to the field of tunnel defect recognition. Due to the wide variety of underground objects, the wide distribution and the complex and changeable geological conditions, the radar spectrogram There are a lot of fringes and noise. These interference information greatly affect the recognition of geological defects.
  • the present invention provides an image processing method, a target recognition model training method and a target recognition method , Used to identify the geological defects in the radar spectrogram.
  • the present invention adopts the following technical solutions.
  • An image processing method includes the following steps:
  • the convolution fusion operation is performed on the multi-feature mosaic image Fc to obtain the multi-feature fusion image Ff.
  • the gradient transformation algorithm is used to extract the edge features of the image to be processed, and the method of the gradient transformation algorithm to extract the edge features is:
  • f(x,y) be the gray value of the (x,y) point on the image to be processed
  • G(x) is the image gray value of the edge detection in the x direction of the image to be processed
  • G(y) is the image to be processed
  • a is the convolution factor of G(x)
  • b is the convolution factor of G(y)
  • the image gray value of each pixel on the gradient prior image Fg is :
  • the edge feature on the gradient priori map Fg is presented by the gray value of each pixel.
  • the texture feature map Fv is obtained by adding the convolutional layer v1 and the convolutional layer v1'.
  • S12 Perform two consecutive m-1 convolution operations on the convolutional layer s1 to obtain the convolutional layer s2, convolutional layer s3,..., convolutional layer sm, convolutional layer s2', and convolutional layer s3, respectively ',..., convolutional layer sm';
  • step S13 The convolutional layer sm and the convolutional layer sm' in step S12 are spliced through contact to obtain the convolutional layer sm+1;
  • the first convolution operation is performed on the image to be processed, and the convolution kernel is a 3*3 matrix when the convolution layer v1 is obtained, and the convolution kernel of the convolution operation in the step S2 is 5* 5 Matrix.
  • the convolution kernel is a 1*1 matrix for the convolution operation of the convolution layer s2m.
  • the radar spectrogram has a lot of redundant information such as noise, this interference information greatly affects the recognition of geological defects. Therefore, in order to improve the recognition accuracy, the original image needs to be feature-enhanced, taking into account the image of the radar spectrogram Features and the shape and structure of geological defects.
  • the edge feature, texture feature and structural feature in the image are separately extracted and merged into a single image for recognition operation, which greatly improves the recognition efficiency of geological defects.
  • the present invention provides a method for training a target recognition model, which includes the following steps:
  • S24 Select a priori boxes with different aspect ratios on Ff and Ff', and use the position information (Xi, Yi, Wi, Hi) of each a priori box in the figure and the corresponding abnormal state type Ci as a label mark;
  • S25 Input a priori box with a label as a training sample into a pre-recognition model for training to obtain a target recognition model.
  • the pre-identification model is a machine self-learning model, including one or more combinations of neural network, convolutional neural network, deep neural network, and feedback neural network.
  • the multi-feature fusion map Ff is down-sampled and doubled to obtain Ff'.
  • the target recognition model needs to be imported to recognize the geological defects.
  • the target recognition model requires a lot of training in the early stage.
  • the training samples are the images that have been labeled with the defect location and type.
  • the size is different, so here we take two different sizes of the multi-feature fusion map Ff and the down-sampled feature map Ff' for recognition training.
  • the target feature is classified and recognized by the classifier, and then non-maximum The value suppression calculation is large to realize the determination of the most approximate a priori box.
  • the present invention also provides a target recognition method, which includes the following steps:
  • a target recognition model is used to recognize the target object in the image to be recognized; wherein, the target recognition model is pre-trained using the above-mentioned target recognition model training method.
  • the target recognition method proposed here is mainly based on the target recognition model trained in this technical solution, which can perform target recognition on the original image similar to the radar spectrogram, compared with the Faster-R-CNN, SSD and Yolov3 in the prior art. And other algorithm models have faster recognition speed.
  • the present invention obtains the gradient prior map Fg, the texture feature map Fv and the structural feature map Fs by feature extraction of the image to be processed, and then stitches the above three images and convolution fusion to obtain the multi-dimensional feature fusion map Ff, which can reduce
  • the influence of noise in the radar spectrogram can also display the edges of the fringe more clearly, and it can also include texture features that are greatly affected by the size of the image, such as fringes in the radar spectrogram, and structural features that are less affected by size changes. , Help to improve the accuracy of target recognition;
  • the present invention uses a 5*5 convolution kernel to perform convolution operation on conv_v1 in the texture feature extraction to obtain conv_v1', and add conv_v1' and conv_vn to obtain the texture feature map Fv, where conv_v1' retains more
  • conv_v1' retains more
  • conv_v1' retains more
  • the original features of, and the receptive field is larger than conv_v1, which is convenient to better retain the texture features in the original image and improve the accuracy of target recognition
  • the original image is first convolved to obtain conv_s1, Then conv_s1 is processed with an hourglass structure that shrinks and then expands, which better preserves the structural features in the original image file, and further improves the accuracy of target recognition.
  • Figure 1 is a schematic flow diagram of an image processing method
  • Figure 2 is a schematic diagram of the process of texture feature extraction in an image processing method
  • Fig. 3 is a schematic diagram of the flow of structure feature extraction in an image processing method
  • Figure 4 is a schematic flow diagram of a target recognition method
  • Figure 5 is a schematic diagram of a bad cavity output in the identification of geological defects in tunnels by a target recognition method
  • Fig. 6 is a schematic diagram of poor voids output by a target recognition method in the recognition of geological defects in tunnels.
  • the image is feature-enhanced.
  • the edge feature, texture feature, and structural feature are enhanced by combining the image features of the radar spectrogram.
  • the extraction enhancement can be advanced in parallel without any order. The following respectively introduces the selection of edge feature, texture feature and structure feature extraction.
  • the gradient transformation algorithm When extracting the edge features of the image to be processed, the gradient transformation algorithm is used, and the obtained gradient a priori map is named Fg, where the gradient transformation algorithm is the sobel operator.
  • Fg the gradient transformation algorithm is the sobel operator.
  • G(x) is the image gray value of the edge detection in the x direction of the image to be processed
  • G(y) is the image gray value of the edge detection in the y direction of the image to be processed
  • a G(x
  • b is the convolution factor of G(y)
  • the image gray value of each point on the gradient prior image Fg is:
  • the values of a and b can be selected as a 3*3 matrix, which is specifically:
  • the method used when extracting texture features in the image to be processed is:
  • a scale-invariant multi-layer CNN module based on transitional convolution is designed, because the details of the texture and the The resolution has a strong matching relationship.
  • the resolution is different, the texture feature must be different. Therefore, the resolution is unchanged when the texture feature is extracted in this solution.
  • the image to be processed is the tunnel scan image I.
  • n 5.
  • the original image output by the geological radar is a three-channel black-and-white image
  • the original image needs to be converted into a single-channel grayscale image.
  • the step size of the convolution is 1, and then while ensuring the resolution, the texture feature extraction is performed.
  • the specific information of each convolution is as follows:
  • Convolutional layer Step size Convolution kernel Number of channels conv_v1 1 3*3 48 conv_v2 1 3*3 48 conv_v3 1 3*3 48 conv_v4 1 3*3 48 conv_v5 1 3*3 48 conv_v1' 1 5*5 48
  • the conv_v1' convolution factor parameter used when obtaining conv_v1' is a convolution kernel of 5*5, Therefore, conv_v1' has a larger receptive field than conv_v1, and can learn original features different from conv_v1 in it, and conv_v1' is convolved from conv_v1. Compared with conv_v1' through other convolution layers, conv_v1' More original features are retained. When adding them to conv_vn, the texture features in the original image can be better retained, thereby improving the accuracy of the recognition of geological defects.
  • the texture feature map obtained by the above method is named Fv.
  • the method used when extracting the structural features in the image to be processed is:
  • S12 Perform two consecutive m-1 convolution operations on conv_s1 to obtain conv_s2, conv_s3,..., conv_sm and conv_s2', conv_s3',..., conv_sm' respectively;
  • Convolutional layer Step size Convolution kernel Number of channels conv_s1 1 3*3 32 conv_s2 2 3*3 64 conv_s3 2 3*3 128 conv_s2' 2 3*3 256 conv_s3' 2 3*3 128 conv_s4 2 3*3 64 conv_s5 2 3*3 32 conv_s6 1 3*3 32
  • the structural feature has a certain scale invariance, that is, when the scale changes within a certain range, its structural feature remains basically unchanged. Since structural features of different sizes may be easily identified on different convolutional layers, if only conv_sm or conv_sm' is used for identification, information will be lost, and structural features of different sizes cannot be well recognized.
  • the Fs obtained by the structural feature extraction method in the scheme not only retains the information of all convolutional layers, but also the size of the original input image, which is convenient for subsequent better identification of structural features.
  • a 1*1 matrix convolution kernel is used to convert the multi-channel convolution layer into a single channel.
  • the structural feature map obtained by the above method is named Fs.
  • the gradient prior map Fg after the edge feature enhancement as well as the texture feature map Fv and the structural feature map Fs.
  • the above three feature maps are extracted and enhanced for different features in the original image, in order to be more accurate
  • the above feature maps need to be spliced and fused.
  • the following method is used to sequentially perform the same number of down-sampling convolution operations and up-sampling convolution operations on Fc to obtain Ff, in order to make full use of the three feature information extracted previously ,Ff here is designed to be the same three-channel as Fc.
  • the convolution kernel is a 3*3 matrix
  • the convolution step size is 2
  • the information of each convolution is:
  • Convolutional layer Step size Convolution kernel Number of channels conv_1 2 3*3 64 conv_2 2 3*3 128 conv_3 2 3*3 128 conv_4 2 3*3 64
  • the target recognition model training method in this technical solution includes the following steps:
  • S24 Select a priori boxes with different aspect ratios on Ff and Ff', and use the position information (Xi, Yi, Wi, Hi) of each a priori box in the figure and the corresponding abnormal state type Ci as a label Mark; where i is the sequence number of the picture, X and Y are the position coordinates of the a priori frame, and W and H are the width and height of the a priori frame;
  • S25 Input a priori box with a label as a training sample into a pre-recognition model for training to obtain a target recognition model.
  • the pre-identification model is a self-learning machine model, including one or more combinations of neural networks, convolutional neural networks, deep neural networks, and feedback neural networks. Due to the different sizes of geological defects, two different sizes of image files of multi-feature fusion map Ff and down-sampling feature map Ff' are used for recognition training. During training, the six sets of a priori boxes are respectively detected and regression under the corresponding scale feature map to obtain the target detection frame coordinates, the corresponding state digital type and the confidence level, and finally the results of the detection on different scales are passed through the NMS The non-maximum suppression screening obtains the final target detection result and outputs it.
  • the target recognition method includes the following steps:
  • a target recognition model is used to recognize the target object in the image to be recognized; wherein, the target recognition model is pre-trained using the above method.
  • the tunnel scan map after image processing has greatly reduced noise, and edge features, texture features, and structural features are presented more clearly, so the speed and accuracy of recognition can be improved.

Abstract

An image processing method, which belongs to the technical field of target recognition. The method comprises the following steps: extracting an edge feature of an image to be processed, so as to obtain a gradient a priori map Fg; extracting a texture feature of said image, so as to obtain a texture feature map Fv; extracting a structural feature of said image, so as to obtain a structural feature map Fs; performing contact splicing on the gradient a priori map Fg, the texture feature map Fv and the structural feature map Fs, so as to obtain a multi-feature splicing map Fc; and performing a convolution fusion operation on the multi-feature splicing map Fc, so as to obtain a multi-feature fusion map Ff. By means of the method, edge feature enhancement is first performed on an image to be processed, and a texture feature and a structural feature are then fused, such that a target feature in the image can be recognized more accurately.

Description

一种图像处理方法、目标识别模型训练方法和目标识别方法An image processing method, target recognition model training method and target recognition method 技术领域Technical field
本发明属于目标识别技术领域,更具体地说,涉及一种图像处理方法、目标识别模型训练方法和目标识别方法。The invention belongs to the technical field of target recognition, and more specifically, relates to an image processing method, a target recognition model training method and a target recognition method.
背景技术Background technique
随着信息技术的发展,基于图像的目标检测技术被越来越广泛地应用,常见的有人脸识别和车载障碍识别。由于应用场景的不同,同一个目标识别模型很难通用在所有的需要进行目标识别的领域。例如在对在隧道工程施工中,需要对于隧道横洞内部可能出现的脱空和空洞进行快速和准确的检测定位。目前在对隧道地质结构异常检测的方法中主要是以人为主观的从地质雷达扫描波谱图中手动进行框选和标记计数的方式,这样的标记计数方式对于人为判断的能力要求很高,难点在于雷达波谱图中含有大量的干扰信息,在没有经过培训的情况下人是很难快速分辨出脱空、空洞等隧道内部的结构异常状态,从而传统的方法具有很耗费时间和人力的缺点,现有技术中还没有应用于隧道地质缺陷识别的目标识别模型和方法。With the development of information technology, image-based target detection technology is more and more widely used, such as facial recognition and vehicle obstacle recognition. Due to different application scenarios, the same target recognition model is difficult to be used in all areas that require target recognition. For example, in the construction of tunnel engineering, it is necessary to quickly and accurately detect and locate the voids and voids that may appear inside the tunnel's transverse hole. At present, the methods for detecting abnormal geological structures in tunnels are mainly based on human subjective manual frame selection and mark counting from the geological radar scanning spectrogram. Such mark counting method requires high human judgment ability, and the difficulty lies in The radar spectrogram contains a large amount of interference information. Without training, it is difficult for people to quickly distinguish the abnormal state of the structure inside the tunnel such as voids and cavities. Therefore, the traditional method has the disadvantage of time-consuming and labor-consuming. There are no target recognition models and methods that have been applied to tunnel geological defect recognition in the technology.
经检索,中国发明专利,公开号:102254165A,公开日:2011年11月23日,公开了一种基于结构编码特征和纹理编码特征融合的手背静脉识别方法,包括步骤一、图像采集和图像预处理;步骤二、提取结构编码特征;步骤三、提取手背静脉纹理编码特征;步骤四、结构编码特征与纹理编码特征进行融合;步骤五、通过分类器识别,得到结果。其中先对图像进行特征增强处理,然后通过分类器进行目标识别,该识别方法中采集到的原始图像是光学成像拍摄得到的手背照片,但是隧道检测中需要检测的目标成像是通过雷达探测波段信号然后利用模数转换得到的数字信息以图像的形式呈现,因此不能简单的将以上识别方法应用到隧道缺陷识别领域,由于地下物体种类多、分布广泛且地质情况复杂多变,因此雷达波谱图中具有大量的条纹和噪点,这些干扰信息极大的影响了对地质缺陷的识别。After searching, Chinese invention patent, publication number: 102254165A, publication date: November 23, 2011, disclosed a hand vein recognition method based on the fusion of structure coding features and texture coding features, including step one, image acquisition and image pre-processing. Processing; step two, extract structure coding features; step three, extract hand vein texture coding features; step four, structure coding features and texture coding features fusion; step five, identify through the classifier, and get the result. Among them, the image is first processed for feature enhancement, and then the target is recognized by the classifier. The original image collected in this recognition method is the photo of the back of the hand taken by optical imaging, but the target imaging that needs to be detected in the tunnel detection is through the radar detection band signal Then the digital information obtained by analog-to-digital conversion is presented in the form of images. Therefore, the above recognition methods cannot be simply applied to the field of tunnel defect recognition. Due to the wide variety of underground objects, the wide distribution and the complex and changeable geological conditions, the radar spectrogram There are a lot of fringes and noise. These interference information greatly affect the recognition of geological defects.
因此需要一种对隧道结构异常检测中得到的地质雷达波谱图像进行处理,并对图中隧道地质缺陷进行识别的方法。Therefore, there is a need for a method for processing the geological radar spectrum image obtained in the tunnel structure abnormality detection, and identifying the geological defects of the tunnel in the map.
发明内容Summary of the invention
1、要解决的问题1. The problem to be solved
针对现有隧道地质缺陷检测中的雷达波谱图具有大量的条纹和噪点,极大的影响了对地质缺陷的识别的问题,本发明提供一种图像处理方法、目标识别模型训练方法和目标识别方法,用以对雷达波谱图中的地质缺陷进行识别。Aiming at the problem that the radar spectrogram in the existing tunnel geological defect detection has a large number of stripes and noises, which greatly affects the recognition of geological defects, the present invention provides an image processing method, a target recognition model training method and a target recognition method , Used to identify the geological defects in the radar spectrogram.
2、技术方案2. Technical scheme
为解决上述问题,本发明采用如下的技术方案。To solve the above-mentioned problems, the present invention adopts the following technical solutions.
一种图像处理方法,包括以下步骤:An image processing method includes the following steps:
提取待处理图像的边缘特征,得到梯度先验图Fg;Extract the edge features of the image to be processed to obtain the gradient prior image Fg;
提取待处理图像的纹理特征,得到纹理特征图Fv;Extract the texture feature of the image to be processed to obtain the texture feature map Fv;
提取待处理图像的结构特征,得到结构特征图Fs;Extract the structural features of the image to be processed to obtain the structural feature map Fs;
将梯度先验图Fg、纹理特征图Fv和结构特征图Fs进行contact拼接,得到多特征拼接图Fc;Connect the gradient prior map Fg, the texture feature map Fv and the structural feature map Fs to contact stitching to obtain a multi-feature stitching image Fc;
对多特征拼接图Fc进行卷积融合操作,得到多特征融合图Ff。The convolution fusion operation is performed on the multi-feature mosaic image Fc to obtain the multi-feature fusion image Ff.
进一步的,采用梯度变换算法提取待处理图像的边缘特征,梯度变换算法提取边缘特征的方法为:Further, the gradient transformation algorithm is used to extract the edge features of the image to be processed, and the method of the gradient transformation algorithm to extract the edge features is:
设f(x,y)为待处理图像上(x,y)点的灰度值,G(x)为待处理图像x方向上边缘检测的图像灰度值,G(y)为待处理图像y方向上边缘检测的图像灰度值,a为G(x)的卷积因子,b为G(y)的卷积因子,则梯度先验图Fg上每个像素点的图像灰度值为:Let f(x,y) be the gray value of the (x,y) point on the image to be processed, G(x) is the image gray value of the edge detection in the x direction of the image to be processed, and G(y) is the image to be processed The image gray value of edge detection in the y direction, a is the convolution factor of G(x), b is the convolution factor of G(y), then the image gray value of each pixel on the gradient prior image Fg is :
Figure PCTCN2021097562-appb-000001
通过每个像素点的灰度值来呈现梯度先验图Fg上的边缘特征。
Figure PCTCN2021097562-appb-000001
The edge feature on the gradient priori map Fg is presented by the gray value of each pixel.
进一步的,对待处理图像进行纹理特征提取时采用以下步骤:Further, the following steps are used when extracting texture features of the image to be processed:
S1:对待处理图像进行连续n次卷积操作,依次得到卷积层v1、卷积层v2、……、卷积层vn;S1: Perform n consecutive convolution operations on the image to be processed, and obtain convolutional layer v1, convolutional layer v2, ..., convolutional layer vn in sequence;
S2:再对卷积层v1进行一次卷积操作,得到卷积层v1';S2: Perform another convolution operation on the convolutional layer v1 to obtain the convolutional layer v1';
S3:将卷积层v1与卷积层v1'通过相加运算得到纹理特征图Fv。S3: The texture feature map Fv is obtained by adding the convolutional layer v1 and the convolutional layer v1'.
进一步的,对待处理图像进行结构特征提取时采用以下步骤:Further, the following steps are used when extracting structural features of the image to be processed:
S11:对待处理图像进行1次卷积操作,得到卷积层s1;S11: Perform one convolution operation on the image to be processed to obtain the convolution layer s1;
S12:对卷积层s1进行两次连续m-1次的卷积操作,分别得到卷积层s2、卷积层s3、……、卷积层sm和卷积层s2'、卷积层s3'、……、卷积层sm';S12: Perform two consecutive m-1 convolution operations on the convolutional layer s1 to obtain the convolutional layer s2, convolutional layer s3,..., convolutional layer sm, convolutional layer s2', and convolutional layer s3, respectively ',..., convolutional layer sm';
S13:将步骤S12中的卷积层sm与卷积层sm'通过contact拼接,得到卷积层sm+1;S13: The convolutional layer sm and the convolutional layer sm' in step S12 are spliced through contact to obtain the convolutional layer sm+1;
S14:将步骤S13中的卷积层sm+1与步骤S12中的卷积层sm-1'、卷积层sm-2'、……、卷积层s2'通过contact连续拼接,最终得到卷积层s2m-1;S14: The convolutional layer sm+1 in step S13 and the convolutional layer sm-1', convolutional layer sm-2',..., convolutional layer s2' in step S12 are continuously spliced through contact, and finally the convolutional layer is obtained. Build-up s2m-1;
S15:将S11中的卷积层s1与步骤S14中的卷积层s2m-1通过contact拼接,得到卷积层s2m,并对卷积层s2m进行一次卷积操作,得到Fs。S15: Join the convolutional layer s1 in S11 and the convolutional layer s2m-1 in step S14 through contact to obtain the convolutional layer s2m, and perform a convolution operation on the convolutional layer s2m to obtain Fs.
进一步的,所述步骤S1中,对待处理图像进行第一次卷积操作,得到卷积层v1时卷积核 为3*3矩阵,所述步骤S2中卷积操作的卷积核为5*5矩阵。Further, in the step S1, the first convolution operation is performed on the image to be processed, and the convolution kernel is a 3*3 matrix when the convolution layer v1 is obtained, and the convolution kernel of the convolution operation in the step S2 is 5* 5 Matrix.
进一步的,所述步骤S15中对卷积层s2m的卷积操作采用卷积核为1*1矩阵。Further, in the step S15, the convolution kernel is a 1*1 matrix for the convolution operation of the convolution layer s2m.
由于雷达波谱图中具有大量的噪点等冗余信息,这些干扰信息极大的影响了对地质缺陷的识别,因此为了提高识别准确率,需要对原始图像进行特征增强,考虑到雷达波谱图的图像特点及地质缺陷的形状结构,这里对图像中的边缘特征、纹理特征和结构特征分别进行提取,并融合至一张图中进行识别操作,大大提高了对地质缺陷的识别效率。Because the radar spectrogram has a lot of redundant information such as noise, this interference information greatly affects the recognition of geological defects. Therefore, in order to improve the recognition accuracy, the original image needs to be feature-enhanced, taking into account the image of the radar spectrogram Features and the shape and structure of geological defects. Here, the edge feature, texture feature and structural feature in the image are separately extracted and merged into a single image for recognition operation, which greatly improves the recognition efficiency of geological defects.
本发明提供一种目标识别模型训练方法,包括以下步骤:The present invention provides a method for training a target recognition model, which includes the following steps:
S21:将地质雷达检测获得的若干图像进行裁剪得到隧道扫描图I;S21: Cut several images obtained by the ground penetrating radar to obtain a tunnel scan image I;
S22:将隧道扫描图I作为待处理图像经过以上图像处理方法处理得到多特征融合图Ff;S22: Use the tunnel scan image I as a to-be-processed image to obtain a multi-feature fusion image Ff through the above image processing method;
S23:将多特征融合图Ff进行下采样操作,得到对应的下采样特征图Ff';S23: Perform a down-sampling operation on the multi-feature fusion map Ff to obtain the corresponding down-sampled feature map Ff';
S24:在Ff和Ff'上选择不同长宽比的先验框,并将每个先验框在图中所处位置信息(Xi,Yi,Wi,Hi)和对应异常状态类型Ci作为标签进行标记;S24: Select a priori boxes with different aspect ratios on Ff and Ff', and use the position information (Xi, Yi, Wi, Hi) of each a priori box in the figure and the corresponding abnormal state type Ci as a label mark;
S25:将带有标签的先验框作为训练样本输入预识别模型进行训练,得到目标识别模型。S25: Input a priori box with a label as a training sample into a pre-recognition model for training to obtain a target recognition model.
进一步的,所述预识别模型为机器自学模型,包括神经网络、卷积神经网络、深度神经网络、反馈神经网络中的一种或多种组合。Further, the pre-identification model is a machine self-learning model, including one or more combinations of neural network, convolutional neural network, deep neural network, and feedback neural network.
进一步的,所述步骤S23中多特征融合图Ff进行下采样一倍的操作得到Ff'。Further, in the step S23, the multi-feature fusion map Ff is down-sampled and doubled to obtain Ff'.
在对图像进行特征增强处理后,需要导入目标识别模型进行对地质缺陷的识别,而目标识别模型前期需要经过大量的训练,训练样本是经过标注缺陷位置和类型等标签的图像,由于地质缺陷的大小不同,因此这里取多特征融合图Ff和下采样特征图Ff'两种不同大小的图档来进行识别训练,在预识别模型中通过分类器对目标特征进行分类识别,然后通过非极大值抑制算大来实现对最近似的先验框的确定。After the feature enhancement processing of the image, the target recognition model needs to be imported to recognize the geological defects. The target recognition model requires a lot of training in the early stage. The training samples are the images that have been labeled with the defect location and type. The size is different, so here we take two different sizes of the multi-feature fusion map Ff and the down-sampled feature map Ff' for recognition training. In the pre-recognition model, the target feature is classified and recognized by the classifier, and then non-maximum The value suppression calculation is large to realize the determination of the most approximate a priori box.
本发明还提出一种目标识别方法,包括以下步骤:The present invention also provides a target recognition method, which includes the following steps:
接收待处理图像;Receive the image to be processed;
利用目标识别模型识别所述待识别图像中的目标对象;其中,所述目标识别模型利用上述目标识别模型训练方法预先训练得到。A target recognition model is used to recognize the target object in the image to be recognized; wherein, the target recognition model is pre-trained using the above-mentioned target recognition model training method.
这里提出的目标识别方法主要基于本技术方案中训练而得到的目标识别模型,可以对类似于雷达波谱图的原始图像进行目标识别,相比于现有技术中Faster-R-CNN、SSD和Yolov3等算法模型具有更快的识别速度。The target recognition method proposed here is mainly based on the target recognition model trained in this technical solution, which can perform target recognition on the original image similar to the radar spectrogram, compared with the Faster-R-CNN, SSD and Yolov3 in the prior art. And other algorithm models have faster recognition speed.
3、有益效果3. Beneficial effects
相比于现有技术,本发明的有益效果为:Compared with the prior art, the beneficial effects of the present invention are:
(1)本发明通过对待处理图像进行特征提取得到梯度先验图Fg、纹理特征图Fv和结构 特征图Fs,然后将以上三种图进行拼接并卷积融合得到多维特征融合图Ff,可以降低雷达波谱图中噪点的影响,还可以将条纹的边缘进行更加清晰的展示,同时能够包含雷达波谱图中的条纹等受图像尺寸大小影响较大的纹理特征及受尺寸变化影响较小的结构特征,有助于提高目标识别的准确率;(1) The present invention obtains the gradient prior map Fg, the texture feature map Fv and the structural feature map Fs by feature extraction of the image to be processed, and then stitches the above three images and convolution fusion to obtain the multi-dimensional feature fusion map Ff, which can reduce The influence of noise in the radar spectrogram can also display the edges of the fringe more clearly, and it can also include texture features that are greatly affected by the size of the image, such as fringes in the radar spectrogram, and structural features that are less affected by size changes. , Help to improve the accuracy of target recognition;
(2)本发明在纹理特征提取中利用5*5的卷积核对conv_v1进行卷积操作得到conv_v1',并将conv_v1'与conv_vn进行相加运算得到纹理特征图Fv,这里conv_v1'保留了较多的原始特征,且感受野较conv_v1更大,便于更好的保留原始图像中的纹理特征,提高目标识别的准确率;而在结构特征提取过程中,先将原始图像进行卷积处理得到conv_s1,然后对conv_s1采取先收缩再扩展的沙漏结构进行处理,更好的保存了原始图档中的结构特征,进一步提高目标识别的准确率。(2) The present invention uses a 5*5 convolution kernel to perform convolution operation on conv_v1 in the texture feature extraction to obtain conv_v1', and add conv_v1' and conv_vn to obtain the texture feature map Fv, where conv_v1' retains more The original features of, and the receptive field is larger than conv_v1, which is convenient to better retain the texture features in the original image and improve the accuracy of target recognition; and in the process of structural feature extraction, the original image is first convolved to obtain conv_s1, Then conv_s1 is processed with an hourglass structure that shrinks and then expands, which better preserves the structural features in the original image file, and further improves the accuracy of target recognition.
附图说明Description of the drawings
图1为一种图像处理方法的流程示意图;Figure 1 is a schematic flow diagram of an image processing method;
图2为一种图像处理方法中纹理特征提取的流程示意图;Figure 2 is a schematic diagram of the process of texture feature extraction in an image processing method;
图3为一种图像处理方法中结构特征提取的流程示意图;Fig. 3 is a schematic diagram of the flow of structure feature extraction in an image processing method;
图4为一种目标识别方法的流程示意图;Figure 4 is a schematic flow diagram of a target recognition method;
图5为一种目标识别方法在隧道地质缺陷识别中输出的空洞不良示意图;Figure 5 is a schematic diagram of a bad cavity output in the identification of geological defects in tunnels by a target recognition method;
图6为一种目标识别方法在隧道地质缺陷识别中输出的脱空不良示意图。Fig. 6 is a schematic diagram of poor voids output by a target recognition method in the recognition of geological defects in tunnels.
具体实施方式detailed description
下面结合具体实施例对本发明进一步进行描述。The present invention will be further described below in conjunction with specific embodiments.
实施例1Example 1
如图1和图4所示,在进行地质缺陷识别前,先对图像进行特征增强处理,这里结合雷达波谱图的图像特点对其中的边缘特征、纹理特征和结构特征进行增强,这三种特征的提取增强可并列推进,无先后顺序,下面分别对选择对边缘特征、纹理特征和结构特征的提取进行介绍。As shown in Figure 1 and Figure 4, before performing geological defect recognition, the image is feature-enhanced. Here, the edge feature, texture feature, and structural feature are enhanced by combining the image features of the radar spectrogram. These three features The extraction enhancement can be advanced in parallel without any order. The following respectively introduces the selection of edge feature, texture feature and structure feature extraction.
提取待处理图像的边缘特征时,采用梯度变换算法,并将得到的梯度先验图命名为Fg,这里梯度变换算法为sobel算子设f(x,y)为待处理图像上(x,y)点的灰度值,G(x)为待处理图像x方向上边缘检测的图像灰度值,G(y)为待处理图像y方向上边缘检测的图像灰度值,a为G(x)的卷积因子,b为G(y)的卷积因子,则梯度先验图Fg上每个点的图像灰度值为:When extracting the edge features of the image to be processed, the gradient transformation algorithm is used, and the obtained gradient a priori map is named Fg, where the gradient transformation algorithm is the sobel operator. Let f(x,y) be (x,y) on the image to be processed ) Point gray value, G(x) is the image gray value of the edge detection in the x direction of the image to be processed, G(y) is the image gray value of the edge detection in the y direction of the image to be processed, a is G(x The convolution factor of ), b is the convolution factor of G(y), then the image gray value of each point on the gradient prior image Fg is:
Figure PCTCN2021097562-appb-000002
本技术方案中,a和b的值可选择3*3的矩阵,具体为:
Figure PCTCN2021097562-appb-000002
In this technical solution, the values of a and b can be selected as a 3*3 matrix, which is specifically:
Figure PCTCN2021097562-appb-000003
Figure PCTCN2021097562-appb-000003
在提取待处理图像中的纹理特征时采用的方法为:The method used when extracting texture features in the image to be processed is:
S1:对待处理图像进行连续n次卷积操作,得到卷积层conv_v1、conv_v2、……、conv_vn;S1: Perform n consecutive convolution operations on the image to be processed to obtain convolutional layers conv_v1, conv_v2, ..., conv_vn;
S2:再对卷积层conv_v1进行一次卷积操作,得到conv_v1';S2: Perform another convolution operation on the convolutional layer conv_v1 to obtain conv_v1';
S3:将conv_vn与conv_v1'通过相加运算得到纹理特征图Fv;S3: The texture feature map Fv is obtained by adding conv_vn and conv_v1';
参照图2所示,在本技术方案中,为了更好的提取图像中存在的原始细节纹理特征,本设计了一种基于过渡卷积的尺度不变多层CNN模块,因为纹理本身的细节和分辨率有很强的匹配关系,当分辨率不同时,纹理特征必然存在差异,因此本方案在提取纹理特征时分辨率不变。待处理图像为隧道扫描图I,在深度卷积神经网络中,随着卷积次数的增加,会出现梯度消失的现象,且卷积次数越多,计算效率越慢,本技术方案中n=5,由于地质雷达输出的原始图像为三通道的黑白图,在输入本模型前,需要将原始图像转化为单通道的灰度图,在对转化之后的单通道灰度图进行卷积操作时,采用3*3的卷积核,为保证分辨率不变,卷积时步长为1,然后在保证分辨率的同时,进行纹理特征提取,每次卷积的具体信息如下表:Referring to Figure 2, in this technical solution, in order to better extract the original detailed texture features in the image, a scale-invariant multi-layer CNN module based on transitional convolution is designed, because the details of the texture and the The resolution has a strong matching relationship. When the resolution is different, the texture feature must be different. Therefore, the resolution is unchanged when the texture feature is extracted in this solution. The image to be processed is the tunnel scan image I. In the deep convolutional neural network, as the number of convolutions increases, the gradient disappears, and the more the number of convolutions, the slower the calculation efficiency. In this technical solution, n= 5. Since the original image output by the geological radar is a three-channel black-and-white image, before inputting this model, the original image needs to be converted into a single-channel grayscale image. When performing convolution operations on the converted single-channel grayscale image , Using a 3*3 convolution kernel, in order to ensure the resolution is unchanged, the step size of the convolution is 1, and then while ensuring the resolution, the texture feature extraction is performed. The specific information of each convolution is as follows:
表1 纹理特征提取卷积信息表Table 1 Texture feature extraction convolution information table
卷积层Convolutional layer 步长Step size 卷积核Convolution kernel 通道数Number of channels
conv_v1conv_v1 11 3*33*3 4848
conv_v2conv_v2 11 3*33*3 4848
conv_v3conv_v3 11 3*33*3 4848
conv_v4conv_v4 11 3*33*3 4848
conv_v5conv_v5 11 3*33*3 4848
conv_v1'conv_v1' 11 5*55*5 4848
由于在卷积层conv_v1进行一次短支路单卷积操作并将得到的conv_v1'卷积层作为过渡卷积层,其中得到conv_v1'时采用的卷积因子参数为5*5的卷积核,因此conv_v1'的感受野相比于conv_v1更大,能够在其中学到与conv_v1不同的原始特征,而且conv_v1'由conv_v1卷积而来,相比于通过其他卷积层进行卷积操作,conv_v1'保留的原始特征更多,在将其与conv_vn相加时,可以更好的保留原始图像中的纹理特征,进而提高对地质缺陷的识别准确率。对于经过以上方法得到的纹理特征图命名为Fv。Since a short branch single convolution operation is performed in the conv_v1 conv_v1 and the obtained conv_v1' convolution layer is used as a transitional convolution layer, the conv_v1' convolution factor parameter used when obtaining conv_v1' is a convolution kernel of 5*5, Therefore, conv_v1' has a larger receptive field than conv_v1, and can learn original features different from conv_v1 in it, and conv_v1' is convolved from conv_v1. Compared with conv_v1' through other convolution layers, conv_v1' More original features are retained. When adding them to conv_vn, the texture features in the original image can be better retained, thereby improving the accuracy of the recognition of geological defects. The texture feature map obtained by the above method is named Fv.
在提取待处理图像中的结构特征时采用的方法为:The method used when extracting the structural features in the image to be processed is:
S11:对待处理图像进行1次卷积操作,得到conv_s1;S11: Perform a convolution operation on the image to be processed to obtain conv_s1;
S12:对conv_s1进行两次连续m-1次的卷积操作,分别得到conv_s2、conv_s3、……、 conv_sm和conv_s2'、conv_s3'、……、conv_sm';S12: Perform two consecutive m-1 convolution operations on conv_s1 to obtain conv_s2, conv_s3,..., conv_sm and conv_s2', conv_s3',..., conv_sm' respectively;
S13:将步骤S12中的conv_sm与conv_sm'通过contact拼接,得到conv_sm+1;S13: splicing conv_sm and conv_sm' in step S12 through contact to obtain conv_sm+1;
S14:将步骤S13中的conv_sm+1与步骤S12中的conv_sm-1'、conv_sm-2'、……、conv_s2'通过contact连续拼接,最终得到conv_s2m-1;S14: Conv_sm+1 in step S13 and conv_sm-1', conv_sm-2',..., conv_s2' in step S12 are continuously spliced through contact, and finally conv_s2m-1 is obtained;
S15:将S11中的conv_s1与步骤S14中的conv_s2m-1通过contact拼接,得到conv_s2m,并对conv_s2m进行一次卷积操作,得到Fs。S15: splicing conv_s1 in S11 with conv_s2m-1 in step S14 through contact to obtain conv_s2m, and perform a convolution operation on conv_s2m to obtain Fs.
参照图3的示意,在本技术方案中,待处理图像同样为已经经过灰度处理的隧道扫描图I,这里m=3,每次卷积的具体信息如下表:3, in this technical solution, the image to be processed is also a tunnel scan image I that has undergone gray-scale processing, where m=3, and the specific information of each convolution is as follows:
表2 结构特征提取卷积信息表Table 2 Structure feature extraction convolution information table
卷积层Convolutional layer 步长Step size 卷积核Convolution kernel 通道数Number of channels
conv_s1conv_s1 11 3*33*3 3232
conv_s2conv_s2 22 3*33*3 6464
conv_s3conv_s3 22 3*33*3 128128
conv_s2'conv_s2' 22 3*33*3 256256
conv_s3'conv_s3' 22 3*33*3 128128
conv_s4conv_s4 22 3*33*3 6464
conv_s5conv_s5 22 3*33*3 3232
conv_s6conv_s6 11 3*33*3 3232
结构特征是具有一定的尺度不变性,即尺度在一定的范围内变化,其结构特征是基本保持不变的。由于不同大小的结构特征,可能分别在不同的卷积层上容易识别,如果仅采用conv_sm或conv_sm'进行识别,将会造成信息丢失,并不能很好的识别不同大小的结构特征,而通过本方案中结构特征提取方法得到的Fs,既保留了所有卷积层的信息,又与输入原图大小,便于后续更好的进行结构特征的识别。步骤S15中对conv_s2m进行卷积操作时,采用1*1矩阵的卷积核,将多通道的卷积层转为单通道。对经过以上方法得到的结构特征图,命名为Fs。The structural feature has a certain scale invariance, that is, when the scale changes within a certain range, its structural feature remains basically unchanged. Since structural features of different sizes may be easily identified on different convolutional layers, if only conv_sm or conv_sm' is used for identification, information will be lost, and structural features of different sizes cannot be well recognized. The Fs obtained by the structural feature extraction method in the scheme not only retains the information of all convolutional layers, but also the size of the original input image, which is convenient for subsequent better identification of structural features. When performing convolution operation on conv_s2m in step S15, a 1*1 matrix convolution kernel is used to convert the multi-channel convolution layer into a single channel. The structural feature map obtained by the above method is named Fs.
通过以上操作已经获得了对边缘特征增强后的梯度先验图Fg,以及纹理特征图Fv和结构特征图Fs,以上三种特征图分别针对原始图像中的不同特征进行提取和增强,为了更准确的识别地质缺陷,需要将以上特征图进行拼接融合,这里采用contact算法对三种特征图进行拼接,得到多特征拼接图Fc,其中Fc定义为:Fc=Concat((Fg,Fv,Fs),0),在对Fc进行卷积融合操作时,采用以下方法,将Fc依次进行相同次数的下采样卷积操作和上采样卷积操 作,得到Ff,为了充分利用前面提取到的三种特征信息,这里的Ff设计为与Fc同样的三通道。本技术方案中,由于因为异常目标的图像特征具有多个尺寸级,为了对不同尺寸的目标均进行检测,因此对Fc进行了连续的2次下采样,然后进行了连续的2次上采样进行还原,卷积核为3*3矩阵,卷积步长为2,每次卷积的信息为:Through the above operations, we have obtained the gradient prior map Fg after the edge feature enhancement, as well as the texture feature map Fv and the structural feature map Fs. The above three feature maps are extracted and enhanced for different features in the original image, in order to be more accurate To identify geological defects, the above feature maps need to be spliced and fused. Here, the contact algorithm is used to splice the three feature maps to obtain a multi-feature spliced map Fc, where Fc is defined as: Fc=Concat((Fg,Fv,Fs), 0). When performing convolution and fusion operations on Fc, the following method is used to sequentially perform the same number of down-sampling convolution operations and up-sampling convolution operations on Fc to obtain Ff, in order to make full use of the three feature information extracted previously ,Ff here is designed to be the same three-channel as Fc. In this technical solution, because the image features of the abnormal target have multiple size levels, in order to detect targets of different sizes, the Fc is continuously down-sampled twice, and then the Fc is continuously up-sampled twice. To restore, the convolution kernel is a 3*3 matrix, the convolution step size is 2, and the information of each convolution is:
表3 特征融合卷积信息表Table 3 Feature Fusion Convolution Information Table
卷积层Convolutional layer 步长Step size 卷积核Convolution kernel 通道数Number of channels
conv_1conv_1 22 3*33*3 6464
conv_2conv_2 22 3*33*3 128128
conv_3conv_3 22 3*33*3 128128
conv_4conv_4 22 3*33*3 6464
经过以上卷积操作,舍去了原始图像中的大量冗余信息,并对边缘特征、纹理特征和结构特征都进行了增强操作,提高了后续目标识别的可靠性,也提高了图像中的信噪比,有利于提高后续目标识别的效率。After the above convolution operation, a large amount of redundant information in the original image is discarded, and the edge feature, texture feature and structural feature are all enhanced, which improves the reliability of subsequent target recognition and also improves the confidence in the image. The noise ratio is conducive to improving the efficiency of subsequent target recognition.
在对隧道地质缺陷进行识别前,先要训练目标识别模型,本技术方案中的目标识别模型训练方法,包括以下步骤:Before identifying the geological defects in the tunnel, the target recognition model must be trained. The target recognition model training method in this technical solution includes the following steps:
S21:将地质雷达检测获得的若干图像进行裁剪得到隧道扫描图I,为了方便计算拼接等需求,这里采用等比例固定大小的裁剪;S21: Cut several images obtained by the geological radar to obtain the tunnel scan image I. In order to facilitate the calculation of splicing and other requirements, a fixed-size crop of equal proportions is used here;
S22:将隧道扫描图I作为待处理图像经过以上图像处理方法处理得到多特征融合图Ff;S22: Use the tunnel scan image I as a to-be-processed image to obtain a multi-feature fusion image Ff through the above image processing method;
S23:将多特征融合图Ff进行下采样一倍操作,得到对应的下采样特征图Ff',因为异常目标的图像特征具有多个尺寸级,为了对不同尺寸的目标均进行检测,因此这里将检测回归搜索的特征图采用两个尺度下进行搜索;S23: The multi-feature fusion map Ff is down-sampled and doubled to obtain the corresponding down-sampled feature map Ff'. Because the image features of the abnormal target have multiple size levels, in order to detect targets of different sizes, we will The feature map of the detection regression search is searched on two scales;
S24:在Ff和Ff'上选择不同长宽比的先验框,并将每个先验框在图中所处位置信息(Xi,Yi,Wi,Hi)和对应异常状态类型Ci作为标签进行标记;这里i为图片序号,X、Y为先验框的位置坐标,W、H为先验框的宽和高;S24: Select a priori boxes with different aspect ratios on Ff and Ff', and use the position information (Xi, Yi, Wi, Hi) of each a priori box in the figure and the corresponding abnormal state type Ci as a label Mark; where i is the sequence number of the picture, X and Y are the position coordinates of the a priori frame, and W and H are the width and height of the a priori frame;
S25:将带有标签的先验框作为训练样本输入预识别模型进行训练,得到目标识别模型。S25: Input a priori box with a label as a training sample into a pre-recognition model for training to obtain a target recognition model.
本技术方案中,预识别模型为机器自学模型,包括神经网络、卷积神经网络、深度神经网络、反馈神经网络中的一种或多种组合。由于地质缺陷的大小不同,因此这里取多特征融合图Ff和下采样特征图Ff'两种不同大小的图档来进行识别训练。在进行训练时,将设定的六组先验框分别在对应的尺度特征图下进行检测回归得到目标检测框坐标、对应的状态数字 化种类和置信度,最后将不同尺度上检测的结果经过NMS非极大抑制筛选得到最终的目标检测结果,并进行输出。In this technical solution, the pre-identification model is a self-learning machine model, including one or more combinations of neural networks, convolutional neural networks, deep neural networks, and feedback neural networks. Due to the different sizes of geological defects, two different sizes of image files of multi-feature fusion map Ff and down-sampling feature map Ff' are used for recognition training. During training, the six sets of a priori boxes are respectively detected and regression under the corresponding scale feature map to obtain the target detection frame coordinates, the corresponding state digital type and the confidence level, and finally the results of the detection on different scales are passed through the NMS The non-maximum suppression screening obtains the final target detection result and outputs it.
目标识别模型训练完成后就可以对经过等比例裁剪的雷达波谱图进行地质缺陷的自动识别检测了,这里的目标识别方法,包括以下步骤:After the target recognition model training is completed, the geological defect can be automatically recognized and detected on the radar spectrogram that has been cropped in equal proportions. The target recognition method here includes the following steps:
接收待处理图像,即经过等比例裁剪的隧道扫描图I;Receive the image to be processed, that is, the tunnel scan image I cropped in equal proportions;
利用目标识别模型识别所述待识别图像中的目标对象;其中,所述目标识别模型利用以上方法预先训练得到。参照图5和图6,经过图像处理后的隧道扫描图,其中噪点大大减少,边缘特征、纹理特征和结构特征更加清晰的得到呈现,因此可以提高识别的速度和准确率。A target recognition model is used to recognize the target object in the image to be recognized; wherein, the target recognition model is pre-trained using the above method. Referring to Figures 5 and 6, the tunnel scan map after image processing has greatly reduced noise, and edge features, texture features, and structural features are presented more clearly, so the speed and accuracy of recognition can be improved.
目前在目标检测领域,基于深度学习的方法基本采用的单网络进行特征提取,这一做法虽然能一定程度上解决通用场景的问题,但是对于特殊场景下的目标检测,例如本技术方案的隧道检测中,需要检测的目标成像并非光学成像拍摄获取得到的,而是通过雷达探测波段信号然后利用模数转换得到的数字信息以图像的形式呈现,图中具有大量的条纹和噪点,因此,从其信号源头上结合图像特点,本技术方案设计了一组多维特征的提取,分别针对梯度先验图Fg、纹理特征图Fv、结构特征图Fs进行提取与融合,从下表4中,检测效果对比上可以看出多维特征融合决策对隧道异常检测这一特殊场景下的检测效果有明显的性能提升。At present, in the field of target detection, deep learning-based methods basically use a single network for feature extraction. Although this approach can solve the problem of general scenarios to a certain extent, it is for target detection in special scenarios, such as tunnel detection in this technical solution. In the image, the target imaging that needs to be detected is not obtained by optical imaging, but the digital information obtained by detecting the band signal by radar and then using analog-to-digital conversion is presented in the form of an image. There are a lot of stripes and noise in the picture. Therefore, from it Combining image characteristics at the source of the signal, this technical solution designs a set of multi-dimensional feature extraction, which extracts and merges the gradient prior map Fg, texture feature map Fv, and structural feature map Fs respectively. From Table 4 below, compare the detection results It can be seen from the above that the multi-dimensional feature fusion decision has a significant performance improvement on the detection effect in the special scene of tunnel anomaly detection.
表4 性能结果对比Table 4 Comparison of performance results
Figure PCTCN2021097562-appb-000004
Figure PCTCN2021097562-appb-000004
以上示意性地对本发明创造及其实施方式进行了描述,该描述没有限制性,在不背离本发明的精神或者基本特征的情况下,能够以其他的具体形式实现本发明。附图中所示的也只是本发明创造的实施方式之一,实际的结构并不局限于此,权利要求中的任何附图标记不应限制所涉及的权利要求。所以,如果本领域的普通技术人员受其启示,在不脱离本创造宗旨的情况下,不经创造性的设计出与该技术方案相似的结构方式及实施例,均应属于本专利的保护范围。此外,“包括”一词不排除其他元件或步骤,在元件前的“一个”一词不排除包括“多个”该元件。产品权利要求中陈述的多个元件也可以由一个元件通过软件或者硬件来实现。第一,第二等词语用来表示名称,而并不表示任何特定的顺序。The creation of the present invention and its implementation are described schematically above. The description is not restrictive. The present invention can be implemented in other specific forms without departing from the spirit or basic characteristics of the present invention. What is shown in the drawings is only one of the embodiments created by the present invention, and the actual structure is not limited to this, and any reference signs in the claims should not limit the related claims. Therefore, if a person of ordinary skill in the art receives its enlightenment, and does not deviate from the purpose of this creation, without creative design, structural methods and embodiments similar to the technical solution should fall within the scope of protection of this patent. In addition, the word "comprising" does not exclude other elements or steps, and the word "a" before an element does not exclude the inclusion of "plurality" of the element. Multiple elements stated in the product claims can also be implemented by one element through software or hardware. Words such as first and second are used to denote names, but do not denote any specific order.

Claims (10)

  1. 一种图像处理方法,其特征在于:包括以下步骤:An image processing method, characterized in that it comprises the following steps:
    提取待处理图像的边缘特征,得到梯度先验图Fg;Extract the edge features of the image to be processed to obtain the gradient prior image Fg;
    提取待处理图像的纹理特征,得到纹理特征图Fv;Extract the texture feature of the image to be processed to obtain the texture feature map Fv;
    提取待处理图像的结构特征,得到结构特征图Fs;Extract the structural features of the image to be processed to obtain the structural feature map Fs;
    将梯度先验图Fg、纹理特征图Fv和结构特征图Fs进行contact拼接,得到多特征拼接图Fc;Connect the gradient prior map Fg, the texture feature map Fv and the structural feature map Fs to contact stitching to obtain a multi-feature stitching image Fc;
    对多特征拼接图Fc进行卷积融合操作,得到多特征融合图Ff。The convolution fusion operation is performed on the multi-feature mosaic image Fc to obtain the multi-feature fusion image Ff.
  2. 根据权利要求1所述的一种图像处理方法,其特征在于:采用梯度变换算法提取待处理图像的边缘特征,梯度变换算法提取边缘特征的方法为:An image processing method according to claim 1, wherein the gradient transformation algorithm is used to extract the edge features of the image to be processed, and the method of the gradient transformation algorithm to extract the edge features is:
    设f(x,y)为待处理图像上(x,y)点的灰度值,G(x)为待处理图像x方向上边缘检测的图像灰度值,G(y)为待处理图像y方向上边缘检测的图像灰度值,a为G(x)的卷积因子,b为G(y)的卷积因子,则梯度先验图Fg上每个像素点的图像灰度值为:Let f(x,y) be the gray value of the (x,y) point on the image to be processed, G(x) is the image gray value of the edge detection in the x direction of the image to be processed, and G(y) is the image to be processed The image gray value of edge detection in the y direction, a is the convolution factor of G(x), b is the convolution factor of G(y), then the image gray value of each pixel on the gradient prior image Fg is :
    Figure PCTCN2021097562-appb-100001
    通过每个像素点的灰度值来呈现梯度先验图Fg上的边缘特征。
    Figure PCTCN2021097562-appb-100001
    The edge feature on the gradient priori map Fg is presented by the gray value of each pixel.
  3. 根据权利要求2所述的一种图像处理方法,其特征在于:对待处理图像进行纹理特征提取时采用以下步骤:An image processing method according to claim 2, wherein the following steps are used when extracting texture features of the image to be processed:
    S1:对待处理图像进行连续n次卷积操作,依次得到卷积层v1、卷积层v2、……、卷积层vn;S1: Perform n consecutive convolution operations on the image to be processed, and obtain convolutional layer v1, convolutional layer v2, ..., convolutional layer vn in sequence;
    S2:再对卷积层v1进行一次卷积操作,得到卷积层v1';S2: Perform another convolution operation on the convolutional layer v1 to obtain the convolutional layer v1';
    S3:将卷积层v1与卷积层v1'通过相加运算得到纹理特征图Fv。S3: The texture feature map Fv is obtained by adding the convolutional layer v1 and the convolutional layer v1'.
  4. 根据权利要求3所述的一种图像处理方法,其特征在于,对待处理图像进行结构特征提取时采用以下步骤:An image processing method according to claim 3, wherein the following steps are used when extracting structural features of the image to be processed:
    S11:对待处理图像进行1次卷积操作,得到卷积层s1;S11: Perform one convolution operation on the image to be processed to obtain the convolution layer s1;
    S12:对卷积层s1进行两次连续m-1次的卷积操作,分别得到卷积层s2、卷积层s3、……、卷积层sm和卷积层s2'、卷积层s3'、……、卷积层sm';S12: Perform two consecutive m-1 convolution operations on the convolutional layer s1 to obtain the convolutional layer s2, convolutional layer s3,..., convolutional layer sm, convolutional layer s2', and convolutional layer s3, respectively ',..., convolutional layer sm';
    S13:将步骤S12中的卷积层sm与卷积层sm'通过contact拼接,得到卷积层sm+1;S13: The convolutional layer sm and the convolutional layer sm' in step S12 are spliced through contact to obtain the convolutional layer sm+1;
    S14:将步骤S13中的卷积层sm+1与步骤S12中的卷积层sm-1'、卷积层sm-2'、……、卷积层s2'通过contact连续拼接,最终得到卷积层s2m-1;S14: The convolutional layer sm+1 in step S13 and the convolutional layer sm-1', convolutional layer sm-2',..., convolutional layer s2' in step S12 are continuously spliced through contact, and finally the convolution is obtained Build-up s2m-1;
    S15:将S11中的卷积层s1与步骤S14中的卷积层s2m-1通过contact拼接,得到卷积层s2m,并对卷积层s2m进行一次卷积操作,得到Fs。S15: Join the convolutional layer s1 in S11 and the convolutional layer s2m-1 in step S14 through contact to obtain the convolutional layer s2m, and perform a convolution operation on the convolutional layer s2m to obtain Fs.
  5. 根据权利要求3或4所述的一种图像处理方法,其特征在于,所述步骤S1中,对待处理图像进行第一次卷积操作,得到卷积层v1时卷积核为3*3矩阵,所述步骤S2中卷积操作的卷积核为5*5矩阵。An image processing method according to claim 3 or 4, characterized in that, in the step S1, the image to be processed is subjected to the first convolution operation, and the convolution kernel is a 3*3 matrix when the convolution layer v1 is obtained , The convolution kernel of the convolution operation in the step S2 is a 5*5 matrix.
  6. 根据权利要求4所述的一种图像处理方法,其特征在于,所述步骤S15中对卷积层s2m的卷积操作采用卷积核为1*1矩阵。An image processing method according to claim 4, wherein the convolution operation of the convolution layer s2m in the step S15 adopts a convolution kernel as a 1*1 matrix.
  7. 一种目标识别模型训练方法,其特征在于,包括以下步骤:A method for training a target recognition model is characterized in that it comprises the following steps:
    S21:将地质雷达检测获得的若干图像进行裁剪得到隧道扫描图I;S21: Cut several images obtained by the ground penetrating radar to obtain a tunnel scan image I;
    S22:将隧道扫描图I作为待处理图像经过权利要求1-6中任一项的图像处理方法处理得到多特征融合图Ff;S22: Use the tunnel scan image I as a to-be-processed image to obtain a multi-feature fusion image Ff through the image processing method of any one of claims 1-6;
    S23:将多特征融合图Ff进行下采样操作,得到对应的下采样特征图Ff';S23: Perform a down-sampling operation on the multi-feature fusion map Ff to obtain the corresponding down-sampled feature map Ff';
    S24:在Ff和Ff'上选择不同长宽比的先验框,并将每个先验框在图中所处位置信息和对应异常状态类型作为标签进行标记;S24: Select a priori boxes with different aspect ratios on Ff and Ff', and mark the position information of each a priori box in the figure and the corresponding abnormal state type as a label;
    S25:将带有标签的先验框作为训练样本输入预识别模型进行训练,得到目标识别模型。S25: Input a priori box with a label as a training sample into a pre-recognition model for training to obtain a target recognition model.
  8. 根据权利要求7所述的一种目标识别模型训练方法,其特征在于,所述预识别模型为机器自学模型,包括神经网络、卷积神经网络、深度神经网络、反馈神经网络中的一种或多种组合。The method for training a target recognition model according to claim 7, wherein the pre-recognition model is a machine self-learning model, including one of a neural network, a convolutional neural network, a deep neural network, and a feedback neural network, or Multiple combinations.
  9. 根据权利要求7所述的一种目标识别模型训练方法,其特征在于,所述步骤S23中多特征融合图Ff进行下采样一倍的操作得到Ff'。A target recognition model training method according to claim 7, characterized in that, in the step S23, the multi-feature fusion map Ff is down-sampled and doubled to obtain Ff'.
  10. 一种目标识别方法,其特征在于,包括以下步骤:A target recognition method is characterized in that it comprises the following steps:
    接收待处理图像;Receive the image to be processed;
    利用目标识别模型识别所述待识别图像中的目标对象;其中,所述目标识别模型利用权利要求7-9中任一项所述的方法预先训练得到。A target recognition model is used to recognize the target object in the image to be recognized; wherein the target recognition model is obtained by pre-training using the method of any one of claims 7-9.
PCT/CN2021/097562 2020-06-10 2021-06-01 Image processing method, target recognition model training method, and target recognition method WO2021249233A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010523228.4 2020-06-10
CN202010523228.4A CN111784642B (en) 2020-06-10 2020-06-10 Image processing method, target recognition model training method and target recognition method

Publications (1)

Publication Number Publication Date
WO2021249233A1 true WO2021249233A1 (en) 2021-12-16

Family

ID=72755828

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/097562 WO2021249233A1 (en) 2020-06-10 2021-06-01 Image processing method, target recognition model training method, and target recognition method

Country Status (2)

Country Link
CN (1) CN111784642B (en)
WO (1) WO2021249233A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114584674A (en) * 2022-05-06 2022-06-03 深圳市巨力方视觉技术有限公司 Visual integration system for processing same image
CN115880290A (en) * 2023-02-22 2023-03-31 季华实验室 OLED (organic light emitting diode) wet film defect detection method based on lightweight semantic segmentation network
CN116363031A (en) * 2023-02-28 2023-06-30 锋睿领创(珠海)科技有限公司 Imaging method, device, equipment and medium based on multidimensional optical information fusion
CN116385953A (en) * 2023-01-11 2023-07-04 哈尔滨市科佳通用机电股份有限公司 Railway wagon door hinge breaking fault image identification method
CN116630899A (en) * 2023-07-21 2023-08-22 四川公路工程咨询监理有限公司 Highway side slope disease monitoring and early warning system
CN117237777A (en) * 2023-11-13 2023-12-15 四川观想科技股份有限公司 Ship target identification method based on multi-mode fusion
CN117649917A (en) * 2024-01-29 2024-03-05 北京大学 Training method and device for test report generation model and test report generation method
WO2024066090A1 (en) * 2022-09-26 2024-04-04 上海闻泰电子科技有限公司 Corner detection method and system based on texture features, electronic device, and medium

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111784642B (en) * 2020-06-10 2021-12-28 中铁四局集团有限公司 Image processing method, target recognition model training method and target recognition method
CN112419152B (en) * 2020-11-23 2024-03-29 中国科学院深圳先进技术研究院 Image super-resolution method, device, terminal equipment and storage medium
CN112529106A (en) * 2020-12-28 2021-03-19 平安普惠企业管理有限公司 Method, device and equipment for generating visual design manuscript and storage medium
CN113128404A (en) * 2021-04-20 2021-07-16 中国地质科学院 Intelligent mineral identification method and system
CN113128521B (en) * 2021-04-30 2023-07-18 西安微电子技术研究所 Method, system, computer equipment and storage medium for extracting characteristics of miniaturized artificial intelligent model

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110182469A1 (en) * 2010-01-28 2011-07-28 Nec Laboratories America, Inc. 3d convolutional neural networks for automatic human action recognition
CN103996209A (en) * 2014-05-21 2014-08-20 北京航空航天大学 Infrared vessel object segmentation method based on salient region detection
US20150117760A1 (en) * 2013-10-30 2015-04-30 Nec Laboratories America, Inc. Regionlets with Shift Invariant Neural Patterns for Object Detection
CN106529447A (en) * 2016-11-03 2017-03-22 河北工业大学 Small-sample face recognition method
CN111784642A (en) * 2020-06-10 2020-10-16 中铁四局集团有限公司 Image processing method, target recognition model training method and target recognition method

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107621626B (en) * 2017-10-09 2018-08-31 中国矿业大学(北京) Radar signal Railway Roadbed detection method based on depth convolutional neural networks
CN109490861B (en) * 2018-10-29 2020-06-02 北京科技大学 Blast furnace burden line extraction method
CN113569796A (en) * 2018-11-16 2021-10-29 北京市商汤科技开发有限公司 Key point detection method and device, electronic equipment and storage medium
CN111028235B (en) * 2019-11-11 2023-08-22 东北大学 Image segmentation method for enhancing edge and detail information by utilizing feature fusion

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110182469A1 (en) * 2010-01-28 2011-07-28 Nec Laboratories America, Inc. 3d convolutional neural networks for automatic human action recognition
US20150117760A1 (en) * 2013-10-30 2015-04-30 Nec Laboratories America, Inc. Regionlets with Shift Invariant Neural Patterns for Object Detection
CN103996209A (en) * 2014-05-21 2014-08-20 北京航空航天大学 Infrared vessel object segmentation method based on salient region detection
CN106529447A (en) * 2016-11-03 2017-03-22 河北工业大学 Small-sample face recognition method
CN111784642A (en) * 2020-06-10 2020-10-16 中铁四局集团有限公司 Image processing method, target recognition model training method and target recognition method

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114584674A (en) * 2022-05-06 2022-06-03 深圳市巨力方视觉技术有限公司 Visual integration system for processing same image
WO2024066090A1 (en) * 2022-09-26 2024-04-04 上海闻泰电子科技有限公司 Corner detection method and system based on texture features, electronic device, and medium
CN116385953B (en) * 2023-01-11 2023-12-15 哈尔滨市科佳通用机电股份有限公司 Railway wagon door hinge breaking fault image identification method
CN116385953A (en) * 2023-01-11 2023-07-04 哈尔滨市科佳通用机电股份有限公司 Railway wagon door hinge breaking fault image identification method
CN115880290A (en) * 2023-02-22 2023-03-31 季华实验室 OLED (organic light emitting diode) wet film defect detection method based on lightweight semantic segmentation network
CN115880290B (en) * 2023-02-22 2023-05-30 季华实验室 OLED wet film defect detection method based on lightweight semantic segmentation network
CN116363031A (en) * 2023-02-28 2023-06-30 锋睿领创(珠海)科技有限公司 Imaging method, device, equipment and medium based on multidimensional optical information fusion
CN116363031B (en) * 2023-02-28 2023-11-17 锋睿领创(珠海)科技有限公司 Imaging method, device, equipment and medium based on multidimensional optical information fusion
CN116630899B (en) * 2023-07-21 2023-10-20 四川公路工程咨询监理有限公司 Highway side slope disease monitoring and early warning system
CN116630899A (en) * 2023-07-21 2023-08-22 四川公路工程咨询监理有限公司 Highway side slope disease monitoring and early warning system
CN117237777A (en) * 2023-11-13 2023-12-15 四川观想科技股份有限公司 Ship target identification method based on multi-mode fusion
CN117237777B (en) * 2023-11-13 2024-02-27 四川观想科技股份有限公司 Ship target identification method based on multi-mode fusion
CN117649917A (en) * 2024-01-29 2024-03-05 北京大学 Training method and device for test report generation model and test report generation method

Also Published As

Publication number Publication date
CN111784642A (en) 2020-10-16
CN111784642B (en) 2021-12-28

Similar Documents

Publication Publication Date Title
WO2021249233A1 (en) Image processing method, target recognition model training method, and target recognition method
Liu et al. Improved kiwifruit detection using pre-trained VGG16 with RGB and NIR information fusion
CN111401384B (en) Transformer equipment defect image matching method
CN112232391B (en) Dam crack detection method based on U-net network and SC-SAM attention mechanism
CN109871823B (en) Satellite image ship detection method combining rotating frame and context information
CN111951284B (en) Optical remote sensing satellite image refined cloud detection method based on deep learning
Thanikkal et al. Advanced plant leaf classification through image enhancement and canny edge detection
CN115331087A (en) Remote sensing image change detection method and system fusing regional semantics and pixel characteristics
CN111462140B (en) Real-time image instance segmentation method based on block stitching
CN111652039B (en) Hyperspectral remote sensing ground object classification method based on residual error network and feature fusion module
US20220366682A1 (en) Computer-implemented arrangements for processing image having article of interest
CN112819837A (en) Semantic segmentation method based on multi-source heterogeneous remote sensing image
CN115984550A (en) Automatic segmentation method for eye iris pigmented spot texture
CN116363527A (en) Remote sensing image change detection method based on interaction feature perception
CN114596503A (en) Road extraction method based on remote sensing satellite image
CN115049921A (en) Method for detecting salient target of optical remote sensing image based on Transformer boundary sensing
CN116977387B (en) Deformable medical image registration method based on deformation field fusion
CN111179278B (en) Image detection method, device, equipment and storage medium
TW202225730A (en) High-efficiency LiDAR object detection method based on deep learning through direct processing of 3D point data to obtain a concise and fast 3D feature to solve the shortcomings of complexity and time-consuming of the current voxel network model
Prabhakar et al. Cdnet++: Improved change detection with deep neural network feature correlation
CN115456957B (en) Method for detecting change of remote sensing image by full-scale feature aggregation
CN113723447B (en) End-to-end template matching method for multi-mode image
CN115661451A (en) Deep learning single-frame infrared small target high-resolution segmentation method
CN115861922A (en) Sparse smoke and fire detection method and device, computer equipment and storage medium
CN113192076B (en) MRI brain tumor image segmentation method combining classification prediction and multi-scale feature extraction

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21823150

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21823150

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 21823150

Country of ref document: EP

Kind code of ref document: A1