CN110349134B - Pipeline disease image classification method based on multi-label convolutional neural network - Google Patents
Pipeline disease image classification method based on multi-label convolutional neural network Download PDFInfo
- Publication number
- CN110349134B CN110349134B CN201910565360.9A CN201910565360A CN110349134B CN 110349134 B CN110349134 B CN 110349134B CN 201910565360 A CN201910565360 A CN 201910565360A CN 110349134 B CN110349134 B CN 110349134B
- Authority
- CN
- China
- Prior art keywords
- pipeline
- layer
- neural network
- convolutional neural
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24147—Distances to closest patterns, e.g. nearest neighbour classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
- Endoscopes (AREA)
Abstract
The invention discloses a pipeline disease image classification method based on a multi-label convolutional neural network, which comprises the following steps: step 1: collecting a pipeline endoscopic detection video, and extracting image frames in the video; step 2: calculating the time stamp characteristics of each image; and step 3: sending a part of the image frames collected in the step 1 into a multi-label convolutional neural network model for training to obtain the multi-label convolutional neural network model capable of correctly classifying the types of the pipeline diseases; and 4, step 4: the method comprises the steps that a trained multi-label convolutional neural network model is used for detecting an endoscopic image of a pipeline to be detected, then the multi-label convolutional neural network model outputs a single hot code, the type of the existing pipeline diseases is determined according to the single hot code, a multi-label classification layer is added on the basis of the existing inclusion-ResNet-v 2 network, and the classification function of the multiple pipeline disease images is achieved.
Description
Technical Field
The invention relates to the field of computer digital image processing and deep learning algorithm based on a convolutional neural network, in particular to a pipeline disease image classification method based on a multi-label convolutional neural network.
Background
There are two commonly used methods for acquiring an endoscopic image of a pipeline: one is a pipeline inspection robot technology based on closed circuit television inspection (CCTV), and the other is a pipeline periscope technology based on pipeline quick view inspection (QV).
In the prior art, a processing method after an image is obtained still mainly adopts manual observation at present and then carries out classification and identification. The following difficulties arise from manual observation:
the increase of the pipeline construction length can cause the generation of huge endoscopic videos and image data, for example, a pipeline detection robot operates at the maximum speed of 2.5-3m/s, videos of at least 5 minutes need to be shot for a pipeline with the length of 1000m, 9000 pictures can be generated when the videos are shot according to the frame rate of 30fps, and the detection result is inaccurate due to factors such as work fatigue and subjective consciousness because of the inevitable manual inspection and screening of huge data quantity; therefore, a computer digital image processing technology is needed to realize automatic classification and identification of diseases in the pipeline endoscopic image.
The pipeline is wide in distribution and multiple in types, and the types of generated diseases are relatively complex, but pipeline detection is an important means for maintaining normal operation of the pipeline, from manual submerged inspection to detection of professional instruments such as ground penetrating radar and sonar, and then to the mode of acquiring an endoscopic image of the pipeline by adopting a pipeline robot or a pipeline periscope for observation at present, the safety and universality of the pipeline are greatly improved, and the digital image processing technology successfully realizes automatic identification and monitoring of several pipeline diseases.
At present, two main schemes are available for classifying endoscopic images of pipelines by using a computer digital image processing technology:
(1) subjectively acquiring image features of specific disease types, and classifying by using a traditional machine learning classifier:
aiming at the images of the diseases of cracking, dislocation and disjointing in the pipeline, extracting horizontal and diagonal wavelet components, image entropy and symbiotic correlation as features, and combining the features into a feature vector;
<2> extracting the roundness, compactness and roughness of the outline as features aiming at the images of dirt deposition and tree root diseases in the pipeline, and combining the features into a feature vector;
extracting angular second moment, image entropy, correlation, dissimilarity, contrast and uniformity as features aiming at disease images such as blockage and illegal pipeline access in a pipeline, and combining the features into a feature vector;
and 4, sending the characteristic vectors into a support vector machine, a random forest and a full-connection neural network for training to realize classification of disease types.
(2) Objectively extracting image abstract features by using a convolutional neural network for classification:
normalizing all the shot endoscopic images of the pipeline into the resolution of 256 multiplied by 256;
sending the pictures into a convolutional neural network shown in figure 3 for training aiming at three diseases of root invasion, dirt deposition and pipe wall breakage which are common in pipeline diseases; in the figure, the input layer × 3 indicates that the input image is a color image, three channels of red, green, and blue are shared, the medium size of each convolution layer indicates the size and the number of channels of the original image through convolution operation. For example, 128 × 128 × 32 indicates that the original image becomes an image with a resolution of 128 × 128 and 32 data channels through the first convolution operation, and this image is referred to as a feature map.
<3> the number in the fully connected layer indicates that the feature map is converted into vectors of specific dimensions, for example 1024, or the feature map is converted into vectors of 1024 dimensions. 3 of the output layer represents a category vector and adopts a one-hot coding identifier;
<4>, for example, if a disease image with tree root intrusion is input, the output layer will output a vector in the shape of [0,0,1 ]; similarly, the deposition of dirt will be output [0,1,0], and the rupture of the tube wall will output [1,0,0].
However, the existing computer digital image processing technology needs to subjectively acquire the corresponding characteristics of the disease image and is too dependent on the quality of the shot image; for the inner wall of a pipeline with dark, damp and complex environment, obtaining high-quality images has certain difficulty;
the distinguishing sensitivity of different disease type image characteristics is different, and certain pipeline detection work experience is required by selecting the characteristics as the classification basis, so that the development of an image detection technology is limited;
in addition, the existing convolutional neural network has the advantages of simple model, few classification types, incapability of solving the problem of complicated pipeline disease type classification and low classification accuracy.
Disclosure of Invention
The invention adds a multi-label classification layer on the basis of the existing increment-ResNet-v 2 network, and realizes the classification function of various pipeline disease images.
A pipeline disease image classification method based on a multi-label convolutional neural network comprises the following steps:
step 1: collecting a pipeline endoscopic detection video, and extracting image frames in the pipeline endoscopic detection video;
step 2: calculating the time stamp characteristics of each image;
and step 3: sending a part of the image frames collected in the step 1 into a multi-label convolutional neural network model for training to obtain the multi-label convolutional neural network model capable of correctly classifying the types of the pipeline diseases;
and 4, step 4: and detecting the endoscopic image of the pipeline to be detected by using the trained multi-label convolutional neural network model, outputting a single-hot code by using the multi-label convolutional neural network model, and determining the type of the existing pipeline diseases according to the single-hot code.
Preferably, the multi-label convolutional neural network model comprises an upper-layer inclusion-ResNet-v 2 network structure and a lower-layer multi-label classification layer, and the upper-layer inclusion-ResNet-v 2 network structure further comprises a random deactivation layer.
Preferably, the processing steps of the lower multi-label classification layer are as follows:
the first step is as follows: a timestamp feature TimeFiture is added in a feature vector output by a random inactivation layer in an upper-layer inclusion-ResNet-v 2 network structure,
the second step is that: performing dimension reduction activation processing on the feature vector added with the timestamp feature in the first step to obtain a medium latitude value,
and thirdly, continuing to perform dimensionality reduction on the medium latitude value in the second step, and finally outputting the one-hot code.
Preferably, the calculation formula of the timestamp feature TimeFeature is as follows:
the Current _ Frame _ Index is a position serial number of the to-be-detected pipeline endoscopic image in the whole pipeline endoscopic detection video, and the All _ Frames _ Num is the total Frame number of the whole pipeline endoscopic detection video.
Preferably, the types of the pipeline diseases comprise: swabbing, tree roots, accumulated mud, debris, scaling, plugging, mild corrosion, moderate corrosion, severe corrosion, dislocation, fracture, dislocation, deformation, intrusion, mild leakage, moderate leakage, and severe leakage.
The beneficial effects obtained by the invention are as follows:
(1) by adopting the convolutional neural network with a complex model, the characteristics do not need to be extracted subjectively, and the accuracy of disease classification is improved;
(2) the time stamp characteristics of the image to be detected are combined, the types of the classification characteristics are enriched, and the time stamp characteristics are utilized to help to distinguish non-disease image frames and irrelevant image frames (such as irrelevant images shot when a pipeline detection periscope or a pipeline detection robot is put into a well or recovered by workers) in a detection video;
(3) can distinguish 17 kinds of pipeline diseases, almost covers all kinds of diseases of the current urban pipeline, and has wider application range.
Drawings
The invention will be further understood from the following description in conjunction with the accompanying drawings. The components in the figures are not necessarily to scale, emphasis instead being placed upon illustrating the principles of the embodiments; like reference numerals designate corresponding parts throughout the different views.
FIG. 1 is a block diagram of a network architecture of the present invention;
FIG. 2 is a process flow diagram of the lower multi-label classification layer of the present invention;
fig. 3 is a diagram illustrating a convolutional neural network structure in the prior art.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail below with reference to embodiments thereof; it should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention. Other systems, methods, and/or features of the present embodiments will become apparent to those skilled in the art upon review of the following detailed description. It is intended that all such additional systems, methods, features and advantages be included within this description, be within the scope of the invention, and be protected by the accompanying claims. Additional features of the disclosed embodiments are described in, and will be apparent from, the detailed description that follows.
The same or similar reference numerals in the drawings of the embodiments of the present invention correspond to the same or similar components; in the description of the present invention, it should be understood that if there is an orientation or positional relationship indicated by the terms "upper", "lower", "left", "right", etc. based on the orientation or positional relationship shown in the drawings, it is only for convenience of describing the present invention and simplifying the description, but it is not intended to indicate or imply that the device or component referred to must have a specific orientation, be constructed in a specific orientation, and be operated, and therefore, the terms describing the positional relationship in the drawings are only used for illustrative purposes and are not to be construed as limiting the present patent, and the specific meaning of the terms described above will be understood by those of ordinary skill in the art according to the specific circumstances.
The first embodiment is as follows:
a pipeline disease image classification method based on a multi-label convolutional neural network comprises the following steps:
step 1: collecting a pipeline endoscopic detection video, and extracting image frames in the pipeline endoscopic detection video;
step 2: calculating the time stamp characteristics of each image;
and step 3: sending a part of the image frames collected in the step 1 into a multi-label convolutional neural network model for training to obtain the multi-label convolutional neural network model capable of correctly classifying the types of the pipeline diseases;
and 4, step 4: and detecting the endoscopic image of the pipeline to be detected by using the trained multi-label convolutional neural network model, outputting a single-hot code by using the multi-label convolutional neural network model, and determining the type of the existing pipeline diseases according to the single-hot code.
The multi-label convolutional neural network model comprises an upper-layer inclusion-ResNet-v 2 network structure and a lower-layer multi-label classification layer, and the upper-layer inclusion-ResNet-v 2 network structure further comprises a random inactivation layer.
The processing steps of the lower multi-label classification layer are as follows:
the first step is as follows: adding a timestamp feature Timefeature in a feature vector output by a random inactivation layer in an upper-layer inclusion-ResNet-v 2 network structure;
the second step is that: performing dimension reduction activation processing on the feature vector added with the timestamp feature in the first step to obtain a medium latitude value;
and thirdly, continuing to perform dimensionality reduction on the medium latitude value in the second step, and finally outputting the one-hot code.
The calculation formula of the timestamp feature TimeFeture is as follows:
the Current _ Frame _ Index is the serial number of the position of the to-be-detected pipeline endoscopic image in the whole pipeline endoscopic detection video, and the All _ Frames _ Num is the total Frame number of the whole pipeline endoscopic detection video.
The types of the pipeline diseases comprise: swabbing, tree roots, accumulated mud, debris, scaling, plugging, mild corrosion, moderate corrosion, severe corrosion, dislocation, fracture, dislocation, deformation, intrusion, mild leakage, moderate leakage, and severe leakage.
Example two:
in this embodiment, the content is further described by combining the above embodiments, so that a person skilled in the art can more clearly understand the implementation of the present invention, wherein a multi-label classification layer is added to the present invention on the basis of the existing inclusion-ResNet-v 2 network to implement the classification function of multiple pipe defect images, and the present invention replaces the SoftMax classifier therein with the multi-label classification layer, so that the inclusion-ResNet-v 2 has the multi-label classification function, thereby maximally detecting the types of defects that should be detected, as shown in fig. 2.
Wherein the random deactivation layer of the upper inclusion-ResNet-v 2 network structure outputs a 1792-dimensional feature vector, corresponding to layer X in fig. 2. The invention adds a one-dimensional feature behind the vector, as shown by a light gray frame in fig. 2, the feature is the time stamp feature of the image to be detected, and the calculation is performed by using a formula 1,
the Current _ Frame _ Index in the formula is the serial number of the position of the to-be-detected pipeline endoscopic image in the whole pipeline endoscopic detection video, the All _ Frames _ Num is the total Frame number of the whole pipeline endoscopic detection video, and the value range of the All _ Frame _ Num is 0-1.
The H and C layers in fig. 2 are the main components of the multi-tag classification layer, wherein the H layer contains 1024 active processing units, namely circles; the layer C comprises 17 disease type identifications output in a single-hot coding mode, and the disease type identifications are respectively as follows from left to right:
c = [ hollow water, tree root, accumulated mud, sundries, scaling, plugging, slight corrosion, moderate corrosion, severe corrosion, disjointing, cracking, stagger, deformation, invasion, slight leakage, moderate leakage and severe leakage ].
For example, if there are tree roots and mud accumulation diseases in the image, the C layer outputs a vector [0,1,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0], and if there are no diseases, outputs a vector [0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0].
Example three:
in the embodiment, the content is further described by combining the above embodiments, so that a person skilled in the art can more clearly understand the implementation of the present invention, wherein a multi-label classification layer is added to the present invention on the basis of the existing inclusion-ResNet-v 2 network, so as to implement the classification function of multiple pipeline disease images, and the present invention replaces the SoftMax classifier therein with the multi-label classification layer, so that the inclusion-ResNet-v 2 has the multi-label classification function, thereby maximally detecting the types of diseases that should be detected.
As shown in fig. 2, wherein the X layer is the output vector of the random deactivation layer in the upper inclusion-ResNet-v 2 network structure shown.
The random inactivation layer of the original inclusion-ResNet-v 2 network structure only outputs the information of the former 1792 dimension, and the invention adds the time information of one dimension on the basis.
The process of converting the X layer into the H layer is a forward calculation process of the fully-connected neural network, and aims to properly reduce the dimension of the 1792+ 1-dimensional vector into 1024 dimensions. For convenience of description, several symbols are appropriately added to the drawings, wherein X 1 ~X 1793 Represents the output of the original Incepration-ResNet-v 2 random deactivation layer plus a new dimension in the text; h 1 ~X 1024 Representing the dimension values of 1793-dimensional vector into 1024-dimensional vector after H-layer activation, C 1 ~C 17 Representing the 17-dimensional unique thermal coding obtained after C-layer processing; w x->H Is a weight vector, which represents the connecting line connecting the X layer and the H layer, and the total number of dimensions is 1973 × 1024, because all the dimensions of the X layer are connected with all the dimensions of the H layer (so called full connection), the calculation process is as follows:
with H in H layer 1 For example, as shown in equation 2,
wherein W xi->H1 Indicates that the ith element of the X layer is connected with the H layer H 1 Is the connection weight, and sigmoid is suchA function capable of mapping real numbers to numbers between 0 and 1, the function being expressed asThe image is S-shaped, and the value range is between 0 and 1. H 2 ~H 1024 Is calculated in a similar manner to equation 1.
The process of H-layer to C-layer conversion, similar to the above calculation method, uses equation 3, with C 1 The calculation process of (a) is taken as an example,
all C are calculated i According to the characteristics of sigmoid, at the C layer, a 17-dimensional vector with element values between 0 and 1 is output, if the output vector is C = {0 (i.e. C) 1 = 0), 0.8 (i.e., C) 2 =0.8, and so on thereafter), 0.9,0,0,0,0,0,0,0,0,0,0,0,0,0,0}.
According to C = [ hollow water, tree roots, accumulated mud, sundries, scaling, plugging, slight corrosion, moderate corrosion, severe corrosion, disjointing, cracking, staggered opening, deformation, invasion, slight leakage, moderate leakage and severe leakage ], the represented pipeline endoscopic image has tree root diseases at 80% probability and accumulated mud diseases at 90% probability.
The original inclusion-ResNet-v 2 network structure described in the present invention is the prior art, and the description of each step according to the prior art can be mainly divided into an input layer, a Stem initial convolution layer, a 5 × inclusion × ResNet-a residual layer, a Reduction-a Reduction layer, a 10 × inclusion × ResNet-a residual layer, a Reduction-B Reduction layer, a 5 × inclusion × ResNet-C residual layer, a global average pooling layer, a random deactivation layer, and a Softmax classifier, where the upper layer described in the present invention includes the input layer, the Stem initial convolution layer, the 5 × inclusion × ResNet-a residual layer, the Reduction-a Reduction layer, the 10 × inclusion × ResNet-a residual layer, the Reduction-B Reduction layer, the 5 × inclusion × res-C residual layer, the global average pooling layer, and the random deactivation layer, and the specific actions and relationships of each layer are not described in the prior art one by one, and the present invention is not described any more.
As shown in fig. 2, wherein the X layer is the output vector of the random deactivation layer in the inclusion-ResNet-v 2 network structure shown in fig. 2.
The random inactivation layer of the original inclusion-ResNet-v 2 network structure only outputs the information of the former 1792 dimension, and the invention adds the time information of one dimension on the basis.
The process of converting the X layer into the H layer is a forward calculation process of the fully-connected neural network, and aims to properly reduce the dimension of the 1792+ 1-dimensional vector into 1024 dimensions. For convenience of description, several symbols are appropriately added to the figure, wherein X 1 ~X 1793 Represents the output of the original Incepration-ResNet-v 2 random deactivation layer plus one dimension newly added in the text; h 1 ~X 1024 Representing the dimension values of 1793-dimensional vector into 1024-dimensional vector after H-layer activation, C 1 ~C 17 Representing the 17-dimensional unique hot code obtained after C-layer processing; w x->H Is a weight vector, which represents the connecting line connecting the X layer and the H layer, and the total number of dimensions is 1973 × 1024, because all the dimensions of the X layer are connected with all the dimensions of the H layer (so called full connection), the calculation process is as follows:
with H in H layer 1 For example, as shown in equation 2,
wherein W xi->H1 Indicates that the ith element of the X layer is connected with the H layer H 1 And sigmoid is a function that can map a real number to a number between 0 and 1, the expression of the function beingThe image is S-shaped, and the value range is between 0 and 1. H 2 ~H 1024 Is calculated in a similar manner to equation 1.
H-layer to C-layer transitionUsing equation 3, as C, similar to the above calculation method 1 The calculation process of (a) is taken as an example,
all C are calculated i According to the characteristics of sigmoid, at the C layer, a 17-dimensional vector with element values between 0 and 1 is output, if the output vector is a vector
C = {0 (i.e., C) 1 = 0), 0.8 (i.e., C) 2 =0.8, and so on) 0.9,0,0,0,0,0,0,0,0,0,0,0,0,0,0}
According to C = [ hollow water, tree roots, accumulated mud, sundries, scaling, plugging, slight corrosion, moderate corrosion, severe corrosion, disjointing, cracking, staggered opening, deformation, invasion, slight leakage, moderate leakage and severe leakage ], the represented pipeline endoscopic image has tree root diseases at 80% probability and accumulated mud diseases at 90% probability.
Although the invention has been described above with reference to various embodiments, it should be understood that many changes and modifications may be made without departing from the scope of the invention. That is, the methods, systems, and devices discussed above are examples. Various configurations may omit, substitute, or add various procedures or components as appropriate. For example, in alternative configurations, the methods may be performed in an order different than described, and/or various components may be added, omitted, and/or combined. Moreover, features described with respect to certain configurations may be combined in various other configurations, as different aspects and elements of the configurations may be combined in a similar manner. Further, elements therein may be updated as technology evolves, i.e., many elements are examples and do not limit the scope of the disclosure or claims.
Specific details are given in the description to provide a thorough understanding of the exemplary configurations including implementations. However, configurations may be practiced without these specific details, for example, well-known circuits, processes, algorithms, structures, and techniques have been shown without unnecessary detail in order to avoid obscuring the configurations. This description provides example configurations only, and does not limit the scope, applicability, or configuration of the claims. Rather, the foregoing description of the configurations will provide those skilled in the art with an enabling description for implementing the described techniques. Various changes may be made in the function and arrangement of elements without departing from the spirit or scope of the disclosure.
In conclusion, it is intended that the foregoing detailed description be regarded as illustrative rather than limiting, and that it be understood that these examples are illustrative only and are not intended to limit the scope of the invention. After reading the description of the invention, the skilled person can make various changes or modifications to the invention, and these equivalent changes and modifications also fall into the scope of the invention defined by the claims.
Claims (3)
1. A pipeline disease image classification method based on a multi-label convolutional neural network is characterized by comprising the following steps:
step 1: collecting a pipeline endoscopic detection video, and extracting image frames in the pipeline endoscopic detection video;
step 2: calculating the time stamp characteristics of each image;
and step 3: sending a part of the image frames collected in the step 1 into a multi-label convolutional neural network model for training to obtain the multi-label convolutional neural network model capable of correctly classifying the types of the pipeline diseases;
and 4, step 4: detecting an endoscopic image of a pipeline to be detected by using a trained multi-label convolutional neural network model, outputting a single-hot code by using the multi-label convolutional neural network model, and determining the type of the existing pipeline diseases according to the single-hot code;
the multi-label convolutional neural network model comprises an upper-layer inclusion-ResNet-v 2 network structure and a lower-layer multi-label classification layer, wherein the upper-layer inclusion-ResNet-v 2 network structure further comprises a random inactivation layer;
the processing steps of the lower multi-label classification layer are as follows:
the first step is as follows: adding a timestamp feature Timefeature in a feature vector output by a random inactivation layer in an upper-layer inclusion-ResNet-v 2 network structure;
the second step: performing dimension reduction activation processing on the feature vector added with the timestamp feature in the first step to obtain a medium latitude value;
and thirdly, continuing to perform dimensionality reduction on the medium latitude value in the second step, and finally outputting the one-hot code.
2. The method for classifying the pipeline disease images based on the multi-tag convolutional neural network as claimed in claim 1, wherein the calculation formula of the timestamp feature is as follows:
3. The method for classifying the pipeline disease images based on the multi-label convolutional neural network as claimed in claim 1, wherein the pipeline disease types comprise: swabbing, tree roots, accumulated mud, debris, scaling, plugging, mild corrosion, moderate corrosion, severe corrosion, dislocation, fracture, dislocation, deformation, intrusion, mild leakage, moderate leakage, and severe leakage.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910565360.9A CN110349134B (en) | 2019-06-27 | 2019-06-27 | Pipeline disease image classification method based on multi-label convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910565360.9A CN110349134B (en) | 2019-06-27 | 2019-06-27 | Pipeline disease image classification method based on multi-label convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110349134A CN110349134A (en) | 2019-10-18 |
CN110349134B true CN110349134B (en) | 2022-12-09 |
Family
ID=68183348
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910565360.9A Active CN110349134B (en) | 2019-06-27 | 2019-06-27 | Pipeline disease image classification method based on multi-label convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110349134B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111209794A (en) * | 2019-12-11 | 2020-05-29 | 浙江省交通运输科学研究院 | Underground pipeline identification method based on ground penetrating radar image |
CN111191593A (en) * | 2019-12-30 | 2020-05-22 | 成都云尚物联环境科技有限公司 | Image target detection method and device, storage medium and sewage pipeline detection device |
CN113963212A (en) * | 2021-10-25 | 2022-01-21 | 郑州大学 | Pipeline disease image classification method and device based on increment-Resnet neural network |
CN114549402B (en) * | 2022-01-05 | 2024-08-06 | 江苏海洋大学 | Underwater image quality comparison method without reference image |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108830287A (en) * | 2018-04-18 | 2018-11-16 | 哈尔滨理工大学 | The Chinese image, semantic of Inception network integration multilayer GRU based on residual error connection describes method |
CN108898595A (en) * | 2018-06-27 | 2018-11-27 | 慧影医疗科技(北京)有限公司 | A kind of construction method of thoracopathy detection model and application |
CN109716108A (en) * | 2016-12-30 | 2019-05-03 | 同济大学 | A kind of Asphalt Pavement Damage detection system based on binocular image analysis |
CN109800824A (en) * | 2019-02-25 | 2019-05-24 | 中国矿业大学(北京) | A kind of defect of pipeline recognition methods based on computer vision and machine learning |
CN109871906A (en) * | 2019-03-15 | 2019-06-11 | 西安获德图像技术有限公司 | A kind of classification method of the coil open defect based on depth convolutional neural networks |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7215811B2 (en) * | 2000-11-22 | 2007-05-08 | Osama Moselhi | Method and apparatus for the automated detection and classification of defects in sewer pipes |
-
2019
- 2019-06-27 CN CN201910565360.9A patent/CN110349134B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109716108A (en) * | 2016-12-30 | 2019-05-03 | 同济大学 | A kind of Asphalt Pavement Damage detection system based on binocular image analysis |
CN108830287A (en) * | 2018-04-18 | 2018-11-16 | 哈尔滨理工大学 | The Chinese image, semantic of Inception network integration multilayer GRU based on residual error connection describes method |
CN108898595A (en) * | 2018-06-27 | 2018-11-27 | 慧影医疗科技(北京)有限公司 | A kind of construction method of thoracopathy detection model and application |
CN109800824A (en) * | 2019-02-25 | 2019-05-24 | 中国矿业大学(北京) | A kind of defect of pipeline recognition methods based on computer vision and machine learning |
CN109871906A (en) * | 2019-03-15 | 2019-06-11 | 西安获德图像技术有限公司 | A kind of classification method of the coil open defect based on depth convolutional neural networks |
Non-Patent Citations (4)
Title |
---|
Automated detection of sewer pipe defects in closed-circuit television images using deep learning techniques;Jack C.P. Cheng,et al;《Automation in Construction 95 (2018)》;20180827;期刊第155-171页 * |
基于卷积神经网络的多标签图像分类_-;苏越;《中国优秀硕士论文全文数据库(信息科技辑)》;20190115;全文 * |
基于卷积神经网络的多标签场景分类;陈智;《中国优秀硕士论文全文数据库(信息科技辑)》;20160215;全文 * |
基于深度学习的图像分类关键技术研究;杜珂;《中国优秀硕士论文全文数据库(信息科技辑)》;20190115;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN110349134A (en) | 2019-10-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110349134B (en) | Pipeline disease image classification method based on multi-label convolutional neural network | |
CN109961019B (en) | Space-time behavior detection method | |
JP6759475B2 (en) | Ship detection methods and systems based on multidimensional features of the scene | |
CN113936339B (en) | Fighting identification method and device based on double-channel cross attention mechanism | |
CN112733950A (en) | Power equipment fault diagnosis method based on combination of image fusion and target detection | |
CN113139489B (en) | Crowd counting method and system based on background extraction and multi-scale fusion network | |
CN112561876A (en) | Image-based pond and reservoir water quality detection method and system | |
CN116342894B (en) | GIS infrared feature recognition system and method based on improved YOLOv5 | |
CN112967227B (en) | Automatic diabetic retinopathy evaluation system based on focus perception modeling | |
Zhou et al. | Deep multi-scale features learning for distorted image quality assessment | |
CN114529839A (en) | Unmanned aerial vehicle routing inspection-oriented power transmission line hardware anomaly detection method and system | |
CN115497015A (en) | River floating pollutant identification method based on convolutional neural network | |
CN116386081A (en) | Pedestrian detection method and system based on multi-mode images | |
CN115170985B (en) | Remote sensing image semantic segmentation network and segmentation method based on threshold attention | |
CN110599458A (en) | Underground pipe network detection and evaluation cloud system based on convolutional neural network | |
CN115810123A (en) | Small target pest detection method based on attention mechanism and improved feature fusion | |
CN115861210A (en) | Transformer substation equipment abnormity detection method and system based on twin network | |
CN111767826A (en) | Timing fixed-point scene abnormity detection method | |
CN117671597B (en) | Method for constructing mouse detection model and mouse detection method and device | |
CN110827375A (en) | Infrared image true color coloring method and system based on low-light-level image | |
CN117392668A (en) | Wheat scab state evaluation method and system and electronic equipment | |
CN111881803A (en) | Livestock face recognition method based on improved YOLOv3 | |
CN116823638A (en) | Image defogging method based on improved GAN | |
CN110599460A (en) | Underground pipe network detection and evaluation cloud system based on hybrid convolutional neural network | |
CN115393802A (en) | Railway scene unusual invasion target identification method based on small sample learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |