CN115995056A - Automatic bridge disease identification method based on deep learning - Google Patents

Automatic bridge disease identification method based on deep learning Download PDF

Info

Publication number
CN115995056A
CN115995056A CN202310281075.0A CN202310281075A CN115995056A CN 115995056 A CN115995056 A CN 115995056A CN 202310281075 A CN202310281075 A CN 202310281075A CN 115995056 A CN115995056 A CN 115995056A
Authority
CN
China
Prior art keywords
bridge
representing
training
disease
loss
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310281075.0A
Other languages
Chinese (zh)
Inventor
章羚
王铁鑫
岳大川
于宪政
温晓光
王京杭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Aeronautics and Astronautics
Original Assignee
Nanjing University of Aeronautics and Astronautics
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Aeronautics and Astronautics filed Critical Nanjing University of Aeronautics and Astronautics
Priority to CN202310281075.0A priority Critical patent/CN115995056A/en
Publication of CN115995056A publication Critical patent/CN115995056A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses an automatic bridge defect recognition method based on deep learning, which is characterized by collecting defect videos of related bridge parts, performing screenshot operation on the videos, marking the positions of bridge defects and the types of the defects in pictures, and thus building a training data set; training the bridge diseases according to the training data set to obtain a deep neural network model; and inputting the bridge disease video into the optimized YOLOv5 neural network model for automatic identification, outputting a disease identification and classification result and a disease positioning result, and generating a corresponding XML file. According to the invention, the automatic mode is used for identifying and marking the bridge diseases instead of the manual mode, so that the efficiency of identifying and marking the bridge diseases is effectively improved, the labor cost is reduced, and the output result can be applied to automatic generation training of a large-scale bridge disease data set.

Description

Automatic bridge disease identification method based on deep learning
Technical Field
The invention belongs to the technical field of bridge detection, and particularly relates to an automatic bridge disease identification method based on deep learning.
Background
According to the development statistics publication of the transportation industry in 2021 issued by the China department of transportation, the total amount of national highway bridges reaches 96.11 ten thousand by the end of 2021, and the total length reaches 7380.21 ten thousand of linear meters. Bridges have become an integral part of transportation systems, playing a vital role in national economy operations, all over the world. However, along with the extension of the service life of the bridge, related diseases often exist in bridge engineering under the influence of environment and load, and a plurality of potential safety hazards are caused to the bridge. Once the bridge is in accident, great economic loss is caused, and the life safety of people is even endangered. The bridge is detected regularly at the present stage, the diseases are found early, and the prevention and the control are carried out timely, so that the method is an effective measure for transforming the dangerous old bridge. The daily inspection of bridge diseases is also a correct way for ensuring the safe passing of the bridge and prolonging the service life of the bridge.
The continuous increase of the total length of the bridge and the continuous extension of the service life thereof lead to various diseases of the bridge. These diseases can cause certain potential safety hazards, and the service life of the bridge is reduced, such as cracking of bridge support concrete, surface weathering, exposed reinforcing steel bars and the like. The high-risk conditions for bridge inspection and the enormous cost of bridge maintenance indicate that efficient bridge inspection and assessment is necessary. Before deep learning techniques are widely adopted, most of the detection utilized conventional computer vision methods based on image processing. For example, edge detection methods (e.g., sobel operator, canny edge detection, LOG filter operator) can be used to detect edges between the lesion and the surrounding background, by setting a corresponding static threshold for a specific case, and determining whether there is a relevant lesion in the region of interest in the image based on the threshold, but the method is not applicable to dynamic and complex scenes. To solve this limitation, the acquired images may be detected using a computer vision technique based on deep learning, helping inspectors determine the kind and location of bridge diseases, thereby solving a large number of manpower and improving disease evaluation efficiency. However, the deep learning method needs a large number of data sets, and the current data set production mainly depends on manual work, so that the accuracy is low, the cost is high, the detection work is complex, and the situation that the data sets are difficult to acquire is caused.
Disclosure of Invention
The invention aims to: the invention aims to provide an automatic bridge disease recognition method based on deep learning, which is used for carrying out video recording on the appearance of a bridge through an unmanned aerial vehicle, accurately recognizing the bridge disease condition by combining a computer vision technology, realizing automatic classification of bridge disease categories, outputting an XML file of a classification result, and being capable of being used for subsequent model optimization training.
The technical scheme is as follows: the invention provides an automatic bridge disease identification method based on deep learning, which comprises the following steps:
(1) Preprocessing a bridge disease video obtained in advance to obtain the disease position and the disease type of the bridge, and constructing a data set to obtain a training set and a verification set;
(2) The attention introducing mechanism module optimizes the Darknet-53 network structure and the loss function in the YOLOv5 algorithm and adjusts the data parameters of the Darknet-53 network structure and the loss function; training the bridge disease identification model in batches by weight parameter data in the frozen neural network;
(3) Observing the convergence condition of the training model, and judging whether the loss function is over-fitted or not and whether the loss function meets the training standard or not; if the training standard is met, directly outputting a standard bridge disease recognition model, if the training standard is not met, finding a weight function meeting a preset standard, thawing the data set, performing refined training again, and outputting the standard bridge disease recognition model after the weight function meeting the standard is obtained;
(4) And rapidly identifying the damage condition in the bridge video by using a standard bridge damage identification model, obtaining the bridge damage category and the probability of possibility thereof, and outputting and storing the result as an XML file for training.
Further, the implementation process of the step (1) is as follows:
performing frame cutting operation on the bridge defect video to obtain image data, performing smooth processing, expanding and unifying the format and the size of the image; marking the processed picture data set, and dividing the processed picture data set into two data sets with diseases and without diseases; refining a dataset having a disease, the disease comprising: cracking, peeling, weathering, exposing the steel bars and rusting; the data set is divided into a training set and a validation set.
Further, the YOLOv5 in the step (2) includes a network dark-53 in the convolutional neural network CNN, and a network inclusion batch standardization and LeakyReLU activation function layer.
Further, the implementation process of the attention introducing mechanism module in the step (2) for optimizing the dark-53 network structure in YOLOv5 is as follows:
the attention mechanism module comprises a channel attention part and a space attention part, and the attention mechanism is led between the input connector and the output connector to lead the network model training to be focused on a target area; wherein the channel attention formula is:
Figure SMS_1
wherein ,
Figure SMS_2
activating a function for Sigmoid->
Figure SMS_3
Represented as a multi-layer perceptron->
Figure SMS_4
and />
Figure SMS_5
Separate tableShown as average pooling and maximum pooling operation, < >>
Figure SMS_6
For input features->
Figure SMS_7
Representing a channel attention mechanism obtained after parallel pooling operation and input of a multi-layer perceptron;
the space attention part performs maximum pooling and average pooling splicing on the output result obtained by the channel attention, and continues to perform convolution operation, and after the dimension of the output obtained by the channel attention is reduced to a single channel, the final output result is obtained through activation; the formula of spatial attention is:
Figure SMS_8
Figure SMS_9
convolution operation representing convolution kernel 7*7, < >>
Figure SMS_10
Characteristic results of the spatial attention mechanism after convolution and activation function processing pooling operation are represented.
Further, the loss function in the step (2) is:
Figure SMS_11
Figure SMS_12
Figure SMS_13
wherein :
Figure SMS_14
representing the weight of the coordinate loss to the total loss; />
Figure SMS_15
A weight representing the total loss of category loss;
Figure SMS_16
a weight indicating total loss of the station without the detection target confidence loss;
Figure SMS_17
representing the weight of the confidence loss containing the detection target accounting for the total loss;
Figure SMS_18
whether the jth anchor frame of the ith grid contains the detection target or not is 1, and is not 0;
Figure SMS_19
indicating whether the jth anchor frame of the ith grid contains the detection target, and is not 1, but is 0;
Figure SMS_20
an abscissa value representing the center coordinates of the real target frame;
Figure SMS_21
a ordinate value representing the center coordinates of the real target frame;
Figure SMS_22
representing the width of a real target frame;
Figure SMS_23
representing the height of a real target frame;
Figure SMS_24
representation ofConfidence of the target frame;
Figure SMS_25
a category indicating a detection target;
Figure SMS_26
an abscissa value representing a center coordinate of the predicted target frame;
Figure SMS_27
a ordinate value representing the center coordinates of the predicted target frame;
Figure SMS_28
representing the width of the predicted target frame;
Figure SMS_29
representing the height of the predicted target frame;
Figure SMS_30
representing the confidence of the predicted target frame;
Figure SMS_31
representing the category of the predicted target.
Compared with the prior art, the invention has the beneficial effects that:
according to the method, the bridge diseases are identified by optimizing the You Only Look Once version5 YOLOv5 algorithm, the identification accuracy is superior to that of the common YOLOv5 algorithm, the identification speed is high, the detection time can be greatly shortened, and the automatic marking efficiency is improved; the detection method of the invention not only can identify the position information of the bridge diseases, but also can judge the disease condition of the bridge and mark the disease probability, shortens the manual marking time and obviously improves the generation speed of the disease data set; the network model obtained by training in the invention automatically marks the disease position and disease category in the video, can be used for automatically generating the bridge disease picture data set, reduces the labor cost and improves the detection efficiency.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a schematic diagram of a hybrid attention mechanism module architecture;
FIG. 3 is a schematic view of a channel attention structure;
fig. 4 is a schematic view of a spatial attention structure.
Description of the embodiments
The invention is described in further detail below with reference to the accompanying drawings.
The invention provides an automatic bridge disease recognition method based on deep learning, which comprises the steps of acquiring data through flying around a bridge by an unmanned aerial vehicle, and storing acquired positioning information and video information of bridge appearance conditions into a saving card through a data processing unit; processing the collected video information frame by frame to obtain image data, and preprocessing the picture information by expanding, adjusting the size, smoothing and filtering; marking the processed pictures into a data set to obtain a training set and a verification set; optimizing a YOLOv5 algorithm, performing optimization test on network model parameters, extracting picture characteristic values, and training a bridge disease identification model by freezing a part of training set; judging whether the obtained training model converges or not, whether a loss function is fitted or not, and continuing to train the model by thawing the data set optimization parameters; verifying a final training model, calculating the intersection ratio (Intersection of Union, ioU) of the bridge disease area and the actual marking area obtained by the model, drawing an average value (mean Average Precision, mAP) curve of each class of precision values, and further analyzing whether the model meets the condition; and identifying the bridge diseases by using the trained model, and outputting an XML file. The XML file output by the model can be used for manufacturing a data set in the aspect of bridge detection, the bridge disease condition can be detected, and the XML file capable of being trained can be output, so that the function of automatic marking of the bridge disease is realized, the picture data set in the bridge field can be automatically generated, the detection efficiency is improved, and the labor cost is saved. As shown in fig. 1, the method specifically comprises the following steps:
step 1: preprocessing a bridge disease video obtained in advance, obtaining the disease position and the disease type of the bridge, and constructing a data set.
The bridge unmanned aerial vehicle flies around the bridge, the micro inertial measurement unit in the positioning module collects positioning information, the data processing unit processes the collected measured values, the camera module collects video information of the bridge component, and the positioning information and the video information are stored through the data storage unit.
And collecting video of related components of the bridge, extracting fault frames of the video, obtaining image data, expanding the image data and performing image smoothing treatment. And extracting the stored video data, converting the video data into RGB picture data through a written Python program, and performing filtering smoothing processing to complete the processing of a picture data set.
Taking a frame of video signal K seconds to obtain a plurality of RGB picture data sets, manually screening to obtain a picture with defects, and smoothing noise points in the picture through Gaussian filtering; the gaussian filtering one-dimensional formula is:
Figure SMS_32
respectively carrying out one-dimensional smoothing on RGB three channels of the picture, wherein
Figure SMS_33
Is the mean value of x>
Figure SMS_34
Is the variance of x, because the center point is the origin when the mean is calculated, therefore +.>
Figure SMS_35
Equal to 0;
and marking a real frame of the processed picture data set by using labelImg software in a python environment, selecting a frame and marking a disease area of the picture, storing the frame as an XML file, taking 80% of the marked data set as a training set and the remaining 20% as a verification set.
For example: the bridge Crack defect is marked as 'Crack', the peeling mark is marked as 'spalling', the weathering mark is marked as 'Efflorescence', the exposed reinforcing steel bar is marked as 'Exposedbars', the corrosion mark is marked as 'corrosingain', and a corresponding XML file is obtained after each picture is marked; the main parameters contained in the XML file are described as follows: size: picture size, object: objects contained in a picture, (one picture may contain a plurality of objects, in this method, bridge diseases), name: the label class (in this method, there is a mask, spallation, efficiency, exposedBars, corrosionStain), bndbox: object real frame, difficult: whether it is difficult to identify.
Step 2: the Darknet-53 network structure and the loss function in YOLOv5 are optimized through the attention introducing mechanism module, data parameters of the Darknet-53 network structure and the loss function are optimized, and a part of training set is frozen for training the bridge disease identification model in batches.
The YOLOv5 framework contains network dark-53 in convolutional neural network CNN (Convolutional Neural Network), which contains batch normalization (Batch Normalization) and the LeakyReLU activation function layer.
The network model training loss function adopts a mean square error (Mean Square Error, MSE) loss function aiming at target frame coordinate regression, and the category and the confidence coefficient adopt cross entropy loss functions, wherein the specific forms are as follows:
Figure SMS_36
Figure SMS_37
Figure SMS_38
wherein :
Figure SMS_39
representing the weight of the coordinate loss to the total loss;
Figure SMS_40
a weight representing the total loss of category loss;
Figure SMS_41
a weight indicating total loss of the station without the detection target confidence loss;
Figure SMS_42
representing the weight of the confidence loss containing the detection target accounting for the total loss;
Figure SMS_43
whether the jth anchor frame of the ith grid contains the detection target or not is 1, and is not 0;
Figure SMS_44
indicating whether the jth anchor frame of the ith grid contains the detection target, and is not 1, but is 0;
Figure SMS_45
an abscissa value representing the center coordinates of the real target frame;
Figure SMS_46
a ordinate value representing the center coordinates of the real target frame;
Figure SMS_47
representing the width of a real target frame;
Figure SMS_48
representing the height of a real target frame;
Figure SMS_49
representing the confidence level of the target frame;
Figure SMS_50
a category indicating a detection target;
Figure SMS_51
an abscissa value representing a center coordinate of the predicted target frame;
Figure SMS_52
a ordinate value representing the center coordinates of the predicted target frame;
Figure SMS_53
representing the width of the predicted target frame;
Figure SMS_54
representing the height of the predicted target frame;
Figure SMS_55
representing the confidence of the predicted target frame; />
Figure SMS_56
Representing the category of the predicted target.
The loss function of the whole network is:
Figure SMS_57
wherein ,
Figure SMS_61
, />
Figure SMS_63
, />
Figure SMS_70
, />
Figure SMS_59
respectively represent the coordinate lossesCategory loss, no confidence loss, and weight with confidence loss to total loss; whether the j anchor frame of the i-th grid contains the detection target, namely bridge diseases; (/>
Figure SMS_69
,/>
Figure SMS_62
, />
Figure SMS_71
, />
Figure SMS_60
, />
Figure SMS_68
, />
Figure SMS_58
) The center coordinates and width and height, confidence and category of the real target frame are respectively. And (/ ->
Figure SMS_67
, />
Figure SMS_65
, />
Figure SMS_73
, />
Figure SMS_64
, />
Figure SMS_72
, />
Figure SMS_66
) The central coordinates, width and height, confidence and category of the target frame are predicted.
IoU is the most common indicator in target detection, and measures the similarity between a predicted frame and a target frame by encoding the shape properties (e.g., width, height, position) of the target into normalized metrics, so it has scale invariance. The calculation process is as follows:
Figure SMS_74
wherein :
Figure SMS_75
: is the area of the prediction box.
Figure SMS_76
: is the area of the target frame.
The DarkNet-53 network of the network model has 53 convolution layers, the last layer is 1×1 convolution to realize full connection, 52 convolutions coexist in the main network, wherein the first layer is convolved by a filter consisting of 32 3×3 convolution kernels, the later convolution layer is formed by 5 groups of repeated residual units, and the candidate region of the object (defect) is obtained and the category and the position of the candidate region are marked.
The mixed domain attention mechanism CBAM is introduced to optimize the YOLOv5 model, and as shown in fig. 2, the attention module comprises a channel attention part and a space attention part, and the attention mechanism is introduced between the input connector and the output connector to enable the network model training to be more focused on a target area, so that a better training effect is achieved. Wherein the channel attention is as shown in fig. 3, the formula is:
Figure SMS_77
wherein ,
Figure SMS_78
activating a function for Sigmoid->
Figure SMS_79
Represented as a multi-layer perceptron->
Figure SMS_80
and />
Figure SMS_81
Denoted mean pooling and maximum pooling operations, respectively,>
Figure SMS_82
for input features->
Figure SMS_83
Representing the channel attention mechanism obtained after parallel pooling operation and input to the multi-layer perceptron.
And the spatial attention part is shown in fig. 4, the output result obtained by the channel attention is subjected to maximum pooling and average pooling splicing, convolution operation is continued, the dimension of the output result is reduced to a single channel, and the final output result is obtained through an activation function. The formula of spatial attention is:
Figure SMS_84
wherein :
Figure SMS_85
for the features obtained after the channel attention mechanism, < +.>
Figure SMS_86
and />
Figure SMS_87
Denoted mean pooling and maximum pooling operations, respectively,>
Figure SMS_88
convolution operation representing convolution kernel 7*7, < >>
Figure SMS_89
Activating a function for Sigmoid->
Figure SMS_90
Characteristic results of the spatial attention mechanism after convolution and activation function processing pooling operation are represented.
The network model which is trained by the attention module is added to have higher recognition accuracy.
In the training process of the network model, the initial learning rate is set to be 0.01, the batch-size is set to be 16, a program starts to run after a part of data sets are frozen, and the YOLOv5 model is automatically adjusted to the learning rate in the training process to keep the model in an efficient learning state.
Step 3: observing the convergence condition of the training model, and judging whether the loss function is over-fitted or not and whether the loss function meets the training standard or not; the training standard is to look at the decline condition of the loss function, usually after a certain number of rounds of training, the loss function result is stabilized at a numerical value, the training is less and the fitting is insufficient, the training is more and the fitting is excessive, and the result can be seen through a graph drawn in a training log of the model.
If the training standard is met, directly outputting the model, and if the training standard is not met, finding a more optimized weight function, thawing the data set, performing refinement training again, and outputting the model after the weight function meeting the standard is obtained. In the situation of bridge diseases, the loss function is approximately trained to be about 10, so that the loss function can meet the training standard.
Since there are 5 downsampling processes with a stride of 2 x 2 in dark-53 in the YOLOv5 model, the connector map is scaled down 32 times, so the image size is reprocessed to a multiple of 32. In order to ensure the relative accuracy of model training, inhibit overfitting and improve the generalization capability of the model, a large number of picture data sets should be provided to enable the network model to learn, so that the existing picture data sets are expanded. The method comprises the steps of programming a program to enable a picture to be turned left and right and up and down, and then randomly deleting and adding pixels, adjusting contrast, adjusting brightness and the like to generate similar but not identical samples so as to enlarge a training data set.
Observing whether the numerical value of the loss function meets the requirement, and if the numerical value is stable to about 10 and floats up and down, namely meeting the condition required by the method; if the values float up and down at values greater than 10, the program is terminated prematurely to prevent over-fitting of the model learning process.
The early termination program stores the weight function obtained when the current training is carried out for a period of time, modifies the batch-size into 8, defreezes the frozen data set, finds the optimal value in the stored weight function, and continues the training of the model on the basis after the parameters are adjusted; and finally, storing the final weight function obtained by training, and obtaining a final training result.
And testing the model, obtaining an mAP curve by using the verification set, and observing whether the model meets the requirement. The threshold value is set to 0.5, and is marked as a positive sample when the measured IoU value is greater than the threshold value, and is marked as a negative sample otherwise. Calculate the precision and Recall of each prediction, and calculate AP, mAP:
Figure SMS_91
where p is precision, r is the recovery value corresponding to the precision at this time, and n is the total number of categories. Specifically, there are the following concepts: positive samples (TP) labeled positive samples, positive samples (FP) labeled negative samples, negative samples (TN) labeled negative samples, negative samples (FN) labeled positive samples, and precision and recall (recall); the average precision value (AP) of one type of label can be obtained through calculation through precision and recall rate, the higher the AP value is, the higher the probability of successful identification of the label target is, mAP is the average value of all types of label precision values of the whole data set, the comprehensive accuracy of the network model for identifying the whole target can be judged, and the method is also a judging standard for judging whether the network model is good in identification effect or not.
Step 4: and rapidly identifying bridge disease conditions in the bridge video by using the model, obtaining bridge disease categories and probability, and outputting and storing the results as XML files for training.
And identifying the bridge disease video by using the trained network model, marking the disease position and the disease category in the video, and outputting the result to an XML file. The model reads the video information of the bridge component detected by the bridge unmanned plane, automatically marks the area part of the bridge Liang Binghai in the video, marks the prediction probability of the prediction frame, and stores the result as an XML file. The result output by the model can be used for automatically marking bridge diseases in the video, and the XML file of the marked result can be used for training other models.
The foregoing is merely a preferred embodiment of the method of the present invention, and is not intended to limit the invention in any other way, but the technology according to the present invention can be changed in the specific embodiments and application scope, and is directly or indirectly applied to other technical fields, which are all the same as the protection scope of the patent of the present invention.

Claims (5)

1. The automatic bridge disease identification method based on deep learning is characterized by comprising the following steps of:
(1) Preprocessing a bridge disease video obtained in advance to obtain the disease position and the disease type of the bridge, and constructing a data set to obtain a training set and a verification set;
(2) The attention introducing mechanism module optimizes the Darknet-53 network structure and the loss function in the YOLOv5 algorithm, and optimizes the data parameters of the YOLOv5 algorithm; freezing the weight parameter data in the Darknet-53 network, and training the bridge disease identification model in batches;
(3) Observing the convergence condition of the model, and judging whether the loss function is over-fitted or not and whether the loss function meets the training standard or not; if the training standard is met, directly outputting a standard bridge disease recognition model, if the training standard is not met, finding a weight function meeting a preset standard, thawing the data set, performing refined training again, and outputting the standard bridge disease recognition model after the weight function meeting the standard is obtained;
(4) And rapidly identifying the damage condition in the bridge video according to the standard bridge damage identification model, obtaining the bridge damage category and the probability of possibility thereof, and outputting and storing the result as an XML file for training.
2. The automatic bridge defect recognition method based on deep learning of claim 1, wherein the implementation process of the step (1) is as follows:
performing frame cutting operation on the bridge defect video to obtain image data, performing smooth processing, expanding and unifying the format and the size of the image; marking the processed picture data set, and dividing the processed picture data set into two data sets with diseases and without diseases; refining a dataset having a disease, the disease comprising: cracking, peeling, weathering, exposing the steel bars and rusting; the data set is divided into a training set and a validation set.
3. The automatic bridge defect recognition method based on deep learning according to claim 1, wherein the YOLOv5 in the step (2) comprises a dark-53 network in a convolutional neural network CNN, and the network comprises batch standardization and a LeakyReLU activation function layer.
4. The automatic bridge defect recognition method based on deep learning according to claim 1, wherein the implementation process of optimizing the dark net-53 network structure in YOLOv5 by the attention introducing mechanism module in step (2) is as follows:
the attention mechanism module comprises a channel attention part and a space attention part, and the attention mechanism is led between the input connector and the output connector to lead the network model training to be focused on a target area; wherein the channel attention formula is:
Figure QLYQS_1
wherein ,
Figure QLYQS_2
activating a function for Sigmoid->
Figure QLYQS_3
Represented as a multi-layer perceptron->
Figure QLYQS_4
and />
Figure QLYQS_5
Denoted mean pooling and maximum pooling operations, respectively,>
Figure QLYQS_6
for input features->
Figure QLYQS_7
Representing a channel attention mechanism obtained after parallel pooling operation and input of a multi-layer perceptron;
the space attention part performs maximum pooling and average pooling splicing on the output result obtained by the channel attention, and continues to perform convolution operation, and after the dimension of the output obtained by the channel attention is reduced to a single channel, the final output result is obtained through activation; the formula of spatial attention is:
Figure QLYQS_8
Figure QLYQS_9
convolution operation representing convolution kernel 7*7, < >>
Figure QLYQS_10
Characteristic results of the spatial attention mechanism after convolution and activation function processing pooling operation are represented.
5. The automatic bridge defect recognition method based on deep learning of claim 4, wherein the loss function in the step (2) is:
Figure QLYQS_11
Figure QLYQS_12
Figure QLYQS_13
wherein :
Figure QLYQS_14
representing the weight of the coordinate loss to the total loss;
Figure QLYQS_15
a weight representing the total loss of category loss;
Figure QLYQS_16
a weight indicating total loss of the station without the detection target confidence loss;
Figure QLYQS_17
representing the weight of the confidence loss containing the detection target accounting for the total loss;
Figure QLYQS_18
whether the jth anchor frame of the ith grid contains the detection target or not is 1, and is not 0;
Figure QLYQS_19
indicating whether the jth anchor frame of the ith grid contains the detection target, and is not 1, but is 0;
Figure QLYQS_20
an abscissa value representing the center coordinates of the real target frame;
Figure QLYQS_21
a ordinate value representing the center coordinates of the real target frame;
Figure QLYQS_22
representing the width of a real target frame;
Figure QLYQS_23
representing the height of a real target frame;
Figure QLYQS_24
representing the confidence level of the target frame;
Figure QLYQS_25
a category indicating a detection target;
Figure QLYQS_26
an abscissa value representing a center coordinate of the predicted target frame;
Figure QLYQS_27
a ordinate value representing the center coordinates of the predicted target frame;
Figure QLYQS_28
representing the width of the predicted target frame;
Figure QLYQS_29
representing the height of the predicted target frame;
Figure QLYQS_30
representing the confidence of the predicted target frame;
Figure QLYQS_31
representing the category of the predicted target. />
CN202310281075.0A 2023-03-22 2023-03-22 Automatic bridge disease identification method based on deep learning Pending CN115995056A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310281075.0A CN115995056A (en) 2023-03-22 2023-03-22 Automatic bridge disease identification method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310281075.0A CN115995056A (en) 2023-03-22 2023-03-22 Automatic bridge disease identification method based on deep learning

Publications (1)

Publication Number Publication Date
CN115995056A true CN115995056A (en) 2023-04-21

Family

ID=85992320

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310281075.0A Pending CN115995056A (en) 2023-03-22 2023-03-22 Automatic bridge disease identification method based on deep learning

Country Status (1)

Country Link
CN (1) CN115995056A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116682072A (en) * 2023-08-04 2023-09-01 四川公路工程咨询监理有限公司 Bridge disease monitoring system
CN117058536A (en) * 2023-07-19 2023-11-14 中公高科养护科技股份有限公司 Pavement disease identification method, device and medium based on double-branch network
CN117058536B (en) * 2023-07-19 2024-04-30 中公高科养护科技股份有限公司 Pavement disease identification method, device and medium based on double-branch network

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110222701A (en) * 2019-06-11 2019-09-10 北京新桥技术发展有限公司 A kind of bridge defect automatic identifying method
CN112801972A (en) * 2021-01-25 2021-05-14 武汉理工大学 Bridge defect detection method, device, system and storage medium
CN115082650A (en) * 2022-06-14 2022-09-20 南京航空航天大学 Implementation method of automatic pipeline defect labeling tool based on convolutional neural network
WO2022245936A1 (en) * 2021-05-19 2022-11-24 Rutgers, The State University Of New Jersey Systems and methods for machine learning enhanced railway condition monitoring, assessment and prediction

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110222701A (en) * 2019-06-11 2019-09-10 北京新桥技术发展有限公司 A kind of bridge defect automatic identifying method
CN112801972A (en) * 2021-01-25 2021-05-14 武汉理工大学 Bridge defect detection method, device, system and storage medium
WO2022245936A1 (en) * 2021-05-19 2022-11-24 Rutgers, The State University Of New Jersey Systems and methods for machine learning enhanced railway condition monitoring, assessment and prediction
CN115082650A (en) * 2022-06-14 2022-09-20 南京航空航天大学 Implementation method of automatic pipeline defect labeling tool based on convolutional neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SONGYANG SUN 等: "YOLO Based Bridge Surface Defect Detection Using Decoupled Prediction", 《2022 7TH ASIA-PACIFIC CONFERENCE ON INTELLIGENT ROBOT SYSTEMS》 *
何金宝 等: "基于DCGAN和改进YOLOv5s的桥梁表面缺陷检测识别", 《江西师范大学学报(自然科学版 )》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117058536A (en) * 2023-07-19 2023-11-14 中公高科养护科技股份有限公司 Pavement disease identification method, device and medium based on double-branch network
CN117058536B (en) * 2023-07-19 2024-04-30 中公高科养护科技股份有限公司 Pavement disease identification method, device and medium based on double-branch network
CN116682072A (en) * 2023-08-04 2023-09-01 四川公路工程咨询监理有限公司 Bridge disease monitoring system
CN116682072B (en) * 2023-08-04 2023-10-20 四川公路工程咨询监理有限公司 Bridge disease monitoring system

Similar Documents

Publication Publication Date Title
CN110569837B (en) Method and device for optimizing damage detection result
CN111681240B (en) Bridge surface crack detection method based on YOLO v3 and attention mechanism
CN108257114A (en) A kind of transmission facility defect inspection method based on deep learning
CN111080620A (en) Road disease detection method based on deep learning
CN110992349A (en) Underground pipeline abnormity automatic positioning and identification method based on deep learning
CN113592828B (en) Nondestructive testing method and system based on industrial endoscope
CN114998852A (en) Intelligent detection method for road pavement diseases based on deep learning
CN112330593A (en) Building surface crack detection method based on deep learning network
CN110909657A (en) Method for identifying apparent tunnel disease image
CN112966665A (en) Pavement disease detection model training method and device and computer equipment
CN111144749A (en) Intelligent labeling crowdsourcing platform for power grid images and working method
CN111860106A (en) Unsupervised bridge crack identification method
CN114841920A (en) Flame identification method and device based on image processing and electronic equipment
CN115995056A (en) Automatic bridge disease identification method based on deep learning
CN112329858B (en) Image recognition method for breakage fault of anti-loosening iron wire of railway motor car
CN113516652A (en) Battery surface defect and adhesive detection method, device, medium and electronic equipment
CN111626104B (en) Cable hidden trouble point detection method and device based on unmanned aerial vehicle infrared thermal image
CN115082650A (en) Implementation method of automatic pipeline defect labeling tool based on convolutional neural network
CN115272826A (en) Image identification method, device and system based on convolutional neural network
CN109190489A (en) A kind of abnormal face detecting method based on reparation autocoder residual error
CN116543327A (en) Method, device, computer equipment and storage medium for identifying work types of operators
CN116152674A (en) Dam unmanned aerial vehicle image crack intelligent recognition method based on improved U-Net model
CN113724219A (en) Building surface disease detection method and system based on convolutional neural network
Park et al. Potholeeye+: Deep-learning based pavement distress detection system toward smart maintenance
CN111369508A (en) Defect detection method and system for metal three-dimensional lattice structure

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20230421

RJ01 Rejection of invention patent application after publication