CN115565068B - Full-automatic detection method for breakage of high-rise building glass curtain wall based on light-weight deep convolutional neural network - Google Patents
Full-automatic detection method for breakage of high-rise building glass curtain wall based on light-weight deep convolutional neural network Download PDFInfo
- Publication number
- CN115565068B CN115565068B CN202211210404.4A CN202211210404A CN115565068B CN 115565068 B CN115565068 B CN 115565068B CN 202211210404 A CN202211210404 A CN 202211210404A CN 115565068 B CN115565068 B CN 115565068B
- Authority
- CN
- China
- Prior art keywords
- convolution
- layer
- channel
- network
- detection network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/176—Urban or other man-made structures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/082—Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/17—Terrestrial scenes taken from planes or by drones
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/70—Labelling scene content, e.g. deriving syntactic or semantic representations
Abstract
The invention relates to a full-automatic detection method for high-rise building glass curtain wall damage based on a lightweight deep convolutional neural network, which comprises the steps of constructing a YOLO network, replacing a convolutional layer in the YOLO network by an extended convolutional layer, a channel-by-channel convolutional layer, a point-by-point convolutional layer and a pruning layer which are sequentially connected to obtain a new convolutional layer, and finally using the replaced YOLO network as a constructed first detection network; then the parameter w to be learned in the pruning layer is processed i Carrying out sparse constraint and pruning on the new convolutional layer in the first detection network to obtain a second detection network; and finally, carrying out sparse constraint and pruning on the scaling factor vector of each BN layer in the second detection network to obtain a third detection network. Therefore, the method further reduces the parameter quantity of the model on the premise of ensuring high accuracy, and realizes the full-automatic detection of the glass curtain wall of the urban high-rise building based on the unmanned aerial vehicle.
Description
Technical Field
The invention relates to the field of image detection, in particular to a full-automatic detection method for high-rise building glass curtain wall damage based on a lightweight deep convolutional neural network.
Background
The building glass curtain wall is widely applied to high-rise buildings due to the advantages of beautiful appearance, wide visual field, rapid construction, strong plasticity and the like. In the building glass curtain wall industry, china is subjected to rapid development from none to some market leading positions, and the building glass curtain wall is the first world producing and using big country at present. However, under the influence of factors such as outside (such as bad weather), use time and the like, the glass curtain wall can be damaged in the use process, the method not only influences the beauty and the visual field of the glass curtain wall, but also has huge potential safety hazard, so that the timely and accurate detection of the building glass curtain wall must be enhanced. Therefore, china has made a strict engineering specification document (JGJ 102-2003) of the building glass curtain wall, wherein the following rules are provided: the glass curtain wall needs to be fully checked every five years. This makes the inspection demand for glass curtain wall greatly improve.
At present, the building glass curtain wall detection industry mainly uses the traditional manual visual inspection mode, and this mode is time-consuming and time-consuming, and easily produces the problem of missed measure or false measure because of artifical tired. In addition, the inside of some high-rise glass curtain walls is sheltered from and causes that the detection can only be operated outside the building, and artifical visual needs and personnel high altitude construction combine together to realize the detection this moment. Thus, conventional detection methods also present certain risks. Therefore, an efficient and fast novel glass curtain wall detection method is urgently needed to avoid the defects of the traditional mode.
Along with the rapid development of unmanned aerial vehicle technique, the mode that utilizes unmanned aerial vehicle to carry out the aided detection to building glass curtain wall has appeared in recent years. The picture that this mode was shot with unmanned aerial vehicle transmits to measurement personnel's computer in real time on, measurement personnel only need look over the picture that unmanned aerial vehicle transmitted back on the computer to evaluate building glass curtain wall's state. The mode is labor-saving and time-saving, and effectively avoids dangers brought by high-altitude operation. However, this method requires the human inspection of each image collected by the unmanned aerial vehicle by the inspector, and thus still has the problems of a certain degree of labor and time consumption and missed inspection or false inspection caused by artificial fatigue. Therefore, based on the pictures acquired by the unmanned aerial vehicle, a Convolutional Neural Network (CNN) is introduced to realize a full-automatic detection method of the glass curtain wall of the high-rise building.
CNN is widely used in tasks such as image detection, classification, segmentation, etc. due to its excellent feature extraction capability. In recent years, there are a large number of advanced CNNs used for target detection, for example, the YOLO series, SSD, fast R-CNN, faster R-CNN, and the like. However, these networks have a large number of parameters and are difficult to embed in small devices, drones, which have limited computing power. Further improvements to the prior art are needed for this purpose.
Disclosure of Invention
The invention aims to solve the technical problem of providing a full-automatic detection method for the breakage of a glass curtain wall of a high-rise building based on a light-weight deep convolutional neural network, which has high accuracy and a small parameter number model in the prior art.
The technical scheme adopted by the invention for solving the technical problems is as follows: a full-automatic detection method for high-rise building glass curtain wall damage based on a lightweight deep convolutional neural network is characterized by comprising the following steps:
step 3, constructing a first detection network, and training and verifying the constructed first detection network to obtain the first detection network with optimal parameters;
the method for constructing the first detection network specifically comprises the following steps:
constructing a YOLO network, replacing a convolution layer in the YOLO network by using sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution to obtain a new convolution layer, and finally using the replaced YOLO network as a constructed first detection network;
wherein the structure of the new convolution layer is as follows: judging whether the channel-by-channel convolution step length is 1, if so, connecting a pruning layer after point-by-point convolution, and performing residual connection between the input of the extended convolution and the output of the pruning layer to obtain a new convolution layer; if not, taking the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution as a new convolution layer;
the specific calculation process of the pruning layer is as follows:
the pruning layer comprises a differentiable gate, the output g of which i The calculation formula is as follows:
wherein the content of the first and second substances,is a Sigmoid function, g i ∈[0,1],w i Is the parameter to be learned in the differentiable gate;
output g of the differentiable gate i Normalized to 0 or 1 to obtain
wherein, F i Feature maps output for point-by-point convolution, E (G) i ) Is a spreading function which will G i Extended to and i the same size; as element position-by-position multiplication;
output characteristic diagram F of new convolution layer with convolution step size of 1 channel by channel i3 The calculation formula of (2) is as follows:
wherein the content of the first and second substances,adding elements position by position; f i1 Is the feature map input to the extended convolution; when G is i If =0, then F i3 =F i1 (ii) a When G is i =1, then =>
The training process is as follows: inputting the images in the training set into the first detection network in batches, and performing training each time on the parameter w to be learned in the pruning layer i Carrying out sparse constraint, and calculating Loss function Loss of first detection network sparse training b And through the Loss function Loss b Reversely updating the parameters of the first detection network;
step 5, inputting the images in the training set into a second detection network in batches, carrying out sparse constraint on the scaling factor vector gamma of each BN layer in the second detection network during each training, calculating a Loss function Loss of sparse training of the second detection network, and reversely updating parameters of the second detection network through the Loss function Loss to obtain the second detection network with optimal parameters;
step 7, fine tuning the third detection network by using samples in the training set and the verification set, obtaining a final third detection network if the performance of the fine tuned third detection network meets the requirement, and otherwise, repeating the steps 4-6;
and 8, randomly selecting a glass curtain wall image in the test set, and inputting the selected glass curtain wall image into the final third detection network in the step 7 to obtain a detection result of the glass curtain wall image.
Further, the pruning in the step 4 comprises the following specific steps:
g corresponding to the pruning layer in the new convolution layer with the channel-by-channel convolution step length of 1 i Pruning the new coiling layer to remove the new coiling layer if the coiling layer is less than or equal to 0.5;
g corresponding to the pruned layer in the new convolution layer with channel-by-channel convolution step length of 1 i And if the sum is more than 0.5, removing the pruning layer in the new convolution layer, storing the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution, and performing residual connection between the input of the extended convolution and the output of the point-by-point convolution.
In order to increase the effective receptive field of the detection network, in step 3, after the convolutional layer is replaced by the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution, the method further includes enlarging the size of at least part of convolutional kernels in the new convolutional layer.
Preferably, the YOLO network in step 3 is a YOLO v4 network.
Compared with the prior art, the invention has the advantages that: in order to achieve the aim that the unmanned aerial vehicle carries the detection network to be used for completing full-automatic detection of the damage of the glass curtain wall of the urban high-rise building, firstly, the common convolution layer in the YOLO network is replaced by the sequentially connected extended convolution, channel-by-channel convolution, point-by-point convolution and pruning layer, so that the parameter number of the model is reduced; secondly, in order to increase the effective receptive field of the network, a part of convolution kernels are enlarged, and the detection precision of the network is effectively improved; and then designing a mixed pruning strategy based on the new convolutional layer and the characteristic diagram channel, and further reducing the parameter quantity of the model on the premise of ensuring high accuracy. And finally, deploying the network to the unmanned aerial vehicle, realizing the full-automatic detection of the glass curtain wall of the urban high-rise building based on the unmanned aerial vehicle, and changing the current situations of manual inspection and visual inspection in the industry.
Drawings
FIG. 1 is a flow chart of a fully automatic detection method according to an embodiment of the present invention;
FIG. 2 is a block diagram of a first detection network in accordance with the present invention;
FIG. 3 is a schematic diagram of the extended convolution, channel-by-channel convolution, and point-by-point convolution in connection with the present invention;
FIG. 4 is a schematic diagram of a new convolutional layer with a channel-by-channel convolution step size of 1 in the present invention;
FIG. 5 is a schematic diagram of pruning a new convolutional layer with a channel-by-channel convolution step size of 1 in the present invention;
FIG. 6 is a schematic diagram of channel pruning based on feature maps in the present invention.
Detailed Description
The invention is described in further detail below with reference to the accompanying examples.
As shown in fig. 1 to 6, the method for fully automatically detecting the breakage of a glass curtain wall of a high-rise building based on a lightweight deep convolutional neural network in the embodiment includes the following steps:
in the embodiment, the image of the glass curtain wall of the urban high-rise building is obtained by means of network search, mobile phone shooting, unmanned aerial vehicle shooting and the like; completing image breakage labeling work by using Labelimg;
in this embodiment, the training set, the verification set, and the test set are set according to 7:2:1, the training set is used for training the network, namely network parameters are adjusted; the verification set is used for selecting the optimal network parameters; the test set is used for testing the optimal network generalization capability;
step 3, constructing a first detection network, and training and verifying the constructed first detection network to obtain the first detection network with optimal parameters;
the method for constructing the first detection network specifically comprises the following steps:
constructing a YOLO network, replacing a convolution layer in the YOLO network by sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution to obtain a new convolution layer, and finally using the replaced YOLO network as a constructed first detection network;
wherein the structure of the new convolution layer is as follows: judging whether the channel-by-channel convolution step length is 1, if so, connecting a pruning layer after point-by-point convolution, and performing residual connection between the input of the extended convolution and the output of the pruning layer to obtain a new convolution layer; if not, taking the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution as a new convolution layer;
when the channel-by-channel convolution step length is 1, the size of the feature graph input by the extended convolution is the same as the size of the feature graph output by the pruning layer, so residual connection can be performed between the input of the extended convolution and the output of the pruning layer, and the detection accuracy is improved; on the contrary, when the channel-by-channel convolution step length is not 1, residual connection cannot be performed between the input of the extended convolution and the output of the pruning layer; when the step size of the channel-by-channel convolution is 2 (i.e., stride: 2), no residual connection is made between the input of the spread convolution and the output of the point-by-point convolution, as shown in FIG. 3 (a); when the channel-by-channel convolution step size is 1 (i.e., stride: 1), as shown in fig. 3 (b), residual connection is performed between the input of the spread convolution and the output of the point-by-point convolution; the following new convolutional layer pruning is mainly directed to the case where there is residual concatenation (i.e. to: channel-by-channel convolution step size is 1); as shown in fig. 2, for convenience of description, a module after connection of the extended convolution, the channel-by-channel convolution and the point-by-point convolution is denoted as IRB, and as shown in fig. 4, a new convolution layer after the point-by-point convolution and connected with the pruned layer is denoted as IRB-compact; the first feature map, the second feature map, the third feature map and the fourth feature map as in fig. 4 correspond to an output of the spread convolution, an output of the channel-by-channel convolution, an output of the point-by-point convolution and an output of the pruning layer, respectively;
in this embodiment, the YOLO network is a YOLO v4 network; the YOLO v4 is an improved version of a YOLO v3 target detection algorithm, and compared with the YOLO v3 target detection algorithm, the YOLO v3 target detection algorithm is optimized in the aspects of data enhancement, backbone feature extraction network, feature pyramid activation functions and the like, so that the real-time detection speed and precision of the YOLO v3 target detection algorithm are improved. The YOLO v4 network comprises a feature extraction backbone network, an SPP module, a feature fusion module and a classification regression layer, wherein the feature extraction backbone network is of a CSPDarknet53 network structure and realizes feature extraction of an input image; the SPP module can extract multi-scale depth features with different receptive fields, and the multi-scale depth features are connected in the channel dimension of the feature map for fusion, so that the detection precision is improved; the PANET is a characteristic fusion module which fuses the characteristics of the context by an up-sampling method and a down-sampling method to obtain higher semantic information and improve the accuracy rate of target detection; the YOLO-head structure of YOLO v3 is still used in the classification regression layer. In addition, the feature extraction backbone network and the feature fusion module comprise a plurality of convolution layers, and a BN layer (batch standard layer) is arranged behind each convolution layer; the specific structure of each module of the YOLO v4 network is shown in fig. 2, where the YOLO v4 network is the prior art and is not described herein again;
when the convolution layer is replaced by the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution, the size of at least part of convolution kernels in the new convolution layer is enlarged, namely the IRB with the enlarged size of the convolution kernels is recorded as e-IRB, so that the detection precision of the detection network is effectively improved;
the specific working principle of the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution can refer to the disclosure of the applicant's prior application CN202111395954.3 "an automatic detection method for breakage of glass curtain walls of high-rise buildings", and details thereof are not repeated herein.
The specific calculation process of the pruning layer is as follows:
the pruning layer comprises a differentiable gate, the output g of which i The calculation formula is as follows:
wherein the content of the first and second substances,is a Sigmoid function, g i ∈[0,1],w i Is the parameter to be learned in the differentiable gate;
g obtained here i Is continuous, and in order to accurately generate the pruned network, the output g of the differentiable gate is used in this embodiment i Normalized to 0 or 1 to obtainWherein G is i ∈{0,1},/>Represents: />Then G i =1;Then G i =0; due to the indicating function>Indifferent, straight-through estimator (STE) was used to calculate the gradient;
wherein, F i Feature maps output for point-by-point convolution, G (G) i ) Is a spreading function which will G i Extend toAnd F i The same size; as element position-by-position multiplication; for example:then->
Output profile F of the new convolutional layer i3 The calculation formula of (2) is as follows:
wherein the content of the first and second substances,adding elements position by position; f i1 Is the feature map input to the extended convolution; when G is i When =0, F i3 =F i1 (ii) a When G is i =1, based on the signal strength of the signal strength measured by the sensor>
The training process is as follows: inputting the images in the training set into the first detection network in batches, and performing training each time on the parameter w to be learned in the pruning layer i Carrying out sparse constraint, and calculating Loss function Loss of first detection network sparse training b And through a Loss function Loss b Reversely updating the parameters of the first detection network;
Loss b =λ b ∑σ(w i )+L YOLo
where σ (·) is a Sigmoid function; w is a i Is a parameter that can be trained in the pruning layer; lambda [ alpha ] b The sparse factor is used for controlling the sparse degree of the new convolution layer and is a manually set hyper-parameter; l is YOLo Is a loss function of the YOLO v4, and the calculation method uses a loss function calculation method of the existing YOLO v4 network;
the pruning method comprises the following specific steps:
g corresponding to the pruning layer in the new convolution layer with the channel-by-channel convolution step length of 1 i Pruning the new coiling layer to remove the new coiling layer if the coiling layer is less than or equal to 0.5;
g corresponding to the pruning layer in the new convolution layer with the channel-by-channel convolution step length of 1 i If the number of the pruning layers in the new convolution layer is more than 0.5, removing the pruning layers in the new convolution layer, storing the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution, and performing residual connection between the input of the extended convolution and the output of the point-by-point convolution;
when Gate =0.274, as shown in fig. 5, the new convolutional layer is removed, as shown in fig. 5, denoted as IRB removal module; when Gate =0.631, the pruning layer is removed, and the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution are saved, as shown in fig. 5, and are recorded as IRB retention modules;
step 5, inputting the images in the training set into a second detection network in batches, carrying out sparse constraint on the scaling factor vector gamma of each BN layer in the second detection network during each training, calculating a Loss function Loss of the sparse training of the second detection network, and reversely updating the parameters of the second detection network through the Loss function Loss to obtain the second detection network with optimal parameters;
the calculation formula of Loss is:
wherein, lambda is a balance factor and is a parameter set manually; g (γ) = | γ | L1 regularization expressed as γ; q is a set of scaling factor vectors γ for each BN layer;
the detailed pruning method can refer to the content disclosed in the applicant's prior application CN202111395954.3 "an automatic detection method for breakage of glass curtain walls of high-rise buildings", which is not described herein again; as shown in fig. 6, the numerical values with smaller scaling factor vector γ are pruned, and it can be seen from the figure that the parameters of the model can be obviously reduced after the pruning operation, and the large numerical values in the scaling factors can be retained, so as to accelerate the training speed and improve the training precision of the model;
step 7, fine tuning the third detection network by using samples in the training set and the verification set, obtaining a final third detection network if the performance of the fine tuned third detection network meets the requirement, and otherwise, repeating the steps 4-6;
and 8, randomly selecting a glass curtain wall image in the test set, and inputting the selected glass curtain wall image into the final third detection network in the step 7 to obtain a detection result of the glass curtain wall image.
And deploying the network obtained in the final third detection network on the unmanned aerial vehicle, and realizing full-automatic detection of the glass curtain wall of the urban high-rise building by utilizing the function of the unmanned aerial vehicle automatically surrounding the building. In this embodiment, the unmanned plane selects and uses the PHANTOM 4PRO in the world.
Claims (4)
1. A full-automatic detection method for high-rise building glass curtain wall damage based on a lightweight deep convolutional neural network is characterized by comprising the following steps:
step 1, obtaining a certain number of glass curtain wall images of urban high-rise buildings, and carrying out damage marking on the glass curtain wall images to obtain labels of the glass curtain wall images so as to form a sample set;
step 2, dividing the sample set into a training set, a verification set and a test set;
step 3, constructing a first detection network, and training and verifying the constructed first detection network to obtain the first detection network with optimal parameters;
the method for constructing the first detection network specifically comprises the following steps:
constructing a YOLO network, replacing a convolution layer in the YOLO network by using sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution to obtain a new convolution layer, and finally using the replaced YOLO network as a constructed first detection network;
wherein the structure of the new convolution layer is as follows: judging whether the channel-by-channel convolution step length is 1, if so, connecting a pruning layer after point-by-point convolution, and performing residual connection between the input of the extended convolution and the output of the pruning layer to obtain a new convolution layer; if not, taking the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution as a new convolution layer;
the specific calculation process of the pruning layer comprises the following steps:
the pruning layer comprises a differentiable gate, the output g of which i The calculation formula is as follows:
wherein the content of the first and second substances,is a Sigmoid function, g i ∈[0,1],w i Is the parameter to be learned in the differentiable gate;
will differentiate the output g of the gate i Normalized to 0 or 1 to obtain
wherein, F i Feature maps output for point-by-point convolution, E (G) i ) Is a spreading function which will G i Is extended to be AND i The same size; as element position-by-position multiplication;
output characteristic diagram F of new convolution layer with convolution step length of 1 channel by channel i3 The calculation formula of (2) is as follows:
wherein, the first and the second end of the pipe are connected with each other,adding elements position by position; f i1 Is the feature map input to the extended convolution; when G is i If =0, then F i3 =F i1 (ii) a When G is i If =1, then = live>
The training process is as follows: inputting the images in the training set into the first detection network in batches, and performing training each time on the parameter w to be learned in the pruning layer i Carrying out sparse constraint, and calculating Loss function Loss of first detection network sparse training b And through the Loss function Loss b Reversely updating the parameters of the first detection network;
step 4, pruning the new convolution layer with the channel-by-channel convolution step length of 1 in the first detection network with the optimal parameters in the step 3 to obtain a second detection network;
step 5, inputting the images in the training set into a second detection network in batches, carrying out sparse constraint on the scaling factor vector gamma of each BN layer in the second detection network during each training, calculating a Loss function Loss of sparse training of the second detection network, and reversely updating parameters of the second detection network through the Loss function Loss to obtain the second detection network with optimal parameters;
step 6, pruning the second detection network with the optimal parameters in the step 5 according to the scaling factor vectors gamma of all BN layers to obtain a third detection network;
step 7, fine tuning the third detection network by using samples in the training set and the verification set, obtaining a final third detection network if the performance of the fine tuned third detection network meets the requirement, and otherwise, repeating the steps 4-6;
and 8, randomly selecting a glass curtain wall image in the test set, and inputting the selected glass curtain wall image into the final third detection network in the step 7 to obtain a detection result of the glass curtain wall image.
2. The fully automatic detection method according to claim 1, characterized in that: the specific method for pruning in the step 4 comprises the following steps:
g corresponding to the pruning layer in the new convolution layer with the channel-by-channel convolution step length of 1 i Pruning the new coiling layer to remove the new coiling layer if the coiling layer is less than or equal to 0.5;
g corresponding to the pruning layer in the new convolution layer with the channel-by-channel convolution step length of 1 i And if the sum of the calculated average value and the calculated average value is more than 0.5, removing a pruning layer in the new convolution layer, storing the sequentially connected expansion convolution, channel-by-channel convolution and point-by-point convolution, and performing residual connection between the input of the expansion convolution and the output of the point-by-point convolution.
3. The fully automatic detection method according to claim 1, characterized in that: and in the step 3, after the convolution layer is replaced by the sequentially connected extended convolution, channel-by-channel convolution and point-by-point convolution, the method also comprises the step of expanding the size of at least part of convolution kernels in the new convolution layer.
4. The fully automatic detection method according to any one of claims 1 to 3, characterized in that: the YOLO network in step 3 is a YOLO v4 network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211210404.4A CN115565068B (en) | 2022-09-30 | 2022-09-30 | Full-automatic detection method for breakage of high-rise building glass curtain wall based on light-weight deep convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211210404.4A CN115565068B (en) | 2022-09-30 | 2022-09-30 | Full-automatic detection method for breakage of high-rise building glass curtain wall based on light-weight deep convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115565068A CN115565068A (en) | 2023-01-03 |
CN115565068B true CN115565068B (en) | 2023-04-18 |
Family
ID=84743900
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211210404.4A Active CN115565068B (en) | 2022-09-30 | 2022-09-30 | Full-automatic detection method for breakage of high-rise building glass curtain wall based on light-weight deep convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115565068B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110009565A (en) * | 2019-04-04 | 2019-07-12 | 武汉大学 | A kind of super-resolution image reconstruction method based on lightweight network |
CN114120154A (en) * | 2021-11-23 | 2022-03-01 | 宁波大学 | Automatic detection method for breakage of glass curtain wall of high-rise building |
CN114581868A (en) * | 2022-03-04 | 2022-06-03 | 京东鲲鹏(江苏)科技有限公司 | Image analysis method and device based on model channel pruning |
CN114898327A (en) * | 2022-03-15 | 2022-08-12 | 武汉理工大学 | Vehicle detection method based on lightweight deep learning network |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220101494A1 (en) * | 2020-09-30 | 2022-03-31 | Nvidia Corporation | Fourier transform-based image synthesis using neural networks |
-
2022
- 2022-09-30 CN CN202211210404.4A patent/CN115565068B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110009565A (en) * | 2019-04-04 | 2019-07-12 | 武汉大学 | A kind of super-resolution image reconstruction method based on lightweight network |
CN114120154A (en) * | 2021-11-23 | 2022-03-01 | 宁波大学 | Automatic detection method for breakage of glass curtain wall of high-rise building |
CN114581868A (en) * | 2022-03-04 | 2022-06-03 | 京东鲲鹏(江苏)科技有限公司 | Image analysis method and device based on model channel pruning |
CN114898327A (en) * | 2022-03-15 | 2022-08-12 | 武汉理工大学 | Vehicle detection method based on lightweight deep learning network |
Non-Patent Citations (2)
Title |
---|
Shangqian Gao等.Network Pruning via Performance Maximization. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.2021,9270-9280. * |
刘济樾.基于轻量化网络的实时人脸检测方法研究.中国优秀硕士学位论文全文数据库信息科技辑.2020,I138-1236. * |
Also Published As
Publication number | Publication date |
---|---|
CN115565068A (en) | 2023-01-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111507990B (en) | Tunnel surface defect segmentation method based on deep learning | |
CN110378222B (en) | Method and device for detecting vibration damper target and identifying defect of power transmission line | |
CN111178206B (en) | Building embedded part detection method and system based on improved YOLO | |
CN109472284A (en) | A kind of battery core defect classification method based on zero sample learning of unbiased insertion | |
CN111832607B (en) | Bridge disease real-time detection method based on model pruning | |
CN116402821B (en) | Aircraft skin gluing quality defect detection method based on neural network | |
CN111178438A (en) | ResNet 101-based weather type identification method | |
CN111724358A (en) | Concrete quality detection method and system based on image and convolutional neural network | |
CN113780200A (en) | Computer vision-based pavement multi-disease area detection and positioning method | |
CN111985325A (en) | Aerial small target rapid identification method in extra-high voltage environment evaluation | |
CN111104850B (en) | Remote sensing image building automatic extraction method and system based on residual error network | |
CN116823800A (en) | Bridge concrete crack detection method based on deep learning under complex background | |
CN114612803A (en) | Transmission line insulator defect detection method for improving CenterNet | |
CN114120154B (en) | Automatic detection method for breakage of glass curtain wall of high-rise building | |
CN115565068B (en) | Full-automatic detection method for breakage of high-rise building glass curtain wall based on light-weight deep convolutional neural network | |
CN112966698A (en) | Freshwater fish image real-time identification method based on lightweight convolutional network | |
CN110796716B (en) | Image coloring method based on multiple residual error network and regularized transfer learning | |
CN113762247A (en) | Road crack automatic detection method based on significant instance segmentation algorithm | |
CN117132546A (en) | Concrete surface defect detection method under defect-free sample condition | |
CN115273009A (en) | Road crack detection method and system based on deep learning | |
CN112270370B (en) | Vehicle apparent damage assessment method | |
CN114998251A (en) | Air multi-vision platform ground anomaly detection method based on federal learning | |
CN115272826A (en) | Image identification method, device and system based on convolutional neural network | |
CN107886049A (en) | A kind of visibility identification method for early warning based on camera probe | |
CN114463340B (en) | Agile remote sensing image semantic segmentation method guided by edge information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |