CN113591773A - Power distribution room object detection method, device and equipment based on convolutional neural network - Google Patents

Power distribution room object detection method, device and equipment based on convolutional neural network Download PDF

Info

Publication number
CN113591773A
CN113591773A CN202110914881.8A CN202110914881A CN113591773A CN 113591773 A CN113591773 A CN 113591773A CN 202110914881 A CN202110914881 A CN 202110914881A CN 113591773 A CN113591773 A CN 113591773A
Authority
CN
China
Prior art keywords
image
neural network
convolutional neural
classification
level
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110914881.8A
Other languages
Chinese (zh)
Other versions
CN113591773B (en
Inventor
程津
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Zhongdian Smart Technology Co ltd
Original Assignee
Wuhan Zhongdian Smart Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Zhongdian Smart Technology Co ltd filed Critical Wuhan Zhongdian Smart Technology Co ltd
Priority to CN202110914881.8A priority Critical patent/CN113591773B/en
Publication of CN113591773A publication Critical patent/CN113591773A/en
Application granted granted Critical
Publication of CN113591773B publication Critical patent/CN113591773B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2132Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02EREDUCTION OF GREENHOUSE GAS [GHG] EMISSIONS, RELATED TO ENERGY GENERATION, TRANSMISSION OR DISTRIBUTION
    • Y02E40/00Technologies for an efficient electrical power generation, transmission or distribution
    • Y02E40/70Smart grids as climate change mitigation technology in the energy generation sector
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y04INFORMATION OR COMMUNICATION TECHNOLOGIES HAVING AN IMPACT ON OTHER TECHNOLOGY AREAS
    • Y04SSYSTEMS INTEGRATING TECHNOLOGIES RELATED TO POWER NETWORK OPERATION, COMMUNICATION OR INFORMATION TECHNOLOGIES FOR IMPROVING THE ELECTRICAL POWER GENERATION, TRANSMISSION, DISTRIBUTION, MANAGEMENT OR USAGE, i.e. SMART GRIDS
    • Y04S10/00Systems supporting electrical power generation, transmission or distribution
    • Y04S10/50Systems or methods supporting the power network operation or management, involving a certain degree of interaction with the load-side end user applications

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a power distribution room object detection method based on a convolutional neural network, which comprises the following steps: constructing an image data set of the detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image; constructing a network structure of a convolutional neural network model; inputting the first image and the second image into the convolutional neural network model alternately for training, and continuously optimizing the network structure parameters of the convolutional neural network model; and detecting the target object of the power distribution room by using the optimized convolutional neural network model. According to the method, the convolutional neural network model is constructed, and the labeled first image and the unlabeled second image are alternately input into the convolutional neural network model for training, so that the second image does not need to be labeled, and time cost, labor cost and material cost for labeling a large number of images are reduced; and meanwhile, the accuracy rate of detecting objects in the power distribution room can be ensured.

Description

Power distribution room object detection method, device and equipment based on convolutional neural network
Technical Field
The invention belongs to the technical field of object detection, and particularly relates to a method, a device and equipment for detecting objects in a power distribution room based on a convolutional neural network.
Background
In recent years, with the development of deep learning technology, a plurality of objects based on power distribution room target detection are researched, and the algorithm performs optimal fitting on a model by combining a large amount of object labels on a camera image of a power distribution room and the deep learning technology, so that the aim of assisting object inspection is fulfilled. However, in the automatic inspection process of the target in the distribution room, due to the fact that scenes of the distribution room are varied, the edge distribution difference of images around the day and night is large, and the existing detection method of the target object in the distribution room based on the deep learning technology has the following defects:
(1) the image heterogeneity is serious, images often have different edge distributions due to the influence of illumination and focal length in a power distribution room, and the image heterogeneity can directly cause the reduction of the generalization performance of the model, namely, similar objects are detected as different classes under different edge distribution conditions.
(2) If the detection quality of the night image is to be improved, a large amount of labels need to be added to the night image, so that the cost of image labeling is high.
Disclosure of Invention
The invention aims to provide a power distribution room object detection method, device and equipment based on a convolutional neural network, which are used for solving at least one technical problem in the prior art.
In order to achieve the purpose, the invention adopts the following technical scheme:
in a first aspect, the present invention provides a method for detecting objects in a power distribution room based on a convolutional neural network, including:
constructing an image data set of the detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image;
constructing a network structure of a convolutional neural network model;
inputting the first image and the second image into the convolutional neural network model alternately for training, and continuously optimizing the network structure parameters of the convolutional neural network model;
and detecting the target object of the power distribution room by using the optimized convolutional neural network model.
In one possible design, the convolutional neural network model includes: the system comprises an encoder, a region generation network, an example-level coordinate regressor, an example-level classifier, an image-level discriminator, an example-level discriminator, an image-level category regularization module and a classification consistency regularization module.
In one possible design, the alternately inputting the first image and the second image into the convolutional neural network model for training, and continuously optimizing the network structure parameters of the convolutional neural network model includes:
alternately inputting the first image and the second image into the encoder, and extracting a first feature GE(xi s) And a second feature GE(xi t) (ii) a Wherein x isi sRepresenting the ith first image, xi tRepresenting the ith second image;
the first feature GE(xi s) And a second feature GE(xi t) Inputting the area generation network to obtain a first candidate frame R (G)E(xi s) And a second frame candidate R (G)E(xi t));
Separately aligning the first candidate box R (G) using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) ) performing coordinate regression and classification to obtain a first positioning result
Figure BDA0003205259550000021
First classification result
Figure BDA0003205259550000022
Second positioning result
Figure BDA0003205259550000023
And a second classification result
Figure BDA0003205259550000024
And calculating the first positioning result
Figure BDA0003205259550000025
And first classification result
Figure BDA0003205259550000026
A first loss function of;
using the image-level discriminator to pair the first features GE(xi s) And a second feature GE(xi t) Classifying and judging the first characteristics GE(xi s) And a second feature GE(xi t) And calculating a second loss function of the image-level discriminator;
in the first feature GE(xi s) On the basis of the first image, predicting a multi-label classification result C (G) of the first image by utilizing the image-level classification regularization moduleE(xi s));
Utilizing the classification consistency regularization module to classify the first classification result
Figure BDA0003205259550000031
With the multi-label classification result C (G)E(xi s) Carrying out consistency judgment to obtain a classification consistency weight matrix lambda;
using the instance-level discriminator to pair the first candidate frames R (G)E(xi s) And the second frame candidate R (G)E(xi t) Classifying and discriminating the first frame candidate R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) and computing a third loss function for the instance-level arbiter based on the classification consistency weight matrix λ;
calculating a total loss function of the convolutional neural network model according to the first loss function, the second loss function and the third loss function;
and continuously optimizing the network structure parameters of the convolutional neural network model by using the total loss function.
In one possible design, the first candidate box R (G) is separately aligned using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) Before performing coordinate regression and classification, further comprising:
using non-maximum value suppression method to the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) the region confidence of the image data are respectively sorted;
respectively selecting the first T candidate frames as the final first candidate frame Rf(GE(xi s) ) and the final second candidate frame Rf(GE(xi t))。
In one possible design, the first positioning result is calculated
Figure BDA0003205259550000032
And first classification result
Figure BDA0003205259550000033
Comprises:
aligning the first positioning result with a cross-entropy penalty
Figure BDA0003205259550000034
And first classification result
Figure BDA0003205259550000035
And carrying out strong supervision, and calculating a strong supervision loss function, wherein the calculation formula is as follows:
Figure BDA0003205259550000041
Figure BDA0003205259550000042
Ls=Ls_class+Ls_pos; (3)
wherein L issIs the first positioning result
Figure BDA0003205259550000043
And first classification result
Figure BDA0003205259550000044
N represents the total number of the first images, i represents the number of the ith first image, M represents the number of labels contained in the ith first image, j represents the jth label in the ith first image, C represents the total number of prediction categories, K represents the number of coordinates of the current label frame, K represents the current kth category or kth label coordinate, y _ class and y _ pos and represents the label value of the current label;
wherein L iss_classAnd Ls_posThe more towards 0, the more accurate the detection classification result and the positioning result.
In one possible design, the second loss function of the image-level discriminator is calculated as follows:
Figure BDA0003205259550000045
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure BDA0003205259550000046
an example level discriminator representing a first image is shown,
Figure BDA0003205259550000047
an example level discriminator representing a second image, DimageA representation of the image-level discriminator,
Figure BDA0003205259550000048
represents the loss function of the image-level discriminator when
Figure BDA0003205259550000049
When the image level is close to 0.5, it indicates that the image level discriminator cannot distinguish the first feature GE(xi s) And the second feature GE(xi t)。
In one possible design, the first classification result is normalized by the classification consistency regularization module
Figure BDA00032052595500000410
And carrying out consistency judgment on the multi-label classification result to obtain a classification consistency weight matrix lambda, wherein the method comprises the following steps:
utilizing the classification consistency regularization module to classify the result into a first classification result
Figure BDA0003205259550000051
As a basis, traverse the multi-label classification result C (G)E(xi s) Finding the multi-label classification result C (G)E(xi s) Whether or not to contain the first classification result
Figure BDA0003205259550000052
And obtaining the classification consistency weight matrix lambda according to the search result.
In one possible design, a third loss function of the example-level discriminator is calculated based on the classification consistency weight matrix λ, as follows:
Figure BDA0003205259550000053
wherein nt and ns respectively represent the total number of the first image samples and the total number of the second image samples,
Figure BDA0003205259550000054
an example level discriminator representing the first image,
Figure BDA0003205259550000055
an example level discriminator representing a second image, DinstanceThere is shown an example level of the discriminator,
Figure BDA0003205259550000056
represents the penalty function of the instance level discriminator when
Figure BDA0003205259550000057
Approaching 0.5, it means that the example-level discriminator has been unable to distinguish the first frame candidate feature R (G)E(xi s) And a second frame candidate feature R (G)E(xi t))。
In one possible design, the first image is a distribution room daytime image and the second image is a distribution room nighttime image.
In a second aspect, the present invention provides a power distribution room object detection apparatus based on a convolutional neural network, the apparatus comprising:
the data set construction module is used for constructing an image data set of the detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image;
the neural network model building module is used for building a network structure of the convolutional neural network model;
the training module is used for inputting the first image and the second image into the convolutional neural network model alternately for training and optimizing the network structure parameters of the convolutional neural network model continuously;
and the detection module is used for detecting the target object of the power distribution room by using the optimized convolutional neural network model.
In one possible design, the convolutional neural network model includes: the system comprises an encoder, a region generation network, an example-level coordinate regressor, an example-level classifier, an image-level discriminator, an example-level discriminator, an image-level category regularization module and a classification consistency regularization module.
In one possible design, the training module specifically includes:
a feature extraction unit for alternately inputting the first image and the second image into the encoder and extracting a first feature GE(xi s) And a second feature GE(xi t) (ii) a Wherein x isi sRepresenting the ith first image, xi tRepresenting the ith second image;
a candidate frame extracting unit for extracting the first feature GE(xi s) And a second feature GE(xi t) Inputting the area generation network to obtain a first candidate frame R (G)E(xi s) And a second frame candidate R (G)E(xi t));
A coordinate regression and classification unit for separately classifying the first candidate frame R (G) using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) ) performing coordinate regression and classification to obtain a first positioning result
Figure BDA0003205259550000061
First classification result
Figure BDA0003205259550000062
Second positioning result
Figure BDA0003205259550000063
And a second classification result
Figure BDA0003205259550000064
And calculating the first positioning result
Figure BDA0003205259550000065
And first classification result
Figure BDA0003205259550000066
A first loss function of;
a first classification discrimination unit for discriminating the first feature G by the image-level discriminatorE(xi s) And a second feature GE(xi t) Classifying and judging the first characteristics GE(xi s) And a second feature GE(xi t) And calculating a second loss function of the image-level discriminator;
a classification result prediction unit for predicting the first feature GE(xi s) On the basis of the first image, predicting a multi-label classification result C (G) of the first image by utilizing the image-level classification regularization moduleE(xi s));
A classification result judgment unit for utilizing the classification consistency regularization module to classify the first classification result
Figure BDA0003205259550000071
With the multi-label classification result C (G)E(xi s) Carrying out consistency judgment to obtain a classification consistency weight matrix lambda;
a second classification discrimination unit for discriminating the first candidate frame R (G) using the instance-level discriminatorE(xi s) And the second frame candidate R (G)E(xi t) Classifying and discriminating the first frame candidate R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) and computing a third loss function for the instance-level arbiter based on the classification consistency weight matrix λ;
a loss function calculation unit, configured to calculate a total loss function of the convolutional neural network model according to the first loss function, the second loss function, and the third loss function;
and the model optimization unit is used for continuously optimizing the network structure parameters of the convolutional neural network model by utilizing the total loss function.
In one possible design, the first candidate box R (G) is separately aligned using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) Before performing coordinate regression and classification, the training module is further configured to:
using non-maximum value suppression method to the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) the region confidence of the image data are respectively sorted;
respectively selecting the first T candidate frames as the final first candidate frame Rf(GE(xi s) ) and the final second candidate frame Rf(GE(xi t))。
In one possible design, the first positioning result is calculated
Figure BDA0003205259550000072
And first classification result
Figure BDA0003205259550000073
The coordinate regression and classification unit is specifically configured to:
aligning the first positioning result with a cross-entropy penalty
Figure BDA0003205259550000074
And first classification result
Figure BDA0003205259550000075
And carrying out strong supervision, and calculating a strong supervision loss function, wherein the calculation formula is as follows:
Figure BDA0003205259550000076
Figure BDA0003205259550000081
Ls=Ls_class+Ls_pos; (3)
wherein L issIs the first positioning result
Figure BDA0003205259550000082
And first classification result
Figure BDA0003205259550000083
N represents the total number of the first images, i represents the number of the ith first image, M represents the number of labels contained in the ith first image, j represents the jth label in the ith first image, C represents the total number of prediction categories, K represents the number of coordinates of the current label frame, K represents the current kth category or kth label coordinate, y _ class and y _ pos and represents the label value of the current label;
wherein L iss_classAnd Ls_posThe more towards 0, the more accurate the detection classification result and the positioning result.
In one possible design, when calculating the second loss function of the image-level discriminator, the first classification discrimination unit uses the following calculation formula:
Figure BDA0003205259550000084
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure BDA0003205259550000085
an example level discriminator representing a first image is shown,
Figure BDA0003205259550000086
an example level discriminator representing a second image, DimageA representation of the image-level discriminator,
Figure BDA0003205259550000087
represents the loss function of the image-level discriminator when
Figure BDA0003205259550000088
When the image level is close to 0.5, it indicates that the image level discriminator cannot distinguish the first feature GE(xi s) And the second feature GE(xi t)。
In one possible design, the first classification result is normalized by the classification consistency regularization module
Figure BDA0003205259550000089
And when consistency judgment is carried out on the multi-label classification result to obtain a classification consistency weight matrix lambda, the classification result judgment unit is specifically used for:
utilizing the classification consistency regularization module to classify the result into a first classification result
Figure BDA00032052595500000810
As a basis, traverse the multi-label classification result C (G)E(xi s) Finding the multi-label classification result C (G)E(xi s) Whether or not to contain the first classification result
Figure BDA0003205259550000091
And obtaining the classification consistency weight matrix lambda according to the search result.
In one possible design, a third loss function of the example-level discriminator is calculated based on the classification consistency weight matrix λ, as follows:
Figure BDA0003205259550000092
wherein nt and ns respectively represent the total number of the first image samples and the total number of the second image samples,
Figure BDA0003205259550000093
an example level discriminator representing the first image,
Figure BDA0003205259550000094
an example level discriminator representing a second image, DinstanceThere is shown an example level of the discriminator,
Figure BDA0003205259550000095
represents the penalty function of the instance level discriminator when
Figure BDA0003205259550000096
Approaching 0.5, it means that the example-level discriminator has been unable to distinguish the first frame candidate feature R (G)E(xi s) And a second frame candidate feature R (G)E(xi t))。
In one possible design, the first image is a distribution room daytime image and the second image is a distribution room nighttime image.
In a third aspect, the present invention provides a computer device, comprising a memory, a processor and a transceiver, which are communicatively connected in sequence, wherein the memory is used for storing a computer program, the transceiver is used for sending and receiving messages, and the processor is used for reading the computer program and executing the convolutional neural network-based object detection method for an electric distribution room as described in any one of the possible designs of the first aspect.
In a fourth aspect, the present invention provides a computer-readable storage medium having stored thereon instructions which, when executed on a computer, perform a method for detecting objects in a power distribution room based on a convolutional neural network as set forth in any one of the possible designs of the first aspect.
In a fifth aspect, the present invention provides a computer program product containing instructions which, when run on a computer, cause the computer to perform a method of convolutional neural network-based object detection in a power distribution room as described in any one of the possible designs of the first aspect.
Has the advantages that:
1. the method comprises the steps of constructing an image data set of a detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image; constructing a network structure of a convolutional neural network model; then, the first image and the second image are alternately input into the convolutional neural network model for training, and the network structure parameters of the convolutional neural network model are continuously optimized; and finally, detecting the target object of the power distribution room by using the optimized convolutional neural network model. According to the method, the convolutional neural network model is constructed, and the labeled first image and the unlabeled second image are alternately input into the convolutional neural network model for training, so that the second image does not need to be labeled, and time cost, labor cost and material cost for labeling a large number of images are reduced; and meanwhile, the accuracy and robustness of object detection of the distribution room can be ensured.
2. The method utilizes the image-level discriminator and the example-level discriminator in the countermeasure generation network to classify the first feature and the second feature, and the first candidate frame and the second candidate frame respectively, so as to extract the consistency feature, and the consistency feature is used in the subsequent model training, so that more accurate image features can be provided for the neural network model, and the applicability of the neural network model is improved; secondly, the weak positioning characteristic of a multi-label classifier in the image-level classification regularization module is utilized to promote the neural network model to pay attention to important areas containing main targets, and the training efficiency is improved; and finally, calculating the consistency between image-level prediction and example-level prediction by using a classification consistency regularization module after obtaining the image-level multi-classification result, and further constraining the target detection classification result, thereby improving the detection capability of the neural network model on the difficult samples.
Drawings
Fig. 1 is a flowchart of a convolutional neural network-based object detection method for a power distribution room in the present embodiment;
fig. 2 is a network configuration block diagram of a neural network model in the present embodiment;
fig. 3 is a block diagram of the structure of the distribution room object detection apparatus based on the convolutional neural network in this embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present disclosure more clear, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the drawings in the embodiments of the present disclosure, and it is obvious that the described embodiments are some, but not all embodiments of the present disclosure. All other embodiments, which can be obtained by a person skilled in the art without inventive step based on the embodiments in the present description, belong to the protection scope of the present invention.
Example one
As shown in fig. 1-2, in a first aspect, the present embodiment provides a method for detecting objects in a power distribution room based on a convolutional neural network, including but not limited to the following steps S101 to S104:
s101, constructing an image data set of a detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image;
the first image is a daytime image of the distribution room, and the second image is a nighttime image of the distribution room, wherein objects in the distribution room, such as a distribution cabinet, a worker's clothes, a worker's hat, and the like, are marked in the nighttime image.
S102, constructing a network structure of a convolutional neural network model;
it should be noted that the convolutional neural network model includes, but is not limited to, an encoder, a region generation network, an example-level coordinate regressor, an example-level classifier, an image-level discriminator, an example-level discriminator, an image-level category regularization module, and a classification consistency regularization module.
Wherein, the encoder is composed of N coding blocks, preferably, the encoder is composed of 5 coding blocks, each coding block can include 1 3 × 3 convolution layers, 1 batch processing normalization layer, one relu activation layer and 1 pooling layer; wherein the area generation network is composed of a series of convolutional layers; wherein, the example-level coordinate regressor and the example-level classifier are composed of 3 full connection layers; the image-level discriminator and the example-level discriminator are respectively composed of three convolution layers, a maximum pooling layer, a full-connection layer and a sigmoid activation layer; the image-level classification regularization module is composed of three convolution layers, a maximum pooling layer, a full connection layer and a softmax activation layer.
S103, alternately inputting the first image and the second image into the convolutional neural network model for training, and continuously optimizing network structure parameters of the convolutional neural network model;
in step S103, alternately inputting the first image and the second image into the convolutional neural network model for training, and continuously optimizing network structure parameters of the convolutional neural network model, including:
step S1031, alternately inputting the first image and the second image into the encoder, and extracting a first feature GE(xi s) And a second feature GE(xi t) (ii) a Wherein x isi sRepresenting the ith first image, xi tRepresenting the ith second image;
for example: when the encoder comprises 5 encoding blocks, if the size of the input image is 512 × 3, the image is subjected to convolution, nonlinear activation and down-sampling processing through the first encoding block to obtain a first scale feature F1_ 1256 × 256 × 128, wherein 256 × 256 is the size, and 128 is the number of channels; the first scale feature F1_1 is subjected to convolution, nonlinear activation and down-sampling processing through a second coding block to obtain a second scale feature F1_ 2128 × 128; the second scale feature F1_2 is subjected to convolution, nonlinear activation and down-sampling processing through a third coding block to obtain a third scale feature F1_ 364 × 64 × 256; the third scale feature F1_3 is subjected to convolution, nonlinear activation and downsampling processing through a fourth coding block, so that a fourth scale feature F1_ 432 × 32 × 512 is obtained; the fourth scale feature F1_4 is subjected to convolution, nonlinear activation, and downsampling processing by the fifth coding block, and a fifth scale feature F1_ 416 × 16 × 1024 is obtained.
The scale of the n-th scale feature extracted by the M-th coding block is larger than the scale of the n-1-th scale feature extracted by the M-1-th coding block, the M-th scale is the largest scale in the multiple scales, M is a positive integer larger than or equal to 2, and M is smaller than or equal to M and larger than or equal to 2. Preferably, M may be 5.
Step S1032. will the first characteristic GE(xi s) And a second feature GE(xi t) Inputting the area generation network to obtain a first candidate frame R (G)E(xi s) And a second frame candidate R (G)E(xi t));
Wherein, the first characteristic G isE(xis) or second feature GE(xit) firstly carrying out convolution by 3 × 3 to obtain a 256 × 16 × 6 feature map, and then carrying out convolution by two times by 1 × 1 to respectively obtain an 18 × 16 × 6 feature map and a 36 × 16 × 6 feature map, wherein each feature map comprises 2 fractional features and 4 coordinate features, two fractional features express that the candidate frame is a target and not a target, and four coordinates are four coordinate features of the candidate frame.
Step S1033, respectively aligning the first candidate frames R (G) by using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) ) performing coordinate regression and classification to obtain a first positioning result
Figure BDA0003205259550000131
First classification result
Figure BDA0003205259550000132
Second positioning result
Figure BDA0003205259550000133
And a second classification result
Figure BDA0003205259550000134
And calculating the first positioning result
Figure BDA0003205259550000135
And first classification result
Figure BDA0003205259550000136
A first loss function of;
wherein, before step S1033, the method further comprises:
using non-maximum value suppression method to the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) the region confidence of the image data are respectively sorted;
respectively selecting the first T candidate frames as the final first candidate frame Rf(GE(xi s) ) and the final second candidate frame Rf(GE(xi t)). Where T is a hyper-parameter, which needs to be set manually, in this example, T is preferably 10.
Wherein, in step S1033, the first positioning result is calculated
Figure BDA0003205259550000137
And first classification result
Figure BDA0003205259550000138
Comprises:
aligning the first positioning result with a cross-entropy penalty
Figure BDA0003205259550000139
And first classification result
Figure BDA00032052595500001310
And carrying out strong supervision, and calculating a strong supervision loss function, wherein the calculation formula is as follows:
Figure BDA0003205259550000141
Figure BDA0003205259550000142
Ls=Ls_class+Ls_pos; (3)
wherein L issIs the first positioning result
Figure BDA0003205259550000143
And first classification result
Figure BDA0003205259550000144
N represents the total number of the first images, i represents the number of the ith first image, M represents the number of labels contained in the ith first image, j represents the jth label in the ith first image, C represents the total number of prediction categories, K represents the number of coordinates of the current label frame, K represents the current kth category or kth label coordinate, y _ class and y _ pos and represents the label value of the current label; wherein L iss_classAnd Ls_posThe more towards 0, the more accurate the detection classification result and the positioning result.
Step S1034, utilizing the image-level discriminator to discriminate the first characteristic G of the first characteristicE(xi s) And a second feature GE(xi t) Classifying and judging the first characteristics GE(xi s) And a second feature GE(xi t) And calculating a second loss function of the image-level discriminator;
wherein the first feature G is aligned using the image-level discriminatorE(xi s) And a second feature GE(xi t) ) classification, including in particular:
based on the antagonistic learning mechanism, the first feature GE(xi s) And a second feature GE(xi t) Is inputted to theIn the image discriminator, a first feature consistent confidence coefficient and a second feature consistent confidence coefficient are respectively obtained; wherein the label of the first feature is 1 and the label of the second feature is 0.
Wherein the final goal of the counterlearning mechanism is when the image-level discriminator cannot discriminate the first feature GE(xi s) And a second feature GE(xi t) Valid classification) indicates that the distributions of the two features are very similar at this point, and thus the feature is referred to as a consistent feature.
Wherein, a second loss function of the image-level discriminator is calculated, and the calculation formula is as follows:
Figure BDA0003205259550000145
Figure BDA0003205259550000151
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure BDA0003205259550000152
an example level discriminator representing a first image is shown,
Figure BDA0003205259550000153
an example level discriminator representing a second image, DimageA representation of the image-level discriminator,
Figure BDA0003205259550000154
represents the loss function of the image-level discriminator when
Figure BDA0003205259550000155
When the image level is close to 0.5, it indicates that the image level discriminator cannot distinguish the first feature GE(xi s) And the second feature GE(xi t)。
Step S1035. in the first characteristic GE(xi s) On the basis of the first image, predicting a multi-label classification result C (G) of the first image by utilizing the image-level classification regularization moduleE(xi s));
Wherein, for the labeled first image, the following categories are included in the first image, but not limited to: the multi-label classification result C (G) of the first image can be obtained after the worker clothes, the wearing safety helmet and the non-wearing safety helmet are predicted by the image-level classification regularization moduleE(xi s) Including but not limited to, wearing work clothes, wearing safety helmets, and not wearing safety helmets. The image-level classification regularization module promotes the neural network model to only concern important areas containing main target objects in the power distribution room by using the weak positioning characteristic of the multi-label classifier, and improves the classification efficiency.
Step S1036, utilizing the classification consistency regularization module to classify the first classification result
Figure BDA0003205259550000156
With the multi-label classification result C (G)E(xi s) Carrying out consistency judgment to obtain a classification consistency weight matrix lambda;
wherein, in step S1036, the method includes:
utilizing the classification consistency regularization module to classify the result into a first classification result
Figure BDA0003205259550000157
As a basis, traverse the multi-label classification result C (G)E(xi s) Finding the multi-label classification result C (G)E(xi s) Whether or not to contain the first classification result
Figure BDA0003205259550000158
Wherein if the weight of the category is 1, otherwise, the weight is 3; and then obtaining the classification consistency weight matrix lambda according to the search result.
Step S1037, utilizing the instance-level discriminator to pair the first candidate framesR(GE(xi s) And the second frame candidate R (G)E(xi t) Classifying and discriminating the first frame candidate R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) and computing a third loss function for the instance-level arbiter based on the classification consistency weight matrix λ;
wherein the first frame candidate R (G) is subjected to the instance-level discriminatorE(xi s) And the second frame candidate R (G)E(xi t) ) classification, including in particular:
based on the antagonistic learning mechanism, the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) Input into the example-level discriminator to respectively obtain a first feature consistent confidence coefficient and a second feature consistent confidence coefficient; wherein the label of the first feature is 1 and the label of the second feature is 0.
Wherein the final goal of the counterlearning mechanism is when the image-level discriminator cannot match the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) Valid classification) indicates that the distributions of the two features are very similar at this point, and thus the feature is referred to as a consistent feature.
In one possible design, a third loss function of the example-level discriminator is calculated based on the classification consistency weight matrix λ, as follows:
Figure BDA0003205259550000161
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure BDA0003205259550000162
an example level discriminator representing the first image,
Figure BDA0003205259550000163
an example level discriminator representing a second image, DinstanceThere is shown an example level of the discriminator,
Figure BDA0003205259550000164
represents the penalty function of the instance level discriminator when
Figure BDA0003205259550000165
Approaching 0.5, it means that the example-level discriminator has been unable to distinguish the first frame candidate feature R (G)E(xi s) And a second frame candidate feature R (G)E(xi t))。
Step S1038, calculating a total loss function of the convolutional neural network model according to the first loss function, the second loss function and the third loss function;
wherein the calculation formula of the total loss function is as follows:
Figure BDA0003205259550000171
and S1038, continuously optimizing network structure parameters of the convolutional neural network model by using the total loss function.
And S104, detecting the target object of the power distribution room by using the optimized convolutional neural network model.
After the application of the embodiment, it is found that when the convolutional neural network model is used for detecting the target object in the distribution room, compared with a method for predicting the object in the night image by directly using the distribution room daytime image training model, when the neural network model of the embodiment is used for predicting the object in the night image, the average accuracy and the recall rate of object detection can be improved by about 10%.
In a second aspect, as shown in fig. 3, the present invention provides a convolutional neural network-based object detection apparatus for a power distribution room, the apparatus comprising:
the data set construction module is used for constructing an image data set of the detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image;
the neural network model building module is used for building a network structure of the convolutional neural network model;
the training module is used for inputting the first image and the second image into the convolutional neural network model alternately for training and optimizing the network structure parameters of the convolutional neural network model continuously;
and the detection module is used for detecting the target object of the power distribution room by using the optimized convolutional neural network model.
In one possible design, the convolutional neural network model includes: the system comprises an encoder, a region generation network, an example-level coordinate regressor, an example-level classifier, an image-level discriminator, an example-level discriminator, an image-level category regularization module and a classification consistency regularization module.
In one possible design, the training module specifically includes:
a feature extraction unit for alternately inputting the first image and the second image into the encoder and extracting a first feature GE(xi s) And a second feature GE(xi t) (ii) a Wherein x isi sRepresenting the ith first image, xi tRepresenting the ith second image;
a candidate frame extracting unit for extracting the first feature GE(xi s) And a second feature GE(xi t) Inputting the area generation network to obtain a first candidate frame R (G)E(xi s) And a second frame candidate R (G)E(xi t));
A coordinate regression and classification unit for separately classifying the first candidate frame R (G) using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) ) performing coordinate regression and classification to obtain a first positioning result
Figure BDA0003205259550000181
First classification result
Figure BDA0003205259550000182
Second positioning result
Figure BDA0003205259550000183
And a second classification result
Figure BDA0003205259550000184
And calculating the first positioning result
Figure BDA0003205259550000185
And first classification result
Figure BDA0003205259550000186
A first loss function of;
a first classification discrimination unit for discriminating the first feature G by the image-level discriminatorE(xi s) And a second feature GE(xi t) Classifying and judging the first characteristics GE(xi s) And a second feature GE(xi t) And calculating a second loss function of the image-level discriminator;
a classification result prediction unit for predicting the first feature GE(xi s) On the basis of the first image, predicting a multi-label classification result C (G) of the first image by utilizing the image-level classification regularization moduleE(xi s));
A classification result judgment unit for utilizing the classification consistency regularization module to classify the first classification result
Figure BDA0003205259550000187
With the multi-label classification result C (G)E(xi s) Carrying out consistency judgment to obtain a classification consistency weight matrix lambda;
a second classification discrimination unit for discriminating the first candidate frame R (G) using the instance-level discriminatorE(xi s) And the second frame candidate R (G)E(xi t) Classifying and discriminating the first frame candidate R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) and computing a third loss function for the instance-level arbiter based on the classification consistency weight matrix λ;
a loss function calculation unit, configured to calculate a total loss function of the convolutional neural network model according to the first loss function, the second loss function, and the third loss function;
and the model optimization unit is used for continuously optimizing the network structure parameters of the convolutional neural network model by utilizing the total loss function.
In one possible design, the first candidate box R (G) is separately aligned using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) Before performing coordinate regression and classification, the training module is further configured to:
using non-maximum value suppression method to the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) the region confidence of the image data are respectively sorted;
respectively selecting the first T candidate frames as the final first candidate frame Rf(GE(xi s) ) and the final second candidate frame Rf(GE(xi t))。
In one possible design, the first positioning result is calculated
Figure BDA0003205259550000191
And first classification result
Figure BDA0003205259550000192
The coordinate regression and classification unit is specifically configured to:
aligning the first positioning result with a cross-entropy penalty
Figure BDA0003205259550000193
And first classification result
Figure BDA0003205259550000194
And carrying out strong supervision, and calculating a strong supervision loss function, wherein the calculation formula is as follows:
Figure BDA0003205259550000195
Figure BDA0003205259550000196
Ls=Ls_class+Ls_pos; (3)
wherein L issIs the first positioning result
Figure BDA0003205259550000197
And first classification result
Figure BDA0003205259550000198
N represents the total number of the first images, i represents the number of the ith first image, M represents the number of labels contained in the ith first image, j represents the jth label in the ith first image, C represents the total number of prediction categories, K represents the number of coordinates of the current label frame, K represents the current kth category or kth label coordinate, y _ class and y _ pos and represents the label value of the current label; wherein L iss_classAnd Ls_posThe more towards 0, the more accurate the detection classification result and the positioning result.
In one possible design, when calculating the second loss function of the image-level discriminator, the first classification discrimination unit uses the following calculation formula:
Figure BDA0003205259550000201
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure BDA0003205259550000202
an example level discriminator representing a first image is shown,
Figure BDA0003205259550000203
an example level discriminator representing a second image, DimageA representation of the image-level discriminator,
Figure BDA0003205259550000204
represents the loss function of the image-level discriminator when
Figure BDA0003205259550000205
When the image level is close to 0.5, it indicates that the image level discriminator cannot distinguish the first feature GE(xi s) And the second feature GE(xi t)。
In one possible design, the first classification result is normalized by the classification consistency regularization module
Figure BDA0003205259550000206
And when consistency judgment is carried out on the multi-label classification result to obtain a classification consistency weight matrix lambda, the classification result judgment unit is specifically used for:
utilizing the classification consistency regularization module to classify the result into a first classification result
Figure BDA0003205259550000207
As a basis, traverse the multi-label classification result C (G)E(xi s) Finding the multi-label classification result C (G)E(xi s) Whether or not to contain the first classification result
Figure BDA0003205259550000208
And obtaining the classification consistency weight matrix lambda according to the search result.
In one possible design, in calculating the third loss function of the example-level discriminator based on the classification consistency weight matrix λ, the second classification discrimination unit uses the following calculation formula:
Figure BDA0003205259550000211
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure BDA0003205259550000212
an example level discriminator representing the first image,
Figure BDA0003205259550000213
an example level discriminator representing a second image, DinstanceThere is shown an example level of the discriminator,
Figure BDA0003205259550000214
represents the penalty function of the instance level discriminator when
Figure BDA0003205259550000215
Approaching 0.5, it means that the example-level discriminator has been unable to distinguish the first frame candidate feature R (G)e(xi s) And a second frame candidate feature R (G)E(xi t))。
In one possible design, the first image is a distribution room daytime image and the second image is a distribution room nighttime image.
In a third aspect, the present invention provides a computer device, comprising a memory, a processor and a transceiver, which are communicatively connected in sequence, wherein the memory is used for storing a computer program, the transceiver is used for sending and receiving messages, and the processor is used for reading the computer program and executing the convolutional neural network-based object detection method for an electric distribution room as described in any one of the possible designs of the first aspect.
In a fourth aspect, the present invention provides a computer-readable storage medium having stored thereon instructions which, when executed on a computer, perform a method for detecting objects in a power distribution room based on a convolutional neural network as set forth in any one of the possible designs of the first aspect.
In a fifth aspect, the present invention provides a computer program product containing instructions which, when run on a computer, cause the computer to perform a method of convolutional neural network-based object detection in a power distribution room as described in any one of the possible designs of the first aspect.
Finally, it should be noted that: the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (11)

1. A method for detecting objects in a power distribution room based on a convolutional neural network is characterized by comprising the following steps:
constructing an image data set of the detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image;
constructing a network structure of a convolutional neural network model;
inputting the first image and the second image into the convolutional neural network model alternately for training, and continuously optimizing the network structure parameters of the convolutional neural network model;
and detecting the target object of the power distribution room by using the optimized convolutional neural network model.
2. A method of convolutional neural network based object detection in a power distribution room, as claimed in claim 1, wherein said convolutional neural network model comprises: the system comprises an encoder, a region generation network, an example-level coordinate regressor, an example-level classifier, an image-level discriminator, an example-level discriminator, an image-level category regularization module and a classification consistency regularization module.
3. The convolutional neural network-based electric distribution room object detection method as claimed in claim 2, wherein the alternately inputting the first image and the second image into the convolutional neural network model for training and continuously optimizing the network structure parameters of the convolutional neural network model comprises:
alternately inputting the first image and the second image into the encoder, and extracting a first feature GE(xi s) And a second feature GE(xi t) (ii) a Wherein x isi sRepresenting the ith first image, xi tRepresenting the ith second image;
the first feature GE(xi s) And a second feature GE(xi t) Inputting the area generation network to obtain a first candidate frame R (G)E(xi s) And a second frame candidate R (G)E(xi t));
Separately aligning the first candidate box R (G) using the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) ) performing coordinate regression and classification to obtain a first positioning result
Figure RE-FDA0003238628440000021
First classification result
Figure RE-FDA0003238628440000022
Second positioning result
Figure RE-FDA0003238628440000023
And a second classification result
Figure RE-FDA0003238628440000024
And calculating the first positioning result
Figure RE-FDA0003238628440000025
And first classification result
Figure RE-FDA0003238628440000026
A first loss function of;
using the image-level discriminator to pair the first features GE(xi s) And a second feature GE(xi t) Classifying and judging the first characteristics GE(xi s) And a second feature GE(xi t) And calculating a second loss function of the image-level discriminator;
in the first feature GE(xi s) On the basis of the first image, predicting a multi-label classification result C (G) of the first image by utilizing the image-level classification regularization moduleE(xi s));
Utilizing the classification consistency regularization module to classify the first classification result
Figure RE-FDA0003238628440000027
With the multi-label classification result C (G)E(xi s) Carrying out consistency judgment to obtain a classification consistency weight matrix lambda;
using the instance-level discriminator to pair the first candidate frames R (G)E(xi s) And the second frame candidate R (G)E(xi t) Classifying and discriminating the first frame candidate R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) and based onCalculating a third loss function of the example-level discriminator by the classification consistency weight matrix lambda;
calculating a total loss function of the convolutional neural network model according to the first loss function, the second loss function and the third loss function;
and continuously optimizing the network structure parameters of the convolutional neural network model by using the total loss function.
4. The convolutional neural network-based object detection method for a power distribution room of claim 3, wherein the first candidate frame R (G) is separately processed by the example-level coordinate regressor and the example-level classifierE(xi s) And the second frame candidate R (G)E(xi t) Before performing coordinate regression and classification, further comprising:
using non-maximum value suppression method to the first candidate frame R (G)E(xi s) And the second frame candidate R (G)E(xi t) ) the region confidence of the image data are respectively sorted;
respectively selecting the first T candidate frames as the final first candidate frame Rf(GE(xi s) ) and the final second candidate frame Rf(GE(xi t))。
5. The convolutional neural network-based electrical distribution room object detection method of claim 3, wherein the first positioning result is calculated
Figure RE-FDA0003238628440000031
And first classification result
Figure RE-FDA0003238628440000032
Comprises:
aligning the first positioning result with a cross-entropy penalty
Figure RE-FDA0003238628440000033
And first classification result
Figure RE-FDA0003238628440000034
And carrying out strong supervision, and calculating a strong supervision loss function, wherein the calculation formula is as follows:
Figure RE-FDA0003238628440000035
Figure RE-FDA0003238628440000036
Ls=Ls_class+Ls_pos; (3)
wherein L issIs the first positioning result
Figure RE-FDA0003238628440000037
And first classification result
Figure RE-FDA0003238628440000038
N represents the total number of the first images, i represents the number of the ith first image, M represents the number of labels contained in the ith first image, j represents the jth label in the ith first image, C represents the total number of prediction categories, K represents the number of coordinates of the current label frame, K represents the current kth category or kth label coordinate, y _ class and y _ pos and represents the label value of the current label;
wherein L iss_classAnd Ls_posThe more towards 0, the more accurate the detection classification result and the positioning result.
6. The convolutional neural network-based electric distribution room object detection method of claim 3, wherein the second loss function of the image-level discriminator is calculated as follows:
Figure RE-FDA0003238628440000041
wherein n ist,nsRespectively representing a first total number of image samples and a second total number of image samples,
Figure RE-FDA0003238628440000042
an example level discriminator representing a first image is shown,
Figure RE-FDA0003238628440000043
an example level discriminator representing a second image, DimageA representation of the image-level discriminator,
Figure RE-FDA0003238628440000044
represents the loss function of the image-level discriminator when
Figure RE-FDA0003238628440000045
When the image level is close to 0.5, it indicates that the image level discriminator cannot distinguish the first feature GE(xi s) And the second feature GE(xi t)。
7. The convolutional neural network-based electric distribution room object detection method of claim 3, wherein the first classification result is normalized by the classification consistency regularization module
Figure RE-FDA0003238628440000046
And carrying out consistency judgment on the multi-label classification result to obtain a classification consistency weight matrix lambda, wherein the method comprises the following steps:
utilizing the classification consistency regularization module to classify the result into a first classification result
Figure RE-FDA0003238628440000047
As a basis, traverse theMulti-label classification result C (G)E(xi s) Finding the multi-label classification result C (G)E(xi s) Whether or not to contain the first classification result
Figure RE-FDA0003238628440000051
And obtaining the classification consistency weight matrix lambda according to the search result.
8. The convolutional neural network-based object detection method for a power distribution room, as claimed in claim 3, wherein the third loss function of the instance-level discriminator is calculated based on the classification consistency weight matrix λ, and the calculation formula is as follows:
Figure RE-FDA0003238628440000052
wherein nt and ns respectively represent the total number of the first image samples and the total number of the second image samples,
Figure RE-FDA0003238628440000053
an example level discriminator representing the first image,
Figure RE-FDA0003238628440000054
an example level discriminator representing a second image, DinstanceThere is shown an example level of the discriminator,
Figure RE-FDA0003238628440000055
represents the penalty function of the instance level discriminator when
Figure RE-FDA0003238628440000056
Approaching 0.5, it means that the example-level discriminator has been unable to distinguish the first frame candidate feature R (G)E(xi s) And a second frame candidate feature R (G)E(xi t))。
9. The method of claim 1, wherein the first image is a distribution room daytime image and the second image is a distribution room nighttime image.
10. A power distribution room object detection apparatus based on a convolutional neural network, the apparatus comprising:
the data set construction module is used for constructing an image data set of the detected target object; wherein the image dataset comprises a labeled first image and an unlabeled second image;
the neural network model building module is used for building a network structure of the convolutional neural network model;
the training module is used for inputting the first image and the second image into the convolutional neural network model alternately for training and optimizing the network structure parameters of the convolutional neural network model continuously;
and the detection module is used for detecting the target object of the power distribution room by using the optimized convolutional neural network model.
11. A computer device comprising a memory, a processor and a transceiver communicatively connected in sequence, wherein the memory is used for storing a computer program, the transceiver is used for transmitting and receiving messages, and the processor is used for reading the computer program and executing the method for detecting objects in the power distribution room based on the convolutional neural network as claimed in any one of claims 1 to 9.
CN202110914881.8A 2021-08-10 2021-08-10 Distribution room object detection method, device and equipment based on convolutional neural network Active CN113591773B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110914881.8A CN113591773B (en) 2021-08-10 2021-08-10 Distribution room object detection method, device and equipment based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110914881.8A CN113591773B (en) 2021-08-10 2021-08-10 Distribution room object detection method, device and equipment based on convolutional neural network

Publications (2)

Publication Number Publication Date
CN113591773A true CN113591773A (en) 2021-11-02
CN113591773B CN113591773B (en) 2024-03-08

Family

ID=78256921

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110914881.8A Active CN113591773B (en) 2021-08-10 2021-08-10 Distribution room object detection method, device and equipment based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN113591773B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114066884A (en) * 2022-01-11 2022-02-18 南京医科大学眼科医院 Retinal blood vessel segmentation method and device, electronic device and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109614512A (en) * 2018-11-29 2019-04-12 亿嘉和科技股份有限公司 A kind of power equipment search method based on deep learning
US20200257984A1 (en) * 2019-02-12 2020-08-13 D-Wave Systems Inc. Systems and methods for domain adaptation
CN112132042A (en) * 2020-09-24 2020-12-25 西安电子科技大学 SAR image target detection method based on anti-domain adaptation
CN113158943A (en) * 2021-04-29 2021-07-23 杭州电子科技大学 Cross-domain infrared target detection method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109614512A (en) * 2018-11-29 2019-04-12 亿嘉和科技股份有限公司 A kind of power equipment search method based on deep learning
US20200257984A1 (en) * 2019-02-12 2020-08-13 D-Wave Systems Inc. Systems and methods for domain adaptation
CN112132042A (en) * 2020-09-24 2020-12-25 西安电子科技大学 SAR image target detection method based on anti-domain adaptation
CN113158943A (en) * 2021-04-29 2021-07-23 杭州电子科技大学 Cross-domain infrared target detection method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHANG-DONG XU ET.AL: "Exploring Categorical Regularization for Domain Adaptive Object Detection", 《ARXIV: 2003.09152V1 [CS.CV]》, pages 2 - 7 *
王露荻 等: "基于域适应Faster RCNN 的复杂背景目标检测", 《航天控制》, vol. 38, no. 1, pages 66 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114066884A (en) * 2022-01-11 2022-02-18 南京医科大学眼科医院 Retinal blood vessel segmentation method and device, electronic device and storage medium

Also Published As

Publication number Publication date
CN113591773B (en) 2024-03-08

Similar Documents

Publication Publication Date Title
CN108108657B (en) Method for correcting locality sensitive Hash vehicle retrieval based on multitask deep learning
CN113011319B (en) Multi-scale fire target identification method and system
CN108171136A (en) A kind of multitask bayonet vehicle is to scheme to search the system and method for figure
Ding et al. Vehicle and parking space detection based on improved yolo network model
CN112861635B (en) Fire disaster and smoke real-time detection method based on deep learning
CN108960124B (en) Image processing method and device for pedestrian re-identification
CN110298265A (en) Specific objective detection method in a kind of elevator based on YOLO neural network
CN109919223B (en) Target detection method and device based on deep neural network
CN114694178A (en) Method and system for monitoring safety helmet in power operation based on fast-RCNN algorithm
CN112241692B (en) Channel foreign matter intelligent detection and classification method based on aerial image super-pixel texture
CN112163572A (en) Method and device for identifying object
CN108664875A (en) Underground belt-conveying monitoring method based on image recognition
Rentao et al. Indoor smoking behavior detection based on yolov3-tiny
CN111145222A (en) Fire detection method combining smoke movement trend and textural features
CN111680759B (en) Power grid inspection insulator detection classification method
CN115512387A (en) Construction site safety helmet wearing detection method based on improved YOLOV5 model
CN115131747A (en) Knowledge distillation-based power transmission channel engineering vehicle target detection method and system
CN114140750A (en) Filling station safety helmet wearing real-time detection method based on YOLOv4-Tiny
CN113591773A (en) Power distribution room object detection method, device and equipment based on convolutional neural network
CN113379603B (en) Ship target detection method based on deep learning
CN113936299A (en) Method for detecting dangerous area in construction site
CN117523437A (en) Real-time risk identification method for substation near-electricity operation site
CN116205905B (en) Power distribution network construction safety and quality image detection method and system based on mobile terminal
CN115240394B (en) Method and system for monitoring and early warning water level of accident oil pool of transformer substation
CN110765900A (en) DSSD-based automatic illegal building detection method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant