CN115620107A - Transformer substation bird-involved fault related bird species identification method based on deep learning - Google Patents

Transformer substation bird-involved fault related bird species identification method based on deep learning Download PDF

Info

Publication number
CN115620107A
CN115620107A CN202211386862.3A CN202211386862A CN115620107A CN 115620107 A CN115620107 A CN 115620107A CN 202211386862 A CN202211386862 A CN 202211386862A CN 115620107 A CN115620107 A CN 115620107A
Authority
CN
China
Prior art keywords
bird
module
convnext
bird species
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211386862.3A
Other languages
Chinese (zh)
Inventor
李帆
万诚
邱志斌
饶斌斌
华桦
周志彪
詹涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
State Grid Jiangxi Electric Power Co ltd
State Grid Corp of China SGCC
Nanchang University
Electric Power Research Institute of State Grid Jiangxi Electric Power Co Ltd
Original Assignee
State Grid Jiangxi Electric Power Co ltd
State Grid Corp of China SGCC
Nanchang University
Electric Power Research Institute of State Grid Jiangxi Electric Power Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by State Grid Jiangxi Electric Power Co ltd, State Grid Corp of China SGCC, Nanchang University, Electric Power Research Institute of State Grid Jiangxi Electric Power Co Ltd filed Critical State Grid Jiangxi Electric Power Co ltd
Priority to CN202211386862.3A priority Critical patent/CN115620107A/en
Publication of CN115620107A publication Critical patent/CN115620107A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method for identifying bird species related to bird-involved faults of a transformer substation based on deep learning, which comprises the steps of firstly, aiming at a transformer substation polling image and a public bird species image set, expanding bird species images by adopting motion blur and defocusing blur, and constructing a transformer substation bird-involved fault related bird species image data set; then, making a label for each image in the data set to obtain a real rectangular frame containing bird species categories and coordinate information, and dividing the frame into a test set, a training set and a verification set; secondly, improving a YOLOv5 target detection model structure, replacing an original main feature extraction network with ConvNeXt-T, and extracting richer bird image features; and finally, training the improved YOLOv5 model by combining a plurality of training skills, and detecting the divided test set images by using the trained model. The invention can provide reference for differential prevention and control of bird-involved faults of the transformer substation.

Description

Transformer substation bird-involved fault related bird species identification method based on deep learning
Technical Field
The invention belongs to the technical field of transformer substation monitoring image data processing, and particularly relates to a transformer substation bird-related fault related bird species identification method based on deep learning.
Background
The transformer substation is a core place for bearing electric energy conversion and distribution in a power grid, and safe and stable operation of the transformer substation has great significance to the power grid. With the continuous development of power grids, power transmission lines, transformer substations and the like are widely spread, and part of the transformer substations are in remote areas with relatively good ecological environments, so that a proper production environment is brought to birds. However, birds nest, fly, defecate and other activities in the open-type transformer substation, easily cause insulation flashover, disconnecting link switching failure, switch tripping, even equipment burning loss and other faults, and seriously threaten the safe and stable operation of the transformer equipment. However, due to the characteristics of randomness and burstiness, the problems of manpower and material resource consumption, low efficiency and the like are caused because the system cannot be found in time depending on manual routing inspection. In recent years, equipment such as a robot and an unmanned aerial vehicle are used for carrying an intelligent detection model to patrol a transformer substation, bird species images related to bird-related faults of the transformer substation are collected, and intelligent detection is achieved. According to the bird species category that detects out, carry out differentiation prevention and cure, very big improvement patrols and examines efficiency.
At present, with the further development of machine learning, the deep learning intelligent detection algorithm is applied to various fields. Compared with the traditional image recognition method, the method has higher detection precision and speed. The convolutional neural network can automatically extract the image features, so that the complex process of manually extracting the image features is avoided, and the efficiency is greatly improved. However, intelligent identification of bird species related to bird-related faults in a transformer substation is less, most of bird species are concentrated in the field of bird ecology, and an intelligent bird identification method is urgently needed to assist inspection personnel.
Disclosure of Invention
In view of this, the invention aims to provide a method for identifying bird species related to bird-involved faults of a transformer substation based on deep learning, which can realize accurate intelligent identification of bird species related to bird-involved faults of the transformer substation.
In order to achieve the purpose, the invention provides the following technical scheme: a transformer substation bird-related fault related bird species identification method based on deep learning comprises the following steps:
s1, constructing a bird species image data set related to bird-related faults of a transformer substation;
s2, constructing an improved Yolov5 bird species target detection model: constructing four parts, namely an input end, a backbone section network ConvNeXt-T for generating characteristics, a neck section network FPN + PAN for carrying out characteristic fusion and a head section network for carrying out prediction; updating parameters through back propagation of a loss function, wherein the loss function consists of detection frame loss, confidence coefficient loss and classification loss, and finally obtaining an improved Yolov5 bird target detection model;
s3, training an improved Yolov5 bird species target detection model;
and S4, detecting the images of the test set by using the trained improved Yolov5 bird species target detection model.
Preferably, in step S1, the substation patrol image and the bird species image set are expanded by using motion blur and defocus blur, a substation bird-related fault-related bird species image data set including multiple bird species is constructed, a tool Labeling is applied to label each image in the substation bird-related fault-related bird species image data set, the obtained bird species and coordinate information are prepared into a standard Pascal VOC data set format, and the standard Pascal VOC data set format is divided into a test set, a training set and a verification set.
More preferably, in step S1, the bird species image is motion blurred by a motion blur function (PSF):
Figure 223274DEST_PATH_IMAGE001
in the formulaf(x, y) Andg(x, y) Respectively representing the input bird species image and the motion-blurred bird species image,h(x, y) For a motion blur function (PSF) that convolves the input bird species image,n(x, y) For an additive noise function, the formula for the motion blur function is expressed as:
Figure 609518DEST_PATH_IMAGE002
in the formulaLAnd
Figure 673289DEST_PATH_IMAGE003
is the length and direction of motion blur; defocus blur is obtained by substituting the motion blur function with the defocus functionh r (x, y) The result of the convolution:
Figure 934506DEST_PATH_IMAGE004
in the formularIs the defocus blur radius.
Further preferably, the input end performs mosaic data enhancement, adaptive anchor frame calculation and adaptive picture scaling technical processing on the input image to obtain an appropriate anchor frame and an input image with a uniform size.
It is further preferred that the backbone segment network of the original YOLOv5 is replaced by ConvNeXt-T, which in turn consists of 1 separate 4 × 4 convolutional layer, a normative layer (LayerNorm), a ConvNeXt _1 module, a ConvNeXt _2 module, a ConvNeXt _3 module, a downsampling module i, a ConvNeXt _4 module, a ConvNeXt _5 module, a ConvNeXt _6 module, a downsampling module ii, a ConvNeXt _7 module, a ConvNeXt _8 module, a ConvNeXt _9 module, a ConvNeXt _10 module, a ConvNeXt _11 module, a ConvNeXt _12 module, a ConvNeXt _13 module, a ConvNeXt _14 module, a ConvNeXt _15 module, a downsampling module, a ConvNeXt _16 module, a ConvNeXt _17, a normative _ iii module, a linvre _16 module, a global averaging layer (layernom _ 18), and a global averaging layer.
Further preferably, the first features F are extracted through a ConvNeXt _3 module, a ConvNeXt _6 module, a ConvNeXt _15 module and a ConvNeXt _18 module in the backbone segment network ConvNeXt-T respectively 1 A second feature F 2 The third feature F 3 And fourth feature F 4 And the convolution characteristic extraction of the image from shallow to deep is realized, and then the image is input to a neck segment network FPN + PAN for fusion.
Further preferably, the first feature F extracted from the neck segment network FPN + PAN pair 1 Second characteristic F 2 A third feature F 3 Fourth feature F 4 Carrying out characteristic fusion from deep to shallow and from shallow to deep, specifically: fourth feature F 4 Obtaining a first feature map P through a CSPN2 module and a CBL convolution block 1 First characteristic diagram P 1 After upsampling and the second feature F 2 Performing fusion to obtain a second feature map P 2 (ii) a Second characteristic diagram P 2 Obtaining a third characteristic diagram P through a CSPN2 module and a CBL convolution block 3 Third characteristic diagram P 3 After up-sampling and the first characteristic F 1 Performing fusion and obtaining a fourth characteristic diagram P through a CSPN2 module 4 (ii) a Fourth feature map P 4 Obtaining a fifth feature map P through the CBL convolution block 5 Fifth characteristic diagram P 5 And the third characteristic diagram P 3 Merging and obtaining a sixth feature map P through a CSPN2 module 6 (ii) a Sixth characteristic diagramP 6 Obtaining a seventh feature map P by CBL convolution block 7 Seventh feature diagram P 7 And the third feature F 3 Performing fusion and obtaining an eighth characteristic diagram P through a CSPN2 module 8
More preferably, convNeXt _iThe module is used for carrying out the following steps,i
Figure 665702DEST_PATH_IMAGE005
1, 2, 3, \ 823018, which consists of 17 × 7 layer-by-layer convolutional layer, a normative layer, 1 convolutional layer, 1 GELU activation function, 1 convolutional layer, a scaling and regularization layer in sequence, and adds input and output by a residual connection.
Further preferably, the CSPN2 module has 2 branches, 1 branch is 2 CBL convolution blocks and 1 convolution layer, the other 1 branch is 1 convolution layer, and the output of the two branches is subjected to feature fusion and then sequentially passes through the batch normalization layer, the leak RuLU activation function and the CBL convolution block to obtain the final output.
Preferably, three detection frames with different prediction sizes exist in the head segment network, and the detection frames respectively correspond to large, medium and small detection targets; the obtained fourth feature map P with strengthened fusion 4 The sixth characteristic diagram P 6 And an eighth feature map P 8 After convolution, the data are input into detection boxes with large, medium and small prediction sizes.
Loss function loss by detection box in head segment networkLossRect, loss of confidenceLossConf and classification lossLossA cls component, updating the weight and the deviation by a loss function through back propagation, and finally obtaining an optimal detection frame, a most accurate confidence score and a classification result at the same time; the detection frame Loss adopts a CIOU _ Loss function, confidence coefficient Loss and classification Loss as cross entropy Loss functions.
Firstly, pre-training an improved Yolov5 bird species target detection model by using a COCO public data set to obtain pre-training weights; and then loading pre-training weights, and retraining the improved Yolov5 bird species target detection model by using the divided training set and verification set, wherein the training skills comprise: the method comprises the following steps of multi-stage transfer learning, mixed precision training fp16, mosaic and mixed data enhancement, label smoothing, SGDM model parameter optimization and Step length (Step) learning rate reduction, wherein the training skills can improve the speed, generalization capability and robustness of model training; secondly, evaluating the trained improved YOLOv5 bird species target detection model through a verification set during training, and optimizing parameters of the improved YOLOv5 bird species target detection model until the training is finished; finally, obtaining the optimal weight of the improved Yolov5 bird target detection model;
further preferably, a multi-stage migration learning method is adopted to train the improved Yolov5 bird species target detection model, a first stage freezes a main stem to train, and a second stage finely tunes all layer parameters; and (4) optimizing parameters by adopting an SGDM model.
Preferably, in step S4, the optimal weight is loaded, the test set image is scaled to a uniform size by using an adaptive picture scaling technique, and then input to the improved YOLOv5 bird species target detection model for detection, and a flexible non-maximum value is used for suppressing and selecting a positioning accurate frame, so as to realize detection of the bird species target, wherein the detected image includes the species and specific position information of the bird species and the confidence of the prediction result, and the number of the bird species is statistically output.
Further preferably, in step S4, the adaptive image scaling technique (lettbex) scales the test set image to 640 × 640, inputs the scaled test set image into the improved yollov 5 bird species target detection model for detection, and selects a positioning accurate frame through flexible non-maximum suppression: prediction candidate boxY r The higher the coincidence degree with the highest-score prediction candidate frame M, the higher the prediction candidate frame MY r The lower the score of (d); prediction candidate boxY r The degree of coincidence with the highest-score prediction candidate frame M is less than the threshold valueN t The prediction candidate block is retainedY r
Compared with the prior art, the invention has the beneficial effects that: the bird species image is expanded through motion blur and defocusing blur, the improved YOLOv5 bird species target detection model is trained on the basis of the improved YOLOv5 bird species target detection model and in combination with various training skills, the trained improved YOLOv5 bird species target detection model is used for detecting the divided test set image, and accurate intelligent identification of bird species related to bird-related faults of the transformer substation can be achieved. The technical scheme provided by the invention has better detection precision and detection speed, is more favorable for practical application, and can be used for carrying out differential prevention and control on bird-involved faults of the transformer substation for transformer substation operation and maintenance personnel.
Drawings
FIG. 1 is a flow chart of the method of the present invention.
FIG. 2 is a schematic diagram of an improved Yolov5 bird species target detection model.
Fig. 3 is a schematic diagram of the ConvNeXt _1 module.
Fig. 4 is a schematic diagram of a CSPN2 module.
Detailed Description
The present invention is further described in the following examples, which should not be construed as limiting the scope of the invention, but rather as providing the following examples which are set forth to illustrate and not limit the scope of the invention.
A method for identifying bird species related to bird-involved faults in a transformer substation based on deep learning is disclosed, wherein a flow chart of the method is shown in figure 1, and comprises the following steps:
s1, constructing a bird species image data set related to bird-involved faults of a transformer substation: aiming at a transformer substation inspection image and a bird species image set, extending bird species images by adopting motion blur and defocusing blur, constructing a transformer substation bird-related fault related bird species image data set containing 10 bird species, labeling each image in the transformer substation bird-related fault related bird species image data set by using a tool Labeling, and making the obtained bird species and coordinate information into a standard Pascal VOC data set format and dividing the standard bird species and coordinate information into a test set, a training set and a verification set;
in the embodiment, 10 birds such as pied magpie, ash pied magpie, rhododendron, vernonia nucifera, white head 40526, swallow, red falcon and black collar 26891, silkete 26891, bird and big mouth crow are selected as research objects from the bird species image, each bird species comprises 200 bird species images, and 2000 bird species image samples are counted. The bird species image is augmented with motion blur and defocus blur: motion blurring the bird species image by a motion blur function:
Figure 721382DEST_PATH_IMAGE001
(1)
in the formulaf(x, y) Andg(x, y) Respectively representing the input bird species image and the motion-blurred bird species image,h(x, y) For the motion blur function of convolving the input bird species image,n(x, y) For an additive noise function, the formula for the motion blur function is expressed as:
Figure 39493DEST_PATH_IMAGE006
(2)
in the formulaLAnd
Figure 838822DEST_PATH_IMAGE003
is the length and direction of motion blur; defocus blur is obtained by substituting the motion blur function with the defocus functionh r (x, y) The result of the convolution:
Figure 424524DEST_PATH_IMAGE007
(3)
in the formularIs the defocus blur radius. The existing bird species image is expanded through motion blur and defocusing blur, a real environment is simulated, the diversity of the bird species image is enhanced, and overfitting is prevented. And expanding 100 bird species images for each bird, adding the previously collected bird species images, and constructing a bird species image data set related to bird-related faults in the transformer substation, wherein each bird comprises 300 bird species images, and 3000 bird species image samples are calculated. Labeling each image in the data set by using a tool Labeling, wherein the obtained bird species and coordinate information are prepared into a standard Pascal VOC data set format according to the following steps of 9:1 into training and test sets, 10% of the test sets beingAnd the number of the pictures in the verification set is 2430, 300 and 270 respectively.
S2, constructing an improved Yolov5 bird species target detection model: constructing four parts of an input end, a backbone segment network ConvNeXt-T for generating characteristics, a neck segment network FPN + PAN for carrying out characteristic fusion and a head segment network for carrying out prediction; updating parameters through back propagation of a loss function, wherein the loss function consists of detection frame loss, confidence coefficient loss and classification loss, and finally obtaining an improved Yolov5 bird target detection model; as shown in fig. 2, the improved YOLOv5 bird species target detection model constructed in this embodiment is composed of four parts, a first part is an input end, a second part is a backbone segment network ConvNeXt-T, a third part is a neck segment network FPN + PAN, and a fourth part is a head segment network:
the input end carries out mosaic data enhancement, adaptive Anchor frame (Anchor) calculation and adaptive picture scaling technology (Letterbox) processing on an input image to obtain an appropriate Anchor frame (Anchor) and an input image with uniform size, wherein the input size is 640 multiplied by 3, the mosaic data enhancement can enrich a data set and greatly improve the training speed of a network at the same time, the memory requirement of a model is reduced, the adaptive Anchor frame (Anchor) calculation can avoid artificially defining Anchor frames with different sizes and aspect ratios, the quality of the Anchor frame (Anchor) is improved, the adaptive picture scaling technology (Letterbox) can reduce a large amount of information redundancy, and the reasoning speed of an algorithm is improved;
replacing the backbone segment network of the original YOLOv5 with ConvNeXt-T, the backbone segment network ConvNeXt-T being composed of 1 separate 4 × 4 convolutional layer, a normative layer (LayerNorm), a ConvNeXt _1 module, a ConvNeXt _2 module, a ConvNeXt _3 module, a downsampling module i, a ConvNeXt _4 module, a ConvNeXt _5 module, a ConvNeXt _6 module, a downsampling module ii, a ConvNeXt _7 module, a ConvNeXt _8 module, a ConvNeXt _9 module, a ConvNeXt _10 module, a ConvNeXt _11 module, a convnexxt _12 module, a ConvNeXt _13 module, a convnextt _14 module, a convnextt _15 module, a downsampling module, a convnextt _16 module, a ConvNeXt _17 module, a normative layer (layerjiii) module, a global averaging layer (layerjt) and a global averaging layer (global averaging layer) in sequenceTo avoid pictures that are too large to be ambiguous, the global average pooling layer, the normalization layer (LayerNorm), and the linearization layer (Linear) are not plotted in fig. 2). As shown in fig. 3, the ConvNeXt _1 module is composed of 17 × 7 Layer-by-Layer convolutional Layer, a normative Layer (LayerNorm), 1 × 1 convolutional Layer, 1 GELU activation function, 1 × 1 convolutional Layer, scaling (Layer Scale), and a regularization Layer (Drop Path) in sequence, and adds the input and output by a residual connection; the down-sampling module (a down-sampling module I, a down-sampling module II and a down-sampling module III) consists of a standard layer (LayerNorm) and 1 convolution layer of 2 multiplied by 2. ConvNeXt _2, convNeXt _3, \8230, convNeXt _18, with the same structure as ConvNeXt _ 1; extracting first characteristics F through a ConvNeXt _3 module, a ConvNeXt _6 module, a ConvNeXt _15 module and a ConvNeXt _18 module in the backbone segment network ConvNeXt-T respectively 1 And a second feature F of 80X 192 2 And a third feature F of 40X 384 3 And a fourth feature F of 20X 768 4 The method comprises the steps of extracting the convolution characteristics of an image from shallow to deep, inputting the image into a neck segment network FPN + PAN for fusion, wherein the selected lightweight skeleton segment network ConvNeXt-T also reduces the parameter quantity of a model while keeping the characteristic quality, and only the convolution characteristic extraction is realized but not the classification function is realized;
first feature F extracted from neck segment network FPN + PAN pair 1 Second characteristic F 2 A third feature F 3 Fourth feature F 4 Carrying out characteristic fusion from deep to shallow and from shallow to deep, specifically: fourth feature F 4 Obtaining a first characteristic diagram P through a CSPN2 module and a CBL convolution block 1 First characteristic diagram P 1 After up-sampling and the second characteristic F 2 Fusing to obtain a second characteristic diagram P 2 (ii) a Second characteristic diagram P 2 Obtaining a third characteristic diagram P through a CSPN2 module and a CBL convolution block 3 Third characteristic diagram P 3 After up-sampling and the first characteristic F 1 Merging and obtaining a fourth characteristic diagram P through a CSPN2 module 4 (ii) a Fourth feature map P 4 Obtaining a fifth feature map P through the CBL convolution block 5 Fifth characteristic diagram P 5 And the third characteristic diagram P 3 Go on to meltCombining the six characteristic maps P obtained by the CSPN2 module 6 (ii) a Sixth feature map P 6 Obtaining a seventh feature map P by CBL convolution block 7 Seventh feature diagram P 7 And the third feature F 3 Merging and obtaining an eighth feature map P through a CSPN2 module 8 (ii) a As shown in fig. 4, the CSPN2 module has 2 branches, 1 branch is 2 CBL convolution blocks, 1 convolution layer, and the other 1 branch is 1 convolution layer, and the output of the two branches is subjected to feature fusion and then sequentially passes through a batch normalization layer (BN), a leak RuLU activation function and a CBL convolution block to obtain a final output, wherein the CBL convolution block is composed of a normal convolution, a batch normalization layer (BN) and a leak RuLU activation function, and a neck segment network FPN + PAN shortens an information path between features of a lower layer and a top layer, so that information of the lower layer is more easily transmitted;
three detection frames with different prediction sizes of 20 multiplied by 45, 40 multiplied by 45 and 80 multiplied by 45 exist in the head segment network, and the detection frames respectively correspond to a large detection target, a medium detection target and a small detection target; the obtained fourth feature map P with strengthened fusion 4 The sixth characteristic diagram P 6 And an eighth feature map P 8 After convolution, inputting the data into detection frames with large, medium and small prediction sizes; loss function in head segment network is lost by detection box (LossRect), loss of confidence (LossConf) and classification loss (LossCls), updating the weight and deviation by a loss function through back propagation, and finally obtaining an optimal detection frame, a most accurate confidence score and a classification result at the same time; the detection frame Loss adopts a CIOU _ Loss function, confidence coefficient Loss and classification Loss as cross entropy Loss functions, and the improved YOLOv5 bird target detection model Loss is shown as a formula (4):
Figure 916685DEST_PATH_IMAGE008
(4)
s3, training an improved Yolov5 bird species target detection model: firstly, pre-training an improved Yolov5 bird species target detection model by using a COCO public data set to obtain pre-training weights; then loading pre-training weights, and retraining the improved Yolov5 bird species target detection model by using the divided training set and verification set, wherein the training skills comprise: the method comprises the following steps of multi-stage transfer learning, mixed precision training fp16, mosaic and mixed data enhancement, label smoothing, SGDM model parameter optimization and Step length (Step) learning rate reduction, wherein the training skills can improve the speed, generalization capability and robustness of model training; secondly, evaluating the trained improved YOLOv5 bird species target detection model through a verification set during training, and optimizing parameters of the improved YOLOv5 bird species target detection model until the training is finished; finally, obtaining the optimal weight of the improved Yolov5 bird target detection model;
in the embodiment, a multistage migration learning is adopted to train an improved yollov 5 bird species target detection model, a first stage freezes a trunk to train, the number of training rounds is 50 rounds, the batch size (batch size) is 8, a second stage finely tunes all layer parameters, and the number of training rounds is 50 rounds, and the batch size (batch size) is 4; the mosaic data enhancement is set to be used at a 50% probability in each iteration, the mixed data enhancement is set to be used after the mosaic data enhancement at a 50% probability, and the value of the label smoothing is set to be 0.01; adopting SGDM model parameter optimization, setting momentum value to be 0.9, and using Step length (Step) learning rate reduction strategy, the maximum learning rate is 1 multiplied by 10 -2 The minimum learning rate is 1 × 10 -4 (ii) a The training model was evaluated through a validation set during training, set to evaluate once every 10 rounds, optimizing the model parameters.
S4, detecting the images of the test set by using the trained improved Yolov5 bird species target detection model: the method comprises the steps of loading optimal weight, utilizing a self-adaptive picture scaling technology (Letterbox) to scale a test set image to a uniform size, inputting the test set image to an improved YOLOv5 bird species target detection model for detection, selecting a positioning accurate frame through flexible non-maximum suppression (Soft-NMS), realizing the detection of a bird species target, wherein the detected image comprises bird species, specific position information and confidence coefficient of a prediction result, and carrying out statistical output on bird species quantity;
in this embodiment, the adaptive image scaling technique (lettbex) is used to scale the test set image to 640 × 640, and then the scaled test set image is input into the improved YOLOv5 bird species target detection modelDetecting, and selecting a positioning accurate frame through flexible non-maximum suppression (Soft-NMS): prediction candidate boxY r The higher the coincidence degree with the highest-score prediction candidate frame M is, the prediction candidate frame isY r The lower the score of (a); prediction candidate boxY r The coincidence degree with the highest-score prediction candidate frame M is less than the threshold valueN t Then the prediction candidate box is retainedY r
Figure 444795DEST_PATH_IMAGE009
(5)
In the formulaS i Is a set of prediction candidate box scores,Mthe candidate box is predicted for the highest score,Y r in order to predict the candidate block(s),N t for threshold, the threshold is set to 0.3 and IOU is the cross-over ratio.
In this embodiment, the trained improved YOLOv5 bird species target detection model is used to detect the test set images, so as to realize the bird classification identification related to the bird-related fault of the transformer substation. The accuracy and detection speed of the model are evaluated by mean average precision (mAP) and Frame Per Second (FPS) to evaluate the classification performance. The result shows that the average precision mean value of the detection of 300 test sets containing 10 types of bird species can reach 97.47 percent, the FPS is 23, and the detection speed is related to the experimental environment. The effectiveness of the method for identifying the bird species related to the bird-involved fault of the transformer substation based on deep learning is verified. The bird-involved fault differential prevention and control improvement reference can be developed for transformer substation operation and maintenance personnel.
The foregoing description is of the preferred embodiment of the present invention only, and is not intended to limit the invention in any way, so that those skilled in the art, having the benefit of this disclosure, may modify and/or adapt the same to equivalent embodiments without departing from the scope of the present invention. However, any simple modification, equivalent change and modification of the above embodiments according to the technical essence of the present invention are still within the protection scope of the technical solution of the present invention.

Claims (10)

1. A method for identifying bird species related to bird-related faults of a transformer substation based on deep learning is characterized by comprising the following steps:
s1, constructing a bird species image data set related to bird-involved faults of a transformer substation;
s2, constructing an improved Yolov5 bird species target detection model: constructing four parts, namely an input end, a backbone section network ConvNeXt-T for generating characteristics, a neck section network FPN + PAN for carrying out characteristic fusion and a head section network for carrying out prediction; updating parameters through back propagation of a loss function, wherein the loss function consists of detection frame loss, confidence coefficient loss and classification loss, and finally obtaining an improved Yolov5 bird target detection model;
s3, training an improved Yolov5 bird species target detection model;
and S4, detecting the test set image by using the trained improved YOLOv5 bird species target detection model.
2. The method for identifying the bird species related to the bird-related fault of the transformer substation based on the deep learning of claim 1, wherein in the step S1, for a transformer substation patrol image and a bird species image set, the bird species image is expanded by adopting motion blur and defocus blur, a transformer substation bird-related fault related bird species image data set containing a plurality of bird species is constructed, a tool Labeling is used for Labeling each image in the transformer substation bird-related fault related bird species image data set, and the obtained bird species and coordinate information are manufactured into a standard Pascal VOC data set format and are divided into a test set, a training set and a verification set.
3. The method for identifying the bird species related to the bird-involved fault in the transformer substation based on the deep learning as claimed in claim 2, wherein in the step S1, the bird species image is subjected to motion blurring through a motion blurring function:
Figure 448483DEST_PATH_IMAGE001
in the formulaf(x, y) Andg(x, y) Respectively representing the input bird species image and the motion-blurred bird species image,h(x, y) For the motion blur function of convolving the input bird species image,n(x, y) For an additive noise function, the formula for the motion blur function is expressed as:
Figure 303306DEST_PATH_IMAGE002
in the formulaLAnd
Figure 784841DEST_PATH_IMAGE003
is the length and direction of motion blur; defocus blur is obtained by substituting the motion blur function with the defocus functionh r (x, y) The result of the convolution:
Figure 117733DEST_PATH_IMAGE004
in the formularIs the defocus blur radius.
4. The deep learning-based identification method for bird-involved fault-related bird species of transformer substation according to claim 1, wherein the input end performs mosaic data enhancement, adaptive anchor frame calculation and adaptive picture scaling on the input image to obtain an appropriate anchor frame and input image with uniform size.
5. The deep learning based substation bird-related fault related bird species identification method of claim 1, wherein the backbone segment network of the original YOLOv5 is replaced by ConvNeXt-T, which in turn consists of 1 individual 4 x 4 convolutional layer, a normative layer, a ConvNeXt _1 module, a ConvNeXt _2 module, a ConvNeXt _3 module, a downsampling module i, a ConvNeXt _4 module, a ConvNeXt _5 module, a ConvNeXt _6 module, a downsampling module ii, a ConvNeXt _7 module, a ConvNeXt _8 module, a ConvNeXt _9 module, a ConvNeXt _10 module, a convnexxt _11 module, a convnexxt _12 module, a convnexxt _13 module, a convnexxt _14 module, a vvnext _15 module, a ConvNeXt _ iii, a ConvNeXt _16 module, a normative layer, a global average layer, a ConvNeXt _17 module, and a global average layer.
6. The deep learning-based substation bird-related fault related bird species identification method according to claim 5, wherein first features F are extracted through a ConvNeXt _3 module, a ConvNeXt _6 module, a ConvNeXt _15 module and a ConvNeXt _18 module in a backbone segment network ConvNeXt-T respectively 1 Second characteristic F 2 A third feature F 3 And a fourth feature F 4 And the convolution characteristic extraction of the image from shallow to deep is realized, and then the image is input to a neck segment network FPN + PAN for fusion.
7. The deep learning-based bird species identification method related to bird-involved fault in transformer substation according to claim 6, characterized in that the first feature F extracted from the neck segment network FPN + PAN pair 1 A second feature F 2 The third feature F 3 Fourth feature F 4 Carrying out characteristic fusion from deep to shallow and from shallow to deep, specifically: fourth feature F 4 Obtaining a first characteristic diagram P through a CSPN2 module and a CBL convolution block 1 First characteristic diagram P 1 After up-sampling and the second characteristic F 2 Fusing to obtain a second characteristic diagram P 2 (ii) a Second characteristic diagram P 2 Obtaining a third characteristic diagram P through a CSPN2 module and a CBL convolution block 3 Third characteristic diagram P 3 After up-sampling and the first characteristic F 1 Merging and obtaining a fourth characteristic diagram P through a CSPN2 module 4 (ii) a Fourth characteristic diagram P 4 Obtaining a fifth feature map P through the CBL convolution block 5 Fifth characteristic diagram P 5 And the third characteristic diagram P 3 Performing fusion and obtaining a sixth characteristic diagram P through a CSPN2 module 6 (ii) a Sixth feature map P 6 Obtaining a seventh feature map P by CBL convolution block 7 Seventh feature diagram P 7 And the third feature F 3 Merging and obtaining an eighth feature map P through a CSPN2 module 8
8. The deep learning-based method for identifying bird related faults of transformer substations according to claim 7, wherein three detection frames with different prediction sizes are present in the head section network, and are respectively corresponding to prediction of large, medium and small detection targets; the obtained fourth feature map P for enhancing fusion 4 The sixth characteristic diagram P 6 And an eighth feature map P 8 After convolution, the data are input into detection boxes with large, medium and small prediction sizes.
9. The deep learning-based identification method for bird-involved fault-related bird species of transformer substation according to claim 8, wherein in the head segment network, the loss function updates the weight and the deviation through back propagation, and finally, the optimal detection frame and the most accurate confidence score and classification result are obtained simultaneously; the detection frame Loss adopts a CIOU _ Loss function, confidence coefficient Loss and classification Loss as cross entropy Loss functions.
10. The method for identifying bird species related to bird-involved fault in the transformer substation based on deep learning of claim 9, wherein in step S4, optimal weights are loaded, the image of the test set is scaled to a uniform size by using an adaptive picture scaling technique and then input to an improved YOLOv5 bird species target detection model for detection, and an accurate positioning frame is selected by suppressing a flexible non-maximum value to realize detection of bird species targets, the detected image includes bird species, specific position information and confidence of prediction results, and statistical output is performed on the number of birds.
CN202211386862.3A 2022-11-07 2022-11-07 Transformer substation bird-involved fault related bird species identification method based on deep learning Pending CN115620107A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211386862.3A CN115620107A (en) 2022-11-07 2022-11-07 Transformer substation bird-involved fault related bird species identification method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211386862.3A CN115620107A (en) 2022-11-07 2022-11-07 Transformer substation bird-involved fault related bird species identification method based on deep learning

Publications (1)

Publication Number Publication Date
CN115620107A true CN115620107A (en) 2023-01-17

Family

ID=84879155

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211386862.3A Pending CN115620107A (en) 2022-11-07 2022-11-07 Transformer substation bird-involved fault related bird species identification method based on deep learning

Country Status (1)

Country Link
CN (1) CN115620107A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116229380A (en) * 2023-05-10 2023-06-06 国网江西省电力有限公司电力科学研究院 Method for identifying bird species related to bird-related faults of transformer substation
CN116912890A (en) * 2023-09-14 2023-10-20 国网江苏省电力有限公司常州供电分公司 Method and device for detecting birds in transformer substation
CN117095411A (en) * 2023-10-16 2023-11-21 青岛文达通科技股份有限公司 Detection method and system based on image fault recognition
CN117874173A (en) * 2024-03-11 2024-04-12 腾讯科技(深圳)有限公司 Training method and related device of vector model

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111104906A (en) * 2019-12-19 2020-05-05 南京工程学院 Transmission tower bird nest fault detection method based on YOLO
CN112215822A (en) * 2020-10-13 2021-01-12 北京中电兴发科技有限公司 Face image quality evaluation method based on lightweight regression network
CN113486865A (en) * 2021-09-03 2021-10-08 国网江西省电力有限公司电力科学研究院 Power transmission line suspended foreign object target detection method based on deep learning
CN114842208A (en) * 2022-06-08 2022-08-02 南昌大学 Power grid harmful bird species target detection method based on deep learning
CN115082401A (en) * 2022-06-22 2022-09-20 桂林电子科技大学 SMT production line chip mounter fault prediction method based on improved YOLOX and PNN

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111104906A (en) * 2019-12-19 2020-05-05 南京工程学院 Transmission tower bird nest fault detection method based on YOLO
CN112215822A (en) * 2020-10-13 2021-01-12 北京中电兴发科技有限公司 Face image quality evaluation method based on lightweight regression network
CN113486865A (en) * 2021-09-03 2021-10-08 国网江西省电力有限公司电力科学研究院 Power transmission line suspended foreign object target detection method based on deep learning
CN114842208A (en) * 2022-06-08 2022-08-02 南昌大学 Power grid harmful bird species target detection method based on deep learning
CN115082401A (en) * 2022-06-22 2022-09-20 桂林电子科技大学 SMT production line chip mounter fault prediction method based on improved YOLOX and PNN

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
TIANYU WANG等: "An improved YOLOv5-based bird detection algorithm for transmission line" *
孙卫红等: "基于MSRCR与注意力机制的群体蚕茧智能识别算法" *
李悦言等: "一种改进的轻量化目标检测算法" *
陈金广等: "改进YOLOv5网络的轻量级服装目标检测方法" *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116229380A (en) * 2023-05-10 2023-06-06 国网江西省电力有限公司电力科学研究院 Method for identifying bird species related to bird-related faults of transformer substation
CN116229380B (en) * 2023-05-10 2023-09-12 国网江西省电力有限公司电力科学研究院 Method for identifying bird species related to bird-related faults of transformer substation
CN116912890A (en) * 2023-09-14 2023-10-20 国网江苏省电力有限公司常州供电分公司 Method and device for detecting birds in transformer substation
CN116912890B (en) * 2023-09-14 2023-11-24 国网江苏省电力有限公司常州供电分公司 Method and device for detecting birds in transformer substation
CN117095411A (en) * 2023-10-16 2023-11-21 青岛文达通科技股份有限公司 Detection method and system based on image fault recognition
CN117095411B (en) * 2023-10-16 2024-01-23 青岛文达通科技股份有限公司 Detection method and system based on image fault recognition
CN117874173A (en) * 2024-03-11 2024-04-12 腾讯科技(深圳)有限公司 Training method and related device of vector model

Similar Documents

Publication Publication Date Title
CN115620107A (en) Transformer substation bird-involved fault related bird species identification method based on deep learning
CN113486865B (en) Power transmission line suspended foreign object target detection method based on deep learning
CN114627360B (en) Substation equipment defect identification method based on cascade detection model
CN110968866B (en) Defense method for resisting attack for deep reinforcement learning model
CN109446925A (en) A kind of electric device maintenance algorithm based on convolutional neural networks
CN115167442A (en) Power transmission line inspection path planning method and system
CN115861853A (en) Transmission line bird nest detection method in complex environment based on improved yolox algorithm
CN114943689A (en) Method for detecting components of steel cold-rolling annealing furnace based on semi-supervised learning
CN115019302A (en) Improved YOLOX target detection model construction method and application thereof
CN115410087A (en) Transmission line foreign matter detection method based on improved YOLOv4
CN112861646B (en) Cascade detection method for oil unloading worker safety helmet in complex environment small target recognition scene
Yan et al. Learning probabilistic multi-modal actor models for vision-based robotic grasping
CN113011308A (en) Pedestrian detection method introducing attention mechanism
CN115690730A (en) High-speed rail contact net foreign matter detection method and system based on single classification and abnormal generation
CN114862768A (en) Improved YOLOv5-LITE lightweight-based power distribution assembly defect identification method
CN114677357A (en) Model, method and equipment for detecting self-explosion defect of aerial photographing insulator and storage medium
CN117974985A (en) High-voltage line foreign matter detection method and system based on improvement YOLOv s
CN114359167A (en) Insulator defect detection method based on lightweight YOLOv4 in complex scene
CN112837281A (en) Pin defect identification method, device and equipment based on cascade convolutional neural network
CN116152699B (en) Real-time moving target detection method for hydropower plant video monitoring system
CN110110651B (en) Method for identifying behaviors in video based on space-time importance and 3D CNN
CN116385950A (en) Electric power line hidden danger target detection method under small sample condition
CN112199980A (en) Overhead line robot obstacle identification method
CN110991366A (en) Shipping monitoring event identification method and system based on three-dimensional residual error network
CN112233071A (en) Multi-granularity hidden danger detection method and system based on power transmission network picture in complex environment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20230117