CN114998124A - Image sharpening processing method for target detection - Google Patents

Image sharpening processing method for target detection Download PDF

Info

Publication number
CN114998124A
CN114998124A CN202210567780.2A CN202210567780A CN114998124A CN 114998124 A CN114998124 A CN 114998124A CN 202210567780 A CN202210567780 A CN 202210567780A CN 114998124 A CN114998124 A CN 114998124A
Authority
CN
China
Prior art keywords
image
loss
generator
domain
degraded
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210567780.2A
Other languages
Chinese (zh)
Inventor
胡海苗
赵子琛
黎世豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN202210567780.2A priority Critical patent/CN114998124A/en
Publication of CN114998124A publication Critical patent/CN114998124A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • G06T5/73
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

An image sharpening processing method for target detection comprises the following steps: training loop generation confrontation network, obtaining generator 1 from degraded image domain to clear image domain (S1); model parameters are loaded to a new generator 3 from a generator with training convergence, a confrontation network is generated by training degraded images with target labeling information and clear image sample data, a generator is obtained (S2), wherein the generator model parameters are restricted through the detection loss of a detector and the confrontation loss of the extracted features of the detector, the domain migration of a target semantic level is realized, and the generator model parameters are restricted through the loss of bottom visual features such as color, scattering, gradient and the like, so that the reduction of the bottom visual features of the degraded images is realized. The invention not only improves the image visual quality, but also obviously improves the performance of the subsequent target detection algorithm, can be used as a preprocessing plug-in of a video monitoring system, and provides a clear image for the subsequent target detection algorithm as input.

Description

Image sharpening processing method for target detection
Technical Field
The invention relates to an image sharpening processing method for target detection, and belongs to the technical field of digital image processing.
Background
With the wide application of video monitoring, a large amount of video data is generated every day, a video monitoring system gradually develops towards the direction of intellectualization, and high-level computer vision tasks such as target detection and the like play an increasingly greater role as artificial assistance. However, intelligent algorithms such as object detection generally require high quality input images. The imaging quality of the outdoor monitoring system is easily affected by a plurality of degradation factors: for example, at night, the ambient light illumination is low, and the scene reflected light signal is weak, so that the imaging contrast is reduced; in foggy weather, the scene signal is attenuated by atmospheric particle scattering in the air propagation process and mixed into atmospheric light, so that the imaging contrast is reduced. This limits the effective application scenarios of the intelligent algorithm. Therefore, efforts are made to enhance or restore degraded images by image sharpening methods to provide sufficient information for monitoring systems.
The existing image sharpening methods mainly comprise methods based on physical models and data driving. The method based on the physical model depends on physical modeling of specific degradation factors, and the specific degradation factors are removed by solving characteristic components in the physical model and reversely deducing the physical model. For example, at night, the illumination component is solved according to the theory of the retinal cortex, and then the brightness is remapped; and solving the illumination component and the transmissivity component during dense fog, and then performing reverse thrust according to the atmospheric scattering model. The method has a good visual effect due to deep knowledge of the degradation process, but still has a plurality of degradation factors which can not be modeled at present. The data-driven based approach relies primarily on generative models, with removal of the degradation process being achieved in a supervised or semi-supervised data-driven manner, with the degradation process being modeled implicitly by the design of the loss function.
It can be seen that the existing image sharpening method mainly restores or enhances the weakened bottom layer visual characteristics such as color, edge, texture and the like so as to improve the visual quality of human eyes; but neglects the reduction of high-level semantic features, and therefore has limited effect on the improvement of subsequent high-level visual tasks such as target detection. Therefore, the performance of intelligent algorithms such as target detection and the like on various degraded images is limited, and the intelligent performance of the whole video monitoring system is weakened.
Disclosure of Invention
Under the background, the invention considers the improvement effect on the performance of the target detection task, restores the degraded image from the aspects of scene perception, target semantics and bottom visual characteristics, further provides an image sharpening processing method for target detection, realizes sharpening of images under different severe environments such as dense fog, low illumination and the like, improves the performance of the target detection task under the scenes, and has important significance for an intelligent monitoring system.
The invention aims to improve the performance of the target detection task by image sharpening processing. An image sharpening processing method facing target detection is provided, and the sharpening problem of a degraded image under the influence of dense fog, low illumination and the like is converted into a domain adaptation problem from a degraded image domain A to a clear image domain B. And (3) by designing a constraint network model through a loss function, realizing the reduction of the degraded image on the level of bottom visual features and high-level semantic features so as to improve the performance of a subsequent target detection task on a clear processing result.
According to one aspect of the present invention, there is provided an image sharpening processing method for object detection, comprising:
A) training cycle generation countermeasure network to obtain scene domain adaptation generator G from degraded image domain A to clear image domain B 1 The loop generation countermeasure network model comprises two generators G with the same structure 1 、G 2 Two discriminators D of the same construction 1 ,D 2 The method comprises the following steps:
A1) selecting a sample image from the degraded image domain A and the clear image domain B;
A2) to generator G 1 And discriminator D 1 G, generator 2 And discriminator D 2 Performing countermeasure training, and performing multiple rounds of iterative training to obtain a generator G with convergence 1 Carrying out domain migration on the degraded image from a scene perception level;
B) training the Generation of an antagonistic network model, from the Generator G which has been trained to converge 1 In the network, network parameters are loaded, and then the determined detector is aimed atTraining generator G with degraded image domain and clear image domain sample images labeled with target positions and categories 3 Wherein the generating of the countermeasure network model includes a generator G 3 And a discriminator D 3 The method comprises the following steps:
B1) by generators G 3 To degrade the sample image a in the image domain A o For input, an image a is generated 2 And combining the images a 2 Sending the predicted frame loss value and the classification loss value of the target detector and the head feature F of the feature extraction module to a discriminator D 3 Obtaining the countermeasure loss, forming a pair generator G with the predicted frame loss, the classification loss and the countermeasure loss 3 Constraint at the semantic level;
B2) by generators G 3 To degrade the sample image a in the image domain A o For input, an image a is generated 2 And to a 2 Extracting visual characteristics of bottom layer including color, scattering and gradient to obtain loss function value, and forming pair generator G 3 The bottom characteristic layer is restrained to ensure the human eye visual quality of the generated image;
B3) selecting sample images from the degraded image domain A and the clear image domain B, and generating the image 3 And discriminator D 3 Training is carried out, and a generator G for convergence is obtained through multiple rounds of iteration 3 The degraded image is restored from the level of target semantics and bottom visual features;
C) the testing phase takes the degraded image as a generator G 3 To obtain the sharpening process result J of the present invention.
Drawings
Fig. 1 is a schematic network structure diagram of an image sharpening processing method for object detection according to an embodiment of the present invention.
FIG. 2 is a generator G according to one embodiment of the present invention 1 、G 2 、G 3 All generators have the same lightweight network structure.
FIG. 3 is a decision according to one embodiment of the inventionPin D 1 、D 2 、D 3 The network structure of (1) is schematic, and all the discriminators have the same lightweight network structure.
FIGS. 4(a) to 4(c) are experimental comparisons of an image sharpening method according to an embodiment of the present invention with a typical image sharpening method in the prior art on a test image, and boxes on the experimental comparisons indicate target detection results of the Faster RCNN detector; fig. 4(a) shows a dense fog image and a low-illumination image, fig. 4(b) shows a result of sharpening the dense fog image by the scatter-removal algorithm CEP and a result of sharpening the low-illumination image by the low-illumination reduction algorithm KinD, respectively, and fig. 4(c) shows a sharpened image obtained by the method of the present invention.
Detailed Description
According to one aspect of the invention, an image sharpening processing method for object detection is provided. The method realizes the improvement of the performance of the target detection algorithm by the sharpening process, and the sharpened image obtained by the method has the advantages of good human visual quality and accurate detection of the machine target.
In order to realize the image sharpening for the target detection, the invention provides an image sharpening processing method for the target detection, which is characterized by comprising the following steps:
A) training cycle generation countermeasure network to obtain scene domain adaptation generator G from degraded image domain A to clear image domain B 1 Wherein the cyclic generation countermeasure network model comprises two structurally identical generators G 1 、G 2 Two discriminators D of the same construction 1 ,D 2 The method comprises the following steps:
A1) selecting a sample image from the degraded image domain A and the clear image domain B;
A2) to generator G 1 And discriminator D 1 Generator G 2 And discriminator D 2 Performing countermeasure training, and performing multiple rounds of iterative training to obtain a generator G with convergence 1 Carrying out domain migration on the degraded image from a scene perception level;
B) training the Generation of an antagonistic network model, from the Generator G which has been trained to converge 1 Mid-loading network parametersThen, for the determined detector, training generator G with sample image of degraded image domain and clear image domain labeled with target position and category 3 Wherein said generating a countermeasure network model comprises a generator G 3 And a discriminator D 3 The method comprises the following steps:
B1) by generators G 3 To degrade the sample image a in the image domain A o For input, an image a is generated 2 And combining the images a 2 Sending the predicted frame loss value and the classification loss value of the target detector and the head feature F of the feature extraction module to a discriminator D 3 Obtaining the countermeasure loss, forming a pair generator G with the predicted frame loss, the classification loss and the countermeasure loss 3 Constraint at the semantic level;
B2) by generators G 3 To degrade the sample image a in the image domain A o For input, an image a is generated 2 And to a 2 Extracting visual characteristics of bottom layer including color, scattering and gradient to obtain loss function value, and forming pair generator G 3 Constraining at the bottom characteristic layer to ensure the human visual quality of the generated image;
B3) selecting sample images from the degraded image domain A and the clear image domain B, and generating the image 3 And discriminator D 3 Training is carried out, and a convergent generator G is obtained through multiple rounds of iteration 3 The degraded image is restored from the level of target semantics and bottom visual features;
C) using degraded image as generator G in test stage 3 To obtain the sharpening process result J of the present invention.
According to a further embodiment of the invention, in the step a) above, the two image domains are non-paired degraded image and clear image, and the training loop generates a confrontation model to obtain the generator G from the degraded image domain to the clear image domain 1 The method has the main function of zooming in the feature difference of the whole perception level of the scene and taking the feature difference as the pre-training weight of the subsequent target semantic level domain migration.
According to the inventionIn a further embodiment of the above step a), as shown in fig. 2, a generator G 1 And G 2 The network structure is the same, only include 2 convolutional layers and 6 residual convolutional layers, the convolutional kernel size is 3, the number of channels is 64, the filling mode is refelection pad2d, except that the last 1 convolutional layers are activated by Tanh and have no normalization layer, the other convolutional layer activation functions are LeakRelu, and normalization is adopted. Generator G 1 With sample image a in degraded image domain A o For input, an image a is generated which approximates a sharp image field B 1 (ii) a Generator G 2 With a sample image B in the clear image domain B o For input, an image b is generated that approximates the degraded image domain A 1 . As shown in fig. 3, a discriminator D 1 And D 2 The network structure is the same, including only 6 convolutional layers, the convolutional kernel size is 4, the step size of the first 4 convolutional layers is 2, and the step size of the last 2 convolutional layers is 1. Discriminator D 1 With a generator G 1 Generated image a of 1 And a clear image field image b o For input, each pixel point value in the output single-channel image indicates whether the image belongs to a clear image domain B; discriminator D 2 With a generator G 2 Generated image b 1 And degraded image domain image a o For input, each pixel point value in the output single-channel image indicates whether the image belongs to the degraded image domain A or not.
According to a further embodiment of the present invention, in the above step a2), two sets of generators \ discriminators optimize the network parameters by opposing training loss: the optimization goal of the discriminator is to be able to accurately distinguish whether the input image belongs to the sharp image domain B or the degraded image domain a, and the optimization goal of the generator is to generate an image that can confuse the discriminator. The generator G is listed as formula (1) 1 And a discriminator D 1 Loss function of resistance training, D 1 For G 1 The closer the generated image discrimination result is to the clear image region B, the more G is indicated 1 The better the generation, the smaller the loss function value, corresponding to D 1 The larger the loss function value of (D), the more D 1 The accuracy is not sufficient.
Loss g1 =E z~A [log(D 1 (G 1 (z)))] (1)
Loss d1 =E x~B [logD 1 (x)]+E z~A [log(1-D 1 (G 1 (z)))]
Where B and a represent the sharp image set and the degraded image set, respectively. Clear image samples in the training data are from the VOC2007 data set, dense fog degraded image samples are from the RTTS data set, and low-illumination degraded image samples are from the Darkface data set. The model parameter optimizer is Adam, the learning rate is 1e-4, and the model parameter initialization mode is Xavier. Through counterstudy and mutual constraint between the generator and the discriminator, the generator can generate images which are closer to a target domain, the discriminator can more accurately discriminate the domain to which the images belong, and the model convergence of about 50 rounds of iterative training is achieved to obtain a generator G 1 And carrying out domain migration on the degraded image from a perception level.
According to a further embodiment of the present invention, in the step B), the two input image domains are a degraded image and a clear image with target position and category labels, and the loss function constraint mainly includes two aspects: through the target semantic feature constraint and the unsupervised constraint of the underlying visual feature introduced by the fast RCNN detector, as shown in formula (2), the loss functions of the two parts respectively form a generator G from the two aspects of target semantic and visual quality 3 The optimization objective of (1). Meanwhile, the content of the invention is explained by taking the Faster RCNN as a detection module, but the invention is not only directed to one detector of the Faster RCNN, and has a remarkable improvement effect on the performances of different detectors such as Yolo and the like.
Loss=Loss con +Loss low (2)
Therein, Loss con Is Loss of target semantics, Loss low Is a loss of underlying visual features.
According to a further embodiment of the invention, in the above step B), the generator G 3 The network structure of (2) is shown in FIG. 2, the input is a degraded image, the output is a generated clarified image, and a discriminator D 3 In a network structure such asAs shown in FIG. 3, the input is G 3 And (3) enabling the generated image and clear image domain image samples to be in a head characteristic diagram of a Faster RCNN characteristic extraction module, and enabling each pixel point value in an output single-channel image to indicate whether the image belongs to a clear image domain B or not.
According to a further embodiment of the present invention, in the above step B1), the domain migration at semantic level is trained using degraded sample images with target location labeling, and the loss function is defined as shown in formula (3):
Loss con =Loss detect +Loss g3 (3)
therein, loss detect And Loss g3 The detection loss of the detector and the countering loss of the detector characteristic, respectively, are defined as follows:
Loss detect =Loss cls (G 3 (z))+Loss loc (G 3 (z)) (4)
wherein, G 3 (z) generator G with degraded image as input 3 Output image of (2), Loss cls And Loss loc The target classification penalty and the prediction box offset penalty of fast RCNN, respectively.
Loss g3 =E z~N1 [log(D 3 (F(G 3 (z))))] (5)
Wherein, F represents the 7 th layer feature map of the Faster RCNN feature extraction module, and the dimension is 512.
According to a further embodiment of the present invention, in the above step B2), the visual quality is restored mainly by the constraints of the underlying visual features such as color, scattering, gradient, etc., so that the resulting image of the generator is more consistent with the visual quality of human eyes, and the loss function is defined as shown in equation (6):
loss low =loss recon +α·loss color +β·loss haze +γ·loss edge (6)
therein, loss recon 、loss color 、loss haze 、loss edge The image is restored from the aspects of structure, color, scattering and edge gradient features respectivelyThe equation is constrained, α is 8, β and γ are both 0.1, and the respective loss functions are defined as follows:
Figure BDA0003658162830000071
wherein the content of the first and second substances,
Figure BDA0003658162830000072
and
Figure BDA0003658162830000073
respectively representing degraded images and G 3 Each pixel position in the generated image is at the maximum of three channels, and the L1 loss constraint network generation result and the original image are structurally consistent.
Figure BDA0003658162830000074
Wherein, the first and the second end of the pipe are connected with each other,
Figure BDA0003658162830000075
respectively generate an image G 3 (z) pixel mean at three color channels. According to the gray world assumption, the three-channel mean values of the white-balanced image are approximately equal, and the color cast problem of the whole scene is restrained by using the difference of the maximum and minimum values in the three-channel mean values.
loss haz4 =-var(Gray(G 3 (z))-Gray mean )/Gray std ) (9)
Wherein, Gray (G) 3 (z)) is the Gray scale map, Gray, that generates the image transitions mean And Gray std Mean and variance, Gray (G), respectively, of the grayscale image 3 (z))-Gray mean )/Gray std The gray scale map is normalized and var represents the variance of the entire image. The texture is constrained using the penalty function, which is smaller the richer the texture.
Figure BDA0003658162830000081
Wherein, grad represents the image gradient, and the edge of the generated image which keeps clear is kept by restricting the mean value of the gradient.
According to a further embodiment of the invention, in step B3), the generator G 3 And optimizing model parameters through target semantic loss and bottom visual characteristic loss, optimizing the model parameters through the confrontation loss of detector characteristics by the discriminator, wherein clear image samples in the training data come from a VOC2007 data set, dense fog degraded image samples come from an RTTS data set, and low-illumination degraded image samples come from a DarkFace data set. Adam as model parameter optimizer, 2e-5 as learning rate, and G as generator 3 Directly from the generator G 1 Medium load, discriminator D 3 The model parameters are initialized by using an Xavier mode, iterative training is carried out for about 80 model convergence, and a generator G is obtained 3 And the clarity of the degraded image is realized from the level of target semantics and bottom visual characteristics.
According to a further embodiment of the invention, in step C), the testing phase directly takes the degraded image as generator G 3 And inputting the model, and using the model inference result as a sharpening processing image J.
Therefore, the invention provides an image sharpening processing method for target detection. The method realizes the domain adaptation from the degraded image domain to the clear image domain from the scene perception and target semantic level, thereby closing the difference between target characteristics in the degraded image and the clear image and improving the performance of a target detection task. Meanwhile, the constraint of the bottom layer visual characteristics also enables the clear image obtained by the method to have better human visual quality. Therefore, the method is suitable for degraded images under various complex illumination and scattering conditions, can be used as a preprocessing module to provide a clear image for the subsequent target detection algorithm to call, and improves the performance of the target detection algorithm.
The image sharpening processing method for target detection provided by the invention constitutes substantial significant improvement on the existing image sharpening method. Compared with the prior art, the method has the beneficial effects of bringing the difference between the degraded image and the clear image at the scene perception and target semantic level closer, and obviously improving the performance of tasks such as target detection and the like.
Fig. 4 shows the comparison of the test results of the sharpening method of the present invention with the test results of the prior sharpening method (note that the original figures shown in fig. 4(a) to 4(c) are all in color), and the boxes on the figure are the detection results of the fast RCNN target detector. Fig. 4(a) shows a dense fog image and a low illumination image, the edge of the object, etc. is severely attenuated due to strong scattering and weak signals, resulting in the performance of the object detector being severely affected; fig. 4(b) is a result of the scattering removal algorithm CEP for clarifying the dense fog image and a result of the low-illumination reduction algorithm KinD for clarifying the low-illumination image, which respectively improve the detection accuracy, but the improvement effect is very limited; fig. 4(c) is the image sharpening processing result of the present invention, and it can be seen that many objects are detected by the detector, showing that the image sharpening method of the present invention significantly improves the object detection performance.
It is to be understood that the above disclosure is only illustrative of specific embodiments of the invention. According to the technical idea provided by the invention, the changes which can be thought by the ordinary skilled person in the field shall fall into the protection scope of the invention.

Claims (3)

1. An image sharpening processing method for target detection is characterized by comprising the following steps:
A) training cycle generation countermeasure network to obtain scene domain adaptation generator G from degraded image domain A to clear image domain B 1 Wherein the loop-generated countermeasure network model includes two structurally identical generators G 1 、G 2 Two discriminators D of the same construction 1 ,D 2 The method comprises the following steps:
A1) selecting a sample image from the degraded image domain A and the clear image domain B;
A2) to generator G 1 And discriminator D 1 G, generator 2 And discriminator D 2 Performing countermeasure training, and performing multiple rounds of iterative training to obtain a generator G with convergence 1 Face down from scene perception levelCarrying out domain migration on the texture image;
B) training the Generation of an antagonistic network model, from the Generator G which has been trained to converge 1 Then, aiming at the determined detector, a generator G is trained by the sample images with the degraded image domain and the clear image domain labeled by the target position and the category 3 Wherein: the generating of the antagonistic network model comprises a generator G 3 And a discriminator D 3 The method comprises the following steps:
B1) by generators G 3 To degrade the sample image a in the image domain A o For input, an image a is generated 2 And combining the images a 2 Sending the predicted frame loss value and the classification loss value of the target detector and the head feature F of the feature extraction module to a discriminator D 3 Obtaining the countermeasure loss, forming a pair generator G with the predicted frame loss, the classification loss and the countermeasure loss 3 Constraint at the semantic level;
B2) by generators G 3 To degrade the sample image a in the image domain A o For input, an image a is generated 2 And to a 2 Extracting bottom visual characteristics including color, scattering and gradient to obtain loss function value, and forming a pair generator G 3 The bottom characteristic layer is restrained to ensure the human eye visual quality of the generated image;
B3) selecting sample images from the degraded image domain A and the clear image domain B to the generator G 3 And discriminator D 3 Training is carried out, and a generator G for convergence is obtained through multiple rounds of iteration 3 The degraded image is restored from the level of target semantics and bottom visual features;
C) using degraded image as generator G in test stage 3 To obtain the sharpening process result J of the present invention.
2. The image sharpening processing method according to claim 1, wherein:
in the step A), the input two image fields are unpaired degraded image and clear image, and a training cycle is generatedGenerator G for forming a confrontational model to obtain a range from a degraded image domain to a clear image domain 1 Generator G 1 And G 2 The network structure is the same, including 2 convolutional layers and 6 residual convolutional layers, the convolutional kernel size is 3, the number of channels is 64, the filling mode is refelection pad2d, except that the last 1 convolutional layer is activated by Tanh and has no normalization layer, the activation functions of the other convolutional layers are LeakRelu, and the normalization is adopted. Generator G 1 By degrading the sample image a in the image domain A o For input, an image a is generated which approximates a sharp image field B 1 (ii) a Generator G 2 With a sample image B in the clear image domain B o For input, an image b is generated that approximates the degraded image domain A 1 . Discriminator D 1 And D 2 The network structure is the same, including only 6 convolutional layers, the convolutional kernel size is 4, the step size of the first 4 convolutional layers is 2, and the step size of the last 2 convolutional layers is 1. Discriminator D 1 With a generator G 1 Generated image a of 1 And a clear image field image b o For input, each pixel point value in the output single-channel image indicates whether the image belongs to a clear image domain B; discriminator D 2 With a generator G 2 Generated image b 1 And degraded image domain image a o For inputting, each pixel point value in the outputted single-channel image indicates whether the image belongs to the degraded image domain A or not,
in the step a2), network parameters are optimized between two groups of generators \ discriminators through the training loss: the optimization goal of the discriminator is to be able to accurately distinguish whether the input image belongs to the sharp image domain B or the degraded image domain a, and the optimization goal of the generator is to generate an image that can confuse the discriminator. Equation (1) lists the generator G 1 And a discriminator D 1 Loss function against training:
Loss g1 =E z~A [log(D 1 (G 1 (z)))] (1)
Loss d1 =E x~B [logD 1 (x)]+E z~A [log(1-D 1 (G 1 (z)))]
where B and a represent the sharp image and the degraded image set, respectively. Clarity in training dataThe image samples are from a VOC2007 dataset, the dense fog degraded image samples are from an RTTS dataset, and the low-illumination degraded image samples are from a DarkFace dataset. The model parameter optimizer is Adam, the learning rate is 1e-4, the model parameter initialization mode is Xavier, the iterative training is about 50 model convergence rounds, and a generator G is obtained 1 Performing domain migration on the degraded image from the perception level,
in the step B), the input two image domains are a degraded image (such as dense fog or low illumination) with target position and category labels and a clear image, and the loss function constraint comprises two aspects: through target semantic feature constraint and unsupervised constraint of underlying visual features introduced by the Faster RCNN detector, as shown in formula (2), a generator G is formed from two aspects of target semantics and visual quality respectively 3 The optimization goal of the invention is to be noted, meanwhile, the content of the invention is explained by taking the fast RCNN as a detection module, but the invention not only aims at one detector of the fast RCNN, but also has significant improvement effect on the performances of different detectors such as Yolo and the like,
Loss=Loss con +Loss low (2)
therein, Loss con Is Loss of target semantics, Loss low Is loss of underlying visual features, generator G 3 Input as a degraded image, output as a generated sharpened image, and a discriminator D 3 The input is G 3 The generated image and clear image domain image samples are subjected to head feature map extraction by a fast RCNN feature extraction module, each pixel value in an output single-channel image indicates whether the image belongs to a clear image domain B or not,
in the above step B1), the domain migration at semantic level is trained using degraded sample images with target location labeling,
Loss con =Loss detect +Loss g3 (3)
therein, Loss detect And Loss g3 Respectively detection loss and counter-loss of detector characteristics,
Loss detect =Loss cls (G 3 (z))+Loss loc (G 3 (z)) (4)
wherein G is 3 (z) generator G with degraded image as input 3 Output image of (2), Loss cls And Loss loc Respectively target classification penalty and prediction box offset penalty for fast RCNN,
Loss g3 =E z~N1 [log(D 3 (F(G 3 (z))))] (5)
wherein F represents a 7-layer feature diagram of the fast RCNN feature extraction module, is 512-dimensional,
in the step B2), the reduction of the visual quality depends on the constraints of the bottom layer visual features such as color, scattering, gradient, etc., so that the resulting image of the generator better conforms to the visual quality of human eyes, and the loss function is defined as follows:
loss low =loss recon +α·loss color +β·loss haze +γ·loss edge (6)
among them, loss recon 、loss color 、loss haze 、loss edge The image restoration process is constrained in terms of structure, color, scattering and edge gradient features, respectively, α is 8, β and γ are both 0.1, and the respective loss functions are defined as follows:
Figure FDA0003658162820000041
wherein the content of the first and second substances,
Figure FDA0003658162820000042
and
Figure FDA0003658162820000043
respectively representing degraded images and G 3 The maximum value of each pixel position in the generated image in three channels, the L1 loss constraint network generation result and the original image are structurally consistent,
Figure FDA0003658162820000044
wherein, the first and the second end of the pipe are connected with each other,
Figure FDA0003658162820000045
respectively generate an image G 3 (z) in the pixel mean values of the three color channels, according to the gray world assumption, the three channel mean values of the white balanced image are approximately equal, the difference of the maximum value and the minimum value in the three channel mean values is used for restricting the color cast problem of the whole scene,
loss haze =-var(Gray(G 3 (z))-Gray mean )/Gray std ) (9)
wherein, Gray (G) 3 (z)) is the Gray map, Gray, that generates the image transitions mean And Gray std Mean and variance, Gray (G), respectively, of the grayscale image 3 (z))-Gray mean )/Gray std The gray level image is normalized, var represents the variance of the whole image, the loss function is used for constraining texture, the loss function is smaller when the texture is richer,
Figure FDA0003658162820000046
wherein, grad represents the image gradient, the mean value of the constrained gradient ensures that the result image keeps clear edge,
in said step B3), the generator G 3 Optimizing model parameters through target semantic loss and bottom visual characteristic loss, optimizing the model parameters through the confrontation loss of detector characteristics by a discriminator, wherein clear image samples in training data come from a VOC2007 data set, dense fog degraded image samples come from an RTTS data set, low-illumination degraded image samples come from a DarkFace data set, the model parameter optimizer is Adam, the learning rate is 2e-5, and a generator G 3 Directly from the generator G 1 Medium load, discriminator D 3 The model parameters are initialized by using an Xavier mode, iterative training is carried out for about 80 model convergence, and a generator G is obtained 3 The clarity of the degraded image is realized from the level of target semantics and bottom visual characteristics,
said step C) In the testing stage, the degraded image is directly used as the generator G 3 And inputting the model, and using the model inference result as a sharpening processing image J.
3. The method for image sharpening processing for object detection according to claim 2, characterized in that:
the steps A2), B1) and B2) are respectively designed with different loss functions to constrain the network model parameters, so as to realize the domain adaptation of the scene perception level, the domain adaptation of the target semantic level and the restoration of the bottom visual feature level.
CN202210567780.2A 2022-05-23 2022-05-23 Image sharpening processing method for target detection Pending CN114998124A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210567780.2A CN114998124A (en) 2022-05-23 2022-05-23 Image sharpening processing method for target detection

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210567780.2A CN114998124A (en) 2022-05-23 2022-05-23 Image sharpening processing method for target detection

Publications (1)

Publication Number Publication Date
CN114998124A true CN114998124A (en) 2022-09-02

Family

ID=83027367

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210567780.2A Pending CN114998124A (en) 2022-05-23 2022-05-23 Image sharpening processing method for target detection

Country Status (1)

Country Link
CN (1) CN114998124A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115424119A (en) * 2022-11-04 2022-12-02 之江实验室 Semantic fractal-based interpretable GAN image generation training method and device

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115424119A (en) * 2022-11-04 2022-12-02 之江实验室 Semantic fractal-based interpretable GAN image generation training method and device

Similar Documents

Publication Publication Date Title
Li et al. Single image dehazing via conditional generative adversarial network
CN110472627B (en) End-to-end SAR image recognition method, device and storage medium
CN109685072B (en) Composite degraded image high-quality reconstruction method based on generation countermeasure network
CN111209952B (en) Underwater target detection method based on improved SSD and migration learning
CN113554089B (en) Image classification countermeasure sample defense method and system and data processing terminal
CN106599883B (en) CNN-based multilayer image semantic face recognition method
CN111145123B (en) Image denoising method based on U-Net fusion retention details
CN111915525A (en) Low-illumination image enhancement method based on improved depth separable generation countermeasure network
Zhang et al. Underwater image enhancement using improved generative adversarial network
CN114998124A (en) Image sharpening processing method for target detection
Zhao et al. A multi-scale U-shaped attention network-based GAN method for single image dehazing
CN111160354B (en) Ship image segmentation method based on joint image information under sea and sky background
CN117495718A (en) Multi-scale self-adaptive remote sensing image defogging method
CN116385293A (en) Foggy-day self-adaptive target detection method based on convolutional neural network
CN113627342B (en) Method, system, equipment and storage medium for video depth feature extraction optimization
Wu et al. DHGAN: Generative adversarial network with dark channel prior for single‐image dehazing
Guan et al. DiffWater: Underwater Image Enhancement Based on Conditional Denoising Diffusion Probabilistic Model
Li et al. Image object detection algorithm based on improved Gaussian mixture model
CN114155165A (en) Image defogging method based on semi-supervision
Kumar et al. Underwater Image Enhancement using deep learning
Martin et al. UD-ETR Based Restoration & CNN Approach for Underwater Object Detection from Multimedia Data
Wang et al. Is Underwater Image Enhancement All Object Detectors Need?
Li et al. Single image defogging method based on improved generative adversarial network
CN113781325B (en) Underwater image restoration method for generating countermeasure network based on layering condition
CN115272140B (en) Infrared image quantization and enhancement method, system and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination