CN115376066A - Airport scene target detection multi-weather data enhancement method based on improved cycleGAN - Google Patents

Airport scene target detection multi-weather data enhancement method based on improved cycleGAN Download PDF

Info

Publication number
CN115376066A
CN115376066A CN202210989237.1A CN202210989237A CN115376066A CN 115376066 A CN115376066 A CN 115376066A CN 202210989237 A CN202210989237 A CN 202210989237A CN 115376066 A CN115376066 A CN 115376066A
Authority
CN
China
Prior art keywords
network
cyclegan
target detection
improved
weather
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210989237.1A
Other languages
Chinese (zh)
Inventor
阳媛
王紫航
严如强
徐佳文
杨浩然
况余进
蔡杰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202210989237.1A priority Critical patent/CN115376066A/en
Publication of CN115376066A publication Critical patent/CN115376066A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an airport scene target detection multi-weather data enhancement method based on improved cycleGAN. On the basis of an original CycleGAN network, a generator network is improved by expanding a residual error network block, cross-layer connection of a coder-decoder and introducing an attention mechanism, a dynamic weighted multi-scale discriminator network is designed, an airport scene small sample target detection data set in normal weather and an auxiliary data set in typical severe weather are used for training, a multi-weather image generation network is obtained, high-quality target detection images in various weather conditions such as rainy days, foggy days and nights are effectively generated, and diversity of data samples is increased. The method provided by the invention is used as an effective data enhancement method for saving the labeling cost, and the performance of the airport scene target detection network under the condition of small samples is improved.

Description

一种基于改进CycleGAN的机场场面目标检测多天气数据增强 方法A Multi-weather Data Enhancement for Airport Scene Target Detection Based on Improved CycleGAN method

技术领域technical field

本发明属于计算机视觉中的图像生成领域,涉及一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,尤其涉及一种通过改进的循环一致性生成对抗网络来生成典型恶劣天气下的机场场面目标检测数据从而进行数据增强的方法。The invention belongs to the field of image generation in computer vision, and relates to an airport scene target detection multi-weather data enhancement method based on improved CycleGAN, in particular to an airport scene under typical severe weather generated by an improved cycle consistency generation confrontation network A method for data augmentation based on target detection data.

背景技术Background technique

目标检测作为计算机视觉领域中的典型任务之一,近年来已取得了飞速发展,涌现出大量出色的网络模型。但此类模型往往需要大规模数据集才能表现出较好的性能,这一条件在实际应用场景中难以满足,成本极高,因此数据增强方法成为了小样本数据下提高网络性能的有效手段。在视觉领域中,常见的数据增强方法包括图片的旋转、缩放、翻转、裁剪等,这些基础数据增强方法对图像分类任务效果良好,但作用于目标检测任务时效果却不够理想。生成对抗网络以其优异的表现在近年也成为一大热点,通过生成器与判别器之间的博弈过程,能够生成以假乱真的样本,达到数据增强的效果。Object detection, as one of the typical tasks in the field of computer vision, has achieved rapid development in recent years, and a large number of excellent network models have emerged. However, such models often require large-scale data sets to show good performance. This condition is difficult to meet in practical application scenarios, and the cost is extremely high. Therefore, data enhancement methods have become an effective means to improve network performance under small sample data. In the field of vision, common data enhancement methods include image rotation, scaling, flipping, cropping, etc. These basic data enhancement methods work well for image classification tasks, but are not ideal for target detection tasks. Generative adversarial networks have also become a hot topic in recent years due to their excellent performance. Through the game process between the generator and the discriminator, it can generate fake samples to achieve the effect of data enhancement.

在机场场面目标检测任务中,各类目标如特种车辆、飞行器等,随机性强,尺度多变,且图像光照等随季节和天气变化,难以采集大量数据样本进行标注,小样本条件下目标检测效果不佳。使用改进的CycleGAN网络进行数据生成,增加目标检测数据的多样性,是提高目标检测网络性能的一种有效可行手段。In the airport scene target detection task, various targets such as special vehicles, aircraft, etc., have strong randomness, variable scale, and image illumination changes with seasons and weather, so it is difficult to collect a large number of data samples for labeling. Target detection under small sample conditions not effectively. Using the improved CycleGAN network for data generation and increasing the diversity of target detection data is an effective and feasible means to improve the performance of target detection networks.

发明内容Contents of the invention

针对上述问题,本发明提供了一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,在循环一致性生成对抗网络的基础上,分别改进生成器网络和判别器网络部分,同时调整整体损失函数,使得该模型可以生成多种典型恶劣天气下的高质量图片数据,将原始真实数据与生成的高质量数据结合,增加训练样本的多样性从而提高目标检测网络的泛化性能,以便最终实现更高效的小样本目标检测技术。In view of the above problems, the present invention provides an airport scene target detection multi-weather data enhancement method based on improved CycleGAN. On the basis of cycle consistency generation confrontation network, the generator network and the discriminator network are respectively improved, and the overall loss is adjusted at the same time. function, so that the model can generate a variety of high-quality image data under typical severe weather, combine the original real data with the generated high-quality data, increase the diversity of training samples and improve the generalization performance of the target detection network, so as to finally achieve More efficient small sample target detection technology.

为实现上述目的,本发明采取的技术方案是:For realizing above-mentioned object, the technical scheme that the present invention takes is:

一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,所述方法包括如下步骤:A kind of airport scene object detection multi-weather data enhancement method based on improved CycleGAN, described method comprises the steps:

步骤S1、图像数据采集和预处理,数据集制作;Step S1, image data collection and preprocessing, data set production;

所述步骤S1具体过程为:The specific process of the step S1 is:

通过机场场面监控摄像头采集源图像数据集,获取一段时间内正常天气下的目标检测数据并进行标注,作为原始小样本目标检测数据集;The source image data set is collected through the airport scene surveillance camera, and the target detection data under normal weather for a period of time is obtained and marked as the original small sample target detection data set;

采集典型恶劣天气下的图像数据,作为辅助天气数据集;Collect image data under typical severe weather as an auxiliary weather data set;

对上述两个原始数据集进行基础的数据增强操作,得到改进CycleGAN网络的训练数据集;Perform basic data enhancement operations on the above two original data sets to obtain a training data set for the improved CycleGAN network;

步骤S2、改进的注意力机制生成器网络构建;Step S2, construction of an improved attention mechanism generator network;

所述步骤S2具体过程为:The specific process of the step S2 is:

对原始CycleGAN网络生成器里的残差风格转换网络进行扩充,在编码和解码部分的每个卷积块后均加入一个两层卷积残差网络块;Expand the residual style conversion network in the original CycleGAN network generator, and add a two-layer convolutional residual network block after each convolution block in the encoding and decoding parts;

在原始CycleGAN网络的生成器中,引入U-Net中的跳跃连接思想,即编解码对应部分进行跨层连接操作;In the generator of the original CycleGAN network, the idea of skip connection in U-Net is introduced, that is, the corresponding part of the codec is connected across layers;

使用密集连接网络DenseNet替代原始CycleGAN网络生成器中的残差风格转换网络;Use the densely connected network DenseNet to replace the residual style transfer network in the original CycleGAN network generator;

在风格转换网络DenseNet左右两端各添加一个注意力机制模块;Add an attention mechanism module to the left and right ends of the style transfer network DenseNet;

步骤S3、动态加权多尺度判别器网络构建;Step S3, constructing a dynamic weighted multi-scale discriminator network;

所述步骤S3具体过程为:The specific process of the step S3 is:

通过两个不同尺度的PatchGAN全卷积网络支路对图像进行下采样操作,得到不同尺寸的图像判别结果;The image is down-sampled through two PatchGAN full convolutional network branches of different scales to obtain image discrimination results of different sizes;

步骤S4、对抗性训练;Step S4, confrontational training;

所述步骤S4具体过程为:The specific process of the step S4 is:

在上述改进的生成器与判别器网络基础上调整原始CycleGAN网络的损失函数,即:Adjust the loss function of the original CycleGAN network based on the above improved generator and discriminator network, namely:

L(G,F,DX,DY)=LGAN(G,DY,X,Y)+LGAN(F,DX,Y,X)+λLcyc(G,F)+LIdentity(G,F)L(G,F,D X ,D Y )=L GAN (G,D Y ,X,Y)+L GAN (F,D X ,Y,X)+λL cyc (G,F)+L Identity ( G,F)

其中

Figure BDA0003803253210000021
为循环一致性损失,
Figure BDA0003803253210000022
为恒等损失,LGAN(G,DY,X,Y)为修正的生成对抗网络损失,此处对多尺度判别器网络损失进行动态加权,具体地:in
Figure BDA0003803253210000021
is the cycle consistency loss,
Figure BDA0003803253210000022
is the constant loss, L GAN (G,D Y ,X,Y) is the modified generation confrontation network loss, where the multi-scale discriminator network loss is dynamically weighted, specifically:

Figure BDA0003803253210000023
Figure BDA0003803253210000023

Figure BDA0003803253210000024
Figure BDA0003803253210000024

由dA1=2(1-2(L1))和dA2=2(1-2(L2))分别计算A-distance,从而得到动态加权因子

Figure BDA0003803253210000025
最后有Calculate the A-distance from d A1 = 2(1-2(L 1 )) and d A2 = 2(1-2(L 2 )) to obtain the dynamic weighting factor
Figure BDA0003803253210000025
Finally there

LGAN(G,DY,X,Y)=αL1+(1-α)L2 L GAN (G,D Y ,X,Y)=αL 1 +(1-α)L 2

LGAN(F,DX,Y,X)同理;L GAN (F, D X , Y, X) is the same;

上述损失函数中G,F,DX,DY分别为原始小样本目标检测数据X到辅助天气数据Y的生成器,辅助天气数据Y到原始小样本目标检测数据X的生成器,以原始小样本目标检测数据X为真实样本的判别器,以辅助天气数据Y为真实样本的判别器;DY1、DX1和DY2、DX2分别为上述两个不同尺度的判别器,λ为惩罚系数,网络参数由下式通过梯度下降训练得到:G*,F*=argminG,F maxDX,DY L(G,F,DX,DY)。In the above loss function, G, F, D X , and D Y are the generators from the original small-sample target detection data X to the auxiliary weather data Y, and the generators from the auxiliary weather data Y to the original small-sample target detection data X. The sample target detection data X is the discriminator of the real sample, and the auxiliary weather data Y is the discriminator of the real sample; D Y1 , D X1 and D Y2 , D X2 are the discriminators of the above two different scales respectively, and λ is the penalty coefficient , the network parameters are obtained by the following formula through gradient descent training: G * , F * = argmin G, F max DX, DY L(G, F, D X , D Y ).

进一步地,所述步骤S1中正常天气为白天且晴天。Further, the normal weather in the step S1 is daytime and sunny.

进一步地,所述步骤S1中目标包括特种车辆和飞行器。Further, the targets in the step S1 include special vehicles and aircrafts.

进一步地,所述步骤S1中典型恶劣天气包括阴天、雨天、雾天和夜晚。Further, the typical severe weather in the step S1 includes cloudy days, rainy days, foggy days and nights.

进一步地,所述步骤S1中数据增强操作包括旋转、缩放和翻转。Further, the data enhancement operation in the step S1 includes rotation, scaling and flipping.

进一步地,所述步骤S2中注意力机制模块为空间通道混合注意力机制。Further, the attention mechanism module in step S2 is a spatial channel mixed attention mechanism.

与现有技术相比,本发明的有益效果为:Compared with prior art, the beneficial effect of the present invention is:

通过扩充残差网络块、编解码器的跨层连接以及引入注意力机制的方式改进生成器网络,同时设计动态加权多尺度判别器网络,在机场场面小样本目标检测数据集以及辅助天气数据集条件下训练,可以有效生成雨天、雾天、夜晚等天气下的高质量目标检测图像,增加目标检测数据样本的多样性,从而有效提升小样本条件下的目标检测网络性能。Improve the generator network by expanding the residual network block, the cross-layer connection of the codec and the introduction of the attention mechanism, and at the same time design a dynamic weighted multi-scale discriminator network, a small sample target detection data set in the airport scene and an auxiliary weather data set Training under conditions can effectively generate high-quality target detection images in rainy, foggy, night and other weather conditions, increase the diversity of target detection data samples, and effectively improve the performance of target detection networks under small sample conditions.

附图说明Description of drawings

图1为本发明提出方法的网络结构图;Fig. 1 is the network structure diagram of the method proposed by the present invention;

图2为循环一致性生成对抗网络示意图;Figure 2 is a schematic diagram of the cycle consistency generation confrontation network;

图3为本发明提出方法的生成器网络结构图;Fig. 3 is the generator network structural diagram of the proposed method of the present invention;

图4为本发明提出方法的判别器网络结构图。Fig. 4 is a network structure diagram of the discriminator of the method proposed by the present invention.

具体实施方式Detailed ways

下面结合附图与具体实施方式对本发明作进一步详细描述:Below in conjunction with accompanying drawing and specific embodiment the present invention is described in further detail:

如图1所示,将原始小样本目标检测数据集和辅助天气数据集分别作为源域数据X和目标域数据Y,对改进的CycleGAN网络模型进行训练,获得多天气图像生成网络,进而生成多天气目标检测图像。改进的CycleGAN模型包括生成器G和生成器F,分别对应数据X到数据Y和数据Y到数据X的映射;两种多尺度判别器Dx和Dy分别对两个方向的生成进行判别,且Dx和Dy均包含两个子判别器。通过迭代训练即可获得高质量的多天气图像生成网络。图2为原始的循环一致性生成对抗网络示意图。As shown in Figure 1, the original small-sample target detection dataset and auxiliary weather dataset are used as source domain data X and target domain data Y respectively, and the improved CycleGAN network model is trained to obtain a multi-weather image generation network, and then generate multiple Weather object detection images. The improved CycleGAN model includes generator G and generator F, which correspond to the mapping of data X to data Y and data Y to data X respectively; two multi-scale discriminators Dx and Dy respectively discriminate the generation of two directions, and Dx Both D and Dy contain two sub-discriminators. A high-quality multi-weather image generation network can be obtained through iterative training. Figure 2 is a schematic diagram of the original cycle consistency generation confrontation network.

具体来说,主要步骤如下:Specifically, the main steps are as follows:

步骤S1、图像数据采集和预处理,数据集制作;Step S1, image data collection and preprocessing, data set production;

通过机场场面监控摄像头采集源图像数据集,获取一段时间内正常天气下的目标检测数据并进行标注,作为原始小样本目标检测数据集;采集雾天状况下的图像数据,作为辅助天气数据集;对上述两个原始数据集进行基础的数据增强操作,得到改进CycleGAN网络的训练数据集;Collect the source image data set through the airport scene monitoring camera, obtain the target detection data under normal weather for a period of time and mark it as the original small sample target detection data set; collect the image data under foggy conditions as the auxiliary weather data set; Perform basic data enhancement operations on the above two original data sets to obtain a training data set for the improved CycleGAN network;

步骤S2、改进的注意力机制生成器网络构建;Step S2, construction of an improved attention mechanism generator network;

图3为改进的生成器网络结构图。如图所示,对原始CycleGAN网络生成器里的残差风格转换网络进行扩充,在编码和解码部分的每个卷积块后均加入一个两层卷积残差网络块;在原始CycleGAN网络的生成器中,引入U-Net中的跳跃连接思想,即编解码对应部分进行跨层连接操作;使用密集连接网络DenseNet替代原始CycleGAN网络生成器中的残差风格转换网络;在风格转换网络DenseNet左右两端各添加一个注意力机制模块;Figure 3 is a structural diagram of the improved generator network. As shown in the figure, the residual style conversion network in the original CycleGAN network generator is expanded, and a two-layer convolutional residual network block is added after each convolution block in the encoding and decoding parts; in the original CycleGAN network In the generator, the skip connection idea in U-Net is introduced, that is, the corresponding part of the codec is connected across layers; the densely connected network DenseNet is used to replace the residual style conversion network in the original CycleGAN network generator; around the style conversion network DenseNet Add an attention mechanism module at both ends;

步骤S3、动态加权多尺度判别器网络构建;Step S3, constructing a dynamic weighted multi-scale discriminator network;

图4为改进的判别器网络结构图。如图所示,通过两个不同尺度的PatchGAN全卷积网络支路对图像进行下采样操作,得到不同尺寸的图像判别结果;Figure 4 is a structural diagram of the improved discriminator network. As shown in the figure, the image is down-sampled through two PatchGAN full convolutional network branches of different scales to obtain image discrimination results of different sizes;

步骤4,对抗性训练;Step 4, confrontational training;

依据调整后的网络损失函数,即:According to the adjusted network loss function, namely:

L(G,F,DX,DY)=LGAN(G,DY,X,Y)+LGAN(F,DX,Y,X)+λLcyc(G,F)+LIdentity(G,F)L(G,F,D X ,D Y )=L GAN (G,D Y ,X,Y)+L GAN (F,D X ,Y,X)+λL cyc (G,F)+L Identity ( G,F)

优化目标G*,F*=argminG,F maxDX,DY L(G,F,DX,DY),设置初始化参数对模型进行训练,交替更新判别器和生成器,最终训练完成得到生成网络,将正常天气下的目标检测数据输入即可得到对应生成的雾天目标检测数据。Optimize the target G * , F * = argmin G, F max DX, DY L(G, F, D X , D Y ), set the initialization parameters to train the model, update the discriminator and generator alternately, and finally the training is completed to generate Network, input the target detection data in normal weather to get the corresponding foggy target detection data.

以上所述,仅是本发明的较佳实施例而已,并非是对本发明作任何其他形式的限制,而依据本发明的技术实质所作的任何修改或等同变化,仍属于本发明所要求保护的范围。The above is only a preferred embodiment of the present invention, and is not intended to limit the present invention in any other form, and any modification or equivalent change made according to the technical essence of the present invention still belongs to the scope of protection required by the present invention .

Claims (6)

1.一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,其特征在于,所述方法包括如下步骤:1. a kind of airport scene target detection multi-weather data enhancement method based on improved CycleGAN, it is characterized in that, described method comprises the steps: 步骤S1、图像数据采集和预处理,数据集制作;Step S1, image data collection and preprocessing, data set production; 所述步骤S1具体过程为:The specific process of the step S1 is: 通过机场场面监控摄像头采集源图像数据集,获取一段时间内正常天气下的目标检测数据并进行标注,作为原始小样本目标检测数据集;The source image data set is collected through the airport scene surveillance camera, and the target detection data under normal weather for a period of time is obtained and marked as the original small sample target detection data set; 采集典型恶劣天气下的图像数据,作为辅助天气数据集;Collect image data under typical severe weather as an auxiliary weather data set; 对上述两个原始数据集进行基础的数据增强操作,得到改进CycleGAN网络的训练数据集;Perform basic data enhancement operations on the above two original data sets to obtain a training data set for the improved CycleGAN network; 步骤S2、改进的注意力机制生成器网络构建;Step S2, construction of an improved attention mechanism generator network; 所述步骤S2具体过程为:The specific process of the step S2 is: 对原始CycleGAN网络生成器里的残差风格转换网络进行扩充,在编码和解码部分的每个卷积块后均加入一个两层卷积残差网络块;Expand the residual style conversion network in the original CycleGAN network generator, and add a two-layer convolutional residual network block after each convolution block in the encoding and decoding parts; 在原始CycleGAN网络的生成器中,引入U-Net中的跳跃连接思想,即编解码对应部分进行跨层连接操作;In the generator of the original CycleGAN network, the idea of skip connection in U-Net is introduced, that is, the corresponding part of the codec is connected across layers; 使用密集连接网络DenseNet替代原始CycleGAN网络生成器中的残差风格转换网络;Use the densely connected network DenseNet to replace the residual style transfer network in the original CycleGAN network generator; 在风格转换网络DenseNet左右两端各添加一个注意力机制模块;Add an attention mechanism module to the left and right ends of the style transfer network DenseNet; 步骤S3、动态加权多尺度判别器网络构建;Step S3, constructing a dynamic weighted multi-scale discriminator network; 所述步骤S3具体过程为:The specific process of the step S3 is: 通过两个不同尺度的PatchGAN全卷积网络支路对图像进行下采样操作,得到不同尺寸的图像判别结果;The image is down-sampled through two PatchGAN full convolutional network branches of different scales to obtain image discrimination results of different sizes; 步骤S4、对抗性训练;Step S4, confrontational training; 所述步骤S4具体过程为:The specific process of the step S4 is: 在上述改进的生成器与判别器网络基础上调整原始CycleGAN网络的损失函数,即:Adjust the loss function of the original CycleGAN network based on the above improved generator and discriminator network, namely: L(G,F,DX,DY)=LGAN(G,DY,X,Y)+LGAN(F,DX,Y,X)+λLcyc(G,F)+LIdentity(G,F)L(G,F,D X ,D Y )=L GAN (G,D Y ,X,Y)+L GAN (F,D X ,Y,X)+λL cyc (G,F)+L Identity ( G,F) 其中
Figure FDA0003803253200000011
为循环一致性损失,
Figure FDA0003803253200000012
为恒等损失,LGAN(G,DY,X,Y)为修正的生成对抗网络损失,此处对多尺度判别器网络损失进行动态加权,具体地:
in
Figure FDA0003803253200000011
is the cycle consistency loss,
Figure FDA0003803253200000012
is the constant loss, L GAN (G,D Y ,X,Y) is the modified generation confrontation network loss, where the multi-scale discriminator network loss is dynamically weighted, specifically:
Figure FDA0003803253200000013
Figure FDA0003803253200000013
Figure FDA0003803253200000014
Figure FDA0003803253200000014
由dA1=2(1-2(L1))和dA2=2(1-2(L2))分别计算A-distance,从而得到动态加权因子
Figure FDA0003803253200000021
最后有
Calculate the A-distance from d A1 = 2(1-2(L 1 )) and d A2 = 2(1-2(L 2 )) to obtain the dynamic weighting factor
Figure FDA0003803253200000021
Finally there
LGAN(G,DY,X,Y)=αL1+(1-α)L2 L GAN (G,D Y ,X,Y)=αL 1 +(1-α)L 2 LGAN(F,DX,Y,X)同理;L GAN (F, D X , Y, X) is the same; 上述损失函数中G,F,DX,DY分别为原始小样本目标检测数据X到辅助天气数据Y的生成器,辅助天气数据Y到原始小样本目标检测数据X的生成器,以原始小样本目标检测数据X为真实样本的判别器,以辅助天气数据Y为真实样本的判别器;DY1、DX1和DY2、DX2分别为上述两个不同尺度的判别器,λ为惩罚系数,网络参数由下式通过梯度下降训练得到:In the above loss function, G, F, D X , and D Y are the generators from the original small-sample target detection data X to the auxiliary weather data Y, and the generators from the auxiliary weather data Y to the original small-sample target detection data X. The sample target detection data X is the discriminator of the real sample, and the auxiliary weather data Y is the discriminator of the real sample; D Y1 , D X1 and D Y2 , D X2 are the discriminators of the above two different scales respectively, and λ is the penalty coefficient , the network parameters are obtained by the following formula through gradient descent training: G*,F*=arg minG,FmaxDX,DYL(G,F,DX,DY)。G * ,F * =arg min G,F max DX,DY L(G,F,D X ,D Y ).
2.根据权利要求1所述的一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,其特征在于:2. a kind of airport scene target detection multi-weather data enhancement method based on improved CycleGAN according to claim 1, is characterized in that: 所述步骤S1中正常天气为白天且晴天。The normal weather in the step S1 is daytime and sunny. 3.根据权利要求1所述的一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,其特征在于:3. a kind of airport scene target detection multi-weather data enhancement method based on improved CycleGAN according to claim 1, is characterized in that: 所述步骤S1中目标包括特种车辆和飞行器。The targets in the step S1 include special vehicles and aircraft. 4.根据权利要求1所述的一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,其特征在于:4. a kind of airport scene object detection multi-weather data enhancement method based on improved CycleGAN according to claim 1, is characterized in that: 所述步骤S1中典型恶劣天气包括阴天、雨天、雾天和夜晚。Typical severe weather in the step S1 includes cloudy days, rainy days, foggy days and nights. 5.根据权利要求1所述的一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,其特征在于:5. a kind of airport scene object detection multi-weather data enhancement method based on improved CycleGAN according to claim 1, is characterized in that: 所述步骤S1中数据增强操作包括旋转、缩放和翻转。The data enhancement operations in step S1 include rotation, scaling and flipping. 6.根据权利要求1所述的一种基于改进CycleGAN的机场场面目标检测多天气数据增强方法,其特征在于:6. a kind of airport scene object detection multi-weather data enhancement method based on improved CycleGAN according to claim 1, is characterized in that: 所述步骤S2中注意力机制模块为空间通道混合注意力机制。The attention mechanism module in the step S2 is a spatial channel hybrid attention mechanism.
CN202210989237.1A 2022-08-17 2022-08-17 Airport scene target detection multi-weather data enhancement method based on improved cycleGAN Pending CN115376066A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210989237.1A CN115376066A (en) 2022-08-17 2022-08-17 Airport scene target detection multi-weather data enhancement method based on improved cycleGAN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210989237.1A CN115376066A (en) 2022-08-17 2022-08-17 Airport scene target detection multi-weather data enhancement method based on improved cycleGAN

Publications (1)

Publication Number Publication Date
CN115376066A true CN115376066A (en) 2022-11-22

Family

ID=84064726

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210989237.1A Pending CN115376066A (en) 2022-08-17 2022-08-17 Airport scene target detection multi-weather data enhancement method based on improved cycleGAN

Country Status (1)

Country Link
CN (1) CN115376066A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116912680A (en) * 2023-06-25 2023-10-20 西南交通大学 SAR ship identification cross-modal domain migration learning and identification method and system
CN118154467A (en) * 2024-05-11 2024-06-07 华东交通大学 An image deraining method and system based on improved CycleGAN network model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116912680A (en) * 2023-06-25 2023-10-20 西南交通大学 SAR ship identification cross-modal domain migration learning and identification method and system
CN118154467A (en) * 2024-05-11 2024-06-07 华东交通大学 An image deraining method and system based on improved CycleGAN network model

Similar Documents

Publication Publication Date Title
Cui et al. Semantic segmentation of remote sensing images using transfer learning and deep convolutional neural network with dense connection
CN111340122B (en) Multi-modal feature fusion text-guided image restoration method
CN110119780B (en) A Generative Adversarial Network-Based Super-resolution Reconstruction Method for Hyperspectral Images
Wang et al. Object-scale adaptive convolutional neural networks for high-spatial resolution remote sensing image classification
CN108764063A (en) A kind of pyramidal remote sensing image time critical target identifying system of feature based and method
Yi et al. Efficient and accurate multi-scale topological network for single image dehazing
CN112733656B (en) Skeleton action recognition method based on multiflow space attention diagram convolution SRU network
CN110110599B (en) Remote sensing image target detection method based on multi-scale feature fusion
CN111401436B (en) Streetscape image segmentation method fusing network and two-channel attention mechanism
CN110675462B (en) Gray image colorization method based on convolutional neural network
CN110232394A (en) A kind of multi-scale image semantic segmentation method
CN113870160B (en) Point cloud data processing method based on transformer neural network
CN112560865B (en) A Semantic Segmentation Method for Point Clouds in Large Outdoor Scenes
CN110781773A (en) Road extraction method based on residual error neural network
CN115376066A (en) Airport scene target detection multi-weather data enhancement method based on improved cycleGAN
CN114581560A (en) Multi-scale neural network infrared image colorization method based on attention mechanism
CN116645598A (en) Remote sensing image semantic segmentation method based on channel attention feature fusion
Hou et al. Fe-fusion-vpr: Attention-based multi-scale network architecture for visual place recognition by fusing frames and events
Wang et al. Hierarchical kernel interaction network for remote sensing object counting
CN117173022A (en) Remote sensing image super-resolution reconstruction method based on multi-path fusion and attention
Yao et al. ModeRNN: Harnessing spatiotemporal mode collapse in unsupervised predictive learning
CN116630704A (en) A network model for object classification based on attention enhancement and dense multi-scale
CN119128453B (en) Federal domain generalized fault diagnosis method and system
CN111680667A (en) A classification method of remote sensing images based on deep neural network
Yang et al. Remote sensing image object detection based on improved yolov3 in deep learning environment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination