CN115482433A - Small-scale data set amplification method based on learnable data enhancement - Google Patents

Small-scale data set amplification method based on learnable data enhancement Download PDF

Info

Publication number
CN115482433A
CN115482433A CN202211121663.XA CN202211121663A CN115482433A CN 115482433 A CN115482433 A CN 115482433A CN 202211121663 A CN202211121663 A CN 202211121663A CN 115482433 A CN115482433 A CN 115482433A
Authority
CN
China
Prior art keywords
image
change
probability
small
data set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211121663.XA
Other languages
Chinese (zh)
Inventor
郑艳伟
黄博文
王鹏
于东晓
孙恩涛
杜超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong University
Shanghai Step Electric Corp
Original Assignee
Shandong University
Shanghai Step Electric Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong University, Shanghai Step Electric Corp filed Critical Shandong University
Priority to CN202211121663.XA priority Critical patent/CN115482433A/en
Publication of CN115482433A publication Critical patent/CN115482433A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a small-scale data set amplification method based on learnable data enhancement, which comprises the following steps: performing data enhancement on an original image in the small-scale data set by using a probability p; inputting the image after data enhancement into a discriminator for generating a confrontation network, randomly sampling a normally distributed noise as the input of a generator for generating the confrontation network, generating an image by the generator, inputting the image into the discriminator, and alternately training the generator and the discriminator to optimize a target function; in the training process, continuously learning and updating the probability p and the parameters for generating the countermeasure network; and randomly sampling a plurality of normally distributed noises, and respectively inputting the noises into a generator for generating the confrontation network after training to respectively generate corresponding images so as to realize the amplification of the small-scale data set. The method disclosed by the invention can generate a large amount of high-quality images, realizes the amplification of small-scale data sets, and avoids the condition that the generated images are inconsistent with the original data.

Description

Small-scale data set amplification method based on learnable data enhancement
Technical Field
The invention belongs to the field of computer vision, and particularly relates to a small-scale data set amplification method based on learnable data enhancement.
Background
The most typical method of image augmentation is to generate a countermeasure network, which contains two modules, a generator and a discriminator. The goal of the generator network is to try to generate a true picture to fool the discriminator, i.e. to make it impossible for the discriminator to tell whether an image is true or synthetic. The goal of the discriminator is to separate the picture generated by the generator and the real picture as much as possible, namely to enhance the discrimination capability of the discriminator. By continuously training the generator and the discriminator, the model can finally produce images with high trueness.
But generating a competing network requires a large amount of training data, and for small data sets the network is easily over-fitted, leading to a training crash.
In order to solve the problem of overfitting, the invention designs a series of data enhancement methods to increase the diversity of images, and simultaneously, in order to avoid the generated images from being influenced by enhancement data, a learnable parameter p is used as the probability of various enhancement modes to achieve the aim of leading a network to identify correct source images.
Disclosure of Invention
In order to solve the technical problems, the invention provides a small-scale data set amplification method based on learnable data enhancement, which can generate a large number of high-quality images, realize the amplification of small-scale data sets and avoid the situation that the generated images are inconsistent with the original data.
In order to achieve the purpose, the technical scheme of the invention is as follows:
a method for amplification of small-scale data sets based on reinforcement of learnable data, comprising the steps of:
step one, data enhancement: performing data enhancement on an original image in the small-scale data set by using probability p to obtain an image after data enhancement;
step two, generating confrontation network training: inputting the image after data enhancement into a discriminator for generating a countermeasure network, randomly sampling normally distributed noise as the input of a generator for generating the countermeasure network, generating an image by the generator, inputting the image into the discriminator, judging whether the input image is a real image or a generated image by the discriminator, and alternately training the generator and the discriminator to optimize a target function; in the training process, continuously learning and updating the probability p and parameters for generating the countermeasure network, and repeating the first step and the second step to finish the training for generating the countermeasure network;
step three, amplifying the small-scale data set: and randomly sampling a plurality of normally distributed noises, and respectively inputting the noises into a generator for generating a countermeasure network after training to respectively generate corresponding images so as to realize the amplification of a small-scale data set.
In the above scheme, in the first step, the data enhancement method includes geometric transformation, pixel transformation and image filtering.
In a further technical solution, the geometric transformation includes displacement transformation, equal scaling and unequal scaling.
In a further technical solution, the pixel transformation includes brightness variation, contrast variation, saturation variation, noise addition, and random erasure.
Preferably, the specific method of step one is as follows: and sequentially carrying out displacement transformation, equal scaling, unequal scaling, brightness change, contrast change, saturation change, noise addition, random erasure and image filtering on the original image according to the probability p to obtain an image with enhanced data.
In a further technical solution, the displacement transformation is to perform overall displacement on the image with probability p, and the transformation process is as follows:
t x ,t y ~U(-0.1,0.1)
Figure BDA0003847360250000021
wherein, the probability p is a learnable parameter with an initial value of 0.5; t is t x ,t y Respectively representing the times of image width and height scaling; i.e. i 11 U (0, 1), a random number from 0 to 1, for determining whether a shift transform operation is performed; u () represents uniform distribution, translate represents a shift transform operation, and w and h represent the width and height of an image; round () means round; x 0 Representing an original image; x 11 The image after displacement transformation is obtained;
the equal ratio scaling is to scale the width and height of the image according to the same ratio by probability p, and the transformation process is as follows:
s~U(0.5,2)
Figure BDA0003847360250000022
wherein i 12 U (0, 1), a random number from 0 to 1, for determining whether an equal ratio scaling operation is performed; scale represents the zoom operation, s is the zoom multiple of the width and height of the image; crop is a Crop operation to change the width and height of the image back to w and h to keep the image size unchanged, X 12 Is an image scaled by an equal ratio;
the non-equal scaling scales the width and height of the image according to different proportions by probability p, and the transformation process is as follows:
s1~U(0.5,2),s2~U(0.5,2)
Figure BDA0003847360250000023
wherein i 13 U (0, 1) of 0 to 1Is used for determining whether the non-equal ratio scaling operation is executed, s1 and s2 are the scaling times of the width and the height of the image respectively, X 13 Representing an anisometric scaled image.
In a further technical solution, the luminance change is a change of the luminance of the image with a probability p, and the change process is as follows:
b~U(0.5,1.5)
Figure BDA0003847360250000031
wherein i 21 U (0, 1) is a random number of 0 to 1 for determining whether a brightness change operation is performed, U () represents a uniform distribution, bright represents a brightness change operation, b is a multiple of the brightness change, X 13 Representing an unequally scaled image, X 21 Representing the image after brightness change;
the contrast change is to change the contrast of the image by a probability p, and the change process is as follows:
c~U(0.7,1.2)
Figure BDA0003847360250000032
wherein i 22 U (0, 1) is a random number of 0 to 1 for determining whether a Contrast change operation is performed, contrast denotes a Contrast change operation, c is a multiple of the Contrast change, X 22 Representing the image after the contrast is changed;
the saturation change is to change the saturation of the image by a probability p, and the change process is as follows:
s~U(0.6,1.2)
Figure BDA0003847360250000033
wherein i 23 U (0, 1), a random number from 0 to 1, for determining whether a saturation change operation is performed,saturration denotes the Saturation change operation, s is the multiple of the Saturation change, X 23 Representing the image after saturation change;
the noise addition is to add a random noise to the image by using the probability p, and the change process is as follows:
r,g,b~N(0,1)
Figure BDA0003847360250000034
wherein i 24 U (0, 1) is a random number from 0 to 1 for determining whether to add noise, (m, n) represents a pixel coordinate on the image, and 0. Ltoreq. M < w, 0. Ltoreq. N < h, w and h represent the width and height of the image, R, G, B are three normally distributed random numbers, corresponding to the three components R, G, B of the pixel, respectively; n () represents a normal distribution; x 24 Representing the image after the noise addition operation;
the random erasing is to randomly select a region in the image with probability p to remove the region, and the change process is as follows:
c x ,c y ~U(0.3,0.6)
left=round((c x -0.25)×w)
low=round((c y -0.25)×h)
right=round((c x +0.25)×w)
high=round((c y +0.25)×h)
Figure BDA0003847360250000041
Figure BDA0003847360250000042
wherein i 25 U (0, 1) is a random number of 0 to 1 for determining whether or not to perform random erasure, (left, high) denotes the coordinates of the upper left vertex of the erasure area, (right, low) denotes the coordinates of the lower right vertex of the erasure area, and round () represents rounding; (c) x ,c y ) Coordinate indicating center point of erase region, mask is mask of erase region, an | _ indicates an OR operation, X 25 Representing the image after the random erasure is performed.
In a further technical scheme, the image filtering is to perform filtering operation on an image with probability p, and 4 filters with different sizes are adopted, and the process is as follows:
size∈{(3,3),(5,5),(7,7),(9,9)}
Figure BDA0003847360250000043
wherein i 31 U (0, 1) is a random number of 0 to 1 for determining whether to perform an image filtering operation, filter represents the image filtering operation, size represents the size of the four filters, X 25 Representing images after random erasure, X 31 Is to filter the enhanced image and this operation randomly selects one of four different filters.
Preferably, in step two, the generation countermeasure network adopts StyleGAN.
In a further technical scheme, in the second step, when the generator is trained, the parameters of the discriminator are fixed, and at this time, the objective function is minimized:
Figure BDA0003847360250000044
when the arbiter trains, the parameters of the generator are fixed, at which time the objective function is maximized:
Figure BDA0003847360250000045
wherein V (D, G) is a training function containing probability p and parameters for generating a confrontation network, x is a real image subjected to data enhancement, z is noise subjected to random sampling, D and G respectively represent a discriminator and a generator, D () and G () respectively represent outputs of the discriminator and the generator, and E represents an expected value.
By the technical scheme, the small-scale data set amplification method based on the reinforcement of the learnable data has the following beneficial effects:
(1) According to the method, a series of data enhancement is performed on the images of the small-scale data set, so that the overfitting problem in the process of training a network by using the small-scale data set is solved, a large number of high-quality images are successfully generated, and the purpose of data set amplification is achieved.
(2) According to the method, a learnable parameter p is used as the probability of data enhancement, so that the network can learn the distribution of the original data (without data enhancement) in the training process, and the condition that the generated image is inconsistent with the original data is avoided.
(3) The invention realizes the training of generating the confrontation network by maximizing or minimizing the objective function at different stages of the training, and then generates a large amount of data with higher truth from the small-scale data by utilizing the trained generated confrontation network, thereby realizing the amplification of the small-scale data set.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below.
Fig. 1 is a schematic flow chart of a small-scale data set amplification method based on learnable data enhancement according to an embodiment of the present invention.
Detailed Description
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention.
The invention provides a small-scale data set amplification method based on learnable data enhancement, which comprises the following steps as shown in figure 1:
step one, data enhancement: and performing data enhancement on the original image in the small-scale data set by using the probability p to obtain an image after data enhancement.
The data enhancement method comprises geometric transformation, pixel transformation and image filtering, wherein the geometric transformation comprises displacement transformation, equal ratio scaling and unequal ratio scaling, and the pixel transformation comprises brightness change, contrast change, saturation change, noise addition and random erasure. The sequence of the method is not fixed and can be adjusted.
In this embodiment, a fresh flower data set is used as a training image, and the data set is a small-scale data set. It contains 17 types of flowers, each with 80 images, and flowers in the dataset with obvious pose and light changes. Data enhancement is performed in the following order: and sequentially carrying out displacement transformation, equal ratio scaling, unequal ratio scaling, brightness change, contrast change, saturation change, noise addition, random erasure and image filtering on the original image according to the probability p to obtain an image with enhanced data.
The specific data enhancement process is as follows:
1. transformation of displacement
The displacement transformation is to displace the whole image by the probability p, and the transformation process is as follows:
t x ,t y ~U(-0.1,0.1)
Figure BDA0003847360250000061
wherein, the probability p is a learnable parameter with an initial value of 0.5; t is t x ,t y Respectively representing the times of image width and height scaling; i.e. i 11 U (0, 1), a random number from 0 to 1, for determining whether a shift transform operation is performed; u () represents uniform distribution, translate represents a shift transform operation, and w and h represent the width and height of an image; round () represents rounding; x 0 Representing an original image; x 11 The transformed image is displaced.
2. Geometric scaling
Scaling is to scale the width and height of an image by the same ratio with probability p, and the transformation process is as follows:
s~U(0.5,2)
Figure BDA0003847360250000062
wherein i 12 U (0, 1), a random number from 0 to 1, for determining whether an equal ratio scaling operation is performed; scale represents the zoom operation, s is the zoom multiple of the width and height of the image; crop is a Crop operation to change the width and height of the image back to w and h to keep the image size unchanged, X 12 Is an image scaled by equal ratio.
3. Non-equal ratio scaling
Non-equal scaling scales the width and height of an image differently with probability p, and the transformation process is as follows:
s1~U(0.5,2),s2~U(0.5,2)
Figure BDA0003847360250000063
wherein i 13 U (0, 1), a random number from 0 to 1, for determining whether a non-equal ratio scaling operation is performed, s1, s2 are the scaling factors of image width and height, X, respectively 13 Representing an anisometric scaled image.
4. Variation of brightness
The brightness change is to change the brightness of the image with probability p, and the change process is as follows:
b~U(0.5,1.5)
Figure BDA0003847360250000064
wherein i 21 U (0, 1), which is a random number from 0 to 1, for determining whether a brightness change operation is performed, U () denotes a uniform distribution, bright denotes a brightness change operation, b is a multiple of the brightness change, X 13 Representing an unequally scaled image, X 21 Representing the image after the brightness change.
5. Contrast variation
The contrast change is to change the contrast of the image with a probability p, and the change process is as follows:
c~U(0.7,1.2)
Figure BDA0003847360250000071
wherein i 22 U (0, 1) is a random number of 0 to 1 for determining whether a Contrast change operation is performed, contrast denotes a Contrast change operation, c is a multiple of the Contrast change, X 22 Representing the image after the contrast change.
6. Change in saturation
The saturation change is to change the saturation of the image by a probability p, and the change process is as follows:
s~U(0.6,1.2)
Figure BDA0003847360250000072
wherein i 23 U (0, 1) is a random number from 0 to 1 for determining whether a Saturation change operation is performed, saturration denotes the Saturation change operation, s is a multiple of the Saturation change, X 23 Representing the image after saturation change.
7. Noise addition
The noise addition is to add a random noise to the image by using the probability p, and the change process is as follows:
r,g,b~N(0,1)
Figure BDA0003847360250000073
wherein i 24 U (0, 1), a random number from 0 to 1, for determining whether to add noise, (m, n) represents a pixel coordinate on the image, and 0. Ltoreq. M < w, 0. Ltoreq. N < h, w and h represent the width and height of the image, R, G, B are three normally distributed random numbers, corresponding to the three components R, G, B of the pixel, respectively; n () represents a normal distribution; x 24 Representing the image after the noise addition operation.
8. Random erase
Random erasure is to remove a region in the image by randomly selecting it with a probability p, and the process of variation is as follows:
c x ,c y ~U(0.3,0.6)
left=round((c x -0.25)×w)
low=round((c y -0.25)×h)
right=round((c x +0.25)×w)
high=round((c y +0.25)×h)
Figure BDA0003847360250000074
Figure BDA0003847360250000075
wherein i 25 U (0, 1) is a random number of 0 to 1 for determining whether or not random erasure is performed, (left, high) denotes the coordinates of the upper left vertex of the erasure area, (right, low) denotes the coordinates of the lower right vertex of the erasure area, and round () denotes rounding; (c) x ,c y ) Coordinate indicating center point of erase region, mask is mask of erase region, an | _ indicates an OR operation, X 25 Representing the image after the random erasure is performed.
9. Image filtering
The image filtering is to perform filtering operation on an image with a probability p, and 4 filters with different sizes are adopted, and the process is as follows:
size∈{(3,3),(5,5),(7,7),(9,9)}
Figure BDA0003847360250000081
wherein i 31 U (0, 1), a random number from 0 to 1, for determining whether to perform an image filtering operation, filter representing an image filtering operation, and size representing four kinds of filtersSize of wave filter, X 25 Representing images after random erasure, X 31 Is filtering the enhanced image. This operation randomly selects one of four different filters.
Step two, generating confrontation network training: inputting the image after data enhancement into a discriminator for generating a confrontation network, randomly sampling a normally distributed noise as the input of a generator for generating the confrontation network, generating an image by the generator, inputting the image into the discriminator, judging whether the input image is a real image or a generated image by the discriminator, and alternately training the generator and the discriminator to optimize a target function; in the training process, continuously learning and updating the probability p and the parameters for generating the confrontation network, and repeating the first step and the second step to finish the training for generating the confrontation network;
in this embodiment, the generation countermeasure network adopts StyleGAN, which adopts a progressively growing network as a generator and a discriminator, that is, a shallow layer of the network generates an image with a lower resolution, and as the network is continuously deepened, the resolution of the generated image gradually increases, and finally a composite image with a higher resolution is obtained. The specific process is as follows:
(1) Image X to be data enhanced 31 Put into a discriminator D, the discriminator judges the class of the image to be a real image or a composite image, namely, class = D (X) 31 );
(2) Randomly sampling a noise z as input to a generator G, placing it in the generator, and generating an image
Figure BDA0003847360250000085
Namely, it is
Figure BDA0003847360250000082
(3) The generated image is input to a discriminator to determine whether the image is a real image or a composite image, i.e.
Figure BDA0003847360250000083
And (3) repeating the steps (1) to (3) continuously, so that the generator generates an image similar to the real image as much as possible, and the discriminator discriminates whether the input image is synthesized or real as much as possible, and finally the discriminator cannot discriminate the source of the image.
The overall training target is as follows:
Figure BDA0003847360250000084
during generator training, the parameters of the arbiter are fixed, and the objective function is minimized:
Figure BDA0003847360250000091
when the arbiter trains, the parameters of the generator are fixed, at which time the objective function is maximized:
Figure BDA0003847360250000092
wherein V (D, G) is a training function containing probability p and parameters for generating a confrontation network, x is a real image subjected to data enhancement, z is noise subjected to random sampling, D and G respectively represent a discriminator and a generator, D () and G () respectively represent outputs of the discriminator and the generator, and E represents an expected value.
The learnable probability p is updated in the training stage of the discriminator, during initial training, due to the fact that training data are too small, overfitting of the discriminator is serious, the discrimination capability of the discriminator is strong, no matter how the generator is optimized, the discriminator can distinguish the truth of an image, and therefore the generator behind cannot be trained effectively, and therefore the probability p can be increased continuously to enhance the diversity of the training data. At the later stage of model training, p will start to decrease slowly because p is too large, which causes the image generated by the generator to be closer to the data-enhanced image than the original image. After the model training is finished, the value of p is between 0.8 and 0.9.
Step three, amplifying the small-scale data set: and randomly sampling a plurality of normally distributed noises, and respectively inputting the noises into a generator for generating a countermeasure network after training to respectively generate corresponding images so as to realize the amplification of a small-scale data set.
The method comprises the following specific steps:
(1) Loading the generator G trained in the previous step;
(2) A noise z is randomly sampled from a normal distribution and input to the generator G to generate the corresponding Image.
Image=G(z)
A plurality of noises are randomly sampled, and a plurality of corresponding images can be respectively generated, so that the amplification of the images is realized.
In this embodiment, through the above process, 10000 images are finally generated by using the method, and the amplification of the data set is realized.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (10)

1. A method for amplification of a small data set based on reinforcement of learnable data, comprising the steps of:
step one, data enhancement: performing data enhancement on an original image in the small-scale data set by using probability p to obtain an image after data enhancement;
step two, generating confrontation network training: inputting the image after data enhancement into a discriminator for generating a countermeasure network, randomly sampling normally distributed noise as the input of a generator for generating the countermeasure network, generating an image by the generator, inputting the image into the discriminator, judging whether the input image is a real image or a generated image by the discriminator, and alternately training the generator and the discriminator to optimize a target function; in the training process, continuously learning and updating the probability p and the parameters for generating the confrontation network, and repeating the first step and the second step to finish the training for generating the confrontation network;
step three, amplifying the small-scale data set: and randomly sampling a plurality of normally distributed noises, and respectively inputting the noises into a generator for generating a countermeasure network after training to respectively generate corresponding images so as to realize the amplification of a small-scale data set.
2. The method for amplifying the small-scale data set based on the learnable data enhancement of claim 1, wherein in the first step, the data enhancement method includes geometric transformation, pixel transformation and image filtering.
3. The method of claim 2, wherein the geometric transformation comprises a shift transformation, an equal scaling and an unequal scaling.
4. The method of claim 2, wherein the pixel transformation comprises brightness variation, contrast variation, saturation variation, noise addition, and random erasure.
5. The method for amplifying the small-scale data set based on the learnable data enhancement, according to the claim 1, is characterized in that the specific method of the first step is as follows: and sequentially carrying out displacement transformation, equal ratio scaling, unequal ratio scaling, brightness change, contrast change, saturation change, noise addition, random erasure and image filtering on the original image according to the probability p to obtain an image with enhanced data.
6. The method for small-scale data set amplification based on learnable data enhancement of claim 5, wherein the displacement transformation is a whole displacement of the image with probability p, and the transformation process is as follows:
t x ,t y ~U(-0.1,0.1)
Figure FDA0003847360240000011
wherein, the probability p is a learnable parameter with an initial value of 0.5; t is t x ,t y Representing the multiple of image width and height scaling respectively; i.e. i 11 U (0, 1), a random number from 0 to 1, for determining whether a shift transform operation is performed; u () denotes uniform distribution, translate denotes a shift transform operation, and w and h denote the width and height of an image; round () represents rounding; x 0 Representing an original image; x 11 The image after displacement transformation is obtained;
the equal ratio scaling is to scale the width and height of the image according to the same ratio by probability p, and the transformation process is as follows:
s~U(0.5,2)
Figure FDA0003847360240000021
wherein i 12 U (0, 1), a random number from 0 to 1, for determining whether an equal ratio scaling operation is performed; scale represents the zoom operation, s is the zoom multiple of the width and height of the image; crop is a Crop operation to change the width and height of the image back to w and h to keep the image size unchanged, X 12 Is an image after geometric scaling;
the non-equal scaling scales the width and height of the image according to different proportions by a probability p, and the transformation process is as follows:
s1~U(0.5,2),s2~U(0.5,2)
Figure FDA0003847360240000022
wherein i 13 U (0, 1), a random number from 0 to 1, for determining whether a non-equal ratio scaling operation is performed, s1, s2 being the scaling of the image width and height, respectivelyMultiple, X 13 Representing an anisometric scaled image.
7. The method for amplifying the small-scale data set based on the learnable data enhancement, according to the claim 5, wherein the brightness variation is the variation of the brightness of the image with probability p, and the variation process is as follows:
b~U(0.5,1.5)
Figure FDA0003847360240000023
wherein i 21 U (0, 1) is a random number of 0 to 1 for determining whether a brightness change operation is performed, U () represents a uniform distribution, bright represents a brightness change operation, b is a multiple of the brightness change, X 13 Representing an unequally scaled image, X 21 Representing the image after brightness change;
the contrast change is to change the contrast of the image by a probability p, and the change process is as follows:
c~U(0.7,1.2)
Figure FDA0003847360240000024
wherein i 22 U (0, 1) is a random number of 0 to 1 for determining whether a Contrast change operation is performed, contrast denotes a Contrast change operation, c is a multiple of the Contrast change, X 22 Representing the image after the contrast is changed;
the saturation change is to change the saturation of the image by a probability p, and the change process is as follows:
s~U(0.6,1.2)
Figure FDA0003847360240000025
wherein i 23 U (0, 1), is a random number from 0 to 1,for determining whether a Saturation change operation is performed, saturration denotes the Saturation change operation, s is a multiple of the Saturation change, X 23 Representing the image after saturation change;
the noise addition is to add a random noise to the image by using the probability p, and the change process is as follows:
r,g,b~N(0,1)
Figure FDA0003847360240000031
wherein i 24 U (0, 1) is a random number of 0 to 1 for determining whether or not to add noise, and (m, n) represents a pixel coordinate on an image, and 0 ≦ m<w,0≤n<h, w and h represent the width and height of the image, and R, G and B are three normally distributed random numbers which respectively correspond to three components R, G and B of the pixel; n () represents a normal distribution; x 24 Representing the image after the noise addition operation;
the random erasing is to randomly select a region in the image with probability p to remove the region, and the change process is as follows:
c x ,c y ~U(0.3,0.6)
left=round((c x -0.25)×w)
low=round((c y -0.25)×h)
right=round((c x +0.25)×w)
high=round((c y +0.25)×h)
Figure FDA0003847360240000032
Figure FDA0003847360240000033
wherein i 25 U (0, 1), a random number from 0 to 1, is used to determine whether to perform random erasure, (left, high) denotes the coordinates of the top left vertex of the erasure area, (right)Low) represents the coordinates of the lower right vertex of the erased area, and round () represents rounding; (c) x ,c y ) Coordinate indicating center point of erase region, mask is mask of erase region, an | _ indicates an OR operation, X 25 Representing the image after the random erasure is performed.
8. The method for amplifying the small-scale data set based on the learnable data enhancement is characterized in that the image filtering is to perform filtering operation on the image with probability p, and 4 filters with different sizes are adopted, and the process is as follows:
size∈{(3,3),(5,5),(7,7),(9,9)}
Figure FDA0003847360240000034
wherein i 31 U (0, 1), a random number from 0 to 1, for determining whether to perform an image filtering operation, filter representing the image filtering operation, size representing the size of the four filters, X 25 Representing images after random erasure, X 31 Is to filter the enhanced image and this operation randomly selects one of four different filters.
9. The method for small-scale data set amplification based on learnable data enhancement of claim 1, wherein in the second step, the generation countermeasure network employs StyleGAN.
10. The method for amplifying the small-scale data set based on the learnable data enhancement as set forth in claim 1 or 9, wherein in the second step, the parameters of the discriminator are fixed during the training of the generator, and the objective function is minimized:
Figure FDA0003847360240000041
when the arbiter trains, the parameters of the generator are fixed, at which time the objective function is maximized:
Figure FDA0003847360240000042
wherein V (D, G) is a training function containing probability p and parameters for generating a confrontation network, x is a real image subjected to data enhancement, z is noise subjected to random sampling, D and G respectively represent a discriminator and a generator, D () and G () respectively represent outputs of the discriminator and the generator, and E represents an expected value.
CN202211121663.XA 2022-09-15 2022-09-15 Small-scale data set amplification method based on learnable data enhancement Pending CN115482433A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211121663.XA CN115482433A (en) 2022-09-15 2022-09-15 Small-scale data set amplification method based on learnable data enhancement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211121663.XA CN115482433A (en) 2022-09-15 2022-09-15 Small-scale data set amplification method based on learnable data enhancement

Publications (1)

Publication Number Publication Date
CN115482433A true CN115482433A (en) 2022-12-16

Family

ID=84423536

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211121663.XA Pending CN115482433A (en) 2022-09-15 2022-09-15 Small-scale data set amplification method based on learnable data enhancement

Country Status (1)

Country Link
CN (1) CN115482433A (en)

Similar Documents

Publication Publication Date Title
CN110473154B (en) Image denoising method based on generation countermeasure network
CN113313657B (en) Unsupervised learning method and system for low-illumination image enhancement
CN112614077B (en) Unsupervised low-illumination image enhancement method based on generation countermeasure network
CN109978807B (en) Shadow removing method based on generating type countermeasure network
CN112288658A (en) Underwater image enhancement method based on multi-residual joint learning
CN103186894B (en) A kind of multi-focus image fusing method of self-adaptation piecemeal
CN113011567B (en) Training method and device of convolutional neural network model
CN111768326A (en) High-capacity data protection method based on GAN amplification image foreground object
CN109978804B (en) Human eye sight line correction method and system based on deep learning
CN112560034B (en) Malicious code sample synthesis method and device based on feedback type deep countermeasure network
CN113256538B (en) Unsupervised rain removal method based on deep learning
CN110675337B (en) Diffusion type traversal method for image noise reduction
CN108596840B (en) Data set enhancement method for deep learning evaluation of vascular network development level
CN115482433A (en) Small-scale data set amplification method based on learnable data enhancement
CN115018729B (en) Content-oriented white box image enhancement method
CN107369138A (en) Image based on higher order statistical model optimizes display methods
CN108447066B (en) Biliary tract image segmentation method, terminal and storage medium
CN115330637A (en) Image sharpening method and device, computing device and storage medium
CN112541866B (en) Human face image restoration model based on evolutionary generation countermeasure network
CN114758021A (en) Earth surface image generation method and system based on generation countermeasure network
CN110009082B (en) Three-dimensional code optimization method, medium, computer device and apparatus
CN113469904A (en) General image quality enhancement method and device based on cycle consistency loss
US8649059B2 (en) Mutual optimization system for class matrix and diffusion weighting used in halftone image processing technique
CN112381849B (en) Image edge detection method based on adaptive differential evolution
Mujtaba et al. A Fast HDR Image TMO based on a Simplified Eye Sensitivity Model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination