CN111815529B - Low-quality image classification enhancement method based on model fusion and data enhancement - Google Patents

Low-quality image classification enhancement method based on model fusion and data enhancement Download PDF

Info

Publication number
CN111815529B
CN111815529B CN202010607913.5A CN202010607913A CN111815529B CN 111815529 B CN111815529 B CN 111815529B CN 202010607913 A CN202010607913 A CN 202010607913A CN 111815529 B CN111815529 B CN 111815529B
Authority
CN
China
Prior art keywords
image
classified
enhancement
neural network
probability vector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010607913.5A
Other languages
Chinese (zh)
Other versions
CN111815529A (en
Inventor
王道累
张天宇
朱瑞
孙嘉珺
李明山
李超
韩清鹏
袁斌霞
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Electric Power University
Original Assignee
Shanghai Electric Power University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Electric Power University filed Critical Shanghai Electric Power University
Priority to CN202010607913.5A priority Critical patent/CN111815529B/en
Publication of CN111815529A publication Critical patent/CN111815529A/en
Application granted granted Critical
Publication of CN111815529B publication Critical patent/CN111815529B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/254Fusion techniques of classification results, e.g. of results related to same input data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Abstract

The invention relates to a low-quality image classification enhancement method based on model fusion and data enhancement, which comprises the following steps of: s1: establishing an image set; s2: performing data enhancement on the image set; s3: constructing and training a VGG16 convolutional neural network model; s4: constructing and training a ResNet convolution neural network model; s5: inputting an image to be classified; s6: acquiring a first probability vector and a second probability vector of an image to be classified; s7: acquiring a fusion probability vector, acquiring the image type of an image to be classified, and if the image type is clear, entering a step S9, otherwise, entering a step S8; s8: enhancing the image to be classified to obtain an enhanced image, and inputting the enhanced image serving as the image to be classified into the step S6; s9: and outputting the image. Compared with the prior art, the method and the device have the advantages that two modes of data enhancement and model fusion are adopted to improve the accuracy of classification, the images can be effectively classified, and the enhancement effect is good.

Description

Low-quality image classification enhancement method based on model fusion and data enhancement
Technical Field
The invention relates to a low-quality image classification enhancement method, in particular to a low-quality image classification enhancement method based on model fusion and data enhancement.
Background
When a camera shoots a picture, a low-quality image is usually shot due to poor lighting environment or unstable self-stability, and the like, so that the image details are lost due to the fact that shadow, low brightness or blurred pictures exist on the low-quality image, and subsequent operations such as identification and analysis are difficult to perform on the image.
In a low-quality image enhancement method under extreme weather conditions disclosed in chinese patent No. cn201610079472.X, etc., a classification enhancement method for distinguishing a haze image and a rain and snow image according to chromaticity component values is proposed. The method adopts a pure physics algorithm, can distinguish a few image types, has poor robustness, and is very easy to have misjudgment and the like. In an imaging identification method and system in severe weather disclosed in chinese patent CN201811484514.3, a method for enhancing and then identifying collected low-quality images in severe weather is provided, an electronic image stabilization algorithm is used to eliminate motion blur of the images or an adaptive image defogging algorithm is used to eliminate cloud and fog interference of the images, and then a VGG16 convolutional neural network is used to identify and classify the images. However, the invention does not mention how to distinguish whether the acquired image is a low-quality image, and only one enhancement algorithm is adopted to perform enhancement processing on the low-quality images of all categories, so that the enhancement effect is general.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provide a low-quality image classification enhancement method based on model fusion and data enhancement.
The purpose of the invention can be realized by the following technical scheme:
a low-quality image classification enhancement method based on model fusion and data enhancement comprises the following steps:
s1: establishing an image set, wherein the image set comprises a clear image, a fuzzy image and a low-brightness image;
s2: performing data enhancement on the images in the image set;
s3: constructing a VGG16 convolutional neural network model, and training the VGG16 convolutional neural network model through an image set;
s4: constructing a ResNet convolutional neural network model, and training the ResNet convolutional neural network model through an image set;
s5: inputting an image to be classified;
s6: inputting an image to be classified into a VGG16 convolutional neural network model to obtain a first probability vector of the image to be classified, inputting the image to be classified into a ResNet convolutional neural network model to obtain a second probability vector of the image to be classified;
s7: fusing the first probability vector and the second probability vector to obtain a fused probability vector, and acquiring the image type of the image to be classified, if the image type is clear, entering a step S9, otherwise entering a step S8;
s8: selecting an image enhancement algorithm corresponding to the image type to enhance the image to be classified to obtain an enhanced image, and inputting the enhanced image serving as the image to be classified into the step S6;
s9: and outputting the image.
Preferably, the first probability vector is:
a1 ,λ a2 ,λ a3 ]
wherein λ is a1 ,λ a2 ,λ a3 Probability that the images acquired in the VGG16 convolutional neural network model are sharp images, blurred images and low-brightness images respectively,
the second probability vector is:
b1 ,λ b2 ,λ b3 ]
wherein λ is b1 ,λ b2 ,λ b3 Respectively the probability that the images acquired in the ResNet convolutional neural network model are sharp images, blurred images and low-brightness images,
the fusion probability vector is as follows:
a1b1 ,λ a2b2 ,λ a3b3 ]
and the image type represented by the maximum value in the fusion probability vector is the image type of the image to be classified.
Preferably, the classifier of the VGG16 convolutional neural network model comprises two fully-connected layers, wherein the second fully-connected layer adopts a Softmax activation function, and the second fully-connected layer maps the input values of the first fully-connected layer between intervals (0,1) to obtain the first probability vector.
Preferably, the classifier of the ResNet convolutional neural network model includes two fully-connected layers, wherein the second fully-connected layer uses a Softmax activation function, and the second fully-connected layer maps the input value of the first fully-connected layer to an interval (0,1) to obtain a second probability vector.
Preferably, the data enhancement is one or more of flipping, rotating, scaling, cropping, shifting, adding noise, and modifying contrast, at random, of the image data.
Preferably, the data enhancement in step S2 is to perform horizontal rotation and vertical rotation on the image data, and step S2 specifically is to:
s21: taking the vertex of the lower left corner of the image as an origin O, establishing an xyz space coordinate system, wherein the size of the image is a multiplied by b, and the image is superposed with an xy plane;
s22: randomly judging whether the image is horizontally rotated or not, wherein the probability of horizontal rotation or not is 0.5, if the image is horizontally rotated, rotating the image 180 degrees by taking (x = a/2, z = 0) as a rotating shaft, and if the image is not horizontally rotated, entering the step S22;
s23: and randomly judging whether the image is vertically rotated or not, wherein the probability of vertical rotation or not is 0.5, if the image is vertically rotated, rotating the image by 180 degrees by taking (y = b/2, z = 0) as a rotating shaft, and finishing data enhancement, otherwise, the image is not vertically rotated, and finishing data enhancement.
Preferably, when the number of times of enhancement of the image is greater than the maximum number of times of enhancement, the image is output:
s5: inputting an image to be classified, and setting the enhancement times p =0;
s6: inputting an image to be classified into a VGG16 convolutional neural network model to obtain a first probability vector of the image to be classified, inputting the image to be classified into a ResNet convolutional neural network model to obtain a second probability vector of the image to be classified;
s7: fusing the first probability vector and the second probability vector to obtain a fused probability vector, and acquiring the image type of the image to be classified, if the image type is clear, entering a step S9, otherwise entering a step S8;
s8: selecting an image enhancement algorithm corresponding to the image type to enhance the image to be classified, obtaining an enhanced image, setting the enhancement times p = p +1, if p is larger than or equal to the maximum enhancement times, entering the step S9, and otherwise, inputting the enhanced image serving as the image to be classified into the step S6;
s9: and outputting the image.
Preferably, the VGG16 convolutional neural network model comprises 13 convolutional layers, 3 fully-connected layers and 5 pooling layers.
Preferably, in S8, a GAN-based blind motion blur removal algorithm is used to perform enhancement processing on the image with the blurred image as the image type.
Preferably, in S8, a low-brightness image enhancement algorithm based on a corresponding camera model is used to perform enhancement processing on an image of which the image type is a low-brightness image.
Compared with the prior art, the invention has the following advantages:
(1) Compared with the prior art, the image classification method based on the convolutional neural network has the advantages that the convolutional neural network is adopted for classifying the image, the accuracy rate is higher compared with that of a traditional classifier based on physics, and meanwhile, the classification accuracy rate is improved by adopting two modes of data enhancement and model fusion;
(2) The invention integrates two algorithms special for enhancing two types of images, namely fuzzy images and low-brightness images, and can effectively classify and restore the detailed information of the images;
(3) In order to prevent two types of conditions from existing on one image, the method sets the maximum enhancement times, and performs classification enhancement on the image for multiple times to realize the maximum enhancement effect;
(4) The random image data enhancement mode is utilized to perform data enhancement of horizontal rotation and vertical rotation on the image data, and the classification accuracy is effectively improved.
Drawings
FIG. 1 is a flow chart of the present invention;
FIG. 2 is a flow chart of an embodiment of the present invention;
FIG. 3 is a block diagram of the VGG16 model;
FIG. 4 is a block diagram of the VGG16 model;
FIG. 5 is a block diagram of a residual unit;
FIG. 6 is a block diagram of the ResNet model;
fig. 7 is a structural diagram of the DeblurGAN model.
Detailed Description
The invention is described in detail below with reference to the figures and specific embodiments. Note that the following description of the embodiments is merely a substantial example, and the present invention is not intended to be limited to the application or the use thereof, and is not limited to the following embodiments.
Examples
A low-quality image classification enhancement method based on model fusion and data enhancement is disclosed, as shown in FIG. 1, and comprises the following steps:
s1: an image set is established.
In the image set of the present embodiment, the image set includes three types of images, i.e., a sharp image, a blurred image, and a low-brightness image, and includes a training set, a test set, and a verification set, where 300 images of each type are included in the training set, 50 images of each type are included in the test set, and 50 images of each type are included in the verification set.
S2: data enhancement is performed on the images in the image set.
Data enhancement is one or more of flipping, rotating, scaling, cropping, shifting, adding noise, and modifying contrast on the image data at random.
In this embodiment, the data enhancement in step S2 is to perform horizontal rotation and vertical rotation on the image data, where step S2 specifically is:
s21: taking the vertex of the lower left corner of the image as an origin O, establishing an xyz space coordinate system, wherein the size of the image is a multiplied by b, and the image is superposed with an xy plane;
s22: randomly judging whether the image is horizontally rotated or not, wherein the probability of horizontal rotation or not is 0.5, if the image is horizontally rotated, rotating the image 180 degrees by taking (x = a/2, z = 0) as a rotating shaft, and if the image is not horizontally rotated, entering the step S22;
s23: and randomly judging whether the image is vertically rotated or not, wherein the probability of vertical rotation or not is 0.5, if the image is vertically rotated, rotating the image by 180 degrees by taking (y = b/2, z = 0) as a rotating shaft, and finishing data enhancement, otherwise, the image is not vertically rotated, and finishing data enhancement.
S3: and constructing a VGG16 convolutional neural network model, and training the VGG16 convolutional neural network model through the image set.
As shown in fig. 3 and 4, the model structure of the VGG16 includes 13 Convolutional layers (volumetric Layer), 3 Fully connected layers (full connected Layer), and 5 pooling layers (Pool Layer). The convolutional layers and the pooling layers can be divided into different blocks (blocks), which are numbered from front to back as blocks 1-5, and each Block comprises a plurality of convolutional layers and pooling layers, for example, block4 comprises 3 convolutional layers (Conv 3-512) and 1 pooling layer (Maxpool); and within the same block, the number of channels of the convolutional layers is the same, for example: block2 contains 2 convolutional layers, each convolutional layer is represented by Conv2-128, namely the convolutional core is 3 x 3, and the number of channels is 128; block3 contains 3 convolutional layers, each denoted by Conv3-256, i.e., the convolutional kernel is 3 × 3 × 3, and the number of channels is 256. The input structure of the VGG16 model is 224 × 224 × 3, where the number of channels is doubled, from 64 to 128 to 256 in sequence, until 512 remains the same and is no longer doubled, and the height and width of the image is halved, from 224 → 112 → 56 → 28 → 14 → 7.
S4: and constructing a ResNet convolutional neural network model, and training the ResNet convolutional neural network model through the image set.
The ResNet convolution neural network solves the degradation problem of the deep network through residual learning: as network depth increases, network accuracy saturates and even decreases. As shown in fig. 5 and 6, the structure of the ResNet model refers to the VGG19 structure, on the basis of which a residual unit is modified and added through a short circuit mechanism.
The residual unit has two layers, the first layer:
F(x)=W 2 σ(W 1 x)
where σ represents the nonlinear activation function ReLu, W 1 And W 2 Two weights in the first layer are represented.
Then the first layer is connected with the second layer through a short circuit, the ReLu function of the second layer obtains output y,
y=F(x)+x
where F (x) is the output of the first layer.
The ResNet is mainly characterized in that the convolution with the convolution kernel size of 2 is directly used for down-sampling, and a global average pooling layer is adopted to replace a full-connection layer. Compared with the common network, the ResNet network imposes a short circuit mechanism between every two layers, so that residual error learning is formed.
S5: inputting an image to be classified;
s6: and inputting the image to be classified into a VGG16 convolutional neural network model to obtain a first probability vector of the image to be classified, inputting the image to be classified into a ResNet convolutional neural network model to obtain a second probability vector of the image to be classified.
S7: and fusing the first probability vector and the second probability vector to obtain a fused probability vector, and acquiring the image type of the image to be classified, if the image type is clear, entering a step S9, and if not, entering a step S8.
The invention adopts model fusion of a VGG16 convolution neural network model and a ResNet convolution neural network model. Model fusion is to directly add probability vectors predicted by two or more basic models, and take the class with the highest probability after addition as the prediction class of the image.
In this embodiment, the probability vector is represented in a Onehot encoding form. The Onehot coding is a common class label coding form in a neural network, and converts a class label with an integer value into a binary value, wherein the index of the label value is marked as 1, and the other indexes are 0. In this embodiment, there are 3 category labels for the sharp image, the blurred image, and the low-luminance image, which correspond to label values 0,1, and 2, respectively. The label value 0 is converted into [1,0,0] after being subjected to One-Hot encoding, the label value 1 is converted into [0,1,0], and the label value 2 is converted into [0,0,1].
In this embodiment, the two model classifiers each include two fully-connected layers, and the second fully-connected layer employs a Softmax activation function, and the second fully-connected layer maps the input values from the first fully-connected layer to intervals (0,1), and obtains a first probability vector and a second probability vector. Specifically, the Softmax function maps the input values of the first fully-connected layer between intervals (0,1), forming, for example, [ λ [ [ λ ] 1 ,λ 2 ,λ 3 ]Of the probability vector of i For each class of probability, the sum of which is 1, usually taken as λ i The largest median is taken as the predicted class.
The formula of the Softmax function is:
Figure BDA0002561455890000071
wherein Z is i And C is the output value of the ith node, and the number of output nodes, namely the number of classified categories.
In this embodiment, the pre-trained VGG16 convolutional neural network model and ResNet convolutional neural network model are adopted to output the first probability vector [ λ [ lambda ] ], respectively a1 ,λ a2 ,λ a3 ]And a second probability vector [ lambda ] b1 ,λ b2 ,λ b3 ]。
Wherein, λ a1 ,λ a2 ,λ a3 Probability of the images acquired in the VGG16 convolutional neural network model being respectively a sharp image, a blurred image and a low-brightness image, lambda b1 ,λ b2 ,λ b3 Respectively, the probability that the image obtained in the ResNet convolutional neural network model is a sharp image, a blurred image, and a low-brightness image, respectively.
Adding the two probability vectors to obtain a fusion probability vector
a1b1 ,λ a2b2 ,λ a3b3 ]
And the image type represented by the maximum value in the fusion probability vector is the image type of the image to be classified.
S8: and (6) selecting an image enhancement algorithm corresponding to the image type to enhance the image to be classified to obtain an enhanced image, and inputting the enhanced image serving as the image to be classified into the step S6.
Specifically, a GAN-based blind motion blur removal algorithm is adopted to enhance an image with a blurred image as an image type, and a corresponding camera model-based low-brightness image enhancement algorithm is adopted to enhance an image with a low-brightness image as an image type.
Further, a GAN-based blind motion blur removal algorithm:
the algorithm treats image blurring removal as an end-to-end task through a GAN network, and a clear image is automatically generated according to a blurred image by using a generator through learning the blurred image and the clear image. The blind deblurring in the algorithm aims to restore a clear image Is by giving a blurred image IB under the condition of no blur kernel, deblurring Is carried out by adopting a generator, and a discrimination network Is introduced in the training process to carry out training and learning in a countermeasure mode.
As shown IN fig. 7, the generator contains two downsampling convolution modules, 9 residual modules (containing one convolution, IN and ReLU), and two upsampling transposed convolution modules, while also introducing global residual concatenation. This structure may allow faster training and at the same time better performance. Besides the generator, the algorithm also defines a discriminator in the training process, and adopts WGAN (Wasserstein GAN) with a penalty term to carry out countertraining.
The loss function of the algorithm comprises two parts: content loss and antagonism loss, the formula is:
L=L CAN +λL x
wherein L is GAN To combat losses, L x For content loss, λ is the weight.
In countering loss, using WGAN-GP as a penalty function, the loss is calculated as:
Figure BDA0002561455890000081
wherein, I B In order to input a blurred image, the image is displayed,
Figure BDA0002561455890000085
in order to be a generator,
Figure BDA0002561455890000086
to discriminate the network.
The content loss adopts the perception loss, and is based on the difference between the generated and target image CNN feature mapping, and the formula is as follows:
Figure BDA0002561455890000082
wherein, I s For the generated image, I B In order to input a blurred image, the image is displayed,
Figure BDA0002561455890000083
to be a generator, phi i,j Is a feature map obtained by the jth convolution (after activation) before the ith max pooling layer in the VGG19 network, W i,j And H i,j Is a dimension function graph.
Further, a low brightness picture enhancement algorithm based on the corresponding model of the camera:
for a picture taken by a camera, the pixel value is not proportional to the brightness value reflected by the object. The nonlinear transformation Function from the luminance value sensed by the Camera photosensitive element to the actual pixel value of the image is called a Camera Response Function (CRF). The algorithm firstly waits for a camera response model by analyzing the relation between pictures with different exposure degrees, then obtains an exposure contrast image of the image by using a picture brightness component estimation method, and finally enhances the low-brightness picture by using the camera response model and the exposure contrast image.
The camera response model is defined as:
P=f(E)
where E is the irradiance of the picture, P is the pixel value of the picture, and f should satisfy:
Figure BDA0002561455890000084
the algorithm uses a luminance Transform Function (BTF) to estimate f.
BTF is a mapping function between two graphs of the same scene but different exposure levels:
R=g(P 0 ,k)
wherein P is 1 And P 0 Is an image with different exposure degrees in the same scene, and k is an exposure ratio.
CRM can therefore be calculated from g (f (E), k) = f (kE).
S9: and outputting the image.
The experimental environment in this example is: the accuracy of the system Win10, the display card Tesla P100, the deep learning framework Tensorflow2.0, the VGG16, the ResNet model and the fused model is shown in the following table, and the accuracy of the fused model is higher than that of any single model, which is shown in the following table.
Figure BDA0002561455890000091
In an embodiment of the present invention, in order to prevent the image processed by the two enhancement algorithms from being blurred or with low brightness, as shown in fig. 2, a maximum enhancement number is set, and when the enhancement number of the image is greater than or equal to the maximum enhancement number, the image is output:
s1: establishing an image set, wherein the image set comprises a clear image, a fuzzy image and a low-brightness image;
s2: performing data enhancement on the images in the image set;
s3: constructing a VGG16 convolutional neural network model, and training the VGG16 convolutional neural network model through an image set;
s4: constructing a ResNet convolutional neural network model, and training the ResNet convolutional neural network model through an image set;
s5: inputting an image to be classified, and setting the enhancement times p =0;
s6: inputting an image to be classified into a VGG16 convolutional neural network model to obtain a first probability vector of the image to be classified, inputting the image to be classified into a ResNet convolutional neural network model to obtain a second probability vector of the image to be classified;
s7: fusing the first probability vector and the second probability vector to obtain a fused probability vector, and acquiring the image category of the image to be classified, if the image type is clear, entering a step S9, otherwise entering a step S8;
s8: selecting an image enhancement algorithm corresponding to the image type to enhance the image to be classified, obtaining an enhanced image, setting the enhancement times p = p +1, if p is larger than or equal to the maximum enhancement times, entering the step S9, and otherwise, inputting the enhanced image serving as the image to be classified into the step S6;
s9: and outputting the image.
The above embodiments are merely examples and do not limit the scope of the present invention. These embodiments may be implemented in other various manners, and various omissions, substitutions, and changes may be made without departing from the technical spirit of the present invention.

Claims (8)

1. A low-quality image classification enhancement method based on model fusion and data enhancement is characterized by comprising the following steps:
s1: establishing an image set, wherein the image set comprises a clear image, a fuzzy image and a low-brightness image;
s2: performing data enhancement on the image in the image set, wherein the data enhancement is one or more operations of turning, rotating, scaling, cutting, shifting, adding noise and modifying contrast on the image data randomly;
s3: constructing a VGG16 convolutional neural network model, and training the VGG16 convolutional neural network model through an image set;
s4: constructing a ResNet convolutional neural network model, and training the ResNet convolutional neural network model through an image set;
s5: inputting an image to be classified;
s6: inputting an image to be classified into a VGG16 convolutional neural network model to obtain a first probability vector of the image to be classified, inputting the image to be classified into a ResNet convolutional neural network model to obtain a second probability vector of the image to be classified;
s7: fusing the first probability vector and the second probability vector to obtain a fused probability vector, and acquiring the image type of the image to be classified, if the image type is clear, entering a step S9, otherwise entering a step S8;
s8: selecting an image enhancement algorithm corresponding to the image type to enhance the image to be classified to obtain an enhanced image, and inputting the enhanced image serving as the image to be classified into the step S6;
s9: outputting an image;
and in the S8, the image with the image type of the blurred image is enhanced by adopting a GAN-based blind motion blur removal algorithm.
2. The method of claim 1, wherein the first probability vector is:
a1 ,λ a2 ,λ a3 ]
wherein λ is a1 ,λ a2 ,λ a3 Probability that the images acquired in the VGG16 convolutional neural network model are sharp images, blurred images and low-brightness images respectively,
the second probability vector is:
b1 ,λ b2 ,λ b3 ]
wherein λ is b1 ,λ b2 ,λ b3 Respectively the probability that the images acquired in the ResNet convolutional neural network model are sharp images, blurred images and low-brightness images,
the fusion probability vector is as follows:
a1b1 ,λ a2b2 ,λ a3ba ]
and the image type represented by the maximum value in the fusion probability vector is the image type of the image to be classified.
3. The method of claim 1, wherein the classifier of the VGG16 convolutional neural network model comprises two fully-connected layers, wherein the second fully-connected layer uses a Softmax activation function, and the second fully-connected layer maps the input values from the first fully-connected layer to the intervals (0,1) to obtain the first probability vector.
4. The method of claim 1, wherein the classifier of the ResNet convolutional neural network model comprises two fully-connected layers, wherein the second fully-connected layer uses a Softmax activation function, and the second fully-connected layer maps the input values from the first fully-connected layer to the intervals (0,1) to obtain the second probability vector.
5. The method according to claim 1, wherein the data enhancement in step S2 is performed by horizontally rotating and vertically rotating the image data, and the step S2 specifically comprises:
s21: taking the vertex of the lower left corner of the image as an origin O, establishing an xyz space coordinate system, wherein the size of the image is a multiplied by b, and the image is superposed with an xy plane;
s22: randomly judging whether the image is horizontally rotated or not, wherein the probability of horizontal rotation or not is 0.5, if the image is horizontally rotated, rotating the image by 180 degrees by taking (x = a/2, z = 0) as a rotating shaft, and otherwise, entering the step S22;
s23: and randomly judging whether the image is vertically rotated or not, wherein the probability of vertical rotation or not is 0.5, if the image is vertically rotated, rotating the image by 180 degrees by taking (y = b/2, z = 0) as a rotating shaft, and finishing data enhancement, otherwise, the image is not vertically rotated, and finishing data enhancement.
6. The method for classifying and enhancing the low-quality image based on the model fusion and the data enhancement as claimed in claim 1, wherein when the enhancing times of the image is greater than or equal to the maximum enhancing times, the image is output:
s5: inputting an image to be classified, and setting the enhancement times p =0;
s6: inputting the image to be classified into a VGG16 convolution neural network model to obtain a first probability vector of the image to be classified, inputting the image to be classified into a ResNet convolution neural network model to obtain a second probability vector of the image to be classified;
s7: fusing the first probability vector and the second probability vector to obtain a fused probability vector, and acquiring the image type of the image to be classified, if the image type is clear, entering a step S9, otherwise entering a step S8;
s8: selecting an image enhancement algorithm corresponding to the image type to enhance the image to be classified, obtaining an enhanced image, setting the enhancement times p = p +1, if p is larger than or equal to the maximum enhancement times, entering the step S9, and otherwise, inputting the enhanced image serving as the image to be classified into the step S6;
s9: and outputting the image.
7. The method of claim 1, wherein the VGG16 convolutional neural network model comprises 13 convolutional layers, 3 fully-connected layers and 5 pooling layers.
8. The method according to claim 1, wherein in S8, an image with a low-brightness image type is enhanced by using a low-brightness image enhancement algorithm based on a corresponding camera model.
CN202010607913.5A 2020-06-30 2020-06-30 Low-quality image classification enhancement method based on model fusion and data enhancement Active CN111815529B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010607913.5A CN111815529B (en) 2020-06-30 2020-06-30 Low-quality image classification enhancement method based on model fusion and data enhancement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010607913.5A CN111815529B (en) 2020-06-30 2020-06-30 Low-quality image classification enhancement method based on model fusion and data enhancement

Publications (2)

Publication Number Publication Date
CN111815529A CN111815529A (en) 2020-10-23
CN111815529B true CN111815529B (en) 2023-02-07

Family

ID=72856666

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010607913.5A Active CN111815529B (en) 2020-06-30 2020-06-30 Low-quality image classification enhancement method based on model fusion and data enhancement

Country Status (1)

Country Link
CN (1) CN111815529B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112541877B (en) * 2020-12-24 2024-03-19 广东宜教通教育有限公司 Defuzzification method, system, equipment and medium for generating countermeasure network based on condition
CN113469083A (en) * 2021-07-08 2021-10-01 西安电子科技大学 SAR image target classification method and system based on anti-sawtooth convolution neural network

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650721A (en) * 2016-12-28 2017-05-10 吴晓军 Industrial character identification method based on convolution neural network
CN106875352A (en) * 2017-01-17 2017-06-20 北京大学深圳研究生院 A kind of enhancement method of low-illumination image
CN107169450A (en) * 2017-05-15 2017-09-15 中国科学院遥感与数字地球研究所 The scene classification method and system of a kind of high-resolution remote sensing image
CN108734667A (en) * 2017-04-14 2018-11-02 Tcl集团股份有限公司 A kind of image processing method and system
CN109522945A (en) * 2018-10-31 2019-03-26 中国科学院深圳先进技术研究院 One kind of groups emotion identification method, device, smart machine and storage medium
CN109801224A (en) * 2018-12-04 2019-05-24 北京奇艺世纪科技有限公司 A kind of image processing method, device, server and storage medium
CN109934293A (en) * 2019-03-15 2019-06-25 苏州大学 Image-recognizing method, device, medium and obscure perception convolutional neural networks
CN110264424A (en) * 2019-06-20 2019-09-20 北京理工大学 A kind of fuzzy retinal fundus images Enhancement Method based on generation confrontation network
CN110428011A (en) * 2019-08-06 2019-11-08 华南理工大学 A kind of deep learning image fault classification method towards video transmission quality
CN110956201A (en) * 2019-11-07 2020-04-03 江南大学 Image distortion type classification method based on convolutional neural network
CN111127435A (en) * 2019-12-25 2020-05-08 福州大学 No-reference image quality evaluation method based on double-current convolutional neural network

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650721A (en) * 2016-12-28 2017-05-10 吴晓军 Industrial character identification method based on convolution neural network
CN106875352A (en) * 2017-01-17 2017-06-20 北京大学深圳研究生院 A kind of enhancement method of low-illumination image
CN108734667A (en) * 2017-04-14 2018-11-02 Tcl集团股份有限公司 A kind of image processing method and system
CN107169450A (en) * 2017-05-15 2017-09-15 中国科学院遥感与数字地球研究所 The scene classification method and system of a kind of high-resolution remote sensing image
CN109522945A (en) * 2018-10-31 2019-03-26 中国科学院深圳先进技术研究院 One kind of groups emotion identification method, device, smart machine and storage medium
CN109801224A (en) * 2018-12-04 2019-05-24 北京奇艺世纪科技有限公司 A kind of image processing method, device, server and storage medium
CN109934293A (en) * 2019-03-15 2019-06-25 苏州大学 Image-recognizing method, device, medium and obscure perception convolutional neural networks
CN110264424A (en) * 2019-06-20 2019-09-20 北京理工大学 A kind of fuzzy retinal fundus images Enhancement Method based on generation confrontation network
CN110428011A (en) * 2019-08-06 2019-11-08 华南理工大学 A kind of deep learning image fault classification method towards video transmission quality
CN110956201A (en) * 2019-11-07 2020-04-03 江南大学 Image distortion type classification method based on convolutional neural network
CN111127435A (en) * 2019-12-25 2020-05-08 福州大学 No-reference image quality evaluation method based on double-current convolutional neural network

Also Published As

Publication number Publication date
CN111815529A (en) 2020-10-23

Similar Documents

Publication Publication Date Title
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN112233038B (en) True image denoising method based on multi-scale fusion and edge enhancement
CN111209952B (en) Underwater target detection method based on improved SSD and migration learning
EP3937481A1 (en) Image display method and device
CN107274445B (en) Image depth estimation method and system
CN111402146B (en) Image processing method and image processing apparatus
US10614736B2 (en) Foreground and background detection method
CN111754446A (en) Image fusion method, system and storage medium based on generation countermeasure network
CN111915525B (en) Low-illumination image enhancement method capable of generating countermeasure network based on improved depth separation
CN110866879B (en) Image rain removing method based on multi-density rain print perception
US10706558B2 (en) Foreground and background detection method
CN111260738A (en) Multi-scale target tracking method based on relevant filtering and self-adaptive feature fusion
CN110288550B (en) Single-image defogging method for generating countermeasure network based on priori knowledge guiding condition
CN110148088B (en) Image processing method, image rain removing method, device, terminal and medium
CN109903315B (en) Method, apparatus, device and readable storage medium for optical flow prediction
CN110532959B (en) Real-time violent behavior detection system based on two-channel three-dimensional convolutional neural network
CN111815529B (en) Low-quality image classification enhancement method based on model fusion and data enhancement
WO2010043771A1 (en) Detecting and tracking objects in digital images
CN112819858B (en) Target tracking method, device, equipment and storage medium based on video enhancement
CN111242026A (en) Remote sensing image target detection method based on spatial hierarchy perception module and metric learning
CN112115979A (en) Fusion method and device of infrared image and visible image
CN111582074A (en) Monitoring video leaf occlusion detection method based on scene depth information perception
CN113034417A (en) Image enhancement system and image enhancement method based on generation countermeasure network
CN115272437A (en) Image depth estimation method and device based on global and local features
Chen et al. Visual depth guided image rain streaks removal via sparse coding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant