CN113723231A - Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device - Google Patents

Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device Download PDF

Info

Publication number
CN113723231A
CN113723231A CN202110940177.XA CN202110940177A CN113723231A CN 113723231 A CN113723231 A CN 113723231A CN 202110940177 A CN202110940177 A CN 202110940177A CN 113723231 A CN113723231 A CN 113723231A
Authority
CN
China
Prior art keywords
semantic segmentation
low
image
loss function
illumination
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110940177.XA
Other languages
Chinese (zh)
Inventor
王卓恒
葛琦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Posts and Telecommunications
Original Assignee
Nanjing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Posts and Telecommunications filed Critical Nanjing University of Posts and Telecommunications
Priority to CN202110940177.XA priority Critical patent/CN113723231A/en
Publication of CN113723231A publication Critical patent/CN113723231A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Computing Systems (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a training method of a low-illumination semantic segmentation model, a semantic segmentation method and a device, wherein the training method of the low-illumination semantic segmentation model comprises the following steps: acquiring an image data set under low illumination; respectively carrying out enhancement processing on the acquired image data through an image enhancement model to obtain an enhanced image data set and an enhanced image loss function; respectively performing semantic segmentation on the acquired enhanced image data through a semantic segmentation model to acquire a semantic segmentation image data set and a semantic segmentation loss function; obtaining a total loss function according to the enhanced image loss function and the semantic segmentation loss function; and adjusting the semantic segmentation model according to a total loss function to obtain a low-illumination semantic segmentation model. The method is used for obtaining accurate segmentation precision in a low-illumination scene.

Description

Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device
Technical Field
The invention relates to image processing and semantic segmentation technologies, in particular to a low-illumination semantic segmentation model training method, a semantic segmentation method and a semantic segmentation device.
Background
The semantic segmentation provides semantic information of a driving environment for a control system of the automobile, helps a driving system to judge driving, and is important for applications such as automatic driving and geological survey. At present, the semantic algorithm for the city street view of automatic driving only acts on an ideal scene with sufficient illumination, the robustness of other scenes is not high, and particularly, when low-illumination scenes such as night and the like are processed, an accurate semantic segmentation result is difficult to obtain.
Disclosure of Invention
The purpose of the invention is as follows: in order to overcome the defects of the prior art, the invention provides a low-illumination semantic segmentation model training method, a semantic segmentation method and a semantic segmentation device, so as to improve the semantic segmentation accuracy under a low-illumination scene.
The technical scheme is as follows: in a first aspect, the present invention provides a training method for a low-light semantic segmentation model, where the method includes:
acquiring an image data set under low illumination; respectively carrying out enhancement processing on the acquired image data through an image enhancement model to obtain an enhanced image data set and an enhanced image loss function; respectively performing semantic segmentation on the acquired enhanced image data through a semantic segmentation model to acquire a semantic segmentation image data set and a semantic segmentation loss function; obtaining a total loss function according to the enhanced image loss function and the semantic segmentation loss function; and adjusting the semantic segmentation model according to a total loss function to obtain a low-illumination semantic segmentation model.
Optionally, the acquisition of the low-light image dataset is based on a cityscaps dataset.
Optionally, the function of the image enhancement model is represented as:
Figure BDA0003214518720000021
wherein the content of the first and second substances,
Figure BDA0003214518720000022
representing normal images output after image enhancement, ILAnd (4) representing a mapping function of the low-illumination image f (x) representing the brightness adjustment of the image, wherein a Sigmod activation function is adopted, and theta is a network parameter of the enhancement network.
Optionally, the enhanced image loss function is expressed as:
Figure BDA0003214518720000023
where N represents the total number of pixels in the input image,
Figure BDA0003214518720000024
and
Figure BDA0003214518720000025
and the output pixel result of the network and the label pixel at the corresponding position are respectively represented, theta is a network parameter of the enhancement network, and beta is set to be 0.1 in the training process.
Optionally, the semantic segmentation model is obtained based on a ghostNet neural network model training.
Optionally, the semantic segmentation loss function is expressed as:
Figure BDA0003214518720000026
wherein, yiRepresenting the probability of presoftmax belonging to class i, e being a constant.
Optionally, the total loss function is obtained by performing cascade training on an enhanced image loss function and a semantic segmentation loss function.
Optionally, the total loss function is represented as:
L=βDLDSLS (4)
where L represents the total loss function of the entire network, LDAnd LSRepresenting the loss function, beta, of the enhancement network and the segmentation networkSAnd betaDTo representThe weight of the function is lost.
In a second aspect, a method for low-illumination semantic segmentation, the method comprising:
acquiring a source image, and carrying out image enhancement on the source image through an image enhancement model to obtain an enhanced source image; and performing semantic segmentation on the enhanced source image through a low-illumination semantic segmentation model to obtain segmentation precision, wherein the low-illumination semantic segmentation model is obtained through training by using the method of any one of the first aspect.
In a third aspect, a low-illumination semantic segmentation apparatus, the apparatus comprising:
the enhancement unit is used for carrying out image enhancement on the source image through the image enhancement model to obtain an enhanced source image; and the segmentation unit is used for performing semantic segmentation on the enhanced source image through a low-illumination semantic segmentation model to obtain segmentation precision, wherein the low-illumination semantic segmentation model is obtained through training by using the method of any one of the third aspect.
Has the advantages that: according to the method, the semantic segmentation model is adjusted according to the total loss function, the image loss is enhanced, and the segmentation precision is improved; a semantic segmentation model is established based on a ghostNet neural network model, the total sum of parameters required in a Ghost module and the calculation complexity are reduced, and the segmentation efficiency is improved; and the image enhancement model adopts a self-coding structure with jump links, so that the image brightness is improved, and more image details can be acquired.
Drawings
Fig. 1 is a schematic flow chart of a training method of a low-illumination semantic segmentation model according to an embodiment of the present invention.
Detailed Description
The invention is further described below with reference to the accompanying drawings. The following examples are only for illustrating the technical solutions of the present invention more clearly, and the protection scope of the present invention is not limited thereby.
Example 1:
as shown in FIG. 1, the invention relates to a low-illumination semantic segmentation model training method, which comprises the following steps:
the Cityscapes data set is selected as a main data set of an experiment, and the Gamma value of the image is adjusted firstly, namely the image is subjected to inverse Gamma transformation. Gamma transformation is an image operation for performing nonlinear adjustment on an image input gray value, and can make the gray value of an image before transformation and the gray value of the image after transformation have an exponential nonlinear relationship, as shown in formula (1):
Figure BDA0003214518720000041
wherein IinIoutEach represents an input/output image, β represents a Gamma value, and α represents a transform coefficient. After Gamma conversion, in order to obtain a better low-illumination effect of the image, a PS tool is also used for adjusting the brightness and the contrast of the city maps street view image after Gamma conversion, so that the finally manually adjusted low-illumination street view data is similar to the actual low-illumination dark scene as much as possible in vision to obtain the low-illumination street view image with good visual effect, thus on the basis of the original data set containing the street view image of normal weather and the corresponding semantic segmentation label, a new corresponding low-illumination image is generated, each group of data contains three city street view images with low illumination respectively, and the city street view image and the semantic segmentation label in normal day. The artwork and segmentation labels for these datasets are from the cityscaps dataset.
The obtained low-illumination image is subjected to image enhancement processing by using a self-coding structure with jump connection, the number of parameters can be effectively reduced through dimension reduction of a coder, representative characteristics of the image are obtained, and a decoder achieves the denoising effect, as shown in formula (2):
Figure BDA0003214518720000051
in the formula (2), the reaction mixture is,
Figure BDA0003214518720000052
representing normal images output after image enhancement, ILIndicating low lightThe image f (x) represents a mapping function of image brightness adjustment, wherein a Sigmod activation function is adopted, and theta is a network parameter of the enhancement network.
When a normal image is obtained, the loss of an image enhancement network structure can be generated, and Mean Absolute Error (MAE) and a quadratic norm (L) are adopted in the training of the enhancement network structure2Error) to calculate the total error of the enhancement network, as shown in equation (3):
Figure BDA0003214518720000053
n represents the total number of pixels in the input image,
Figure BDA0003214518720000054
and
Figure BDA0003214518720000055
label pixels respectively representing the output pixel result and the corresponding position of the network; l is2The regularization term can prevent the network from over-fitting, improve the convergence rate of the network, and beta is set to 0.1 during the training process.
The method is characterized in that the method selects the ghostNet as a semantic segmentation network, and the ghostNet is used as a lightweight network which is applied to generating more characteristic maps by a series of simple linear operations used by the convolutional layer. Compared with the common convolutional neural network, the total number of parameters and the calculation complexity required in the Ghost module are reduced under the condition of not changing the size of the output feature map, and all convolutions in the module are replaced by pointwise convolutions in order to improve the efficiency.
The loss function of semantic segmentation, adopting the cross entropy loss function of softmax, comprises two steps for each pixel in the image, and the mathematical expression of the two steps is shown in the formula (4) (5):
Figure BDA0003214518720000056
Figure BDA0003214518720000057
formula (4), yiRepresenting the probability of presoftmax belonging to class i, e being a constant. Firstly, softmax is calculated for the last layer output of the cascade segmentation network to obtain the probability that each pixel belongs to a certain class, then cross entropy loss L is calculated for the predicted probability vector, each pixel is taken as a sample to calculate the cross entropy loss, and then the average value is taken as the final segmentation network loss Ls
The mathematical expression of the final total loss function of two networks which jointly train their losses through the cascade network is shown as the formula (6)
L=βDLDSLS (6)
Where L represents the total loss function of the entire network, LDAnd LSRepresenting the loss function, beta, of the enhancement network and the segmentation networkSAnd betaDRepresenting the weight of the loss function.
And adjusting the semantic segmentation model established based on the ghostNet neural network model according to the total loss function to obtain the low-illumination semantic segmentation model.
Example 2:
a method of low-light semantic segmentation, the method comprising:
acquiring a source image, and carrying out image enhancement on the source image through an image enhancement model to obtain an enhanced source image;
performing semantic segmentation on the enhanced source image through a low-illumination semantic segmentation model to obtain segmentation precision,
wherein the low-light semantic segmentation model is obtained by training according to the method of any one of embodiment 1.
Example 3:
a low-light semantic segmentation apparatus, the apparatus comprising:
the enhancement unit is used for carrying out image enhancement on the source image through the image enhancement model to obtain an enhanced source image;
a segmentation unit for performing semantic segmentation on the enhanced source image through a low-illumination semantic segmentation model to obtain segmentation accuracy,
wherein the low-light semantic segmentation model is obtained by training according to the method of any one of embodiment 1.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.

Claims (10)

1. A training method of a low-illumination semantic segmentation model is characterized by comprising the following steps:
acquiring an image data set under low illumination;
respectively carrying out enhancement processing on the acquired image data through an image enhancement model to obtain an enhanced image data set and an enhanced image loss function;
respectively performing semantic segmentation on the acquired enhanced image data through a semantic segmentation model to acquire a semantic segmentation image data set and a semantic segmentation loss function;
obtaining a total loss function according to the enhanced image loss function and the semantic segmentation loss function;
and adjusting the semantic segmentation model according to a total loss function to obtain a low-illumination semantic segmentation model.
2. The method for training the low-light semantic segmentation model according to claim 1, wherein the low-light image dataset is obtained based on a cityscaps dataset.
3. The method for training the low-light semantic segmentation model according to claim 1, wherein the function of the image enhancement model is expressed as:
Figure FDA0003214518710000011
wherein the content of the first and second substances,
Figure FDA0003214518710000012
representing normal images output after image enhancement, ILAnd (4) representing a mapping function of the low-illumination image f (x) representing the brightness adjustment of the image, wherein a Sigmod activation function is adopted, and theta is a network parameter of the enhancement network.
4. The method for training the low-light semantic segmentation model according to claim 3, wherein the enhanced image loss function is expressed as:
Figure FDA0003214518710000021
where N represents the total number of pixels in the input image,
Figure FDA0003214518710000022
and
Figure FDA0003214518710000023
and the output pixel result of the network and the label pixel at the corresponding position are respectively represented, theta is a network parameter of the enhancement network, and beta is set to be 0.1 in the training process.
5. The method for training the low-illumination semantic segmentation model according to claim 1, wherein the semantic segmentation model is obtained based on a ghostNet neural network model training.
6. The method for training the low-light semantic segmentation model according to claim 5, wherein the semantic segmentation loss function is expressed as:
Figure FDA0003214518710000024
wherein, yiRepresenting the probability of presoftmax belonging to class i, e being a constant.
7. The method for training the low-illumination semantic segmentation model according to claim 1, wherein the total loss function is obtained by training an enhanced image loss function and a semantic segmentation loss function in a cascade manner.
8. The method for training the low-light semantic segmentation model according to claim 7, wherein the total loss function is expressed as:
L=βDLDSLS (4)
where L represents the total loss function of the entire network, LDAnd LSRepresenting the loss function, beta, of the enhancement network and the segmentation networkSAnd betaDRepresenting the weight of the loss function.
9. A low-light semantic segmentation method, comprising:
acquiring a source image, and carrying out image enhancement on the source image through an image enhancement model to obtain an enhanced source image;
performing semantic segmentation on the enhanced source image through a low-illumination semantic segmentation model to obtain segmentation precision,
wherein the low-light semantic segmentation model is trained by the method of any one of claims 1-8.
10. An apparatus for low-light semantic segmentation, the apparatus comprising:
the enhancement unit is used for carrying out image enhancement on the source image through the image enhancement model to obtain an enhanced source image;
a segmentation unit for performing semantic segmentation on the enhanced source image through a low-illumination semantic segmentation model to obtain segmentation accuracy,
wherein the low-light semantic segmentation model is trained by the method of any one of claims 1-8.
CN202110940177.XA 2021-08-17 2021-08-17 Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device Pending CN113723231A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110940177.XA CN113723231A (en) 2021-08-17 2021-08-17 Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110940177.XA CN113723231A (en) 2021-08-17 2021-08-17 Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device

Publications (1)

Publication Number Publication Date
CN113723231A true CN113723231A (en) 2021-11-30

Family

ID=78676694

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110940177.XA Pending CN113723231A (en) 2021-08-17 2021-08-17 Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device

Country Status (1)

Country Link
CN (1) CN113723231A (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109410129A (en) * 2018-09-28 2019-03-01 大连理工大学 A kind of method of low light image scene understanding
US20200218961A1 (en) * 2017-09-27 2020-07-09 Google Llc End to End Network Model for High Resolution Image Segmentation
CN111507343A (en) * 2019-01-30 2020-08-07 广州市百果园信息技术有限公司 Training of semantic segmentation network and image processing method and device thereof
AU2020103901A4 (en) * 2020-12-04 2021-02-11 Chongqing Normal University Image Semantic Segmentation Method Based on Deep Full Convolutional Network and Conditional Random Field
US20210241107A1 (en) * 2019-03-26 2021-08-05 Tencent Technology (Shenzhen) Company Limited Method, apparatus, device, and storage medium for training image semantic segmentation network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200218961A1 (en) * 2017-09-27 2020-07-09 Google Llc End to End Network Model for High Resolution Image Segmentation
CN109410129A (en) * 2018-09-28 2019-03-01 大连理工大学 A kind of method of low light image scene understanding
CN111507343A (en) * 2019-01-30 2020-08-07 广州市百果园信息技术有限公司 Training of semantic segmentation network and image processing method and device thereof
US20210241107A1 (en) * 2019-03-26 2021-08-05 Tencent Technology (Shenzhen) Company Limited Method, apparatus, device, and storage medium for training image semantic segmentation network
AU2020103901A4 (en) * 2020-12-04 2021-02-11 Chongqing Normal University Image Semantic Segmentation Method Based on Deep Full Convolutional Network and Conditional Random Field

Similar Documents

Publication Publication Date Title
CN111882002B (en) MSF-AM-based low-illumination target detection method
CN113052210B (en) Rapid low-light target detection method based on convolutional neural network
CN112184577B (en) Single image defogging method based on multiscale self-attention generation countermeasure network
CN110400275B (en) Color correction method based on full convolution neural network and characteristic pyramid
CN110363770B (en) Training method and device for edge-guided infrared semantic segmentation model
CN104217404A (en) Video image sharpness processing method in fog and haze day and device thereof
CN109903250B (en) Underwater image sharpening processing method based on multi-scale gradient domain contrast stretching
CN113066025B (en) Image defogging method based on incremental learning and feature and attention transfer
CN111709888B (en) Aerial image defogging method based on improved generation countermeasure network
CN107292830A (en) Low-light (level) image enhaucament and evaluation method
CN110807744B (en) Image defogging method based on convolutional neural network
CN113420794B (en) Binaryzation Faster R-CNN citrus disease and pest identification method based on deep learning
CN116402679A (en) Lightweight infrared super-resolution self-adaptive reconstruction method
CN115035171A (en) Self-supervision monocular depth estimation method based on self-attention-guidance feature fusion
CN114677560A (en) Deep learning algorithm-based lane line detection method and computer system
CN112686804A (en) Image super-resolution reconstruction method and device for mine low-illumination environment
CN113723231A (en) Low-illumination semantic segmentation model training method, semantic segmentation method and semantic segmentation device
CN116452450A (en) Polarized image defogging method based on 3D convolution
CN116542865A (en) Multi-scale real-time defogging method and device based on structural re-parameterization
CN115496764A (en) Dense feature fusion-based foggy image semantic segmentation method
CN115631108A (en) RGBD-based image defogging method and related equipment
CN114862724A (en) Contrast type image defogging method based on exponential moving average knowledge distillation
Wang et al. Low-light traffic objects detection for automated vehicles
CN113763261A (en) Real-time detection method for far and small targets under sea fog meteorological condition
CN111986109A (en) Remote sensing image defogging method based on full convolution network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination