CN113034353A - Essential image decomposition method and system based on cross convolution neural network - Google Patents

Essential image decomposition method and system based on cross convolution neural network Download PDF

Info

Publication number
CN113034353A
CN113034353A CN202110385353.8A CN202110385353A CN113034353A CN 113034353 A CN113034353 A CN 113034353A CN 202110385353 A CN202110385353 A CN 202110385353A CN 113034353 A CN113034353 A CN 113034353A
Authority
CN
China
Prior art keywords
layer
neural network
convolutional neural
network
initiation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110385353.8A
Other languages
Chinese (zh)
Other versions
CN113034353B (en
Inventor
权炜
孙燕平
于军琪
董芳楠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Architecture and Technology
Original Assignee
Xian University of Architecture and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Architecture and Technology filed Critical Xian University of Architecture and Technology
Priority to CN202110385353.8A priority Critical patent/CN113034353B/en
Publication of CN113034353A publication Critical patent/CN113034353A/en
Application granted granted Critical
Publication of CN113034353B publication Critical patent/CN113034353B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/04Context-preserving transformations, e.g. by using an importance map
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details
    • G06T2207/20192Edge enhancement; Edge preservation

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an essential image decomposition method and system based on a cross convolution neural network, wherein the method comprises the following steps: inputting an original image to be decomposed into a trained GoogLeNet-VGG19 cross convolution neural network model to obtain a light map and a reflection map obtained by decomposing the original image; the GoogLeNet-VGG19 cross convolution neural network model is formed by cross fusion of an illumination map generation network and a reflection map generation network; the illumination map generation network is constructed on the basis of a GoogleLeNet convolutional neural network, and the reflection map generation network is constructed on the basis of a VGG19 convolutional neural network; and training the light map generation network and the reflection map generation network by adopting an Adam optimization method. In the invention, the images of the result of the intrinsic image decomposition are kept consistent on the reflectivity of the same object, the protection on edge information and the removal on noise are better, the image quality is higher, and the details and the definition are closer to the true value images.

Description

Essential image decomposition method and system based on cross convolution neural network
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to an essential image decomposition method and system based on a cross convolution neural network.
Background
Intrinsic image decomposition was first proposed by Barrow and Tenenbaum in 1978, and the intrinsic image solving problem is to recover brightness and reflectivity information in a scene corresponding to all pixel points from an image to form a light map and a reflection map respectively. The intrinsic image decomposition is mainly divided into two types according to the algorithm type, the first type is the intrinsic image decomposition based on Retinex theory, and the second type is the intrinsic image decomposition based on deep learning. The conventional intrinsic image decomposition method Retinex assumes that the larger gradient in the image is caused by the object reflectivity, and the smaller gradient belongs to the illumination variation. Because the Retinex method is completely based on the gradient, the Retinex method establishes local constraints.
Another common constraint at present is that the natural image contains a small number of colors and the color distribution is structured, which is called global color sparsity, i.e. the image of the reflectivity layer is required to contain only several colors. Because only local constraints can be established by the gradient-based method, the obtained reflectivity layer image may have a phenomenon of global inconsistency, that is, the reflectivity of two pixels of the same material which are far away from each other is inconsistent, and multiple images in the same scene increase to provide strict requirements for the input of the essential image method. After the gradient values of the reflectivity and brightness images are estimated, the gradient images are integrated by means of Weiss to solve the reflection map and the illumination map. However, this method requires a large number of samples to train the classifier, which is time-consuming, and the obtained intrinsic image has a large error at the edge, and the finally obtained intrinsic image is blurred at the edge.
The intrinsic image decomposition method based on deep learning improves the problems to a certain extent, but has a plurality of defects, such as Narihira and the like, because of the defects of network design, the image is down-sampled to an excessively small scale, so that a large amount of recovered information is lost, and the output result is fuzzy; fan et al integrate a filter in the network to flatten the reflective layer, remove residual noise and geometry information, but ignore the protection of image details resulting in jagged edges.
Disclosure of Invention
The present invention is directed to a method and a system for intrinsic image decomposition based on a cross-convolution neural network, so as to solve one or more of the above technical problems. In the invention, the images of the result of the intrinsic image decomposition are kept consistent on the reflectivity of the same object, the protection on edge information and the removal on noise are better, the image quality is higher, and the details and the definition are closer to the true value images.
In order to achieve the purpose, the invention adopts the following technical scheme:
the invention discloses an essential image decomposition method based on a cross convolution neural network, which comprises the following steps:
inputting an original image to be decomposed into a trained GoogLeNet-VGG19 cross convolution neural network model to obtain a light map and a reflection map obtained by decomposing the original image;
the GoogLeNet-VGG19 cross convolution neural network model is formed by cross fusion of an illumination map generation network and a reflection map generation network; the illumination map generation network is constructed on the basis of a GoogleLeNet convolutional neural network, and the reflection map generation network is constructed on the basis of a VGG19 convolutional neural network;
the method for acquiring the well-trained GoogLeNet-VGG19 cross convolution neural network model comprises the step of training a light map generation network and a reflection map generation network by adopting an Adam optimization method.
The invention has the further improvement that the steps of constructing the illumination map generation network based on the GoogleLeNet convolutional neural network specifically comprise:
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3a, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3a together;
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3b, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3b together;
in the GoogleLeNet convolutional neural network initiation 4a, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
in the GoogleLeNet convolutional neural network initiation 4b, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4b to the DepthConcat layer of the initiation 4d in a jumping manner; directly connecting the convolution operation output after the AveragePool operation of the first layer of the GoogleLeNet convolution neural network initiation 4e to the DepthCocatat layer of the initiation 4 e; connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4e to the DepthConcat layer of the initiation 5b in a jumping manner;
respectively adding 1 ReLU activation function after 4 convolution operations on the third layer of the GoogleLeNet convolutional neural network initiation 5a, and outputting the 4 ReLU activation functions to the DepthConcatin layer of the initiation 5a together;
respectively adding 1 ReLU activation function after 4 convolution operations of the second layer of the GoogleLeNet convolutional neural network initiation 5b, wherein the 4 ReLU activation functions are jointly output to the DepthConcat layer of the initiation 3 b;
an FC layer is added after the FC layer of the google lenet convolutional neural network.
The further improvement of the invention is that the step of constructing the reflection map generation network based on the VGG19 convolutional neural network specifically comprises the following steps:
performing Concat operation on a first MaxBoool output result and a second MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results to a fifth layer of the VGG19 convolutional neural network;
performing Concat operation on a third MaxBoool output result and a fourth MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results into a thirteenth layer of the VGG19 convolutional neural network;
deleting the seventeenth layer and the eighteenth layer of the VGG19 convolutional neural network; and adding two layers with the same structure as the sixteenth layer after the sixteenth layer of the VGG19 convolutional neural network to form a seventeenth layer and an eighteenth layer of the modified VGG19 convolutional neural network.
The further improvement of the invention is that the step of cross-fusion of the illumination map generation network and the reflection map generation network specifically comprises:
connecting the depthConcat layer output of the GoogleLeNet convolutional neural network initiation 4e to the thirteenth layer of the VGG19 convolutional neural network;
the convolution operation connecting the fourth MaxPool output of the VGG19 convolutional neural network to the second layer of the google lenet convolutional neural network initiation 5 a.
The invention is further improved in that the Loss function Loss1 of the illumination map generation network has the expression:
Figure BDA0003014530940000041
Figure BDA0003014530940000042
in the formula, X is an input image,
Figure BDA0003014530940000043
for predicting images, H, W and C are respectively the height, width and channel number of the input image, x and y represent the pixel coordinates of the image, C represents the channel, and muiDenotes the weight at the i-th scale, X(i)Representing the image at the i-th scale,
Figure BDA0003014530940000044
representing the predicted image at the i-th scale generated by the improved google convolutional neural network.
The invention is further improved in that the expression of the Loss function Loss2 of the reflection map generation network is as follows:
Figure BDA0003014530940000045
in the formula, Y represents an input image,
Figure BDA0003014530940000046
representing the estimated value, C, of the input image after being processed by the modified VGG19 networkj,Hj,WjRespectively representing the channel number, height and width of the j-th layer output characteristic diagram, Vj(.) indicates the output of the activation function when the layer j processes the picture, j indicating the number of layers.
The further improvement of the invention is that the step of training the light map generation network and the reflection map generation network by adopting an Adam optimization method specifically comprises the following steps:
taking an image in a pre-constructed training image sample library as a sample, and training a light map generation network and a reflection map generation network simultaneously by adopting an Adam optimization method;
in the training process, inputting the illumination pattern output by the illumination pattern generation network into an identification network, identifying the probability that the illumination pattern output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the illumination pattern generation network; inputting the reflection map output by the reflection map generation network into an identification network, identifying the probability that the reflection map output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the reflection map generation network;
when the Loss function Loss1 reaches the minimum, stopping training the illumination map generation network to obtain a final illumination map generation network; when the Loss function Loss2 reaches the minimum, stopping training the reflection map generation network to obtain a final reflection map generation network;
the identification network is a multilayer convolutional neural network and comprises six same layers; each layer is in turn a convolution operation, a Sigmoid activation function, and MaxPool.
A further refinement of the invention consists in that the Adam optimization parameter beta is set to (0.9,0.999), the learning rate is 0.005, the weight attenuation is 0.0001, epoch is 100 and banksize is 20.
The invention discloses an essential image decomposition system based on a cross convolution neural network, which comprises:
the decomposition module is used for inputting an original image to be decomposed into a trained GoogLeNet-VGG19 cross convolution neural network model to obtain a light map and a reflection map obtained by decomposing the original image;
the GoogLeNet-VGG19 cross convolution neural network model is formed by cross fusion of an illumination map generation network and a reflection map generation network; the illumination map generation network is constructed on the basis of a GoogleLeNet convolutional neural network, and the reflection map generation network is constructed on the basis of a VGG19 convolutional neural network;
the method for acquiring the well-trained GoogLeNet-VGG19 cross convolution neural network model comprises the step of training a light map generation network and a reflection map generation network by adopting an Adam optimization method.
A further development of the invention is that,
the steps of constructing the illumination map generation network based on the GoogleLeNet convolutional neural network specifically comprise:
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3a, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3a together;
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3b, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3b together;
in the GoogleLeNet convolutional neural network initiation 4a, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
in the GoogleLeNet convolutional neural network initiation 4b, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4b to the DepthConcat layer of the initiation 4d in a jumping manner; directly connecting the convolution operation output after the AveragePool operation of the first layer of the GoogleLeNet convolution neural network initiation 4e to the DepthCocatat layer of the initiation 4 e; connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4e to the DepthConcat layer of the initiation 5b in a jumping manner;
respectively adding 1 ReLU activation function after 4 convolution operations on the third layer of the GoogleLeNet convolutional neural network initiation 5a, and outputting the 4 ReLU activation functions to the DepthConcatin layer of the initiation 5a together;
respectively adding 1 ReLU activation function after 4 convolution operations of the second layer of the GoogleLeNet convolutional neural network initiation 5b, wherein the 4 ReLU activation functions are jointly output to the DepthConcat layer of the initiation 3 b;
adding an FC layer behind the FC layer of the GoogleLeNet convolutional neural network;
the step of constructing the reflection map generation network based on the VGG19 convolutional neural network specifically comprises the following steps:
performing Concat operation on a first MaxBoool output result and a second MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results to a fifth layer of the VGG19 convolutional neural network;
performing Concat operation on a third MaxBoool output result and a fourth MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results into a thirteenth layer of the VGG19 convolutional neural network;
deleting the seventeenth layer and the eighteenth layer of the VGG19 convolutional neural network; adding two layers with the same structure as the sixteenth layer after the sixteenth layer of the VGG19 convolutional neural network to form a seventeenth layer and an eighteenth layer of the VGG19 convolutional neural network after modification;
the step of performing cross fusion on the illumination map generation network and the reflection map generation network specifically comprises the following steps:
connecting the depthConcat layer output of the GoogleLeNet convolutional neural network initiation 4e to the thirteenth layer of the VGG19 convolutional neural network;
the convolution operation connecting the fourth MaxPool output of the VGG19 convolutional neural network to the second layer of the google lenet convolutional neural network initiation 5 a.
Compared with the prior art, the invention has the following beneficial effects:
the invention provides an essential image decomposition method based on an improved GoogLeNet-VGG19 cross convolution neural network, which comprises the steps of firstly constructing a training image sample library, then carrying out improved construction of an illumination map generation network based on the traditional GoogLeNet convolution neural network, carrying out improved construction of a reflection map generation network based on the traditional VGG19 convolution neural network, and carrying out cross fusion on the illumination map generation network and the reflection map generation network; next, constructing a recognition network; and finally, training the illumination map generation network and the reflection map generation network by adopting an Adam optimization method to obtain the final illumination map generation network and reflection map generation network. The result image of the essential image decomposition keeps consistent on the reflectivity of the same object, has better performance on the aspects of protecting edge information and removing noise, has higher image quality, and is closer to a true value image in the aspects of detail and definition.
Compared with the prior art, the system has the advantages that the images output by the method are consistent in reflectivity of the same object, the system is good in protection of edge information and removal of noise, and the image quality is high; the generated result is closer to the true value image in the aspects of detail and definition.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art are briefly introduced below; it is obvious that the drawings in the following description are some embodiments of the invention, and that for a person skilled in the art, other drawings can be derived from them without inventive effort.
FIG. 1 is a schematic flow chart of an essential image decomposition method based on an improved GoogLeNet-VGG19 cross-convolution neural network according to an embodiment of the present invention;
FIG. 2 is a diagram illustrating an intrinsic image decomposition result according to an embodiment of the present invention; fig. 2 (a) is a schematic diagram of an original image, fig. 2 (b) is a schematic diagram of an illumination map obtained after decomposition, and fig. 2 (c) is a schematic diagram of a reflection map obtained after decomposition.
Detailed Description
In order to make the purpose, technical effect and technical solution of the embodiments of the present invention clearer, the following clearly and completely describes the technical solution of the embodiments of the present invention with reference to the drawings in the embodiments of the present invention; it is to be understood that the described embodiments are only some of the embodiments of the present invention. Other embodiments, which can be derived by one of ordinary skill in the art from the disclosed embodiments without inventive faculty, are intended to be within the scope of the invention.
Referring to fig. 1, an intrinsic image decomposition method based on an improved google lenet-VGG19 cross convolution neural network according to an embodiment of the present invention includes the following steps:
step 1: constructing a training image sample library;
taking out P images and corresponding light map images and reflection images from the essence image database by adopting the public essence image database; then, randomly cutting the P images to cut a plurality of image blocks with specified sizes; then, image processing is carried out on the image blocks, namely horizontal overturning, vertical overturning, rotation and mirror image are carried out randomly for expanding the database; the image blocks after image processing and the illumination patterns and reflection patterns corresponding to the image blocks form a training image sample library;
step 2: the illumination map generation network is constructed by adopting an improved GoogleLeNet convolution neural network, and the method specifically comprises the following steps:
step 2-1: respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3a, wherein 4 ReLU activation functions are totally added, and the 4 ReLU activation functions are jointly output to a DepthConcat layer of the initiation 3 a;
step 2-2: respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3b, wherein 4 ReLU activation functions are totally added, and the 4 ReLU activation functions are jointly output to a DepthConcat layer of the initiation 3 b;
step 2-3: in the GoogleLeNet convolutional neural network initiation 4a, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and MaxPool operation combination on the 2 connection channels respectively, wherein 2 ReLU activation functions and MaxPool operation combinations are added in total; the ReLU activation function is in front, and the MaxPool operation is in back;
step 2-4: in the GoogleLeNet convolutional neural network initiation 4b, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and MaxPool operation combination on the 2 connection channels respectively, wherein 2 ReLU activation functions and MaxPool operation combinations are added in total; the ReLU activation function is in front, and the MaxPool operation is in back;
step 2-5: connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4b to the DepthConcat layer of the initiation 4d in a jumping manner;
step 2-6: directly connecting the convolution operation output after the AveragePool operation of the first layer of the GoogleLeNet convolution neural network initiation 4e to the DepthCocatat layer of the initiation 4 e;
step 2-7: connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4e to the DepthConcat layer of the initiation 5b in a jumping manner;
step 2-8: respectively adding 1 ReLU activation function after 4 convolution operations on the third layer of the GoogleLeNet convolutional neural network initiation 5a, wherein 4 ReLU activation functions are totally added, and the 4 ReLU activation functions are jointly output to a DepthConcat layer of the initiation 5 a;
step 2-9: adding 1 ReLU activation function after 4 convolution operations of the second layer of the GoogleLeNet convolutional neural network initiation 5b, wherein 4 ReLU activation functions are totally added, and the 4 ReLU activation functions are jointly output to the DepthCocatat layer of initiation 3 b;
step 2-10: a new FC layer is added behind the FC layer of the GoogleLeNet convolutional neural network;
step 2-11: forming an improved GoogLeNet convolutional neural network through the operations from the step 2-1 to the step 2-10;
and step 3: the reflection map generation network is constructed by adopting the improved VGG19 convolutional neural network, and the method comprises the following specific steps:
step 3-1: performing Concat operation on a first MaxBoool output result and a second MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results to a fifth layer of the VGG19 convolutional neural network;
step 3-2: performing Concat operation on a third MaxBoool output result and a fourth MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results into a thirteenth layer of the VGG19 convolutional neural network;
step 3-3: deleting the seventeenth layer and the eighteenth layer of the VGG19 convolutional neural network;
step 3-4: adding two same layers after the sixteenth layer of the VGG19 convolutional neural network to form a new seventeenth layer and an eighteenth layer; the structure of the new seventeenth layer and the eighteenth layer is completely the same as that of the sixteenth layer;
step 3-5: constructing a modified VGG19 convolutional neural network through the operations of step 3-1 to step 3-4;
and 4, step 4: the illumination map generation network and the reflection map generation network are crossed and fused;
step 4-1: connecting the depthConcat layer output of the GoogleLeNet convolutional neural network initiation 4e to the thirteenth layer of the VGG19 convolutional neural network;
step 4-2: a convolution operation connecting the fourth MaxPool output of the VGG19 convolutional neural network to the second layer of the google convolutional neural network initiation 5 a;
and 5: constructing a recognition network;
the identification network is a multilayer convolutional neural network and comprises six same layers; each layer is sequentially convolution operation, Sigmoid activation function and MaxPool;
step 6: defining a loss function;
step 6-1: defining an illumination map generation network Loss function Loss 1:
Figure BDA0003014530940000111
Figure BDA0003014530940000112
wherein, X is an input image,
Figure BDA0003014530940000113
for predicting images, H, W and C are respectively the height, width and channel number of the input image, x and y represent the pixel coordinates of the image, C represents the channel, and muiDenotes the weight at the i-th scale, X(i)Representing the image at the i-th scale,
Figure BDA0003014530940000114
representing a predicted image at the ith scale generated by the improved GoogleLeNet convolutional neural network;
step 6-2: defining a reflection map to generate a network Loss function Loss 2:
Figure BDA0003014530940000115
wherein, Y represents an input image,
Figure BDA0003014530940000116
representing the estimated value, C, of the input image after being processed by the modified VGG19 networkj,Hj,WjRespectively representing the channel number, height and width of the j-th layer output characteristic diagram, Vj(.) indicates the output of the activation function when the j-th layer processes the image, j indicates the layer number;
and 7: network training;
taking the images in the training image sample library constructed in the step 1 as samples, and training a light map generation network and a reflection map generation network simultaneously by adopting an Adam optimization method;
in the training process, inputting the illumination pattern output by the illumination pattern generation network into an identification network, identifying the probability that the illumination pattern output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the illumination pattern generation network; inputting the reflection map output by the reflection map generation network into an identification network, identifying the probability that the reflection map output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the reflection map generation network;
when the Loss function Loss1 reaches the minimum, stopping training the illumination map generation network to obtain a final illumination map generation network; when the Loss function Loss2 reaches the minimum, stopping training the reflection map generation network to obtain a final reflection map generation network;
and 8: and (4) respectively inputting the original image to be decomposed into the final illumination map generation network and the final reflection map generation network obtained in the step (7), wherein the output image is the illumination map and the reflection map obtained by decomposing the original image.
In an embodiment of the present invention, the size of the image block with the size specified in step 1 is 224 × 224.
In the embodiment of the present invention, the parameters set when the network is trained in step 7 are as follows: adam optimization parameter beta is set to (0.9,0.999), learning rate is 0.005, weight attenuation is 0.0001, epoch is 100, and banksize is 20.
Compared with the problem that the image decomposed by the existing method has a lot of noises and the image edge is fuzzy, the method provided by the embodiment of the invention has the advantages that the output image keeps consistent on the reflectivity of the same object, the performance in the aspects of protecting the edge information and removing the noises is better, and the image quality is higher; the method of the invention can generate results which are closer to the true value image in the aspects of detail and definition.
The invention provides an essential image decomposition system based on a cross convolution neural network, which comprises:
the decomposition module is used for inputting an original image to be decomposed into a trained GoogLeNet-VGG19 cross convolution neural network model to obtain a light map and a reflection map obtained by decomposing the original image;
the GoogLeNet-VGG19 cross convolution neural network model is formed by cross fusion of an illumination map generation network and a reflection map generation network; the illumination map generation network is constructed on the basis of a GoogleLeNet convolutional neural network, and the reflection map generation network is constructed on the basis of a VGG19 convolutional neural network;
the method for acquiring the well-trained GoogLeNet-VGG19 cross convolution neural network model comprises the step of training a light map generation network and a reflection map generation network by adopting an Adam optimization method.
Referring to fig. 1 and fig. 2, an intrinsic image decomposition method based on the improved google lenet-VGG19 cross-convolution neural network according to an embodiment of the present invention includes the following steps:
(1) constructing a training image sample library
And (3) taking 1000 images from the MPCal essential image data set, randomly cutting 50 image blocks of 224X 224 in each image, and then randomly turning the image blocks horizontally, vertically, rotating and mirroring to change the 50 image blocks into 200 image blocks. At this time, the total number of image blocks is 20 ten thousand. Meanwhile, in the illumination map and the reflection map corresponding to 1000 images, the illumination pattern block and the reflection pattern block corresponding to the 20 ten thousand image blocks are found. The image blocks, the corresponding illumination image blocks and the reflection image blocks form a training image sample library.
(2) The illumination map generation network and the reflection map generation network constructed by the method are trained simultaneously by using a training image sample library, an Adam optimization method is adopted, Adam optimization parameters beta are set to be (0.9,0.999), the learning rate is 0.005, the weight attenuation is 0.0001, epoch is 100, and batch size is 20. And when the loss functions of the two generated networks are minimum, stopping training to obtain a final light map generation network and a final reflection map generation network. In the training process, inputting the illumination pattern output by the illumination pattern generation network into an identification network, identifying the probability that the illumination pattern output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the illumination pattern generation network; inputting the reflection map output by the reflection map generation network into an identification network, identifying the probability that the reflection map output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the reflection map generation network; the generation network and the recognition network adopt a TTUR training method, and the ratio of the training times of the recognition network to the training times of the generation network is 3 to 1.
(3) As shown in fig. 2, the original image to be processed (shown in (a) of fig. 2) is respectively input into the final illumination map generation network and reflection map generation network, and the output image is the illumination map and reflection map obtained by decomposing the original image (shown in (b) and (c) of fig. 2). The decomposition result shows that the intrinsic image decomposition result of the method has less noise, the image edge is clear, and the integral definition and quality of the image reach higher levels, thereby fully explaining the effectiveness and the practicability of the method.
In summary, the embodiment of the present invention provides an essential image decomposition method based on an improved google lenet-VGG19 cross convolution neural network, which includes first constructing a training image sample library, then performing an improved structure on an illumination map generation network based on a conventional google lenet convolution neural network, performing an improved structure on a reflection map generation network based on a conventional VGG19 convolution neural network, and performing cross fusion on the illumination map generation network and the reflection map generation network; next, constructing a recognition network; and finally, training the illumination map generation network and the reflection map generation network by adopting an Adam optimization method to obtain the final illumination map generation network and reflection map generation network. The result image of the essential image decomposition keeps consistent on the reflectivity of the same object, has better performance on the aspects of protecting edge information and removing noise, has higher image quality, and is closer to a true value image in the aspects of detail and definition.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
Although the present invention has been described in detail with reference to the above embodiments, those skilled in the art can make modifications and equivalents to the embodiments of the present invention without departing from the spirit and scope of the present invention, which is set forth in the claims of the present application.

Claims (10)

1. An essential image decomposition method based on a cross convolution neural network is characterized by comprising the following steps:
inputting an original image to be decomposed into a trained GoogLeNet-VGG19 cross convolution neural network model to obtain a light map and a reflection map obtained by decomposing the original image;
the GoogLeNet-VGG19 cross convolution neural network model is formed by cross fusion of an illumination map generation network and a reflection map generation network; the illumination map generation network is constructed on the basis of a GoogleLeNet convolutional neural network, and the reflection map generation network is constructed on the basis of a VGG19 convolutional neural network;
the method for acquiring the well-trained GoogLeNet-VGG19 cross convolution neural network model comprises the step of training a light map generation network and a reflection map generation network by adopting an Adam optimization method.
2. The method for essential image decomposition based on cross-convolution neural network as claimed in claim 1, wherein the step of constructing the illumination map generation network based on the google lenet convolution neural network specifically includes:
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3a, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3a together;
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3b, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3b together;
in the GoogleLeNet convolutional neural network initiation 4a, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
in the GoogleLeNet convolutional neural network initiation 4b, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4b to the DepthConcat layer of the initiation 4d in a jumping manner; directly connecting the convolution operation output after the AveragePool operation of the first layer of the GoogleLeNet convolution neural network initiation 4e to the DepthCocatat layer of the initiation 4 e; connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4e to the DepthConcat layer of the initiation 5b in a jumping manner;
respectively adding 1 ReLU activation function after 4 convolution operations on the third layer of the GoogleLeNet convolutional neural network initiation 5a, and outputting the 4 ReLU activation functions to the DepthConcatin layer of the initiation 5a together;
respectively adding 1 ReLU activation function after 4 convolution operations of the second layer of the GoogleLeNet convolutional neural network initiation 5b, wherein the 4 ReLU activation functions are jointly output to the DepthConcat layer of the initiation 3 b;
an FC layer is added after the FC layer of the google lenet convolutional neural network.
3. The method of claim 2, wherein the step of constructing the reflection map generation network based on the VGG19 convolutional neural network specifically comprises:
performing Concat operation on a first MaxBoool output result and a second MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results to a fifth layer of the VGG19 convolutional neural network;
performing Concat operation on a third MaxBoool output result and a fourth MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results into a thirteenth layer of the VGG19 convolutional neural network;
deleting the seventeenth layer and the eighteenth layer of the VGG19 convolutional neural network; and adding two layers with the same structure as the sixteenth layer after the sixteenth layer of the VGG19 convolutional neural network to form a seventeenth layer and an eighteenth layer of the modified VGG19 convolutional neural network.
4. The method according to claim 3, wherein the step of performing cross fusion on the illumination map generation network and the reflection map generation network specifically comprises:
connecting the depthConcat layer output of the GoogleLeNet convolutional neural network initiation 4e to the thirteenth layer of the VGG19 convolutional neural network;
the convolution operation connecting the fourth MaxPool output of the VGG19 convolutional neural network to the second layer of the google lenet convolutional neural network initiation 5 a.
5. The method for essential image decomposition based on the cross-convolution neural network as claimed in claim 4, wherein the expression of the Loss function Loss1 of the illumination map generation network is:
Figure FDA0003014530930000031
Figure FDA0003014530930000032
in the formula, X is an input image,
Figure FDA0003014530930000033
for predicting images, H, W and C are respectively the height, width and channel number of the input image, x and y represent the pixel coordinates of the image, C represents the channel, and muiDenotes the weight at the i-th scale, X(i)Representing the image at the i-th scale,
Figure FDA0003014530930000034
representing the predicted image at the i-th scale generated by the improved google convolutional neural network.
6. The method for essential image decomposition based on the cross-convolution neural network as claimed in claim 5, wherein the expression of the Loss function Loss2 of the reflection map generation network is as follows:
Figure FDA0003014530930000035
in the formula, Y represents an input image,
Figure FDA0003014530930000036
representing the estimated value, C, of the input image after being processed by the modified VGG19 networkj,Hj,WjRespectively representing the channel number, height and width of the j-th layer output characteristic diagram, Vj(.) indicates the output of the activation function when the layer j processes the picture, j indicating the number of layers.
7. The method of claim 6, wherein the step of training the histogram generation network and the reflectogram generation network by using an Adam optimization method specifically comprises:
taking an image in a pre-constructed training image sample library as a sample, and training a light map generation network and a reflection map generation network simultaneously by adopting an Adam optimization method;
in the training process, inputting the illumination pattern output by the illumination pattern generation network into an identification network, identifying the probability that the illumination pattern output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the illumination pattern generation network; inputting the reflection map output by the reflection map generation network into an identification network, identifying the probability that the reflection map output by the network is consistent with the label image of the training sample, and reversely updating the network parameters of the reflection map generation network;
when the Loss function Loss1 reaches the minimum, stopping training the illumination map generation network to obtain a final illumination map generation network; when the Loss function Loss2 reaches the minimum, stopping training the reflection map generation network to obtain a final reflection map generation network;
the identification network is a multilayer convolutional neural network and comprises six same layers; each layer is in turn a convolution operation, a Sigmoid activation function, and MaxPool.
8. The method of claim 7, wherein Adam optimization parameter beta is set to (0.9,0.999), learning rate is 0.005, weight attenuation is 0.0001, epoch is 100, and banksize is 20.
9. An essential image decomposition system based on a cross-convolution neural network, comprising:
the decomposition module is used for inputting an original image to be decomposed into a trained GoogLeNet-VGG19 cross convolution neural network model to obtain a light map and a reflection map obtained by decomposing the original image;
the GoogLeNet-VGG19 cross convolution neural network model is formed by cross fusion of an illumination map generation network and a reflection map generation network; the illumination map generation network is constructed on the basis of a GoogleLeNet convolutional neural network, and the reflection map generation network is constructed on the basis of a VGG19 convolutional neural network;
the method for acquiring the well-trained GoogLeNet-VGG19 cross convolution neural network model comprises the step of training a light map generation network and a reflection map generation network by adopting an Adam optimization method.
10. The cross-convolution neural network-based intrinsic image decomposition system of claim 9,
the steps of constructing the illumination map generation network based on the GoogleLeNet convolutional neural network specifically comprise:
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3a, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3a together;
respectively adding 1 ReLU activation function after 4 convolution operations on the second layer of the GoogleLeNet convolutional neural network initiation 3b, and outputting the 4 ReLU activation functions to the DepthConcato layer of the initiation 3b together;
in the GoogleLeNet convolutional neural network initiation 4a, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
in the GoogleLeNet convolutional neural network initiation 4b, 2 convolution operations of a first layer are connected with 2 convolution operations of a second layer to form 2 connecting channels; adding 1 ReLU activation function and a MaxPool operation combination on 2 connection channels respectively, wherein the ReLU activation function is in front of the MaxPool operation combination;
connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4b to the DepthConcat layer of the initiation 4d in a jumping manner; directly connecting the convolution operation output after the AveragePool operation of the first layer of the GoogleLeNet convolution neural network initiation 4e to the DepthCocatat layer of the initiation 4 e; connecting the output of the DepthConcat layer of the GoogleLeNet convolutional neural network initiation 4e to the DepthConcat layer of the initiation 5b in a jumping manner;
respectively adding 1 ReLU activation function after 4 convolution operations on the third layer of the GoogleLeNet convolutional neural network initiation 5a, and outputting the 4 ReLU activation functions to the DepthConcatin layer of the initiation 5a together;
respectively adding 1 ReLU activation function after 4 convolution operations of the second layer of the GoogleLeNet convolutional neural network initiation 5b, wherein the 4 ReLU activation functions are jointly output to the DepthConcat layer of the initiation 3 b;
adding an FC layer behind the FC layer of the GoogleLeNet convolutional neural network;
the step of constructing the reflection map generation network based on the VGG19 convolutional neural network specifically comprises the following steps:
performing Concat operation on a first MaxBoool output result and a second MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results to a fifth layer of the VGG19 convolutional neural network;
performing Concat operation on a third MaxBoool output result and a fourth MaxBoool output result of the VGG19 convolutional neural network, and inputting the obtained results into a thirteenth layer of the VGG19 convolutional neural network;
deleting the seventeenth layer and the eighteenth layer of the VGG19 convolutional neural network; adding two layers with the same structure as the sixteenth layer after the sixteenth layer of the VGG19 convolutional neural network to form a seventeenth layer and an eighteenth layer of the VGG19 convolutional neural network after modification;
the step of performing cross fusion on the illumination map generation network and the reflection map generation network specifically comprises the following steps:
connecting the depthConcat layer output of the GoogleLeNet convolutional neural network initiation 4e to the thirteenth layer of the VGG19 convolutional neural network;
the convolution operation connecting the fourth MaxPool output of the VGG19 convolutional neural network to the second layer of the google lenet convolutional neural network initiation 5 a.
CN202110385353.8A 2021-04-09 2021-04-09 Intrinsic image decomposition method and system based on cross convolution neural network Active CN113034353B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110385353.8A CN113034353B (en) 2021-04-09 2021-04-09 Intrinsic image decomposition method and system based on cross convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110385353.8A CN113034353B (en) 2021-04-09 2021-04-09 Intrinsic image decomposition method and system based on cross convolution neural network

Publications (2)

Publication Number Publication Date
CN113034353A true CN113034353A (en) 2021-06-25
CN113034353B CN113034353B (en) 2024-07-12

Family

ID=76456400

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110385353.8A Active CN113034353B (en) 2021-04-09 2021-04-09 Intrinsic image decomposition method and system based on cross convolution neural network

Country Status (1)

Country Link
CN (1) CN113034353B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113657521A (en) * 2021-08-23 2021-11-16 天津大学 Method for separating two mutually exclusive components in image

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108416805A (en) * 2018-03-12 2018-08-17 中山大学 A kind of intrinsic image decomposition method and device based on deep learning
US20180293712A1 (en) * 2017-04-06 2018-10-11 Pixar Denoising monte carlo renderings using generative adversarial neural networks
CN108764250A (en) * 2018-05-02 2018-11-06 西北工业大学 A method of extracting essential image with convolutional neural networks
CN110232661A (en) * 2019-05-03 2019-09-13 天津大学 Low illumination colour-image reinforcing method based on Retinex and convolutional neural networks
US20190304069A1 (en) * 2018-03-29 2019-10-03 Pixar Denoising monte carlo renderings using neural networks with asymmetric loss
CN110675336A (en) * 2019-08-29 2020-01-10 苏州千视通视觉科技股份有限公司 Low-illumination image enhancement method and device
WO2020068158A1 (en) * 2018-09-24 2020-04-02 Google Llc Photo relighting using deep neural networks and confidence learning
CN111242868A (en) * 2020-01-16 2020-06-05 重庆邮电大学 Image enhancement method based on convolutional neural network under dark vision environment
CN111563577A (en) * 2020-04-21 2020-08-21 西北工业大学 Unet-based intrinsic image decomposition method for skip layer frequency division and multi-scale identification
CN111681223A (en) * 2020-06-09 2020-09-18 安徽理工大学 Method for detecting mine well wall under low illumination condition based on convolutional neural network
GB202012490D0 (en) * 2020-08-11 2020-09-23 Toshiba Kk A Computer Vision Method and System
CN111914738A (en) * 2020-07-29 2020-11-10 南京汽车集团有限公司 Fatigue driving behavior detection system based on parallel cross convolution neural network
CN112131975A (en) * 2020-09-08 2020-12-25 东南大学 Face illumination processing method based on Retinex decomposition and generation of confrontation network

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180293712A1 (en) * 2017-04-06 2018-10-11 Pixar Denoising monte carlo renderings using generative adversarial neural networks
CN108416805A (en) * 2018-03-12 2018-08-17 中山大学 A kind of intrinsic image decomposition method and device based on deep learning
US20190304069A1 (en) * 2018-03-29 2019-10-03 Pixar Denoising monte carlo renderings using neural networks with asymmetric loss
CN108764250A (en) * 2018-05-02 2018-11-06 西北工业大学 A method of extracting essential image with convolutional neural networks
WO2020068158A1 (en) * 2018-09-24 2020-04-02 Google Llc Photo relighting using deep neural networks and confidence learning
CN110232661A (en) * 2019-05-03 2019-09-13 天津大学 Low illumination colour-image reinforcing method based on Retinex and convolutional neural networks
CN110675336A (en) * 2019-08-29 2020-01-10 苏州千视通视觉科技股份有限公司 Low-illumination image enhancement method and device
CN111242868A (en) * 2020-01-16 2020-06-05 重庆邮电大学 Image enhancement method based on convolutional neural network under dark vision environment
CN111563577A (en) * 2020-04-21 2020-08-21 西北工业大学 Unet-based intrinsic image decomposition method for skip layer frequency division and multi-scale identification
CN111681223A (en) * 2020-06-09 2020-09-18 安徽理工大学 Method for detecting mine well wall under low illumination condition based on convolutional neural network
CN111914738A (en) * 2020-07-29 2020-11-10 南京汽车集团有限公司 Fatigue driving behavior detection system based on parallel cross convolution neural network
GB202012490D0 (en) * 2020-08-11 2020-09-23 Toshiba Kk A Computer Vision Method and System
CN112131975A (en) * 2020-09-08 2020-12-25 东南大学 Face illumination processing method based on Retinex decomposition and generation of confrontation network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
吴梦婷;李伟红;龚卫国;: "双框架卷积神经网络用于运动模糊图像盲复原", 计算机辅助设计与图形学学报, no. 12, 15 December 2018 (2018-12-15) *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113657521A (en) * 2021-08-23 2021-11-16 天津大学 Method for separating two mutually exclusive components in image
CN113657521B (en) * 2021-08-23 2023-09-19 天津大学 Method for separating two mutually exclusive components in image

Also Published As

Publication number Publication date
CN113034353B (en) 2024-07-12

Similar Documents

Publication Publication Date Title
CN111080620B (en) Road disease detection method based on deep learning
Li et al. Deep retinex network for single image dehazing
CN109978807B (en) Shadow removing method based on generating type countermeasure network
CN110084817B (en) Digital elevation model production method based on deep learning
CN111696110B (en) Scene segmentation method and system
GB2582833A (en) Facial localisation in images
CN114549563A (en) Real-time composite insulator segmentation method and system based on deep LabV3+
CN110210482B (en) Target detection method for improving class imbalance
CN116563250A (en) Recovery type self-supervision defect detection method, device and storage medium
CN113762265A (en) Pneumonia classification and segmentation method and system
CN112861880A (en) Weak supervision RGBD image saliency detection method and system based on image classification
CN115761574A (en) Weak surveillance video target segmentation method and device based on frame labeling
CN113034353B (en) Intrinsic image decomposition method and system based on cross convolution neural network
CN115222750A (en) Remote sensing image segmentation method and system based on multi-scale fusion attention
CN114862642A (en) Method for removing short video visible watermark and computer readable storage medium
CN117876793A (en) Hyperspectral image tree classification method and device
CN112446417B (en) Spindle-shaped fruit image segmentation method and system based on multilayer superpixel segmentation
CN116091784A (en) Target tracking method, device and storage medium
CN114943655B (en) Image restoration system for generating countermeasure network structure based on cyclic depth convolution
CN115376022A (en) Application of small target detection algorithm based on neural network in unmanned aerial vehicle aerial photography
CN116935303A (en) Weak supervision self-training video anomaly detection method
CN115331052A (en) Garbage data labeling system and method based on deep learning
CN112396126A (en) Target detection method and system based on detection of main stem and local feature optimization
CN114648527B (en) Urothelial cell slide image classification method, device, equipment and medium
WO2023241276A1 (en) Image editing method and related device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant