CN108764250B - Method for extracting essential image by using convolutional neural network - Google Patents
Method for extracting essential image by using convolutional neural network Download PDFInfo
- Publication number
- CN108764250B CN108764250B CN201810407424.8A CN201810407424A CN108764250B CN 108764250 B CN108764250 B CN 108764250B CN 201810407424 A CN201810407424 A CN 201810407424A CN 108764250 B CN108764250 B CN 108764250B
- Authority
- CN
- China
- Prior art keywords
- image
- network
- layer
- branch
- output
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention provides a method for extracting an essential image by using a convolutional neural network. Firstly, constructing a double-current convolution network with a parallel structure from an image to an image; then, the network is trained by adopting a specific training data set, network parameters are optimized, so that multi-layer features with environment invariance are extracted, and essential images (a reflection map and a light map) are directly reconstructed. Because the double-current convolutional neural network constructed based on the deep learning theory has strong feature extraction capability, a reflection map and a light map can be directly separated from an original image. Meanwhile, the model is a full convolution network model from an image to an image, comprises two branch flow directions and is respectively used for generating a light map and a reflection map, and the network structure combines a convolution result of a higher layer with a result after deconvolution operation, so that the reconstruction errors of the light map and the reflection map can be reduced to a certain extent, and the capability of network feature reconstruction is improved.
Description
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a method for extracting an essential image by using a convolutional neural network.
Background
Understanding and analyzing images is one of the most fundamental tasks of image processing. Because the image imaging process is affected by various factors such as the characteristics of the target object, the shooting environment, the acquisition equipment and other conditions, the interference of the factors, such as shadow, color discontinuity, target posture transformation and the like, needs to be fully considered in the image processing process. These variable factors bring great challenges to the image processing algorithm, so that the performance of the existing image analysis algorithm is greatly influenced in a complex environment. Therefore, how to improve the robustness of the image analysis algorithm under a complex environment has become a hot research focus in recent years. In fact, if the essential features in the image can be analyzed based on the existing observation image, the above problems encountered in the image analysis process can be solved well. The essential features are features inherent to the target object, which are not related to the surrounding environment, that is, reflection characteristics (including information such as color and texture) of the object and shape characteristics of the object. Although the two intrinsic characteristics of the target object do not change with the change of the surrounding environment, the acquired observation image of the target object is affected by the environment. Therefore, if the essential characteristics of the object can be directly analyzed from the observation image, the information of the inherent shape, color, texture and the like of the object is extracted, and the influence of environmental change on the image is eliminated, so that the image can be more accurately understood, and a more reliable information basis is provided for further realizing robust image analysis in a complex environment.
The existing algorithms can be divided into three categories according to the way of extracting the essential features of the target: one type of algorithm is an implicit essential feature analysis algorithm, that is, a multi-mode (the appearance of the target object under different lighting conditions and different postures) of the target object is learned through a pattern recognition algorithm. In the learning process, the algorithm does not particularly consider the internal connection among various expressions, but directly performs pattern analysis on various observed results so as to try to obtain the distribution function of the target in the feature space. One serious problem encountered by this type of algorithm is the generalization of the objective description function. That is, the distribution of the training samples seriously affects the distribution function finally learned. If the learning sample is only the image of the target object in a single illumination state or posture, the training learning result is difficult to be popularized to the image of the target object in different illumination states or new postures. Therefore, in the case of imperfect samples, the algorithm is difficult to be generalized to the case of the target object in various complex environments. Another class of algorithms is explicit intrinsic feature analysis algorithms. The algorithm analyzes the internal relation of the target object according to the appearance of the target object in different states. Compared with the first kind of implicit algorithm, the algorithm directly analyzes the reflection characteristics and the shape of the object according to the physical imaging principle and the prior knowledge of the reflection characteristics and the shape. Therefore, the image of the target object in the new state can be directly calculated according to the inherent characteristics. Therefore, the result obtained by the explicit analysis algorithm is more accurate and has popularization. However, the algorithm usually adopts the constraint based on structure, texture, color and the like to realize the estimation of essential features, i.e. the signal classification problem is converted into the energy optimization problem by following the Retinex theoretical framework, and the calculation analysis is completed under a single scale. Therefore, the accuracy of the analysis result depends on the performance of the optimization algorithm to a great extent, and meanwhile, the optimal solution cannot be obtained because the process of solving the convexity of the established function to be optimized cannot be guaranteed to be trapped in a local minimum value, or the initialization step needs to be as close to the optimal solution as possible. These limit the performance of this type of algorithm. Still another algorithm is to extract the intrinsic images by using a neural network based on deep learning, and predict the intrinsic images directly from an RGB image by training a convolutional neural network. However, the network structure of the existing algorithm is simple, and the training set is an artificial image synthesized by computer graphics software, so that the extracted essential image is not very clear, especially when the method is applied to a natural image.
Disclosure of Invention
The invention provides a method for extracting an essential image by using a convolutional neural network, aiming at overcoming the defects of insufficient feature extraction capability of the existing implicit and explicit essential feature analysis algorithms and the problems that the deep learning-based neural network algorithm mainly aims at artificial images and the like. Firstly, a double-current convolution network with a parallel structure from an image to an image is constructed, then the network is trained, network parameters are optimized, so that multilayer features with environment invariance are extracted, and an essential image (a reflection map and a light map) is directly reconstructed. By adopting a multi-stream structure, on one hand, tasks can be separated, and different characteristics can be extracted by different shunts; on the other hand, the two are mutually limiting conditions, and the algorithm precision can be improved.
A method for extracting essential images by using a convolutional neural network is characterized by comprising the following steps:
step 1: and constructing a double-current convolutional neural network structure model with a parallel structure, wherein the network structure model is divided into a public branch and two special branches.
Wherein the common branch is composed of 5 convolutional layers, and each convolutional layer is followed by a pooling layer. Convolution kernels of the convolutional layers are all 3 x 3, each layer outputs a characteristic image, the dimensionality of the output characteristic image of the first convolutional layer is 64, the dimensionality of the output characteristic image of the second convolutional layer is 128, the dimensionality of the output characteristic image of the third convolutional layer is 256, the dimensionality of the output characteristic images of the fourth convolutional layer and the dimensionality of the output characteristic image of the fifth convolutional layer are both 512, and the pooling layer adopts average pooling with the size of 2 x 2.
The two special branches have the same structure and respectively comprise 3 deconvolution layers, convolution kernels are 4 multiplied by 4, one branch is used for reconstructing an illumination image, the other branch is used for reconstructing a reflection image, and the output dimensionality of all the deconvolution layers is 256.
The characteristic image output by the third convolution layer of the public branch and the output of the second deconvolution layer of the special branch are used as the input of the third deconvolution layer of the special branch; the characteristic image output by the fourth convolution layer of the public branch and the output of the first deconvolution layer of the special branch are used as the input of the second deconvolution layer of the special branch.
Step 2: and constructing a training data set, intercepting an image with the size of 1280 multiplied by 1280 from the middle part of each image of the BOLD data set, and dividing the intercepted image into five equal parts on a row and a column respectively, so that 25 images with the size of 256 multiplied by 256 are obtained from each image in the original data set, 53720 groups of images are randomly extracted to form a test set, and the rest images form the training set.
And step 3: and (3) training the double-current convolutional neural network constructed in the step (1) by using the training set obtained in the step (2), firstly performing random initialization on weights of all layers of the network, and then training the network by adopting a supervised error back propagation training method to obtain a trained network. Wherein the basic learning rate of the network is 10-13And adopting a fixed learning rate strategy, wherein the batch size of the network is 5, the loss function is SoftmaxWithLoss, and the network convergence condition is that the difference of the loss function values of the two iterations is within +/-5% of the value.
And 4, step 4: and (3) processing the test set obtained in the step (2) by using a trained network to obtain extracted essential images, namely a light map and a reflection map.
The method is also tested on the MIT Intrinsic Images dataset of the common data set of the Intrinsic Images, and the result shows that the method still has effectiveness.
The invention has the beneficial effects that: due to the adoption of the technical route of the intrinsic image extraction based on the deep learning theory, the reflection map and the illumination map can be directly separated from the original image by using the strong feature extraction capability of the neural network constructed based on the deep learning theory. In addition, the double-current convolution neural network provided by the invention is a full convolution network model from an image to an image, and comprises two branch flow directions which are respectively used for generating a light map and a reflection map; in addition, the network structure combines the convolution result of a higher layer with the result after the deconvolution operation, so that the details of the characteristic diagram after the deconvolution operation can be enhanced, the reconstruction errors of the illumination diagram and the reflection diagram are reduced to a certain extent, and the capability of reconstructing the network characteristics is improved.
Drawings
FIG. 1 is a flow chart of a method for extracting an intrinsic image using a convolutional neural network according to the present invention
FIG. 2 is a diagram of a dual-flow convolutional neural network structure constructed by the present invention
FIG. 3 is an example of a data set partial image constructed by the present invention
Detailed Description
The present invention will be further described with reference to the following drawings and examples, which include, but are not limited to, the following examples.
The invention provides a method for extracting an essential image by using a convolutional neural network, which comprises the following main processes as shown in figure 1:
1. construction of double-current convolution neural network structure model with parallel structure
The reconstruction of the image is actually done by giving different weights to the features extracted from the image and combining the same type of features to accomplish the goal of reconstructing the light and reflection maps from the original image. In other words, all required features are present in the same original image. Thus the feature extraction part can be shared, and the reconstruction of two different types of essential images needs to be done separately. Thus, the network constructed by the invention is divided into two parts, namely a public branch and a private branch. After the convolution operation of the common branch, the size of the characteristic graph output by each layer is gradually reduced. In order to keep the input image and the output image in the same size in space structure, three deconvolution layers are respectively designed on two special branches, so that the space size of the feature map is gradually restored to the original size. The invention is inspired by a residual error network structure, and the invention finds that the combination of the later two layers in the public branch and the later two layers in the special branch can lead the network parameters to obtain better optimization effect in the experimental process. For the above reasons, the present invention constructs a dual-stream convolutional neural network structure having a parallel structure as shown in fig. 2. The network model is divided into a public branch and two special branches.
Wherein the common branch is composed of 5 convolutional layers, and each convolutional layer is followed by a pooling layer. Convolution kernels of the convolutional layers are all 3 x 3, each layer outputs a characteristic image, the dimensionality of the output characteristic image of the first convolutional layer is 64, the dimensionality of the output characteristic image of the second convolutional layer is 128, the dimensionality of the output characteristic image of the third convolutional layer is 256, the dimensionality of the output characteristic images of the fourth convolutional layer and the dimensionality of the output characteristic image of the fifth convolutional layer are both 512, and the pooling layer adopts average pooling with the size of 2 x 2. The two special branches have the same structure and respectively comprise 3 deconvolution layers, convolution kernels are 4 multiplied by 4, one branch is used for reconstructing an illumination image, the other branch is used for reconstructing a reflection image, and the output dimensionality of all the deconvolution layers is 256. The feature image output by the third convolution layer of the public branch and the output of the second deconvolution layer of the special branch are used as the input of the third deconvolution layer of the special branch together; the feature image output by the fourth convolution layer of the public branch and the output of the first deconvolution layer of the private branch are used together as the input of the second deconvolution layer of the private branch.
2. Data set construction
The network structure provided by the invention is more complex, and more network parameters need to be trained. In order to make the network exert the optimal performance, the invention constructs a data set for researching the Intrinsic Image Extraction algorithm on the basis of a BOLD data set (Jiang X Y, Schofield A J, Wyatt J L. correlation-Based Intrinsic Image Extraction from a Single Image [ C ]. European Conference on Computer Vision,2010:58-71) created by Jiang et al. The data set contains 268,600 sets of pictures, each set containing an original picture, an illumination map and a reflection map. 53,720 groups were randomly drawn from the test set to test the performance of the intrinsic image extraction algorithm. The remaining 214,880 groups constitute a training set for training the deep-learning neural network. The BOLD database contains a large number of high resolution image sets, all of which are objects photographed under well-tuned lighting conditions, mainly including various complex patterns, faces, and outdoor scenes, and fig. 3 gives an example of a data set image. The purpose of Jiang et al in constructing this database is to provide a test platform for image processing algorithms. Specifically, the method mainly comprises an essential image extraction algorithm, a delumination algorithm, a light source estimation algorithm and the like. To this end, they provide a map of the lighting conditions and a map of the object surface, i.e. a map of the lighting and a map of the reflection, and are both standard RGB color space pictures with linear luminance characteristics. According to the invention, the number of pictures, the quality of the pictures, the complexity of a scene and the like are comprehensively considered, and finally, a picture group taking a complex pattern as a shooting object is selected to construct a data set for the research, wherein the original picture has 1280 pixel points in each dimension in the transverse direction and 1350 pixel points in the longitudinal direction, so that the data size is too large for a common computer, and the problem of over-learning is easily caused, which is very unfavorable for the training of a deep learning neural network. The image category selected by the invention has a very obvious characteristic: the key information is concentrated in the middle of the image. Therefore, the invention selects a feature frame of 1280 × 1280 in the middle of the image, intercepts the original image, and then divides the image into five equal parts on the row and the column respectively. Thus, an original image can be divided into 25 256 × 256 smaller images. By cutting the original image in the mode, key information in the original image is reserved, data utilization maximization is achieved, and meanwhile, various convenience conditions are provided for the research: the data volume of each group of pictures is moderate, and the requirements on the performance of a computer are not high; the image size is relatively reasonable, and the convolutional neural network is more convenient to design; the cut image contains positive and negative samples at the same time, so that overfitting can be avoided to a certain extent.
3. Network training
The embodiment trains the constructed deconvolution neural network under the Caffe framework using the training set in the dataset created based on BOLD. Compared with other frameworks, the Caffe framework is not only simple and convenient to install, but also supports all operating systems, and has good interface support for Python and Matlab. Because the constructed network structure is relatively complex, the data amount required to be learned is large, the number of times of iteration required by the network is large, and meanwhile, the optimal solution is prevented from being missed too fast in the network learning process, the basic learning rate is set to be 10 through repeated experiments in the network training process-13The learning rate policy is set to "fixed," i.e., a fixed learning rate. Considering computer performance, and also to avoid too fast network convergence, the batch size of the network is set to 5 and the loss function is SoftmaxwithLoss.
The loss function is used for calculating the difference value between the output result and the real label, and the network loss is smaller and smaller along with the increase of the iteration times, namely the estimation result is closer to the real label. The loss function in a single dimension can be written as:
wherein { (x)1,y1),...,(xm,ym) Denotes m sets of labeled training data, x denotes input, y denotes corresponding label, and y ∈ [0,255 }]. 1{ F } is an indication function, the function value is 1 when F is true and 0 when F is false, and theta represents a parameter of the convolutional neural network. During the training process, the error between the estimated image and the real label is propagated back to the neural network to optimize the parameters thereof, so that the error is gradually reduced. For an RGB image, the loss function is the sum of the errors of the above loss function in the three dimensions of the image R, G, B.
After 210,000 iterations, the difference between the front and the back of the loss function value fluctuates within the range of +/-5%, the network gradually converges, namely the network parameters tend to be optimal under the current network structure, the capability of the network for extracting the essential image tends to be optimal, and the trained network is obtained. Although the two proprietary branches look the same in structure, the network parameters they learn during network training will be different due to the different real labels provided to them. Therefore, when the network is used for extracting the essence images, the operation of the network on the data has corresponding difference, so that different branches can extract different types of essence images.
4. Extracting essential images by using trained network
And (3) processing the test set contained in the data set established in the step (2) by using a trained network, namely converting RGB pictures contained in the test set into a three-dimensional matrix as the input of the network, and obtaining extracted essential images, namely a light map and a reflection map, after multilayer operation of the network. The method is also tested on the MIT Intrinsic Images dataset of the common data set of the Intrinsic Images, and the result shows that the method still has effectiveness.
Claims (1)
1. A method for extracting essential images by using a convolutional neural network is characterized by comprising the following steps:
step 1: constructing a double-current convolutional neural network structure model with a parallel structure, wherein the network structure model is divided into a public branch and two special branches;
wherein, the public branch is composed of 5 convolution layers, and each convolution layer is followed by a pooling layer; convolution kernels of the convolutional layers are all 3 multiplied by 3, each layer outputs a characteristic image, the dimensionality of the output characteristic image of the first convolutional layer is 64, the dimensionality of the output characteristic image of the second convolutional layer is 128, the dimensionality of the output characteristic image of the third convolutional layer is 256, the dimensionality of the output characteristic images of the fourth convolutional layer and the dimensionality of the output characteristic image of the fifth convolutional layer are both 512, and the pooling layers are subjected to average pooling with the size of 2 multiplied by 2;
the two special branches have the same structure and respectively comprise 3 deconvolution layers, convolution kernels are 4 multiplied by 4, one branch is used for reconstructing an illumination image, the other branch is used for reconstructing a reflection image, and the output dimensionality of all the deconvolution layers is 256;
the characteristic image output by the third convolution layer of the public branch and the output of the second deconvolution layer of the special branch are used as the input of the third deconvolution layer of the special branch; the characteristic image output by the fourth convolution layer of the public branch and the output of the first deconvolution layer of the special branch are used as the input of the second deconvolution layer of the special branch;
step 2: constructing a training data set, intercepting an image with the size of 1280 multiplied by 1280 from the middle part of each image of the BOLD data set, and dividing the intercepted image into five equal parts on a row and a column respectively, so that 25 images with the size of 256 multiplied by 256 are obtained from each image in the original data set, 53720 groups of images in the original data set are randomly extracted to form a test set, and the rest images form a training set;
and step 3: training the double-current convolutional neural network constructed in the step 1 by using the training set obtained in the step 2, firstly randomly initializing weights of all layers of the network, and then training the network by adopting a supervised error back propagation training method to obtain a trained network; wherein the basic learning rate of the network is 10-13By usingThe fixed learning rate strategy is characterized in that the batch size of the network is 5, the loss function is SoftmaxWithLoss, and the network convergence condition is that the difference of the loss function values of the two iterations is within +/-5% of the value;
and 4, step 4: and (3) processing the test set obtained in the step (2) by using a trained network to obtain extracted essential images, namely a light map and a reflection map.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810407424.8A CN108764250B (en) | 2018-05-02 | 2018-05-02 | Method for extracting essential image by using convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810407424.8A CN108764250B (en) | 2018-05-02 | 2018-05-02 | Method for extracting essential image by using convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108764250A CN108764250A (en) | 2018-11-06 |
CN108764250B true CN108764250B (en) | 2021-09-17 |
Family
ID=64008978
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810407424.8A Active CN108764250B (en) | 2018-05-02 | 2018-05-02 | Method for extracting essential image by using convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108764250B (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109658330B (en) * | 2018-12-10 | 2023-12-26 | 广州市久邦数码科技有限公司 | Color development adjusting method and device |
CN109919869B (en) * | 2019-02-28 | 2021-06-04 | 腾讯科技(深圳)有限公司 | Image enhancement method and device and storage medium |
CN110659023B (en) * | 2019-09-11 | 2020-10-23 | 腾讯科技(深圳)有限公司 | Method for generating programming content and related device |
CN111179196B (en) * | 2019-12-28 | 2023-04-18 | 杭州电子科技大学 | Multi-resolution depth network image highlight removing method based on divide-and-conquer |
CN111325221B (en) * | 2020-02-25 | 2023-06-23 | 青岛海洋科技中心 | Image feature extraction method based on image depth information |
CN111489321B (en) * | 2020-03-09 | 2020-11-03 | 淮阴工学院 | Depth network image enhancement method and system based on derivative graph and Retinex |
CN113034353B (en) * | 2021-04-09 | 2024-07-12 | 西安建筑科技大学 | Intrinsic image decomposition method and system based on cross convolution neural network |
CN114742922A (en) * | 2022-04-07 | 2022-07-12 | 苏州科技大学 | Self-adaptive image engine color optimization method, system and storage medium |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104756152A (en) * | 2012-10-26 | 2015-07-01 | Sk电信有限公司 | Image correction device for accelerating image correction and method for same |
CN106778583A (en) * | 2016-12-07 | 2017-05-31 | 北京理工大学 | Vehicle attribute recognition methods and device based on convolutional neural networks |
CN107145857A (en) * | 2017-04-29 | 2017-09-08 | 深圳市深网视界科技有限公司 | Face character recognition methods, device and method for establishing model |
CN107330451A (en) * | 2017-06-16 | 2017-11-07 | 西交利物浦大学 | Clothes attribute retrieval method based on depth convolutional neural networks |
CN107403197A (en) * | 2017-07-31 | 2017-11-28 | 武汉大学 | A kind of crack identification method based on deep learning |
CN107633272A (en) * | 2017-10-09 | 2018-01-26 | 东华大学 | A kind of DCNN textural defect recognition methods based on compressed sensing under small sample |
JP2018018422A (en) * | 2016-07-29 | 2018-02-01 | 株式会社デンソーアイティーラボラトリ | Prediction device, prediction method and prediction program |
-
2018
- 2018-05-02 CN CN201810407424.8A patent/CN108764250B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104756152A (en) * | 2012-10-26 | 2015-07-01 | Sk电信有限公司 | Image correction device for accelerating image correction and method for same |
JP2018018422A (en) * | 2016-07-29 | 2018-02-01 | 株式会社デンソーアイティーラボラトリ | Prediction device, prediction method and prediction program |
CN106778583A (en) * | 2016-12-07 | 2017-05-31 | 北京理工大学 | Vehicle attribute recognition methods and device based on convolutional neural networks |
CN107145857A (en) * | 2017-04-29 | 2017-09-08 | 深圳市深网视界科技有限公司 | Face character recognition methods, device and method for establishing model |
CN107330451A (en) * | 2017-06-16 | 2017-11-07 | 西交利物浦大学 | Clothes attribute retrieval method based on depth convolutional neural networks |
CN107403197A (en) * | 2017-07-31 | 2017-11-28 | 武汉大学 | A kind of crack identification method based on deep learning |
CN107633272A (en) * | 2017-10-09 | 2018-01-26 | 东华大学 | A kind of DCNN textural defect recognition methods based on compressed sensing under small sample |
Non-Patent Citations (3)
Title |
---|
Correlation-Based Intrinsic Image Extraction from a Single Image;Jiang X Y等;《European Conference on Computer Vision》;20100930;全文 * |
Densely Connected Convolutional Networks;Gao Huang等;《2017 IEEE Conference on Computer Vision and Pattern Recognition》;20171109;全文 * |
基于非下采样 Contourlet 变换和稀疏表示的红外与可见光图像融合方法;王珺等;《兵工学报》;20130731;第34卷(第7期);全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN108764250A (en) | 2018-11-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108764250B (en) | Method for extracting essential image by using convolutional neural network | |
CN110135366B (en) | Shielded pedestrian re-identification method based on multi-scale generation countermeasure network | |
CN109815893B (en) | Color face image illumination domain normalization method based on cyclic generation countermeasure network | |
CN112767468B (en) | Self-supervision three-dimensional reconstruction method and system based on collaborative segmentation and data enhancement | |
CN113313657B (en) | Unsupervised learning method and system for low-illumination image enhancement | |
CN111445582A (en) | Single-image human face three-dimensional reconstruction method based on illumination prior | |
CN112489164B (en) | Image coloring method based on improved depth separable convolutional neural network | |
CN111861906A (en) | Pavement crack image virtual augmentation model establishment and image virtual augmentation method | |
CN110532928A (en) | Facial critical point detection method based on facial area standardization and deformable hourglass network | |
CN114463492A (en) | Adaptive channel attention three-dimensional reconstruction method based on deep learning | |
CN115018711B (en) | Image super-resolution reconstruction method for warehouse scheduling | |
CN113284061A (en) | Underwater image enhancement method based on gradient network | |
CN114066955A (en) | Registration method for registering infrared light image to visible light image | |
CN117292117A (en) | Small target detection method based on attention mechanism | |
CN115797561A (en) | Three-dimensional reconstruction method, device and readable storage medium | |
CN116958420A (en) | High-precision modeling method for three-dimensional face of digital human teacher | |
CN114663880A (en) | Three-dimensional target detection method based on multi-level cross-modal self-attention mechanism | |
CN113159158B (en) | License plate correction and reconstruction method and system based on generation countermeasure network | |
Guan et al. | DiffWater: Underwater image enhancement based on conditional denoising diffusion probabilistic model | |
CN117495718A (en) | Multi-scale self-adaptive remote sensing image defogging method | |
CN117036876A (en) | Generalizable target re-identification model construction method based on three-dimensional visual angle alignment | |
CN116823610A (en) | Deep learning-based underwater image super-resolution generation method and system | |
CN113971760B (en) | High-quality quasi-dense complementary feature extraction method based on deep learning | |
CN113205005B (en) | Low-illumination low-resolution face image reconstruction method | |
CN114913433A (en) | Multi-scale target detection method combining equalization feature and deformable convolution |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |