CN112365400A - Rapid super-resolution reconstruction method for light field angle - Google Patents

Rapid super-resolution reconstruction method for light field angle Download PDF

Info

Publication number
CN112365400A
CN112365400A CN202011164974.5A CN202011164974A CN112365400A CN 112365400 A CN112365400 A CN 112365400A CN 202011164974 A CN202011164974 A CN 202011164974A CN 112365400 A CN112365400 A CN 112365400A
Authority
CN
China
Prior art keywords
images
image
preset
view
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011164974.5A
Other languages
Chinese (zh)
Other versions
CN112365400B (en
Inventor
王兴政
昝永强
游森林
邓元龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen University
Original Assignee
Shenzhen University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen University filed Critical Shenzhen University
Priority to CN202011164974.5A priority Critical patent/CN112365400B/en
Publication of CN112365400A publication Critical patent/CN112365400A/en
Application granted granted Critical
Publication of CN112365400B publication Critical patent/CN112365400B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4046Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10052Images from lightfield camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a fast light field angle super-resolution reconstruction method, which comprises the steps of determining two input view angle images according to the position of a preset view angle; inputting the two input visual angle images into a trained image reconstruction network, and processing the two input visual angle images through the image reconstruction network to obtain a new visual angle image; there is only horizontal direction parallax between the view angle of the new view angle image and the view angles of the two input view angle images. Because each new visual angle image only adopts two visual angle images as input and the two input visual angle images and the new visual angle image only have horizontal parallax, the reconstruction speed of the image is improved while the quality of the new visual angle image is ensured, and the calculation resource is saved. Meanwhile, the invention can reconstruct a new view angle image at any position.

Description

Rapid super-resolution reconstruction method for light field angle
Technical Field
The invention relates to the field of light field imaging, in particular to a rapid light field angle super-resolution reconstruction method.
Background
A light field camera can record a plurality of view images at a time, and there is a trade-off between the spatial resolution and the angular resolution of the light field images, so that it is necessary to study how to improve the angular resolution of the light field while ensuring high spatial resolution.
The super-resolution reconstruction of the light field angle is also called the new visual angle synthesis of the light field. The traditional light field angle super-resolution reconstruction method can be divided into two methods, wherein the first method adopts a depth estimation algorithm to obtain an accurate depth map firstly and then maps the known view angle image to a new view angle position to reconstruct a new view angle; the second approach is to consider the light-field angular super-resolution reconstruction as a sampling and reconstruction of the plenoptic function, and to consider each pixel of the known view-angle image as a sampling of the high-dimensional plenoptic function. However, both of the above methods are susceptible to non-lambertian surfaces. In recent years, deep learning has made great progress in image noise reduction, super-resolution, deblurring, and the like, and the application of Convolutional Neural Network (CNN) to the field of light field angle super-resolution reconstruction has also begun to appear. Yoon and the like apply the convolutional neural network to light field angle super-resolution reconstruction for the first time, the method reconstructs a new view angle image by utilizing information between adjacent multi-view angle images, but only reconstructs a middle view angle image of the adjacent view angles, and cannot reconstruct the new view angle image at any position. Kalatari et al propose a learning-based light field new view reconstruction method, which improves the quality of reconstructed views, but because 4 input view images are required for synthesizing each new view, the synthesized view speed is slow.
Accordingly, the prior art is yet to be improved and developed.
Disclosure of Invention
In view of the above-mentioned deficiencies of the prior art, an object of the present invention is to provide a fast super-resolution reconstruction method for light field angle, which is capable of improving the reconstruction speed and saving the computation resource while ensuring the quality of the reconstructed new view angle image. According to the geometric principle of view interpolation, the intermediate view of the two views can be obtained according to the parallax information. Therefore, the input view angles are selected, so that only two input view angle images are needed for synthesizing each new view angle image, the calculation amount is reduced, and the synthesis speed of the new view angle is improved.
The technical scheme of the invention is as follows:
a fast light field angle super-resolution reconstruction method comprises the following steps:
acquiring two input visual angle images according to the position of a preset visual angle image; the preset visual angle and the input visual angles corresponding to the two input visual angle images only have horizontal parallax;
inputting the two input visual angle images into a trained image reconstruction network, and processing the two input visual angle images through the image reconstruction network to obtain a new visual angle image;
and the view angle corresponding to the new view angle image is a preset view angle.
The fast light field angle super-resolution reconstruction method is characterized in that the position of the preset visual angle is between the two input visual angles or the positions of the two input visual angles.
The fast light field angle super-resolution reconstruction method includes that the trained image reconstruction network includes a trained depth estimation network and a trained color estimation network, and the two input view images are processed by the image reconstruction network to obtain a new view image, including:
inputting the two input visual angle images into a trained depth estimation network, and obtaining the depth information of a new visual angle image through the trained depth estimation network;
and inputting the depth information of the new visual angle image and the two input visual angle images into the trained color estimation network, and obtaining the new visual angle image through the trained color estimation network.
The fast light field angle super-resolution reconstruction method, wherein the depth estimation network includes a trained first convolution neural network, and the obtaining of the depth information of the new view angle image through the depth estimation network includes:
presetting a plurality of depth levels according to the maximum parallax range of the preset visual angle and the input visual angle;
according to the depth information of the depth levels, feature extraction is carried out on the two input visual angle images to obtain a depth feature set;
and inputting the depth feature set into a first convolution neural network, and obtaining the depth information of the new visual angle image through the first convolution neural network.
The fast light field angle super-resolution reconstruction method, wherein the extracting features of the two input view images according to the depth information of the plurality of depth levels to obtain a depth feature set, includes:
obtaining two first mapping images corresponding to each depth level according to the depth information of each depth level and the two input view images;
obtaining a mean value and a standard deviation corresponding to each depth grade according to the two first mapping images corresponding to each depth grade;
and forming a set by the mean values and the standard deviations corresponding to all the depth levels to obtain a depth feature set.
The fast light field angle super-resolution reconstruction method, wherein the trained color estimation network includes a second convolutional neural network, and obtaining a new view angle image through the trained color estimation network includes:
obtaining a color feature set according to the depth information of the new view angle image, the position of the preset view angle and the two input view angle images;
and inputting the color feature set into the trained second convolutional neural network, and obtaining a new visual angle image through the trained second convolutional neural network.
The fast light field angle super-resolution reconstruction method, wherein obtaining a color feature set according to the depth information of the new view image, the position of the preset view and the two input view images, comprises:
obtaining two second mapping images according to the depth information of the new view angle image, the position of the preset view angle and the two input view angle images;
and combining the two second mapping images, the depth information of the new view angle image and the position of the preset view angle into a set to obtain a color feature set.
The fast light field angle super-resolution reconstruction method obtains two second mapping images according to the depth information of the new view image, the position of the preset view and the two input view images, and comprises the following steps:
obtaining the parallax between the two input view angle images and the new view angle image according to the depth information of the new view angle image and the position of the preset view angle;
and obtaining two second mapping images according to the parallax and the two input visual angle images.
The fast light field angle super-resolution reconstruction method comprises the following steps of:
preprocessing image samples in a light field data set to obtain a plurality of groups of training image groups, wherein each group of training image groups comprises two first images and second images corresponding to the two first images, the visual angles of the first images and the second images only have parallax in the horizontal direction, and the visual angle position of the second images is between the visual angle positions of the two first images;
inputting the two first images into a preset image reconstruction network to obtain reconstructed images corresponding to the two first images; and the preset image reconstruction network corrects network parameters of the preset image reconstruction network according to the reconstructed images corresponding to the two first images and the second images corresponding to the two first images, and continues to execute the step of obtaining the reconstructed images corresponding to the two first images according to the two first images until the training condition of the preset image reconstruction network meets a preset condition so as to obtain the trained image reconstruction network.
The fast light field angle super-resolution reconstruction method includes the steps of correcting network parameters of a preset image reconstruction network according to reconstructed images corresponding to the two first images and second images corresponding to the two first images, and continuing to execute the step of obtaining reconstructed images corresponding to the two first images according to the two first images until a training condition of the preset image reconstruction network meets a preset condition, so as to obtain a trained image reconstruction network, and includes the steps of:
obtaining error function values according to the reconstructed images corresponding to the two first images and the second images corresponding to the two first images;
and training the preset image reconstruction network according to the error function value, and continuing to execute the step of obtaining reconstructed images corresponding to the two first images according to the two first images until the training condition of the preset image reconstruction network meets a preset condition so as to obtain the trained image reconstruction network.
Has the advantages that: the invention provides a fast light field angle super-resolution reconstruction method, which comprises the steps of determining two input view angle images according to the position of a preset view angle; inputting the two input visual angle images into a trained image reconstruction network, and processing the two input visual angle images through the image reconstruction network to obtain a new visual angle image; there is only horizontal direction parallax between the view angle of the new view angle image and the view angles of the two input view angle images. Because only two view images are used as input and only horizontal parallax exists between the two input view images and the new view image, the reconstruction speed of the image is improved and the operation resource is saved while the quality of the new view image is ensured. Meanwhile, the invention can reconstruct a new view angle image at any position.
Drawings
Fig. 1 is a first flowchart of a fast optical field angle super-resolution reconstruction method according to a preferred embodiment of the present invention.
FIG. 2 is a second flowchart of a fast light field angle super-resolution reconstruction method according to a preferred embodiment of the present invention.
Fig. 3 is a schematic view of parallax.
Fig. 4 is a schematic diagram of a positional relationship between a preset viewing angle and an input viewing angle.
Detailed Description
The invention provides a fast light field angle super-resolution reconstruction method, which is further detailed in the following in order to make the purpose, technical scheme and effect of the invention clearer and more clear. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The invention provides a fast light field angle super-resolution reconstruction method, as shown in fig. 1, fig. 1 is a first flow diagram of a preferred embodiment of the fast light field angle super-resolution reconstruction method of the invention, the method includes the steps:
s10, determining two input visual angle images according to the position of the preset visual angle;
and S20, inputting the two input view images into a trained image reconstruction network, and processing the two input view images through the image reconstruction network to obtain a new view image.
Specifically, as shown in fig. 3, fig. 3 is a schematic diagram of parallax, where X denotes a point in space, C and C' denote left and right cameras, and parallax denotes an imaging difference of the same point in space on different cameras. For the light-field multi-view images, the view images in the same row only have horizontal parallax, and the view images in the same column only have vertical parallax. According to the geometric principle of view interpolation, the input views corresponding to the preset view and the two input view images only have horizontal parallax, and the parallax information can obtain a middle view image of the two view images, namely a new view image corresponding to the preset view. Therefore, two input view images are selected according to the preset view, so that only two input view images are needed for synthesizing the new view image, the calculation amount of image processing is reduced, and the synthesis speed of the new view is improved.
The view angle corresponding to the new view angle image is a preset view angle, and the view angle corresponding to the input view angle image is an input view angle. As shown in fig. 4, fig. 4 is a schematic diagram of a positional relationship between a preset viewing angle and an input viewing angle, a viewing angle position corresponding to a black image is between viewing angle positions corresponding to white images, both the viewing angle positions corresponding to the black image and the white image can be the positions of the preset viewing angle, and the viewing angle position corresponding to the white image is the position of the input viewing angle. Selecting the position of a preset visual angle, determining two input visual angles in the same horizontal parallax direction with the preset visual angle according to the position of the preset visual angle, and acquiring two input images corresponding to the two input visual angles respectively, wherein the two input visual angles and the preset visual angle only have horizontal parallax, and the position of the preset visual angle is between the two input visual angles or is the position of the two input visual angles. For example, if the 3 rd black image on the 1 st row is selected as the new viewing angle image, the two input viewing angle images are both the 1 st row image, the 1 st or 2 nd image on the 1 st row can be used as one input viewing angle image, and the 4 th, 5 th, 6 th, 7 th or 8 th image on the 1 st row can be used as the other input viewing angle image.
In the embodiment, the trained image reconstruction network is adopted, and the images corresponding to all view angle positions between the two input view angles in the horizontal direction can be reconstructed according to the two input view angle images of the same horizontal direction view angle, so that a new view angle image with higher quality is obtained.
Further, the trained image reconstruction network includes a trained depth estimation network and a trained color estimation network, and the step S20 includes:
s21, inputting the two input perspective images into a trained depth estimation network, and obtaining the depth information of the new perspective image through the trained depth estimation network;
and S22, inputting the depth information of the new visual angle image and the two input visual angle images into the trained color estimation network, and obtaining the new visual angle image through the trained color estimation network.
Specifically, the trained depth estimation network is used for extracting depth features at a preset view angle according to two input images, and obtaining depth information at the preset view angle through the depth features. The trained color estimation network is used for obtaining color features at a preset view angle according to depth information at the preset view angle, obtaining a new view angle image according to the color features, and the network structures of the trained depth estimation network and the trained color estimation network are shown in fig. 2.
Further, the trained depth estimation network includes a first convolutional neural network, and step S21 includes:
s211, presetting a plurality of depth levels according to the maximum parallax range of the preset visual angle and the input visual angle;
s212, according to the depth information of the depth levels, performing feature extraction on the two input visual angle images to obtain a depth feature set;
and S213, inputting the depth feature set into a first convolution neural network, and obtaining the depth information of the new visual angle image through the first convolution neural network.
Specifically, to map the input image to the preset view angle position according to the depth information of the new view angle image, so as to obtain the new view angle image, since the depth information × the base line (the distance between the new view angle and the input view angle) is equal to the parallax, the depth information range of the new view angle image can be obtained according to the maximum parallax range between the preset view angle and the input view angle. The maximum parallax refers to the maximum parallax between the preset viewing angle and the parallax between two input viewing angles, for example, as shown in fig. 4, the viewing angle corresponding to the 3 rd black image in the 1 st row is selected as the preset viewing angle, the viewing angle corresponding to the two white images in the 1 st row is selected as the two input viewing angles, the parallax between the white image on the right side and the black image is the maximum, and if the parallax is 21, the maximum parallax range is [ -21,21]. Due to the absence of the true value of the depth information, the depth information estimation needs to be performed within the depth information range. Presetting a plurality of depth levels in the depth information range of the new visual angle image, wherein each depth level corresponds to one depth information d1,d2,···,dlWherein l is the number of depth levels. After the depth level is preset, two input visual angle images are respectively inputAccording to the depth information mapping of different depth levels, two first mapping images corresponding to each depth level are obtained, and the first mapping image formula is as follows:
Figure BDA0002745497560000081
wherein the content of the first and second substances,
Figure BDA0002745497560000082
a first mapped image is represented as a first image,
Figure BDA0002745497560000083
representing the input view image, s represents the position in the pixel coordinate system, piIndicating the position of the input view angle, q indicating the position of the preset view angle, djAnd the depth information of each depth level is represented, i belongs to {1,2}, j belongs to {1,2 · · l }, and l is the number of the depth levels.
To extract features, the mean and standard deviation of the two first mapping images for each depth level (i.e., the mean and standard deviation for each depth level) are calculated, and the mean for each depth level is calculated
Figure BDA0002745497560000084
And standard deviation of
Figure BDA0002745497560000085
Respectively as follows:
Figure BDA0002745497560000086
Figure BDA0002745497560000087
wherein the content of the first and second substances,
Figure BDA0002745497560000091
and
Figure BDA0002745497560000092
respectively representing two first mapping images. Calculating the mean values and standard deviations corresponding to all depth levels, and combining all the mean values and standard deviations into a depth feature set K ═ Md1,Vd1,Md2,Vd3,···,Mdl,VdlBecause there are l depth levels, the depth feature set has 2l feature vectors. Because two input visual angle images are adopted, and only horizontal parallax exists between the input visual angles corresponding to the two input visual angle images and the preset visual angle corresponding to the new visual angle image, the time required for extracting the depth features is less, and the image reconstruction time is prolonged.
After the depth feature set is obtained, the depth feature set is input into a first convolution neural network, and depth information D at a preset visual angle is estimated by using the depth feature setqThe estimation formula is:
Dq=gd(K)
wherein D isqDepth information, g, representing a new view imagedAnd representing the mapping relation of the depth feature set and the depth information of the new view image.
Further, the color estimation network includes a second convolutional neural network, and the step S22 includes:
s221, obtaining a color feature set according to the depth information of the new view angle image, the position of the preset view angle and the two input view angle images;
s222, inputting the color feature set into the second convolutional neural network, and obtaining a new visual angle image through the trained second convolutional neural network.
Specifically, after the depth information of the new view image is obtained, two input view images are respectively mapped to the position of a preset view according to the depth information of the new view image, and two second mapping images corresponding to the depth information are obtained, where the mapping formula is as follows:
Figure BDA0002745497560000093
wherein D isq(s) depth information of the new view angle image, i ∈ {1,2}, (p)i-q)DqIndicating the disparity between the input view angle corresponding to the input image and the preset view angle corresponding to the new view angle image. From two second mapping images
Figure BDA0002745497560000101
And
Figure BDA0002745497560000102
depth information D of new view angle imageqAnd the position q of the preset visual angle forms a color characteristic set
Figure BDA0002745497560000103
The depth information of each pixel of the new view image corresponds to a corresponding pixel of each input image, so the color information of the new view image can be obtained from the input view image estimation. Inputting the color feature set H into a second convolution neural network to obtain a final new visual angle image, wherein the formula is as follows:
Lq=gc(H)
wherein L isqRepresenting a new view angle image, gcRepresenting a set of color features H and a new perspective image LqThe mapping relationship between them.
In one implementation, the trained image reconstruction network is trained by the following steps:
m1, preprocessing image samples in the optical field data set to obtain a plurality of training image groups, where each training image group includes two first images and second images corresponding to the two first images, the viewing angles of the first images and the second images only have horizontal parallax or vertical parallax, and the viewing angle position of the second image is between the viewing angle positions of the two first images.
Specifically, the network training is performed using a Stanford light field dataset, the data in the Stanford light field dataset being light field images, the light field images being acquired by a LytroIllum light field camera, the actual angular resolution of which is 14 × 14, that is, the obtained light field images are multi-view images composed of 14 × 14 view angle images. Since the edge view images of the multi-view image are affected by noise, vignetting effect, and the like, the central 8 × 8 view images are employed as the light field data set. The training set contained 100 light field images and the test set contained 30 light field images. Since the training speed of the complete image is very slow, the image blocks with the size of 60 × 60 are extracted from the complete image, the step size is 16 pixels, and therefore more than 100000 image blocks are obtained to train the network. Before network training, each image block needs to be preprocessed, that is, each image block is processed into a multi-view image, and in the multi-view image, an image with the same horizontal direction view angle is used as a training image group. In each training image group, two pictures corresponding to the leftmost and rightmost visual angles are used as first images, and other images are used as second images.
M2, inputting the two first images into a preset image reconstruction network to obtain reconstructed images corresponding to the two first images; and the preset network model corrects the model parameters of the preset network model according to the reconstructed images corresponding to the two first images and the second images corresponding to the two first images, and continues to execute the step of obtaining the reconstructed images corresponding to the two first images according to the two first images until the training condition of the preset network model meets the preset condition so as to obtain the trained image reconstruction network.
Specifically, the preset image reconstruction network comprises a preset depth estimation network and a preset color estimation network. Inputting the two first images into a preset image reconstruction network to obtain reconstructed images corresponding to the two first images, and reconstructing the reconstructed images corresponding to the two first images
Figure BDA0002745497560000111
Second image L corresponding to two first imagesq,kThe error between the two is used as an error function, a preset depth estimation network and a preset color estimation network are trained simultaneously according to the error function value of the error function, and the error function is as follows:
Figure BDA0002745497560000112
where k represents 3 channels of RGB. In order to use the gradient descent method to minimize the error function, it is necessary to calculate the error E and the partial derivatives of the two network weights, i.e.
Figure BDA0002745497560000113
And
Figure BDA0002745497560000114
wdand wcRespectively representing the weights of the preset depth estimation network and the preset color estimation network.
Since the output of the pre-set color estimation network is a reconstructed image, therefore,
Figure BDA0002745497560000115
directly using standard error back propagation calculation, and
Figure BDA0002745497560000116
it cannot be calculated directly, and it is divided into 3 parts using the chain-derivative rule as follows:
Figure BDA0002745497560000117
wherein the content of the first and second substances,
Figure BDA0002745497560000121
is the partial derivative between the error E and the reconstructed image and can be directly calculated;
Figure BDA0002745497560000122
is depth information (output of a preset depth estimation network) DqAnd weight w of the pre-set depth estimation networkdPartial derivatives between, can be calculated directly;
Figure BDA0002745497560000123
to representPartial derivative between reconstructed image and depth information, its set of color features
Figure BDA0002745497560000124
In that it is based on depth information DqA set of color features H is obtained, which in turn is used to obtain a reconstructed image
Figure BDA0002745497560000125
Thus, it is possible to obtain:
Figure BDA0002745497560000126
wherein the content of the first and second substances,
Figure BDA0002745497560000127
is a reconstructed image
Figure BDA0002745497560000128
And depth information DqThe partial derivatives between can be directly calculated;
Figure BDA0002745497560000129
is a reconstructed image
Figure BDA00027454975600001210
And a color feature set HtThe partial derivative in the t-th channel,
Figure BDA00027454975600001211
is a set of color features HtAnd depth information DqWith the partial derivatives at the t-th channel.
Figure BDA00027454975600001212
And
Figure BDA00027454975600001213
the partial derivatives corresponding to each channel need to be calculated separately, and the first 3N channels are mapping images mapped by the preset color estimation network. The 3N +1 channel refers to depth information, partial derivativeIs 1, the last two channels refer to the location of the reconstructed image, and is independent of depth, and is therefore 0. And performing error back propagation through the steps, then updating the weight, and repeatedly executing the steps to perform simultaneous training on the two networks until the training condition of the preset reconstruction network meets the preset condition, namely, the error function value is minimized, so as to obtain the trained image reconstruction network.
In summary, the present invention provides a fast super-resolution reconstruction method for light field angles, which determines two input view images according to the position of a preset view; inputting the two input visual angle images into a trained image reconstruction network, and processing the two input visual angle images through the image reconstruction network to obtain a new visual angle image; there is only horizontal direction parallax between the view angle of the new view angle image and the view angles of the two input view angle images. Because only two view images are used as input and only horizontal parallax exists between the two input view images and the new view image, the reconstruction speed of the new view image is improved, and the calculation resources are saved. Meanwhile, the invention can reconstruct a new view angle image at any position.
It is to be understood that the invention is not limited to the examples described above, but that modifications and variations may be effected thereto by those of ordinary skill in the art in light of the foregoing description, and that all such modifications and variations are intended to be within the scope of the invention as defined by the appended claims.

Claims (10)

1. A fast light field angle super-resolution reconstruction method is characterized by comprising the following steps:
acquiring two input visual angle images according to the position of a preset visual angle; the preset visual angle and the input visual angles corresponding to the two input visual angle images only have horizontal parallax;
inputting the two input visual angle images into a trained image reconstruction network, and processing the two input visual angle images through the image reconstruction network to obtain a new visual angle image; and the view angle corresponding to the new view angle image is a preset view angle.
2. The fast light-field angular super-resolution reconstruction method according to claim 1, wherein the position of the preset view is between or at the position of the two input views.
3. The fast light-field angular super-resolution reconstruction method according to claim 2, wherein the trained image reconstruction network comprises a trained depth estimation network and a trained color estimation network, and the processing of the two input view images by the image reconstruction network to obtain a new view image comprises:
inputting the two input visual angle images into a trained depth estimation network, and obtaining the depth information of a new visual angle image through the trained depth estimation network;
and inputting the depth information of the new visual angle image and the two input visual angle images into the trained color estimation network, and obtaining the new visual angle image through the trained color estimation network.
4. The fast light-field angular super-resolution reconstruction method according to claim 3, wherein the trained depth estimation network comprises a first convolutional neural network, and the obtaining of the depth information of the new view image through the depth estimation network comprises:
presetting a plurality of depth levels according to the maximum parallax range of the preset visual angle and the input visual angle;
according to the depth information of the depth levels, feature extraction is carried out on the two input visual angle images to obtain a depth feature set;
and inputting the depth feature set into a first convolution neural network, and obtaining the depth information of the new visual angle image through the first convolution neural network.
5. The fast light-field angular super-resolution reconstruction method according to claim 4, wherein said performing feature extraction on the two input view images according to the depth information of the plurality of depth levels to obtain a depth feature set comprises:
obtaining two first mapping images corresponding to each depth level according to the depth information of each depth level and the two input view images;
obtaining a mean value and a standard deviation corresponding to each depth grade according to the two first mapping images corresponding to each depth grade;
and forming a set by the mean values and the standard deviations corresponding to all the depth levels to obtain a depth feature set.
6. The fast light field angular super-resolution reconstruction method according to claim 3, wherein the trained color estimation network comprises a second convolutional neural network, and the obtaining of the new view image through the trained color estimation network comprises:
obtaining a color feature set according to the depth information of the new view angle image, the position of the preset view angle and the two input view angle images;
and inputting the color feature set into the trained second convolutional neural network, and obtaining a new visual angle image through the trained second convolutional neural network.
7. The fast light-field angular super-resolution reconstruction method according to claim 6, wherein said deriving a color feature set from the depth information of the new view image, the position of the preset view and the two input view images comprises:
obtaining two second mapping images according to the depth information of the new view angle image, the position of the preset view angle and the two input view angle images;
and combining the two second mapping images, the depth information of the new view angle image and the position of the preset view angle into a set to obtain a color feature set.
8. The fast light-field angular super-resolution reconstruction method according to claim 7, wherein obtaining two second mapping images according to the depth information of the new view image, the position of the preset view and the two input view images comprises:
obtaining the parallax between the two input view angle images and the new view angle image according to the depth information of the new view angle image and the position of the preset view angle;
and obtaining two second mapping images according to the parallax and the two input visual angle images.
9. The fast light-field angular super-resolution reconstruction method according to claim 1, wherein the trained image reconstruction network is trained by the following steps:
preprocessing image samples in a light field data set to obtain a plurality of groups of training image groups, wherein each group of training image groups comprises two first images and second images corresponding to the two first images, the visual angles of the first images and the second images only have parallax in the horizontal direction, and the visual angle position of the second images is between the visual angle positions of the two first images;
inputting the two first images into a preset image reconstruction network to obtain reconstructed images corresponding to the two first images; and the preset image reconstruction network corrects network parameters of the preset image reconstruction network according to the reconstructed images corresponding to the two first images and the second images corresponding to the two first images, and continues to execute the step of obtaining the reconstructed images corresponding to the two first images according to the two first images until the training condition of the preset image reconstruction network meets a preset condition so as to obtain the trained image reconstruction network.
10. The fast light field angular super-resolution reconstruction method according to claim 9, wherein the step of modifying the network parameters of the preset image reconstruction network according to the reconstructed images corresponding to the two first images and the second images corresponding to the two first images, and continuing to perform the step of obtaining the reconstructed images corresponding to the two first images according to the two first images until a training condition of the preset image reconstruction network satisfies a preset condition to obtain the trained image reconstruction network comprises:
obtaining error function values according to the reconstructed images corresponding to the two first images and the second images corresponding to the two first images;
and training the preset image reconstruction network according to the error function value, and continuing to execute the step of obtaining reconstructed images corresponding to the two first images according to the two first images until the training condition of the preset image reconstruction network meets a preset condition so as to obtain the trained image reconstruction network.
CN202011164974.5A 2020-10-27 2020-10-27 Rapid light field angle super-resolution reconstruction method Active CN112365400B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011164974.5A CN112365400B (en) 2020-10-27 2020-10-27 Rapid light field angle super-resolution reconstruction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011164974.5A CN112365400B (en) 2020-10-27 2020-10-27 Rapid light field angle super-resolution reconstruction method

Publications (2)

Publication Number Publication Date
CN112365400A true CN112365400A (en) 2021-02-12
CN112365400B CN112365400B (en) 2024-05-28

Family

ID=74510409

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011164974.5A Active CN112365400B (en) 2020-10-27 2020-10-27 Rapid light field angle super-resolution reconstruction method

Country Status (1)

Country Link
CN (1) CN112365400B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114926339A (en) * 2022-05-30 2022-08-19 北京拙河科技有限公司 Light field multi-view image super-resolution reconstruction method and system based on deep learning

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102075779A (en) * 2011-02-21 2011-05-25 北京航空航天大学 Intermediate view synthesizing method based on block matching disparity estimation
CN108280814A (en) * 2018-02-08 2018-07-13 重庆邮电大学 Light field image angle super-resolution rate method for reconstructing based on perception loss
KR20200021891A (en) * 2018-08-21 2020-03-02 삼성전자주식회사 Method for the synthesis of intermediate views of a light field, system for the synthesis of intermediate views of a light field, and method for the compression of a light field

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102075779A (en) * 2011-02-21 2011-05-25 北京航空航天大学 Intermediate view synthesizing method based on block matching disparity estimation
CN108280814A (en) * 2018-02-08 2018-07-13 重庆邮电大学 Light field image angle super-resolution rate method for reconstructing based on perception loss
KR20200021891A (en) * 2018-08-21 2020-03-02 삼성전자주식회사 Method for the synthesis of intermediate views of a light field, system for the synthesis of intermediate views of a light field, and method for the compression of a light field

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
NIMA KHADEMI KALANTARI ET AL.: "Learning-Based View Synthesis for Light Field Cameras", ACM TRANSACTIONS ON GRAPHICS, vol. 35, no. 6, pages 3 - 7 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114926339A (en) * 2022-05-30 2022-08-19 北京拙河科技有限公司 Light field multi-view image super-resolution reconstruction method and system based on deep learning
CN114926339B (en) * 2022-05-30 2023-02-03 北京拙河科技有限公司 Light field multi-view image super-resolution reconstruction method and system based on deep learning

Also Published As

Publication number Publication date
CN112365400B (en) 2024-05-28

Similar Documents

Publication Publication Date Title
CN110782490B (en) Video depth map estimation method and device with space-time consistency
CN106780590B (en) Method and system for acquiring depth map
TW202004679A (en) Image feature extraction method and saliency prediction method including the same
CN111105432B (en) Unsupervised end-to-end driving environment perception method based on deep learning
CN110381268B (en) Method, device, storage medium and electronic equipment for generating video
CN110517306B (en) Binocular depth vision estimation method and system based on deep learning
CN113762358B (en) Semi-supervised learning three-dimensional reconstruction method based on relative depth training
CN110910437B (en) Depth prediction method for complex indoor scene
CN106023230B (en) A kind of dense matching method of suitable deformation pattern
CN111047709B (en) Binocular vision naked eye 3D image generation method
CN111899295B (en) Monocular scene depth prediction method based on deep learning
US11727628B2 (en) Neural opacity point cloud
CN111105452B (en) Binocular vision-based high-low resolution fusion stereo matching method
CN106408596A (en) Edge-based local stereo matching method
CN112288788A (en) Monocular image depth estimation method
CN114996814A (en) Furniture design system based on deep learning and three-dimensional reconstruction
CN116310111A (en) Indoor scene three-dimensional reconstruction method based on pseudo-plane constraint
CN111861888A (en) Image processing method, image processing device, electronic equipment and storage medium
CN112365400B (en) Rapid light field angle super-resolution reconstruction method
CN111369435B (en) Color image depth up-sampling method and system based on self-adaptive stable model
WO2024032331A1 (en) Image processing method and apparatus, electronic device, and storage medium
CN112785502A (en) Light field image super-resolution method of hybrid camera based on texture migration
CN111696167A (en) Single image super-resolution reconstruction method guided by self-example learning
CN116385577A (en) Virtual viewpoint image generation method and device
CN115330935A (en) Three-dimensional reconstruction method and system based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant