CN112465796A - Light field feature extraction method fusing focus stack and full-focus image - Google Patents

Light field feature extraction method fusing focus stack and full-focus image Download PDF

Info

Publication number
CN112465796A
CN112465796A CN202011432055.1A CN202011432055A CN112465796A CN 112465796 A CN112465796 A CN 112465796A CN 202011432055 A CN202011432055 A CN 202011432055A CN 112465796 A CN112465796 A CN 112465796A
Authority
CN
China
Prior art keywords
light field
image
focus
coordinate
view
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011432055.1A
Other languages
Chinese (zh)
Other versions
CN112465796B (en
Inventor
金欣
周思瑶
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen International Graduate School of Tsinghua University
Original Assignee
Shenzhen International Graduate School of Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen International Graduate School of Tsinghua University filed Critical Shenzhen International Graduate School of Tsinghua University
Priority to CN202011432055.1A priority Critical patent/CN112465796B/en
Publication of CN112465796A publication Critical patent/CN112465796A/en
Application granted granted Critical
Publication of CN112465796B publication Critical patent/CN112465796B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10052Images from lightfield camera
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20016Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Studio Devices (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a light field characteristic extraction method for fusing a focus stack and a full focus image, which comprises the following steps of: a1: inputting light field data, decoding and preprocessing the light field data to obtain a light field sub-view image array, and obtaining a focus stack and a full focus image at a plurality of view positions according to the light field sub-view image array; a2: respectively cascading the focus stacks and the full-focus images at a plurality of view angle positions to obtain an image set, and generating a Gaussian difference pyramid according to the image set; a3: and searching local extreme points in the Gaussian difference pyramid as feature point positions, and generating corresponding feature point descriptors. The light field feature extraction method provided by the invention can extract the four-dimensional light field feature points with robust depth and scale.

Description

Light field feature extraction method fusing focus stack and full-focus image
Technical Field
The invention relates to the field of computer vision and digital image processing, in particular to a light field characteristic extraction method for fusing a focus stack and a full-focus image.
Background
The point feature of the image is a sparse vector combination, can represent the self feature of the image which is different from other images, and is the beginning of the computer to identify and understand the image. The point features are also used for finding corresponding positions in different images, so that the method is applied to the fields of image splicing, three-dimensional reconstruction, SLAM and the like.
A micro-lens array is arranged in front of an image sensor of the handheld light field camera, and the spatial position and direction information of scene light rays can be recorded simultaneously in one-time shooting, so that assistance can be provided for research in the fields of image refocusing, depth map estimation, virtual reality and the like. A light field is typically parameterized as a combination of two-dimensional spatial information and two-dimensional angular information, described as a four-dimensional light field. Because the light field has four dimensions, and the common image only comprises two dimensions, the four-dimensional light field is difficult to be directly described through the two-dimensional image, and the traditional two-dimensional image feature extraction method does not consider image angle information and cannot completely represent high-dimensional information of the light field.
The above background disclosure is only for the purpose of assisting understanding of the concept and technical solution of the present invention and does not necessarily belong to the prior art of the present patent application, and should not be used for evaluating the novelty and inventive step of the present application in the case that there is no clear evidence that the above content is disclosed at the filing date of the present patent application.
Disclosure of Invention
In order to solve the technical problems, the invention provides a light field feature extraction method for fusing a focus stack and a full-focus image, which can extract four-dimensional light field feature points with robust depth and scale.
In order to achieve the purpose, the invention adopts the following technical scheme:
one embodiment of the invention discloses a light field feature extraction method for fusing a focus stack and a full focus image, which comprises the following steps:
a1: inputting light field data, decoding and preprocessing the light field data to obtain a light field sub-view image array, and obtaining a focus stack and a full focus image at a plurality of view positions according to the light field sub-view image array;
a2: respectively cascading the focus stacks and the full-focus images at a plurality of view angle positions to obtain an image set, and generating a Gaussian difference pyramid according to the image set;
a3: and searching local extreme points in the Gaussian difference pyramid as feature point positions, and generating corresponding feature point descriptors.
Preferably, the step a1 of obtaining the focal stack and the fully focused image at the multiple viewing angle positions according to the optical field sub-viewing angle image array specifically includes: and extracting the light field self-viewing angle image at the diagonal viewing angle in the light field sub-viewing angle image array, and generating the focus stack and the full focus image at the diagonal viewing angle according to the light field self-viewing angle image at the diagonal viewing angle.
Preferably, the focal stack at the diagonal view angle generated is:
Figure BDA0002820954350000021
wherein the content of the first and second substances,
Figure BDA0002820954350000022
is a coordinate in the angular domain of (u)0,v0) λ is a refocusing coefficient, L (u, v, s, t) represents the decoded lightfield from the lightfield data, (u, v) is an angular domain coordinate, (s, t) is a spatial domain coordinate, and U, V is the number of rows and columns of the lightfield sub-view image array;
preferably, the generated fully focused image is:
Figure BDA0002820954350000023
wherein the content of the first and second substances,
Figure BDA0002820954350000024
is a coordinate in the angular domain of (u)0,v0) Depth is the light field Depth map index.
Preferably, the step a2 of respectively cascading the focal stack and the fully focused image at a plurality of view angle positions to obtain an image set specifically includes: cascading the focal stack and the fully focused image at a plurality of view positions, respectively, using:
Figure BDA0002820954350000025
wherein the content of the first and second substances,
Figure BDA0002820954350000026
is a coordinate in the angular domain of (u)0,v0) The set of images at the viewing angle of (c),
Figure BDA0002820954350000031
is a coordinate in the angular domain of (u)0,v0) The focal stack at the viewing angle of (a),
Figure BDA0002820954350000032
is a coordinate in the angular domain of (u)0,v0) Is used to obtain a fully focused image at the viewing angle.
Preferably, the step a2 of generating the gaussian difference pyramid according to the image set specifically includes: constructing a gaussian difference pyramid using:
Figure BDA0002820954350000033
wherein the content of the first and second substances,
Figure BDA0002820954350000034
is a coordinate in the angular domain of (u)0,v0) The gaussian difference pyramid at the view angle of (a),
Figure BDA0002820954350000035
is a coordinate in the angular domain of (u)0,v0) The scale space of the image set at the view angle of (a);
further, the image set is subjected to fuzzy and downsampling processing by adopting a Gaussian function, and an angle domain coordinate (u) is obtained0,v0) Of the image set at the viewing angle
Figure BDA0002820954350000036
Expressed as:
Figure BDA0002820954350000037
wherein G (s, t, σ)i) Is a scale-variable gaussian function and is,
Figure BDA0002820954350000038
is a coordinate in the angular domain of (u)0,v0) A set of images at a viewing angle.
Preferably, the scale-variable Gaussian function G (s, t, σ)i) Expressed as:
Figure BDA0002820954350000039
where (s, t) is the spatial domain coordinate, σiIs a scale;
further, the scale σiExpressed as: sigmai+1=kσiI is more than or equal to 1 and less than or equal to N, N is the number of scales in the Gaussian scale space, and k is a parameter more than 0.
Preferably, the searching for the local extreme point in the gaussian difference pyramid as the feature point position in step a3 specifically includes:
calculating a local Hessian matrix of the Gaussian difference pyramid:
Figure BDA00028209543500000310
wherein the content of the first and second substances,
Figure BDA00028209543500000311
(s, t) is a space domain coordinate, and D is the Gaussian difference pyramid;
and when the local Hessian matrix takes the local maximum value, judging that the current point is the characteristic point, and taking the position of the current point as the position of the characteristic point.
Preferably, the step a3 of generating the corresponding feature point descriptor specifically includes: and partitioning the surrounding space region of the positions of the feature points, and calculating gradient histograms in 8 directions in a 4 x 4 window to obtain a 128-dimensional vector characterization descriptor.
Another embodiment of the present invention discloses a computer-readable storage medium storing computer-executable instructions that, when invoked and executed by a processor, cause the processor to implement the steps of the light field feature extraction method described above.
Compared with the prior art, the invention has the beneficial effects that: according to the light field characteristic extraction method for fusing the focus stack and the full-focus image, provided by the invention, the characteristic that a light field camera takes pictures first and then focuses is fully utilized, and the spatial information and the angle information which are acquired simultaneously and the relation between the spatial information and the angle information are combined to generate a group of sparse focus stack and full-focus image and calculate the extreme point in the Gaussian difference pyramid space corresponding to the cascade image set of the focus stack and the full-focus image, so that the four-dimensional light field characteristic point with robust depth and scale and high accuracy is obtained, and the application of the light field in the fields of panoramic stitching, SLAM and the like is greatly enriched.
Drawings
FIG. 1 is a flow chart of a light field feature extraction method fusing a focal stack and a fully focused image according to a preferred embodiment of the invention.
Detailed Description
In particular embodiments, when performing the above steps, the following may be followed. It should be noted that the specific methods employed in the practice are merely illustrative, and the scope of the present invention includes, but is not limited to, the following methods. The invention will be further described with reference to the accompanying drawings and preferred embodiments.
Scene information contained in the four-dimensional light field is far larger than that of a common image, and if the feature points are directly extracted, the calculation amount is huge. The three-dimensional focal stack is obtained by reducing the dimension of the light field, is a series of images focused on different planes, and can represent the angle information and the spatial information of the light field. From the focal stack, a fully focused image can be generated by stitching, clearly characterizing the texture information of the light field. The preferred embodiment of the invention is to fully utilize the information of the light field space domain and the angle domain to accurately extract the light field characteristics, and utilize the focal stack and the full-focus image to extract the light field characteristic points.
As shown in fig. 1, the preferred embodiment of the present invention discloses a light field feature extraction method for fusing a focal stack and a full-focus image, which makes full use of information of a light field spatial domain and an angular domain to accurately extract a four-dimensional light field feature point with a robust view angle, and includes the following steps:
a1: inputting light field data, decoding and preprocessing the light field data to obtain a light field sub-view image array, and obtaining a focus stack and a full-focus image at a plurality of view positions according to the light field sub-view image array;
specifically, step a1 includes the following steps:
a11: inputting light field data, and decoding and preprocessing the light field data to obtain a light field sub-view image array;
in this embodiment, L (u, v, s, t) is the input light field, where (u, v) is the angular domain coordinate and (s, t) is the spatial domain coordinate. The method for decoding and preprocessing the input light field to obtain the sub-view image of the light field comprises the following steps:
SAI(u0,v0)={L(u,v,s,t)|u=u0,v=v0}. (1)
wherein SAI (u)0,v0) Is the light field at the viewing angle (u)0,v0) A light field sub-view image of (a); obtaining a light field sub-view image array { SAI (u) after decoding the light field L0,v0)|u0∈[1,U],v0∈[1,V]}, U, V are the number of rows and columns of the light-field sub-view image array.
And after the light field sub-view image array is obtained through decoding, carrying out denoising and color correction operation pretreatment on the array, and outputting the light field sub-view image array for extracting the characteristic points.
A12: for sparse light fields at diagonal view angles, multiple three-dimensional focal stacks corresponding to these view angle positions are generated.
The light field has multiple views, and objects at different depths in the scene have different corresponding parallaxes in the multi-view sub-images. After the plurality of light field sub-visual angle images are translated and overlapped, synthetic aperture imaging can be realized, and a series of light field images F (s, t, lambda) with focusing planes changing from near to far, namely focus stacks, are generated. The focal stacks can describe depth information of scene points, the focal stack corresponding to each sub-visual angle is extracted, and then the characteristic points are extracted, so that the light field characteristic points with robust depth can be obtained. The direct parallax of each sub-view image of the handheld light field camera is small, light field data contain a large amount of redundancy, and a light field at a diagonal view angle can represent the information of the whole light field. Therefore, in the embodiment, only a sparse group of focus stacks corresponding to the diagonal view angles are generated, and then the light field feature points are extracted, so that on one hand, the calculation amount can be greatly reduced, and on the other hand, the speed of extracting the feature points can be improved.
The focal stack is obtained by translating the light field in the spatial dimension and then integrating the light field in the angular dimension, and the calculation process is as follows:
Figure BDA0002820954350000061
wherein the content of the first and second substances,
Figure BDA0002820954350000062
is the angle of view (u)0,v0) And λ is the refocusing coefficient. U, V is the number of rows and columns of the light-field sub-view image array.
The above formula (2) can also be expressed as:
when in use
Figure BDA0002820954350000063
The method comprises the following steps:
Figure BDA0002820954350000064
when in use
Figure BDA0002820954350000065
The method comprises the following steps:
Figure BDA0002820954350000066
when in use
Figure BDA0002820954350000067
The method comprises the following steps:
Figure BDA0002820954350000068
when in use
Figure BDA0002820954350000069
The method comprises the following steps:
Figure BDA00028209543500000610
the number of sub-view images used for generating the focal stack at different view angle positions is different, and the closer to the central view angle, the more the number of the used peripheral sub-view images is, the wider the focal plane range contained in the generated focal stack is, and therefore, the more complete the contained light field depth information is.
The full focus image is formed by pixel stitching extracting the in-focus position in the focal stack:
Figure BDA00028209543500000611
wherein the content of the first and second substances,
Figure BDA00028209543500000612
is the angle of view (u)0,v0) The full focus image at, Depth is the light field Depth map index. In this embodiment, the pixels in focus in the focus stack are indexed by the lightfield depth map, and a fully focused image at the view angle position is spliced.
A2: respectively cascading focus stacks and full-focus images at a plurality of visual angle positions to obtain an image set, and generating a Gaussian difference pyramid according to the image set;
specifically, the focal stack and the full-focus image are cascaded in an RGB channel to obtain an image set:
Figure BDA0002820954350000071
wherein the content of the first and second substances,
Figure BDA0002820954350000072
is a coordinate in the angular domain of (u)0,v0) A set of images at a viewing angle.
Then, for the cascaded image set, carrying out fuzzy and down-sampling processing by adopting a Gaussian function to generate a four-dimensional Gaussian difference pyramid so as to further effectively detect an extreme point of the robustness in a Gaussian scale space;
specifically, the gaussian difference pyramid D constructed in this embodiment is:
Figure BDA0002820954350000073
wherein the content of the first and second substances,
Figure BDA0002820954350000074
is a coordinate in the angular domain of (u)0,v0) The gaussian difference pyramid at the view angle of (a),
Figure BDA0002820954350000075
is a coordinate in the angular domain of (u)0,v0) The scale space of the image set at the view angle of (a);
in order to simulate the multi-scale characteristics of light field data, a Gaussian function is used for carrying out fuzzy and down-sampling processing on the three-dimensional image set after cascading, and the scale space of the obtained image set is as follows:
Figure BDA0002820954350000076
wherein G (s, t, σ)i) Is a scale-variable gaussian function, expressed as:
Figure BDA0002820954350000077
this example constructs a gaussian scale space (DoG) with N scales, i.e. using N scales k times different (k > 0):
σi+1=kσi,1≤i≤N (8)
dimension σiThe size of the three-dimensional image set determines the smoothness degree of the three-dimensional image set, the overall characteristics of the image are corresponding to the large-scale space, and the local characteristics of the image are corresponding to the small-scale space.
A3: and searching local extreme points in the Gaussian difference pyramid space as feature point positions and generating corresponding feature point descriptors.
Specifically, step a3 includes the following steps:
a31: searching local extreme points in a Gaussian difference pyramid space as feature point positions;
in the gaussian difference pyramid space, its local Hessian matrix is calculated:
Figure BDA0002820954350000081
wherein the content of the first and second substances,
Figure BDA0002820954350000082
when the Hessian matrix discriminant takes the local four-dimensional (space s, t + gradient lambda + scale sigma) maximum value, the current point is determined to be an extreme point, and the position of the characteristic point is positioned. The position of the feature point is expressed by four-dimensional coordinates defined by a biplane coordinate system, and the position of the view angle (u)0,v0) When searching the Gaussian difference pyramid, the obtained feature point position angle domain coordinate is (u)0,v0) The spatial domain coordinate is an extreme value(s)*,t*) (ii) a That is, the feature point position of the image set is described as { (u)0,v0,s*,t*)}。
A32: and for each feature point, calculating a feature point neighborhood gradient histogram and generating a corresponding feature point descriptor.
The feature point descriptor is a group of vectors, and the descriptor comprises key points and points which are around the key points and contribute to the key points, and is used as a basis for target matching, so that the key points can have more invariant characteristics. The invention divides the space area around the feature point into blocks, then calculates the gradient histogram of 8 directions in the window of 4 multiplied by 4, and obtains the 128-dimensional vector characterization descriptor.
Since the feature descriptors in this embodiment are constructed from the focal stack and the fully focused image, the features are robust to interfering objects at different depths, including partial occlusions and objects reflected from smooth surfaces.
The preferred embodiment of the invention provides a light field characteristic extraction method for fusing a focus stack and a full-focus image, which comprises the steps of converting a light field into a three-dimensional focus stack and a two-dimensional full-focus image on a group of sparse diagonal view angle image positions, and extracting an extreme point in a focus stack Gaussian difference pyramid space as a light field four-dimensional characteristic point position. Firstly, decoding and preprocessing light field data to obtain a light field sub-view image array, and then generating a plurality of three-dimensional focus stacks and two-dimensional full-focusing images corresponding to the view angle positions for a sparse light field at a diagonal view angle; cascading a focus stack and a full focus image, and performing fuzzy and down-sampling processing on the image set after cascading by adopting a Gaussian function to generate a four-dimensional Gaussian difference pyramid; and searching local extreme points in the Gaussian difference pyramid space as feature point positions and generating corresponding feature point descriptors. Experimental results show that compared with the existing algorithm, the method can extract the four-dimensional light field characteristic points with robust depth and scale.
Embodiments of the present invention further provide a computer-readable storage medium, where the computer-readable storage medium stores computer-executable instructions, and when the computer-executable instructions are called and executed by a processor, the computer-executable instructions cause the processor to implement the light field feature extraction method described above, and specific implementation may refer to method embodiments, and is not described herein again.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several equivalent substitutions or obvious modifications can be made without departing from the spirit of the invention, and all the properties or uses are considered to be within the scope of the invention.

Claims (10)

1. A light field feature extraction method fusing a focus stack and a full focus image is characterized by comprising the following steps:
a1: inputting light field data, decoding and preprocessing the light field data to obtain a light field sub-view image array, and obtaining a focus stack and a full focus image at a plurality of view positions according to the light field sub-view image array;
a2: respectively cascading the focus stacks and the full-focus images at a plurality of view angle positions to obtain an image set, and generating a Gaussian difference pyramid according to the image set;
a3: and searching local extreme points in the Gaussian difference pyramid as feature point positions, and generating corresponding feature point descriptors.
2. The light field feature extraction method according to claim 1, wherein the obtaining of the focal stack and the fully focused image at the plurality of view positions from the light field sub-view image array in step a1 specifically includes: and extracting the light field self-viewing angle image at the diagonal viewing angle in the light field sub-viewing angle image array, and generating the focus stack and the full focus image at the diagonal viewing angle according to the light field self-viewing angle image at the diagonal viewing angle.
3. The light field feature extraction method according to claim 2, wherein the focal stack at the diagonal view angle generated is:
Figure FDA0002820954340000011
wherein the content of the first and second substances,
Figure FDA0002820954340000012
is a coordinate in the angular domain of (u)0,v0) λ is a refocusing coefficient, L (u, v, s, t) represents the decoded lightfield from the lightfield data, (u, v) is an angular domain coordinate, (s, t) is a spatial domain coordinate, and U, V is the number of rows and columns of the lightfield sub-view image array.
4. The light field feature extraction method according to claim 3, wherein the generated fully focused image is:
Figure FDA0002820954340000013
wherein the content of the first and second substances,
Figure FDA0002820954340000014
is a coordinate in the angular domain of (u)0,v0) Depth is the light field Depth map index.
5. The light field feature extraction method according to claim 1, wherein the step a2 of respectively cascading the focus stack and the fully focused image at a plurality of view positions to obtain an image set specifically includes: cascading the focal stack and the fully focused image at a plurality of view positions, respectively, using:
Figure FDA0002820954340000021
wherein the content of the first and second substances,
Figure FDA0002820954340000022
is a coordinate in the angular domain of (u)0,v0) To seeThe set of images at the corners of the image,
Figure FDA0002820954340000023
is a coordinate in the angular domain of (u)0,v0) The focal stack at the viewing angle of (a),
Figure FDA0002820954340000024
is a coordinate in the angular domain of (u)0,v0) Is used to obtain a fully focused image at the viewing angle.
6. The light field feature extraction method according to claim 1, wherein the step a2 of generating the gaussian difference pyramid from the image set specifically comprises: constructing a gaussian difference pyramid using:
Figure FDA0002820954340000025
wherein the content of the first and second substances,
Figure FDA0002820954340000026
is a coordinate in the angular domain of (u)0,v0) The gaussian difference pyramid at the view angle of (a),
Figure FDA0002820954340000027
is a coordinate in the angular domain of (u)0,v0) The scale space of the image set at the view angle of (a);
further, the image set is subjected to fuzzy and downsampling processing by adopting a Gaussian function, and an angle domain coordinate (u) is obtained0,v0) Of the image set at the viewing angle
Figure FDA0002820954340000028
Expressed as:
Figure FDA0002820954340000029
wherein G (s, t, σ)i) Is a scale-variable gaussian function and is,
Figure FDA00028209543400000210
is a coordinate in the angular domain of (u)0,v0) A set of images at a viewing angle.
7. The light field feature extraction method according to claim 6, wherein the scale-variable Gaussian function G (s, t, σ)i) Expressed as:
Figure FDA00028209543400000211
where (s, t) is the spatial domain coordinate, σiIs a scale;
further, the scale σiExpressed as: sigmai+1=kσiI is more than or equal to 1 and less than or equal to N, N is the number of scales in the Gaussian scale space, and k is a parameter more than 0.
8. The light field feature extraction method according to claim 1, wherein the step a3 of searching the local extreme points in the gaussian difference pyramid as feature point positions specifically comprises:
calculating a local Hessian matrix of the Gaussian difference pyramid:
Figure FDA0002820954340000031
wherein the content of the first and second substances,
Figure FDA0002820954340000032
(s, t) is a space domain coordinate, and D is the Gaussian difference pyramid;
and when the local Hessian matrix takes the local maximum value, judging that the current point is the characteristic point, and taking the position of the current point as the position of the characteristic point.
9. The light field feature extraction method according to claim 1, wherein the generating of the corresponding feature point descriptor in step a3 specifically includes: and partitioning the surrounding space region of the positions of the feature points, and calculating gradient histograms in 8 directions in a 4 x 4 window to obtain a 128-dimensional vector characterization descriptor.
10. A computer-readable storage medium storing computer-executable instructions that, when invoked and executed by a processor, cause the processor to carry out the steps of the light field feature extraction method of any one of claims 1 to 9.
CN202011432055.1A 2020-12-07 2020-12-07 Light field feature extraction method integrating focal stack and full-focus image Active CN112465796B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011432055.1A CN112465796B (en) 2020-12-07 2020-12-07 Light field feature extraction method integrating focal stack and full-focus image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011432055.1A CN112465796B (en) 2020-12-07 2020-12-07 Light field feature extraction method integrating focal stack and full-focus image

Publications (2)

Publication Number Publication Date
CN112465796A true CN112465796A (en) 2021-03-09
CN112465796B CN112465796B (en) 2023-11-21

Family

ID=74801136

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011432055.1A Active CN112465796B (en) 2020-12-07 2020-12-07 Light field feature extraction method integrating focal stack and full-focus image

Country Status (1)

Country Link
CN (1) CN112465796B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116823914A (en) * 2023-08-30 2023-09-29 中国科学技术大学 Unsupervised focal stack depth estimation method based on all-focusing image synthesis
CN117253054A (en) * 2023-11-20 2023-12-19 浙江优众新材料科技有限公司 Light field significance detection method and related equipment thereof

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102043960A (en) * 2010-12-03 2011-05-04 杭州淘淘搜科技有限公司 Image grey scale and gradient combining improved sift characteristic extracting method
CN110246172A (en) * 2019-06-18 2019-09-17 首都师范大学 A kind of the light field total focus image extraction method and system of the fusion of two kinds of Depth cues
CN110490924A (en) * 2019-07-16 2019-11-22 西安理工大学 A kind of light field image feature point detecting method based on multiple dimensioned Harris
CN110996104A (en) * 2019-12-05 2020-04-10 华中科技大学 Light field focus stack image sequence encoding and decoding method, device and system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102043960A (en) * 2010-12-03 2011-05-04 杭州淘淘搜科技有限公司 Image grey scale and gradient combining improved sift characteristic extracting method
CN110246172A (en) * 2019-06-18 2019-09-17 首都师范大学 A kind of the light field total focus image extraction method and system of the fusion of two kinds of Depth cues
CN110490924A (en) * 2019-07-16 2019-11-22 西安理工大学 A kind of light field image feature point detecting method based on multiple dimensioned Harris
CN110996104A (en) * 2019-12-05 2020-04-10 华中科技大学 Light field focus stack image sequence encoding and decoding method, device and system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
XUANWU YIN ET AL.: "Iteratively reconstructing 4D light fields from focal stacks", APPLIED OPTICS *
苏博妮: "基于梯度域的光场全聚焦图像生成方法", 西南大学学报(自然科学版) *
陈怡良 等: "基于超分辨率焦点堆栈光场成像的全聚焦图像生成研究", 电脑知识与技术 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116823914A (en) * 2023-08-30 2023-09-29 中国科学技术大学 Unsupervised focal stack depth estimation method based on all-focusing image synthesis
CN116823914B (en) * 2023-08-30 2024-01-09 中国科学技术大学 Unsupervised focal stack depth estimation method based on all-focusing image synthesis
CN117253054A (en) * 2023-11-20 2023-12-19 浙江优众新材料科技有限公司 Light field significance detection method and related equipment thereof
CN117253054B (en) * 2023-11-20 2024-02-06 浙江优众新材料科技有限公司 Light field significance detection method and related equipment thereof

Also Published As

Publication number Publication date
CN112465796B (en) 2023-11-21

Similar Documents

Publication Publication Date Title
Zhang et al. LFNet: Light field fusion network for salient object detection
Kalantari et al. Learning-based view synthesis for light field cameras
Ji et al. Deep view morphing
Wang et al. A 4D light-field dataset and CNN architectures for material recognition
CN111598993B (en) Three-dimensional data reconstruction method and device based on multi-view imaging technology
Choi et al. Depth analogy: Data-driven approach for single image depth estimation using gradient samples
Wang et al. High-fidelity view synthesis for light field imaging with extended pseudo 4DCNN
CN108353188B (en) Method for encoding a light field content
Agrafiotis et al. Underwater photogrammetry in very shallow waters: main challenges and caustics effect removal
CN112465796B (en) Light field feature extraction method integrating focal stack and full-focus image
Pan et al. Multi-stage feature pyramid stereo network-based disparity estimation approach for two to three-dimensional video conversion
Zhou et al. AIF-LFNet: All-in-focus light field super-resolution method considering the depth-varying defocus
Yan et al. Deep learning on image stitching with multi-viewpoint images: A survey
CN112329662B (en) Multi-view saliency estimation method based on unsupervised learning
Arampatzakis et al. Monocular Depth Estimation: A Thorough Review
CN117541629A (en) Infrared image and visible light image registration fusion method based on wearable helmet
Fu et al. Image Stitching Techniques Applied to Plane or 3D Models: A Review
CN117115632A (en) Underwater target detection method, device, equipment and medium
Liu et al. Feature matching for texture-less endoscopy images via superpixel vector field consistency
Farhood et al. 3D point cloud reconstruction from a single 4D light field image
CN115410014A (en) Self-supervision characteristic point matching method of fisheye image and storage medium thereof
Zhang et al. Light field salient object detection via hybrid priors
Pouplin et al. Multimodal deep homography estimation using a domain adaptation generative adversarial network
Liu et al. A new stereo matching method for RAW image data based on improved SGBM
Imran et al. Unsupervised deep learning for depth estimation with offset pixels

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant