CN110708532B - Universal light field unit image generation method and system - Google Patents

Universal light field unit image generation method and system Download PDF

Info

Publication number
CN110708532B
CN110708532B CN201910981967.5A CN201910981967A CN110708532B CN 110708532 B CN110708532 B CN 110708532B CN 201910981967 A CN201910981967 A CN 201910981967A CN 110708532 B CN110708532 B CN 110708532B
Authority
CN
China
Prior art keywords
image
camera
display
pixel
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910981967.5A
Other languages
Chinese (zh)
Other versions
CN110708532A (en
Inventor
闫兴鹏
燕展
蒋晓瑜
王晨卿
刘云鹏
汪熙
王子强
屈强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Academy of Armored Forces of PLA
Original Assignee
Academy of Armored Forces of PLA
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Academy of Armored Forces of PLA filed Critical Academy of Armored Forces of PLA
Priority to CN201910981967.5A priority Critical patent/CN110708532B/en
Publication of CN110708532A publication Critical patent/CN110708532A/en
Application granted granted Critical
Publication of CN110708532B publication Critical patent/CN110708532B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/236Image signal generators using stereoscopic image cameras using a single 2D image sensor using varifocal lenses or mirrors

Abstract

The invention discloses a method and a system for generating a universal light field unit image. The method comprises the following steps: determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array; determining a first camera according to the position of the intersection point; is divided intoRespectively acquiring a first image and a second image; determining the pixel position difference degree of the first image and the second image; acquiring nearest neighbor shooting light; determining the distance delta between the display light and the nearest neighbor shooting light; according to the pixel difference degree and the first position offset amount delta idDetermining a second position offset Δ iu(ii) a Determining pixel values of an image displayed by the display lens
Figure DDA0002235489560000011
The method and the system for generating the universal light field unit image can avoid the problem that the pixel at the corresponding position cannot be accurately acquired due to the introduction of the position error of the reference surface, so as to improve the accuracy of generating the unit image.

Description

Universal light field unit image generation method and system
Technical Field
The invention relates to the technical field of image generation, in particular to a method and a system for generating a universal light field unit image.
Background
The unit image is based on integrated imaging light field display, the accuracy of the unit image determines the accuracy of the light field display, the existing method has matching requirements on a display system, one or more reference surfaces are required to be arranged in the process of generating the image, and the position error of the reference surface can be introduced by introducing the reference surface, so that the accuracy of the obtained unit image is poor, and the three-dimensional effect reconstructed by using the unit image is poor. Therefore, the invention provides a method capable of remarkably improving the imaging accuracy of an image, and is a technical problem to be solved in the field.
Disclosure of Invention
The invention aims to provide a method and a system for generating a universal light field unit image, which can accurately generate the unit image.
In order to achieve the purpose, the invention provides the following scheme:
a universal light field unit image generation method is applied to a light field image acquisition device based on a camera array; the light field image acquisition device based on the camera array comprises: a camera array and a display lens array; the camera array and the display lens array are arranged in parallel; an object to be photographed is placed between the camera array and the display lens array;
the method comprises the following steps:
determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
determining the pixel position difference degree of the first image and the second image;
acquiring all the shooting light rays parallel to the display light rays, and correspondingly determining one of the shooting light rays emitted by the first camera which is closest to the display light rays, and recording the shooting light ray as a nearest neighbor shooting light ray;
determining a distance delta between the display ray and the nearest neighbor photographing ray,
Figure GDA0002774044100000021
wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,
Figure GDA0002774044100000022
dP is the center-to-center distance between adjacent cameras;
determining a first position deviation amount delta i according to the pixel position difference degree and the distanced
Figure GDA0002774044100000023
Wherein k is the pixel position difference degree;
according to the pixel difference degree and the first position offset delta idDetermining said second amount of position offset Δ iu,Δiu=k-Δid
Determining pixel values of an image displayed by the display lens
Figure GDA0002774044100000024
Figure GDA0002774044100000025
Wherein the content of the first and second substances,
Figure GDA0002774044100000026
pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,
Figure GDA0002774044100000027
is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,
Figure GDA0002774044100000028
is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
Optionally, the method further includes: acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
Optionally, the intersection point of the display light emitted by the display lens array and the camera array is x:
Figure GDA0002774044100000031
and D is the distance between the display lens array and the camera array.
Optionally, the determining the difference degree between the pixel positions of the first image and the second image includes:
obtaining the sum of the squares of errors R between the first image and the second imagej(k'),
Figure GDA0002774044100000032
According to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,
Figure GDA0002774044100000033
for the p-th pixel value in the first image,
Figure GDA0002774044100000034
for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
A pervasive light field unit image generation system, comprising:
the position determining module is used for determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
the first camera determining module is used for determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
the image acquisition module is used for respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
a pixel position difference degree determining module, configured to determine a pixel position difference degree of the first image and the second image;
a nearest neighbor shooting light ray obtaining module, configured to obtain all the shooting light rays emitted by the first camera and parallel to the display light ray, and correspondingly determine one of the shooting light rays closest to the display light ray, which is recorded as a nearest neighbor shooting light ray;
a distance determining module for determining a distance delta between the display ray and the nearest neighbor photographing ray,
Figure GDA0002774044100000041
wherein x is the position of the intersection point of the display light rays emitted by the display lens array and the camera array, n is the serial number of the first camera, and dP is the center distance between adjacent cameras;
a first position offset determination module, configured to determine a first position offset Δ i according to the pixel position difference and the distanced
Figure GDA0002774044100000042
Wherein k is the pixel position difference degree;
a second position offset determination module for determining the first position offset Δ i according to the pixel differencedDetermining said second amount of position offset Δ iu,Δiu=k-Δid
A pixel value determination module for determining pixel values of an image displayed by the display lens
Figure GDA0002774044100000043
Figure GDA0002774044100000044
Wherein the content of the first and second substances,
Figure GDA0002774044100000045
pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,
Figure GDA0002774044100000046
is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,
Figure GDA0002774044100000047
is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
Optionally, the system further includes: and the initialization parameter acquisition module is used for respectively acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
Optionally, the pixel position difference determining module includes:
a sum of squared errors acquisition unit for acquiring a sum of squared errors R between the first image and the second imagej(k'),
Figure GDA0002774044100000048
A pixel position difference degree determining unit for determining the difference degree according to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,
Figure GDA0002774044100000051
for the p-th pixel value in the first image,
Figure GDA0002774044100000052
for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
Optionally, the intersection point of the display light emitted by the display lens array and the camera array is x:
Figure GDA0002774044100000053
and D is the distance between the display lens array and the camera array.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects: according to the method and the system for generating the universal light field unit image, the pixel position obtained is corrected by adopting the pixel position difference degree and the position offset, so that the position of the obtained pixel has no error in the process of forming the unit image by adopting the pixel, the problem that the pixel at the corresponding position cannot be accurately obtained due to the introduction of the position error of the reference surface is solved, and the accuracy of generating the unit image is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.
FIG. 1 is a schematic diagram of a camera array and a display lens array in a generalized light field unit image generating method disclosed in an embodiment of the present invention;
FIG. 2 is a flowchart of a generalized light field unit image generation method according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram of a generalized light field unit image generation system according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The invention aims to provide a method and a system for generating a universal light field unit image, which can accurately generate the unit image.
In order to make the aforementioned objects, features and advantages of the present invention comprehensible, embodiments accompanied with figures are described in further detail below.
Fig. 1 is a schematic diagram of the operation of a camera array and a display lens array in the method for generating a ubiquitous light field unit image disclosed in the embodiment of the present invention, and fig. 2 is a flowchart of the method for generating a ubiquitous light field unit image disclosed in the embodiment of the present invention.
As shown in fig. 1 and 2, a generalized light field unit image generation method is mainly applied to a light field image acquisition device based on a camera array. The light field image acquisition device based on the camera array comprises: a camera array 1-1 and a display lens array 1-2. The camera array 1-1 and the display lens array 1-2 are arranged in parallel with each other. An object 1-3 to be photographed is placed between the camera array 1-1 and the display lens array 1-2.
The method comprises the following steps:
s100, determining the position of the intersection point of the display light ray emitted by one display lens in the display lens array 1-2 and the camera array 1.
S101, determining a camera at the position of the intersection point of the display light and the camera array according to the position of the intersection point, and recording the camera as a first camera.
And S102, respectively acquiring a first image and a second image. The first image is an image captured by the first camera. The second image is an image shot by a camera closest to the display light ray in cameras adjacent to the first camera.
S103, determining the pixel position difference degree of the first image and the second image.
S104, all the shooting light rays which are emitted by the first camera and are parallel to the display light rays are obtained, and one of the shooting light rays which is closest to the display light rays is correspondingly determined and recorded as a nearest neighbor shooting light ray.
S105, determining the distance delta between the display light ray and the nearest neighbor shooting light ray,
Figure GDA0002774044100000071
wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,
Figure GDA0002774044100000072
dP is the center-to-center spacing between adjacent cameras.
S106, determining a first position deviation amount delta i according to the pixel position difference degree and the distanced
Figure GDA0002774044100000073
And k is the pixel position difference degree.
S107, according to the pixel difference degree and the first position offset delta idDetermining said second amount of position offset Δ iu,Δiu=k-Δid
S108, determining the pixel value of the image displayed by the display lens
Figure GDA0002774044100000074
Figure GDA0002774044100000075
Wherein the content of the first and second substances,
Figure GDA0002774044100000076
pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,
Figure GDA0002774044100000077
is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,
Figure GDA0002774044100000078
is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
Since the displayed unit image is composed of individual pixels, the final display image can be obtained by calculating the pixel at any position in the display lens array.
Second position in the above-described method of the present disclosureOffset amount Δ iuAnd the essence of the position deviation is the position deviation of the shooting light which is parallel to the display light of the display lens and is closest to the display light in the shooting light emitted by the camera which is closest to the display light in the camera adjacent to the first camera.
The method may further comprise: respectively acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
The position of the intersection point of the display light emitted by the display lens array and the camera array is x:
Figure GDA0002774044100000081
and D is the distance between the display lens array and the camera array.
The determining the pixel position difference degree of the first image and the second image comprises:
obtaining the sum of the squares of errors R between the first image and the second imagej(k'),
Figure GDA0002774044100000082
According to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k')。
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,
Figure GDA0002774044100000083
for the p-th pixel value in the first image,
Figure GDA0002774044100000084
for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
The method disclosed above is mainly a method for determining pixels in the vertical direction of a camera array and a display lens array, and the method for determining pixels in the horizontal direction is the same as the method for determining pixels in the vertical direction, and is not repeated here.
The present invention also provides a system for generating a generic light field unit image, as shown in fig. 3, the system includes: the device comprises a position determining module 1, a first camera determining module 2, an image acquiring module 3, a pixel position difference degree determining module 4, a nearest neighbor shooting ray acquiring module 5, a distance determining module 6, a first position offset amount determining module 7, a second position offset amount determining module 8 and a pixel value determining module 9.
The position determining module 1 is used for determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array.
And determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point by using a first camera determination module 2, and recording the camera as a first camera.
The image acquisition module 3 is used for acquiring a first image and a second image respectively. The first image is an image captured by the first camera. The second image is an image shot by a camera closest to the display light ray in cameras adjacent to the first camera.
And determining the pixel position difference degree of the first image and the second image by adopting a pixel position difference degree determining module 4.
And acquiring all the shooting light rays which are parallel to the display light rays and are emitted by the first camera by adopting a nearest neighbor shooting light ray acquisition module 5, correspondingly determining one of the shooting light rays which is closest to the display light rays, and recording the shooting light ray as the nearest neighbor shooting light ray.
Determining the display ray and the nearest neighbor by using a distance determination module 6The interval delta of the photographing light is set to,
Figure GDA0002774044100000091
wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,
Figure GDA0002774044100000092
dP is the center-to-center spacing between adjacent cameras.
Determining a first position offset delta i by adopting a first position offset determination module 7 according to the pixel position difference degree and the distanced
Figure GDA0002774044100000093
And k is the pixel position difference degree.
A second position offset determining module 8 is adopted to determine the first position offset delta i according to the pixel difference degreedDetermining said second amount of position offset Δ iu,Δiu=k-Δid
Determining pixel values of an image displayed by the display lens using a pixel value determining module 9
Figure GDA0002774044100000094
Figure GDA0002774044100000095
Wherein the content of the first and second substances,
Figure GDA0002774044100000096
pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,
Figure GDA0002774044100000097
is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,
Figure GDA0002774044100000098
is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
The system also comprises an initialization parameter acquisition module. The initialization parameter acquisition module respectively acquires the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
The pixel position difference degree determination module may further include: an error square sum acquisition unit and a pixel position difference degree determination unit.
Wherein the error square sum acquisition unit acquires the error square sum R between the first image and the second imagej(k'),
Figure GDA0002774044100000101
The pixel position difference degree determining unit determines the difference degree according to the sum of squares of errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k')。
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,
Figure GDA0002774044100000102
for the p-th pixel value in the first image,
Figure GDA0002774044100000103
for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k ' corresponding to the minimum value is obtained, where k ' is the center position of the start of the pixel search range.
According to the specific embodiment provided by the invention, the invention discloses the following technical effects: according to the method and the system for generating the universal light field unit image, the pixel position obtained is corrected by adopting the pixel position difference degree and the position offset, so that the position of the obtained pixel has no error in the process of forming the unit image by adopting the pixel, the problem that the pixel at the corresponding position cannot be accurately obtained due to the introduction of the position error of the reference surface is solved, and the accuracy of generating the unit image is improved.
In addition, in the method and the system for generating the universal light field unit image, which are provided by the invention, weighting operation is carried out on various pixels in the process of forming the required pixels, so that the continuity and consistency of the unit image generated by the pixels can be obviously improved.
The embodiments in the present description are described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments are referred to each other. For the system disclosed by the embodiment, the description is relatively simple because the system corresponds to the method disclosed by the embodiment, and the relevant points can be referred to the method part for description.
The principles and embodiments of the present invention are explained herein using specific examples, which are presented only to assist in understanding the method and its core concepts. Meanwhile, for a person skilled in the art, according to the idea of the present invention, the specific embodiments and the application range may be changed. In view of the above, the present disclosure should not be construed as limiting the invention.

Claims (8)

1. A universal light field unit image generation method is applied to a light field image acquisition device based on a camera array; the light field image acquisition device based on the camera array comprises: a camera array and a display lens array; the camera array and the display lens array are arranged in parallel; an object to be photographed is placed between the camera array and the display lens array; characterized in that the method comprises:
determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
determining the pixel position difference degree of the first image and the second image;
acquiring all the shooting light rays parallel to the display light rays, and correspondingly determining one of the shooting light rays emitted by the first camera which is closest to the display light rays, and recording the shooting light ray as a nearest neighbor shooting light ray;
determining a distance delta between the display ray and the nearest neighbor photographing ray,
Figure FDA0002814304930000011
wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,
Figure FDA0002814304930000012
dP is the center-to-center distance between adjacent cameras;
determining a first position deviation amount delta i according to the pixel position difference degree and the distanced
Figure FDA0002814304930000013
Wherein k is the pixel position difference degree;
according to the pixel position difference degree and the first position offset delta idDetermining a second position offset Δ iu,Δiu=k-Δid
Determining pixel values of an image displayed by the display lens
Figure FDA0002814304930000014
Figure FDA0002814304930000015
Wherein the content of the first and second substances,
Figure FDA0002814304930000021
pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,
Figure FDA0002814304930000022
is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,
Figure FDA0002814304930000023
is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
2. A method for generating a generic light field unit image according to claim 1, further comprising: acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
3. The method of claim 1, wherein the position of the intersection of the display light rays emitted by the display lens array and the camera array is x:
Figure FDA0002814304930000024
wherein D is the distance between the display lens array and the camera array, and m is the serial number of the display lens.
4. The method of claim 1, wherein the determining the difference degree of pixel positions of the first image and the second image comprises:
obtaining the sum of the squares of errors R between the first image and the second imagej(k'),
Figure FDA0002814304930000025
According to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,
Figure FDA0002814304930000026
for the p-th pixel value in the first image,
Figure FDA0002814304930000027
for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k' corresponding to the minimum value is obtained.
5. A pervasive light field unit image generation system, comprising:
the position determining module is used for determining the position of the intersection point of a display ray emitted by a display lens in the display lens array and the camera array;
the first camera determining module is used for determining a camera at the intersection point of the display ray and the camera array according to the position of the intersection point, and recording the camera as a first camera;
the image acquisition module is used for respectively acquiring a first image and a second image; the first image is an image shot by the first camera; the second image is an image shot by a camera which is closest to the display light ray in the adjacent cameras of the first camera;
a pixel position difference degree determining module, configured to determine a pixel position difference degree of the first image and the second image;
a nearest neighbor shooting light ray obtaining module, configured to obtain all the shooting light rays emitted by the first camera and parallel to the display light ray, and correspondingly determine one of the shooting light rays closest to the display light ray, which is recorded as a nearest neighbor shooting light ray;
a distance determining module for determining a distance delta between the display ray and the nearest neighbor photographing ray,
Figure FDA0002814304930000031
wherein x is the position of the intersection point of the display light emitted by the display lens array and the camera array, n is the serial number of the first camera,
Figure FDA0002814304930000032
dP is the center-to-center distance between adjacent cameras;
a first position offset determination module, configured to determine a first position offset Δ i according to the pixel position difference and the distanced
Figure FDA0002814304930000033
Wherein k isThe pixel position difference degree is obtained;
a second position offset determining module, configured to determine the first position offset Δ i according to the pixel position differencedDetermining a second position offset Δ iu,Δiu=k-Δid
A pixel value determination module for determining pixel values of an image displayed by the display lens
Figure FDA0002814304930000034
Figure FDA0002814304930000035
Wherein the content of the first and second substances,
Figure FDA0002814304930000041
pP is the pixel size of the camera at the intersection, f is the focal length of the camera at the intersection, dE is the center distance between the display lenses, pE is the pixel size of the display lens emitting the display light, g is the distance between the display lens emitting the display light and the display screen, j is the j-th pixel serial number of the display lens emitting the display light, ceil (x) is the minimum integer which returns the calculation result which is greater than or equal to x, a is the weight, a is delta/dP,
Figure FDA0002814304930000042
is the i + Δ i th image in the first imagedThe value of each of the pixels is calculated,
Figure FDA0002814304930000043
is the i- Δ i in the second imageuAnd n 'is the serial number of the camera which is closest to the display light ray in the cameras adjacent to the first camera, and n' is n +/-1.
6. The ubiquitous light field unit image generating system of claim 5, wherein said system further comprises: and the initialization parameter acquisition module is used for respectively acquiring the center distance between every two adjacent cameras in the camera array, the pixel size of every camera, the focal length of every camera, the center distance between every two adjacent display lenses in the display lens array, the pixel size of every display lens, the distance between every display lens and a display screen and the distance between the display lens array and the camera array.
7. The system of claim 5, wherein the pixel position disparity determination module comprises:
a sum of squared errors acquisition unit for acquiring a sum of squared errors R between the first image and the second imagej(k'),
Figure FDA0002814304930000044
A pixel position difference degree determining unit for determining the difference degree according to the sum of squared errors Rj(k ') determining a degree of difference in pixel position between said first image and said second image, k' where Rj(k');
Where th is the set pixel search range, k' is the center position of the start of the pixel search range,
Figure FDA0002814304930000045
for the p-th pixel value in the first image,
Figure FDA0002814304930000046
for the t-th pixel value, k' while R, in the second imagej(k') is when Rj(k ') the value of k' corresponding to the minimum value is obtained.
8. The system of claim 5, wherein the position of the intersection of the display light rays emitted by the display lens array and the camera array is x:
Figure FDA0002814304930000047
wherein D is the distance between the display lens array and the camera array, and m is the serial number of the display lens.
CN201910981967.5A 2019-10-16 2019-10-16 Universal light field unit image generation method and system Active CN110708532B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910981967.5A CN110708532B (en) 2019-10-16 2019-10-16 Universal light field unit image generation method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910981967.5A CN110708532B (en) 2019-10-16 2019-10-16 Universal light field unit image generation method and system

Publications (2)

Publication Number Publication Date
CN110708532A CN110708532A (en) 2020-01-17
CN110708532B true CN110708532B (en) 2021-03-23

Family

ID=69199885

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910981967.5A Active CN110708532B (en) 2019-10-16 2019-10-16 Universal light field unit image generation method and system

Country Status (1)

Country Link
CN (1) CN110708532B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111225221B (en) * 2020-01-15 2021-12-14 未来新视界文化科技(嘉善)有限公司 Panoramic video image processing method and device
CN111595254B (en) * 2020-06-04 2021-09-21 中国人民解放军陆军装甲兵学院 Method and system for measuring axial distance between lens array and LCD display screen

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103475889B (en) * 2013-09-01 2015-02-18 西安电子科技大学 Method for obtaining micro cell image array in real time based on double cameras
CN104735360B (en) * 2013-12-18 2017-12-22 华为技术有限公司 Light field image treating method and apparatus
CN104469343B (en) * 2014-11-26 2017-02-01 北京智谷技术服务有限公司 Optical field display control and device and optical field display device
US10247858B2 (en) * 2015-10-25 2019-04-02 Facebook Technologies, Llc Liquid crystal half-wave plate lens
CN107121786B (en) * 2017-05-18 2019-07-19 中国人民解放军装甲兵工程学院 The calibration method of integration imaging light field display system lens array
CN107580207A (en) * 2017-10-31 2018-01-12 武汉华星光电技术有限公司 The generation method and generating means of light field 3D display cell picture

Also Published As

Publication number Publication date
CN110708532A (en) 2020-01-17

Similar Documents

Publication Publication Date Title
US20230362344A1 (en) System and Methods for Calibration of an Array Camera
EP2751521B1 (en) Method and system for alignment of a pattern on a spatial coded slide image
CN109813251B (en) Method, device and system for three-dimensional measurement
US9715734B2 (en) Image processing apparatus, imaging apparatus, and image processing method
CN109859272B (en) Automatic focusing binocular camera calibration method and device
Genovese et al. Stereo-camera calibration for large-scale DIC measurements with active phase targets and planar mirrors
CN103776419B (en) A kind of binocular distance measurement method improving measurement range
CN107077743A (en) System and method for the dynamic calibration of array camera
US10904512B2 (en) Combined stereoscopic and phase detection depth mapping in a dual aperture camera
EP2849148A2 (en) Three-dimensional printing system and method for three-dimensional printing
CN107808398B (en) Camera parameter calculation device, calculation method, program, and recording medium
CN111080705B (en) Calibration method and device for automatic focusing binocular camera
US8749652B2 (en) Imaging module having plural optical units in which each of at least two optical units include a polarization filter and at least one optical unit includes no polarization filter and image processing method and apparatus thereof
US8817121B2 (en) Image apparatus and imaging method
JP2019082680A (en) Method, device, and method for calibration of three-dimensional display device
CN110708532B (en) Universal light field unit image generation method and system
JP2011149931A (en) Distance image acquisition device
CN104200456B (en) A kind of coding/decoding method for line-structured light three-dimensional measurement
CN106888344A (en) Camera module and its inclined acquisition methods of image planes and method of adjustment
JP2011099779A (en) Distance image acquisition device, and distance image acquisition processing method
JP2015212772A (en) Range-finding device, imaging apparatus, range-finding method, and range-finding parameter calculation method
JP2020153865A (en) Three-dimensional information acquisition device, information processor, and system
CN111127379B (en) Rendering method of light field camera 2.0 and electronic equipment
CN116625258A (en) Chain spacing measuring system and chain spacing measuring method
CN102171533B (en) Distance measuring device and method for manufacturing same

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant