CN112040214A - Double-camera three-dimensional imaging system and processing method - Google Patents
Double-camera three-dimensional imaging system and processing method Download PDFInfo
- Publication number
- CN112040214A CN112040214A CN201910481518.4A CN201910481518A CN112040214A CN 112040214 A CN112040214 A CN 112040214A CN 201910481518 A CN201910481518 A CN 201910481518A CN 112040214 A CN112040214 A CN 112040214A
- Authority
- CN
- China
- Prior art keywords
- image
- camera
- camera lens
- lens
- resolution
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/243—Image signal generators using stereoscopic image cameras using three or more 2D image sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
The invention discloses a self-adaptive three-dimensional imaging system.A light field shooting part comprises a first imaging part, a first camera lens and a second camera or a camera lens; the first camera lens and the second camera or the camera lens are respectively positioned at the back part and the front part of the lens part, an entrance pupil plane and a matching device are arranged between the first camera lens and the second camera or the camera lens, an internal reflection unit is formed between the first camera lens and the entrance pupil plane and between the first camera lens and the matching device, and the high-resolution camera part also comprises a second imaging part and a third camera lens; the light field imaging section and the high resolution imaging section are configured such that the third camera lens obtains a second image that coincides with a front view vertical direction among the plurality of secondary images, and outputs the plurality of secondary images and the second image at the same time. The invention can acquire accurate depth information and can acquire a high-resolution video to analyze the three-dimensional image.
Description
Technical Field
The invention belongs to the field of stereo imaging, and particularly relates to a double-camera three-dimensional stereo imaging system and a processing method based on a light field technology.
Background
There are many design schemes for cameras used to capture three-dimensional images and videos in the prior art. The most common solution is to arrange two camera modules of the same specification at a certain distance, for example, about 60 to 65 mm, in a linear arrangement, so as to simulate the stereoscopic vision principle of human eyes. The image sensors of the two camera modules can respectively record two-dimensional images or videos obtained by respective shooting, a depth map can be established by utilizing the two-dimensional images or two-dimensional videos obtained by software processing, and then the two-dimensional images or two-dimensional videos are converted into three-dimensional images or three-dimensional videos. Another solution is to use a stereo camera directly to take the picture. Two image sensors are arranged in a main body of the stereo camera and respectively record two-dimensional images or two-dimensional videos from two lens groups of a camera lens, and then a system and software attached to the camera can synthesize the two-dimensional images into a three-dimensional image or synthesize the two-dimensional videos into a three-dimensional video. However, the two schemes may not synchronize two-dimensional images or videos, or may be influenced by external factors such as lighting conditions of the environment, etc. to generate three-dimensional images and video quality.
More advanced image capturing devices, such as light field cameras (also known as plenoptic cameras), capture a light field image of a scene at a time using a microlens array lens, and extract depth information of the scene by calculation to create a depth map and convert a two-dimensional image into a three-dimensional image. However, the main disadvantages of such a light field imaging apparatus are that the image resolution significantly drops, the parallax angle is small, and it is not well suited for capturing video. The recent design is to add a reflection unit to capture multi-angle images of the target object, because the parallax angle is large, the processed images can generate more clear depth images and three-dimensional stereo images, and the images are also suitable for shooting videos, but the problem of resolution drop is still not solved by the attempt.
Disclosure of Invention
The invention aims to provide a double-camera three-dimensional stereo imaging system and a processing method for improving the resolution of a three-dimensional video. The imaging system has wide application in a variety of fields, such as medical, biotech research, industrial equipment manufacturing, semiconductor product quality verification, and the like. The invention can acquire accurate depth information and also can acquire high-quality video to analyze the three-dimensional image.
The invention provides a double-camera three-dimensional imaging system, which is characterized by comprising: a light field imaging section that obtains a first image and a high resolution imaging section that obtains a second image; wherein the light field photographing section includes a first imaging section, a first camera lens and a second camera or camera lens; the first camera lens and the second camera or camera lens are respectively positioned at the back and front of the lens portion, an entrance pupil plane and a matching device are positioned between the first camera lens and the second camera or camera lens, the entrance pupil plane and the matching means can be adapted to different focal lengths of the second camera or camera lens, an internal reflection unit is formed between the first camera lens and said entrance pupil plane and matching means, the internal reflection unit is used for decomposing and refracting the captured first image into a plurality of secondary images with different angle shifts, the high-resolution camera part further comprises a second imaging part and a third camera lens, the central axis adjusting device can adjust the double lenses of the first camera lens and the second camera lens or the camera lens and the single lens of the third camera lens, and enables the axes of the double lenses and the single lens to be parallel; the light field imaging section and the high resolution imaging section are configured such that the third camera lens obtains a second image that coincides with a front view vertical direction among the plurality of secondary images, and outputs the plurality of secondary images and the second image at the same time.
In one aspect of the invention, the light field image pick-up section is located as close as possible to the high resolution image pick-up section and both are centered on the same vertical plane.
One aspect of the invention wherein the plurality of secondary images having different angular offsets have an angular offset range of 10-20 degrees.
One aspect of the invention, wherein the angular offset of the elevation view in the plurality of secondary images is 0 degrees. The first imaging section further includes a first image sensor and a fly-eye lens that captures a first image; the fly-eye lens transmitting the captured first image to the first image sensor; and the second imaging section further includes a second image sensor; the second image obtained by the third camera lens is transmitted to the second image sensor.
In one aspect of the invention, the fly-eye lens is a plurality of micro-lens arrays, and the radius, thickness and array pitch of each micro-lens are related to the size of the first image sensor.
In one aspect of the present invention, the first camera lens and the second camera or camera lens have adjustable apertures and focal lengths, the second camera or camera lens and the third camera lens are interchangeable lenses, and the aperture of the second camera or camera lens is larger than the size of the internal reflection unit.
In one aspect of the invention, the entrance pupil plane and the matching means are pupil lenses having a diameter larger than the diameter of the internal reflection unit and allowing the incident light rays of the light field image to be refracted in the internal reflection unit.
In one aspect of the present invention, each of the secondary images has a subtle difference of a scene, and the size of the internal reflection unit and the focal length of each of the secondary images are calculated based on the following equations (1) and (2):
wherein the FOV is a field of view of the second video camera or camera lens;
n is a refractive index of the internal reflection unit;
r is the number of internal reflections;
z is the size of the internal reflection unit;
flensis said second video camera or camera lensThe focal length of the head;
fsubis the focal length of the secondary image.
The invention also provides a processing method of the double-camera three-dimensional imaging, which comprises the following steps: obtaining original depth map data of a first image by a light field imaging part; correcting the original depth map data; obtaining a high-resolution depth map generated by interpolation by using an edge-oriented or directional rendering method; and simultaneously obtaining a second image by using a high-resolution image pick-up part, and correcting the original depth map data of the first image by using a data model and combining the second image as reference data until an optimal interpolation-generated high-resolution depth map is obtained.
The three-dimensional imaging system and the processing method provided by the invention can provide two-dimensional and three-dimensional videos with higher resolution, and meanwhile, compared with a light field camera of a high-resolution image sensor, the cost is increased very limitedly; in addition, since the system of the present invention does not affect the function of the light field camera portion, the information obtained by the light field camera itself can still be used to calculate object depth and build a depth map.
Drawings
In order to more clearly illustrate the technical solution in the embodiments of the present invention, the drawings required to be used in the embodiments will be briefly described below. It is obvious that the drawings in the following description are only examples of the invention, and that for a person skilled in the art, other drawings can be derived from them without making an inventive step.
Fig. 1 is a perspective view of a three-dimensional imaging system of the present invention.
Fig. 2 is a block diagram of a three-dimensional imaging system of the present invention.
Fig. 3 is a schematic diagram of a three-dimensional imaging system of the present invention acquiring a first image 120.
Fig. 4 is a schematic diagram of the three-dimensional imaging system according to the present invention after normalizing the acquired first image 120.
Fig. 5 is a flow chart of the three-dimensional imaging system of the present invention processing the second image 130.
FIG. 6 is a flow chart of obtaining an image of a target by the three-dimensional imaging system of the present invention.
Detailed Description
Specific embodiments of the present invention will now be described with reference to the accompanying drawings. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. These embodiments are provided only for the purpose of exhaustive and comprehensive description of the invention so that those skilled in the art can fully describe the scope of the invention. The terminology used in the detailed description of the embodiments illustrated in the accompanying drawings is not intended to be limiting of the invention.
Fig. 1 is a perspective view of a three-dimensional imaging system of the present invention. The three-dimensional imaging system of the present invention is composed of a light field image capturing section 100 for obtaining a first image 120 (not shown in fig. 1) and a high resolution image capturing section 140 for obtaining a second image 130 (not shown in fig. 1), wherein the light field image capturing section 100 may be a light field camera in chinese patent application 201711080588.6, which includes a first imaging section 110, a first camera lens 101 and a second camera or camera lens 103, wherein the first camera lens 101 is a rear camera lens; with adjustable aperture and focal length. The second camera or camera lens 103 is a front camera or camera lens, which can adjust the focal length of the camera. Intermediate the first camera lens 101 and the second camera or camera lens 103 is an entrance pupil plane and matching device 109, which entrance pupil plane and matching device 109 may be a pupil lens, with an internal reflection unit 102 between the pupil lens 109 and the first camera lens 101. The high-resolution image pickup section 140 and the light field image pickup section 100 are integrally fixed together, and the high-resolution image pickup section 140 includes a second imaging section 116, and the lens central axis 112a (see fig. 2) of the third camera lens 117 of the high-resolution image pickup section 140 and the lens central axis 112b (see fig. 2) of the first camera lens 101 and the second camera lens 103 in the light field image pickup section 100 are held in parallel by the central axis adjusting means 118.
Fig. 2 is a block diagram of a dual camera three-dimensional imaging system of the present invention. The light field imaging part 100 of the three-dimensional imaging system comprises a first imaging part 110 and a lens part 111, wherein the first imaging part 110 comprises a first image sensor 104; a fly-eye lens 105; the first image sensor 104 is an image sensor with higher imaging quality; fly-eye lens 105 is formed by a series of combinations of lenslets that capture information of an image, such as light field image information, from different angles, thereby stripping off three-dimensional information to identify a particular object. Fly-eye lens 105 is composed of an array of micro-lenses designed to produce a depth map in addition to capturing a light field image. Also, the fly-eye lens 105 is served by the first image sensor 104, and thus, it is related to a parameter of the first image sensor 104. For example, each microlens parameter of the fly-eye lens 105 has a radius of 0.5 mm, a thickness of 0.9 micrometers, and an array pitch of each microlens is 60 micrometers. The fly-eye lens is scalable in size with respect to the first image sensor 104. In one embodiment, the size of the C-type image sensor using the advanced photography system is 25 mm × 17 mm; in yet another embodiment, a full-frame image sensor is used having dimensions of 37 mm by 25 mm.
The lens section 111 is detachably connected to the first imaging section 110. The pupil lens 109 may be a single lens that acts as a condenser and compresses the information received by the second camera or camera lens 103. An imaging process is performed at the second camera or camera lens 103, and the imaging angle is different as the second camera or camera lens 103 is replaced or replaced. The first camera lens 101 is a short-focus lens or a macro lens, which is fixed to a housing (not shown in fig. 2), and the design of the first camera lens 101 determines the size of the imaging system of the present invention. The secondary imaging process is performed at the first camera lens 101. The entrance pupil plane and the matching means 109 are designed to correct the light. Between the entrance pupil plane and the matching means 109 and the first lens 101 is an internal reflection unit 102; the internal reflection unit 102 decomposes and reflects an image to be photographed into multi-angle images having independent secondary images with different angle offsets. The internal reflection unit 102 is designed to provide a plurality of virtual images at different viewing angles. The size and scale of the internal reflection unit 102 are determinative of the number of reflections and the ratio of reflected images, resulting in images at different angles. Each reflection produces a secondary image with subtle differences in the scene, and the target image has a slight offset. The size of the internal reflection unit 102 and the focal length of each secondary image may be calculated based on the following equations (1) and (2):
wherein the FOV is a field of view of the second video camera or camera lens;
n is a refractive index of the internal reflection unit;
r is the number of internal reflections;
x, Y, Z are the dimensions of the internal reflection unit, width, height, length, respectively;
flensis the focal length of the second video camera or camera lens;
fsubis the focal length of the secondary image.
The size of the internal reflection unit 102 may be the same as that of the first image sensor 104, and in one embodiment may be 24 mm (width) by 36 mm (height) by 95 mm (length), that is, the ratio of the units is about 2: 3: 8. The pupil lens 109 is used to match the size of the secondary image with the size of the internal reflection unit 102 and to correctly reflect in the internal reflection unit 102. To achieve this, the diameter of the pupil lens 109 should be larger than the internal reflection unit 102. The pupil lens 109 employed in one embodiment is approximately 50 mm in diameter and 50 mm in focal length. The second camera or camera lens 103 is designed to be able to be replaced by any camera or camera lens as long as the aperture of the second camera or camera lens 103 is larger than the size of the internal reflection unit 102.
The high resolution image pickup section 140 includes the second imaging section 116, the second image sensor 119, and the third camera lens 117. The adjusting means 118 for adjusting the center axis of the twin lens of the first camera lens 101 and the second camera or camera lens 103 and the single lens of the third camera lens 117 is located outside the high-resolution imaging section 140 and is independent of the light field imaging section 100 and the high-resolution imaging section 140 as long as the axis 112b of the first camera lens 101, the second camera or camera lens 103 can be made parallel to the axis 112a of the third camera lens 117 by adjusting the adjusting means 118. The second image sensor 119 may adopt a sensor with the same or different specification as the first image sensor 104, but the resolution of the second image sensor 119 should be at least above the resolution 1/9 of the first camera sensor 104, so as to achieve the purpose of improving the light field video resolution of the three-dimensional imaging system of the present invention.
Fig. 3 is a schematic diagram of a first image 120 obtained by the light field imaging section 100 of the three-dimensional imaging system of the present invention. The internal reflection unit 102 in the light field imaging section 100 decomposes the captured first image 120, i.e., the light field image or the video picture, and reflects into a plurality of secondary images or video pictures having different angular offsets, e.g., 9 secondary images or video pictures, which are acquired by the first image sensor 104 of the first imaging section 110 through the fly-eye lens. The secondary image (i) in the middle of the 9 secondary images or video pictures is the front view of the shot scene, and the other 8 secondary images (ii) -nine or video pictures are the secondary images or video pictures shifted by a specific angle. Each image or video frame has on average only 1/9 or less resolution of the first image sensor 104. Before generating the scene depth map, 9 secondary images or video frames are segmented and each of the secondary images is preprocessed.
Fig. 4 is a schematic diagram of the first image 120 obtained by the light field imaging section 100 of the three-dimensional imaging system according to the present invention after normalization processing. Each secondary image is normalized by the following equation (3):
wherein In(n ═ 1, 2.., 9) represents the image before normalization; i'n(n ═ 1, 2.., 9) represents the normalized image; mirror (I)mLeft, right, up, down) (m ═ 1, 2, 3, 4) represents the image mirror flipping left, right, up, and down; rotate (I)kPi) (k ═ 6.., 9) represents image rotation.
After normalization, the shift of the scene among each secondary image can be identified, with each secondary image being a separate original compound eye image. The next step is to perform pre-processing by image processing techniques including, but not limited to, image noise removal, and then decoding by synthetic aperture techniques, so as to obtain the light field information in the original image of the compound eye, and generate a quasi-focused secondary image by digital refocusing techniques. Wherein the synthetic aperture image can be digitally refocused using the following principles:
L′(u,v,x′,y′)=L(u,v,kx′+(1-k)u,ky′+(1-k)v)(6)
I′(x′,y′)=∫∫L(u,v,kx′+(1-k)u,ky′+(1-k)v)dudv(7)
wherein I and I' represent the coordinate systems of the primary and secondary imaging planes;
l and L' represent the energy of the primary and secondary imaging planes.
After the 9 in-focus secondary images are acquired, two secondary images, such as secondary images (c) and (c), can be selected, a Disparity Map (Disparity Map) is calculated by using a binocular Stereo Matching algorithm, and then a scene depth Map with a lower resolution is further established. Among them, Semi-Global Matching (Semi-Global Matching) algorithm is one of commonly used binocular stereo vision Matching algorithms, provides good parallax effect and has ideal calculation speed. By establishing a cost function associated with the disparity map:
wherein D represents a disparity map;
p and q are a certain pixel in the image;
C(p,Dp) The disparity value representing the current pixel is DpA cost value for the pixel;
Nprepresents the pixels adjacent to pixel p, typically 8;
p1 and P2 are penalty coefficients, the disparity value difference of P1 for pixel P and the adjacent pixel is equal to 1, and the disparity value difference of P2 for pixel P and the adjacent pixel is greater than 1;
t [ ] is a function that returns a 1 if the parameter in the function is true, and returns a 0 otherwise.
For each pixel of the image, the minimum cost value of a certain parallax value is calculated, and the cost values of 8 directions are accumulated, and the parallax value with the lowest value after accumulation is equivalent to the final parallax value of the pixel. The whole disparity map can be obtained by calculating each pixel one by one.
Finally, the following formula can be used for converting the disparity value and the depth value:
wherein d ispRepresents a depth value of a certain pixel;
f is the normalized focal length;
b is the baseline distance between the two secondary images;
Dprepresenting the disparity value of the current pixel.
The second image 130 obtained by the second imaging portion 116 is a 2D image or a video picture, and can completely reflect the information of the object, so that the resolution of the second image is not reduced, and the second image 130 does not need to be normalized or refocused for the same reason.
For example, the first image sensor 104 and the second image sensor 119 both use 4K image sensors, that is, the first and second image sensors each have 3840 × 2160 pixels. Assuming that the 9 secondary images or video frames are the same size, each secondary image has on average only a sensor 1/9 or lower resolution, e.g., 1280x 720 pixels per secondary image, which would be similarly limited in resolution by 1280x 720 pixels if the scene depth map and light field video were generated directly. Therefore, a scene depth map of 1280x 720 is created by using the secondary image (i-ninthly), and then the resolution of the depth map is improved to 3840x 2160 by using an Edge-directed interpolation algorithm with reference to the 3840x 2160 pixel high-resolution second image 130 obtained by the second image sensor 119. The formula used for edge-directed interpolation is as follows:
wherein m and n are low resolution and high resolution image grids before and after interpolation;
y [ n ] represents the depth map generated after interpolation;
s and R represent the data model of the second image 130 and the operator of the edge-directed rendering step, respectively;
λ is the gain of the correction process;
k is the iteration index.
Due to the high resolution of the second image 130 as a basis, the accuracy of the correction step of the interpolation calculation is sufficient to cope with the need to generate a 3D image. Then, the high-resolution second image 130, i.e., the 2D image or the video image, is combined with the depth map with the improved resolution to generate a high-resolution 2D + Z3D image or video format, and the 3D image or video format is output to the display, so that the resolution of the light field image or the light field video can be greatly improved to 3840 × 2160 pixels.
FIG. 5 is a flow chart of the three-dimensional imaging system of the present invention processing a target image. In step 501, original depth map data of the first image 120 is obtained by the first imaging section 100; in step 502, correcting the original depth map data; in step 503, using an edge-oriented or directional rendering method, obtaining a high-resolution depth map generated by interpolation in step 504; wherein in step 505, the second image obtained in step 506 as reference data is compared with a data model of the second image; the original depth map data of the first image 120 is modified until the best interpolated high resolution depth map is obtained.
FIG. 6 is a flow chart of obtaining an image of a target by the three-dimensional imaging system of the present invention. In step 601, a first image 120 comprising 9 secondary images or video frames is acquired by a first image sensor 104; in step 602, 9 secondary images or video frames are divided and normalized; in step 603, image noise removal processing is performed on each secondary image; in step 604, the light field information obtained from the 9 secondary images is decoded using a synthetic aperture technique, and then a quasi-focus image is generated using a digital refocusing technique; in step 605, a lower resolution scene depth map is created using the 9 in-focus secondary images; combining the high resolution 2D image obtained by the second image sensor or the second image of the video frame obtained at step 608; in step 606, with reference to the second image, the resolution of the depth map is increased using an edge-directed or edge-directed interpolation algorithm; in step 607, the enhanced resolution depth map and the second image are combined to generate a high resolution 3D image or video picture.
The above description is only for the purpose of illustrating the present invention, and any person skilled in the art can modify and change the above embodiments without departing from the spirit and scope of the present invention. Therefore, the scope of the claims should be accorded the full scope of the claims. The invention has been explained above with reference to examples. However, other embodiments than the above described are equally possible within the scope of this disclosure. The different features and steps of the invention may be combined in other ways than those described. The scope of the invention is limited only by the appended claims. More generally, those of ordinary skill in the art will readily appreciate that all parameters, dimensions, materials, and configurations described herein are exemplary and that actual parameters, dimensions, materials, and/or configurations will depend upon the particular application or applications for which the teachings of the present invention is/are used.
Claims (18)
1. A dual camera three dimensional stereoscopic imaging system, comprising:
a light field imaging section that obtains a first image and a high resolution imaging section that obtains a second image;
wherein the light field photographing section includes a first imaging section, a first camera lens and a second camera or camera lens; the first camera lens and the second camera or camera lens are respectively positioned at the back and front of the lens portion, an entrance pupil plane and a matching device are positioned between the first camera lens and the second camera or camera lens, the entrance pupil plane and the matching means can be adapted to different focal lengths of the second camera or camera lens, an internal reflection unit is formed between the first camera lens and said entrance pupil plane and matching means, the internal reflection unit is used for decomposing and refracting the captured first image into a plurality of secondary images with different angle shifts, the high-resolution camera part further comprises a second imaging part and a third camera lens, the central axis adjusting device can adjust the double lenses of the first camera lens and the second camera lens or the camera lens and the single lens of the third camera lens, and enables the axes of the double lenses and the single lens to be parallel;
the light field imaging section and the high resolution imaging section are configured such that the third camera lens obtains a second image that coincides with a front view vertical direction among the plurality of secondary images, and outputs the plurality of secondary images and the second image at the same time.
2. The system of claim 1, wherein the light field imaging section is as close as possible to the high resolution imaging section and both are centered on the same vertical plane.
3. The system of claim 1, wherein the plurality of secondary images having different angular offsets have an angular offset range of 10-20 degrees.
4. The system of claim 3, wherein the angular offset of the elevation view in the plurality of secondary images is 0 degrees.
5. The system of claims 1-4,
the first imaging section further includes a first image sensor and a fly-eye lens that captures a first image; the fly-eye lens transmitting the captured first image to the first image sensor; and is
The second imaging section further includes a second image sensor; the second image obtained by the third camera lens is transmitted to the second image sensor.
6. The system of claim 5,
the fly-eye lens is a plurality of micro-lens arrays, and the radius, the thickness and the array spacing of each micro-lens are related to the size of the first image sensor.
7. The system of any of claims 1-4, 6,
the first and second camera lenses have apertures and focal lengths that are adjustable, the second and third camera lenses are interchangeable lenses, and the apertures of the second camera or camera lens are larger than the size of the internal reflection unit.
8. The system according to any of claims 1-4, 6,
the entrance pupil plane and the matching means are pupil lenses having a diameter larger than that of the internal reflection unit and allowing incident light rays of the light field image to be refracted in the internal reflection unit.
9. The system according to any of claims 1-4, 6,
each of the secondary images has a subtle difference in scene, and the size of the internal reflection unit and the focal length of each secondary image are calculated based on the following equations (1) and (2):
wherein the FOV is a field of view of the second video camera or camera lens;
n is a refractive index of the internal reflection unit;
r is the number of internal reflections;
z is the size of the internal reflection unit;
flensis the focal length of the second video camera or camera lens;
fsubis thatFocal length of the secondary image.
10. A processing method for three-dimensional imaging of double cameras comprises the following steps:
obtaining original depth map data of a first image by a light field imaging part;
correcting the original depth map data;
obtaining a high-resolution depth map generated by interpolation by using an edge-oriented or directional rendering method;
and simultaneously obtaining a second image by using a high-resolution image pick-up part, and correcting the original depth map data of the first image by using a data model and combining the second image as reference data until an optimal interpolation-generated high-resolution depth map is obtained.
11. The process of claim 10 wherein,
the light field shooting part comprises a first imaging part, a first camera lens and a second camera or a camera lens; the first camera lens and the second camera or camera lens are respectively positioned at the back and front of the lens portion, an entrance pupil plane and a matching device are positioned between the first camera lens and the second camera or camera lens, the entrance pupil plane and the matching means can be adapted to different focal lengths of the second camera or camera lens, an internal reflection unit is formed between the first camera lens and said entrance pupil plane and matching means, the internal reflection unit is used for decomposing and refracting the captured first image into a plurality of secondary images with different angle shifts, the high-resolution camera part further comprises a second imaging part and a third camera lens, the central axis adjusting device can adjust the double lenses of the first camera lens and the second camera lens or the camera lens and the single lens of the third camera lens, and enables the axes of the double lenses and the single lens to be parallel; the light field imaging section and the high resolution imaging section are configured such that the third camera lens obtains a second image that coincides with a front view vertical direction among the plurality of secondary images, and outputs the plurality of secondary images and the second image at the same time.
12. The processing method of claim 11,
the first image is 9 secondary images or video pictures acquired by the first image sensor;
dividing the 9 secondary images or video pictures and carrying out normalized processing on each secondary image;
removing image noise from each secondary image; decoding light field information acquired by the 9 secondary images by utilizing a synthetic aperture technology, and generating a quasi-focus image by utilizing a digital refocusing technology;
establishing a scene depth map with lower resolution by using the secondary images with 9 focuses; combining the second image, which is a high resolution 2D image or video picture obtained by the second image sensor;
with reference to the second image, increasing the resolution of the depth map using an edge-directed or edge-directed interpolation algorithm; and combining the enhanced resolution depth map and the second image to produce a high resolution 3D image or video picture.
13. The processing method of claim 12, wherein each secondary image is normalized by the following equation:
wherein In(n ═ 1, 2.., 9) represents the image before normalization; i isn(n ═ 1, 2.., 9) represents the normalized image; mirror (I)mLeft, right, up, down) (m ═ 1, 2, 3, 4) represents the image mirror flipping left, right, up, and down; rotate (I)kPi) (k ═ 6.., 9) represents image rotation.
14. The process of claim 12 wherein the digital refocusing is performed on the synthetic aperture image using the following principles:
L′(u,v,x′,y′)=L(u,v,kx′+(1-k)u,ky′+(1-k)v) (6)
I′(x′,y′)=∫∫L(u,v,kx′+(1-k)u,ky′+(1-k)v)dudv (7)
wherein I and I' represent the coordinate systems of the primary and secondary imaging planes;
l and L' represent the energy of the primary and secondary imaging planes.
15. The processing method of claim 12, wherein a Disparity Map (Disparity Map) is calculated using a binocular Stereo Matching algorithm, and then a lower resolution scene depth Map is further created.
16. The processing method of claim 15, wherein the binocular stereo vision Matching algorithm is a Semi-Global Matching (Semi-Global Matching) algorithm by setting the following cost function associated with disparity maps:
wherein D represents a disparity map;
p and q are a certain pixel in the image;
C(p,Dp) The disparity value representing the current pixel is DpA cost value for the pixel;
Nprepresents the pixels adjacent to pixel p, typically 8;
p1 and P2 are penalty coefficients, the disparity value difference of P1 for pixel P and the adjacent pixel is equal to 1, and the disparity value difference of P2 for pixel P and the adjacent pixel is greater than 1;
t [ ] is a function, if the parameter in the function is true, 1 is returned, otherwise 0 is returned;
calculating the minimum cost value of each pixel of the image when the pixel is a certain parallax value, accumulating the cost values of 8 directions, and calculating each pixel one by one to obtain the whole parallax image, wherein the accumulated parallax value with the lowest value is equivalent to the final parallax value of the pixel.
17. The processing method as claimed in claim 12, wherein the conversion of the disparity value to the depth value may use the following formula:
wherein d ispRepresents a depth value of a certain pixel;
f is the normalized focal length;
b is the baseline distance between the two secondary images;
Dprepresenting the disparity value of the current pixel.
18. The processing method of claim 10, wherein the edge-directed interpolation uses the formula:
wherein m and n are low resolution and high resolution image grids before and after interpolation;
y [ n ] represents the depth map generated after interpolation;
s and R represent the data model of the second image 130 and the operator of the edge-directed rendering step, respectively;
λ is the gain of the correction process;
k is the iteration index.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910481518.4A CN112040214A (en) | 2019-06-04 | 2019-06-04 | Double-camera three-dimensional imaging system and processing method |
PCT/CN2020/079099 WO2020244273A1 (en) | 2019-06-04 | 2020-03-13 | Dual camera three-dimensional stereoscopic imaging system and processing method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910481518.4A CN112040214A (en) | 2019-06-04 | 2019-06-04 | Double-camera three-dimensional imaging system and processing method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112040214A true CN112040214A (en) | 2020-12-04 |
Family
ID=73576536
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910481518.4A Pending CN112040214A (en) | 2019-06-04 | 2019-06-04 | Double-camera three-dimensional imaging system and processing method |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112040214A (en) |
WO (1) | WO2020244273A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114404084A (en) * | 2022-01-21 | 2022-04-29 | 北京大学口腔医学院 | Scanning device and scanning method |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102595171A (en) * | 2012-02-03 | 2012-07-18 | 浙江工商大学 | Imaging method and imaging system of dynamic optical fields of multichannel space-time coding apertures |
CN102663712A (en) * | 2012-04-16 | 2012-09-12 | 天津大学 | Depth calculation imaging method based on flight time TOF camera |
CN106780383A (en) * | 2016-12-13 | 2017-05-31 | 长春理工大学 | The depth image enhancement method of TOF camera |
CN107689050A (en) * | 2017-08-15 | 2018-02-13 | 武汉科技大学 | A kind of depth image top sampling method based on Color Image Edge guiding |
CN107991838A (en) * | 2017-11-06 | 2018-05-04 | 万维科研有限公司 | Self-adaptation three-dimensional stereo imaging system |
CN108805921A (en) * | 2018-04-09 | 2018-11-13 | 深圳奥比中光科技有限公司 | Image-taking system and method |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150381965A1 (en) * | 2014-06-27 | 2015-12-31 | Qualcomm Incorporated | Systems and methods for depth map extraction using a hybrid algorithm |
CN106651938B (en) * | 2017-01-17 | 2019-09-17 | 湖南优象科技有限公司 | A kind of depth map Enhancement Method merging high-resolution colour picture |
WO2019127192A1 (en) * | 2017-12-28 | 2019-07-04 | 深圳市大疆创新科技有限公司 | Image processing method and apparatus |
-
2019
- 2019-06-04 CN CN201910481518.4A patent/CN112040214A/en active Pending
-
2020
- 2020-03-13 WO PCT/CN2020/079099 patent/WO2020244273A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102595171A (en) * | 2012-02-03 | 2012-07-18 | 浙江工商大学 | Imaging method and imaging system of dynamic optical fields of multichannel space-time coding apertures |
CN102663712A (en) * | 2012-04-16 | 2012-09-12 | 天津大学 | Depth calculation imaging method based on flight time TOF camera |
CN106780383A (en) * | 2016-12-13 | 2017-05-31 | 长春理工大学 | The depth image enhancement method of TOF camera |
CN107689050A (en) * | 2017-08-15 | 2018-02-13 | 武汉科技大学 | A kind of depth image top sampling method based on Color Image Edge guiding |
CN107991838A (en) * | 2017-11-06 | 2018-05-04 | 万维科研有限公司 | Self-adaptation three-dimensional stereo imaging system |
CN108805921A (en) * | 2018-04-09 | 2018-11-13 | 深圳奥比中光科技有限公司 | Image-taking system and method |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114404084A (en) * | 2022-01-21 | 2022-04-29 | 北京大学口腔医学院 | Scanning device and scanning method |
CN114404084B (en) * | 2022-01-21 | 2024-08-02 | 北京大学口腔医学院 | Scanning device and scanning method |
Also Published As
Publication number | Publication date |
---|---|
WO2020244273A1 (en) | 2020-12-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7043085B2 (en) | Devices and methods for acquiring distance information from a viewpoint | |
CN102164298B (en) | Method for acquiring element image based on stereo matching in panoramic imaging system | |
CN106303228B (en) | A kind of rendering method and system of focus type light-field camera | |
KR101514502B1 (en) | Image-capturing device, image processing apparatus, image processing method, and method of manufacturing image-capturing device | |
WO2019100933A1 (en) | Method, device and system for three-dimensional measurement | |
JP5140210B2 (en) | Imaging apparatus and image processing method | |
US9063345B2 (en) | Super light-field lens with doublet lenslet array element | |
CN111182191B (en) | Wide-field high-resolution camera shooting equipment and method based on aberration compensation calculation | |
JP5204350B2 (en) | Imaging apparatus, playback apparatus, and image processing method | |
US20120275667A1 (en) | Calibration for stereoscopic capture system | |
CN107991838B (en) | Self-adaptive three-dimensional imaging system | |
CN201662682U (en) | Stereo pick-up device | |
JP2010128820A (en) | Apparatus, method and program for processing three-dimensional image, and three-dimensional imaging apparatus | |
US9063323B2 (en) | Super light-field lens and image processing methods | |
WO2013108339A1 (en) | Stereo imaging device | |
WO2012029299A1 (en) | Image capture device, playback device, and image-processing method | |
JP2012178693A (en) | Solid-state imaging apparatus and portable information terminal | |
CN108805921B (en) | Image acquisition system and method | |
KR100915039B1 (en) | Method and Device for Transformation from Multi Focused 2D Image to 3D Image, and Recording Media | |
CN111064945A (en) | Naked eye 3D image acquisition and generation method | |
CN112040214A (en) | Double-camera three-dimensional imaging system and processing method | |
CN101916035A (en) | Stereo pick-up device and method | |
CN111292380B (en) | Image processing method and device | |
KR102031485B1 (en) | Multi-view capturing apparatus and method using single 360-degree camera and planar mirrors | |
JP5088973B2 (en) | Stereo imaging device and imaging method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20201204 |