WO2018161883A1 - 虚拟光线跟踪方法及光场动态重聚焦显示系统 - Google Patents
虚拟光线跟踪方法及光场动态重聚焦显示系统 Download PDFInfo
- Publication number
- WO2018161883A1 WO2018161883A1 PCT/CN2018/078098 CN2018078098W WO2018161883A1 WO 2018161883 A1 WO2018161883 A1 WO 2018161883A1 CN 2018078098 W CN2018078098 W CN 2018078098W WO 2018161883 A1 WO2018161883 A1 WO 2018161883A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- virtual reality
- display
- light
- processor
- refocusing
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/34—Stereoscopes providing a stereoscopic pair of separated images corresponding to parallactically displaced views of the same object, e.g. 3D slide viewers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/06—Ray-tracing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10052—Images from lightfield camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
Definitions
- the invention relates to the technical field of image rendering, in particular to a virtual ray tracing method and a light field dynamic refocusing display system.
- the ray tracing algorithm is a realistic image generation method proposed in 1980. It can generate high-quality realistic graphics. It is currently one of the commonly used algorithms for 3D realistic graphics.
- the ray tracing algorithm is actually the approximate inverse of the physical process of light illumination. The process, which tracks specularly reflected light between objects and regular transmission, simulates the propagation of light from an ideal surface.
- Figure 1a shows a ray tracing method orifice image and depth map, assuming the scene depth range [z 0, z 1] ( z 1> z 0> 0), wherein z 1 can be infinity, the photosensitive sheet to The distance between the pinholes is s.
- a virtual thin lens is placed at the position of the pinhole, and the virtual thin lens focal length is f, f is satisfied. That is to say, the farthest plane z 1 is focused on the photosensitive sheet. Further, it is assumed that the diameter of the lens is R. In such a virtual imaging system, the point in the plane z ⁇ [z 0 , z 1 ] will be a blurred image on the photosensitive sheet, and the size of the circle of confusion is:
- the point on the plane z 0 has the largest radius of the circle on the photosensitive sheet.
- the radius of the circle of dispersion also depends on the circular cone of the light cone (large cone in Figure 1b) and the z 0 plane.
- the intersection radius can be or
- we can dynamically adjust the position of the photosensitive s f to render the image focused on the plane z f , where z f satisfies 1/s_f+1/z f 1/f.
- the pixel index is used to calculate the light cone index of the central ray. This ensures that the rendered image has the same resolution as the original image and that the same pixel position corresponds to the same position in the space.
- the final color of each pixel p is obtained by integrating all the rays in the cone.
- each ray needs to be tracked into the scene.
- the ray tracing algorithm in the image space tracks a ray into the scene and backtracks to the original image based on the three-dimensional coordinates in the scene.
- the pixel color can be directly used as the color of the back-tracked light.
- the existing virtual reality device can only display the all-focus image or the all-focus virtual reality scene, only the direction of the head of the experiencer, and cannot provide the principle of human eye imaging.
- the content and the content of the area of interest of the human eye so it is impossible to display the content and achieve refocusing in full accordance with the direction of the human eye.
- the ciliary muscles of the human eye 310 can relax and contract, so that the human eye 310 can be focused on different objects at different distances, and the human eye 310 can selectively focus on a spherical surface of different radii, focusing at a distance (i.e., radius).
- the existing virtual reality device lacks an effective data processing method, so that the human eye 310 can only see the image of the full focus or the fixed focal plane. See the fixed focus image shown in FIG. 12, and cannot refocus, this Not meeting the visual characteristics of the human eye 310 viewing the three-dimensional world reduces the visual experience of the human eye 310 viewing virtual reality content. Therefore, in addition to the need to simplify the computational complexity, there is an urgent need to invent a virtual reality display system that can dynamically focus on the direction of the human eye to see multiple fixed focus planes.
- the present invention aims to solve the existing problems, and aims to provide a virtual ray tracing method and a light field dynamic refocusing display system.
- Light information collection that is, random retrieval of pixels on a full-focus image to reduce the number of light samples; and/or simplify the intersection search by assuming that the surface of the object is smooth;
- Color fusion is performed based on the above ray information to obtain a final display result.
- the light collection is a vibration sampling, that is, the aperture is divided into a plurality of equally divided regions, and then one pixel is collected in each of the equally divided regions.
- intersection search is based on Calculate l pq directly with a fixed d q , sample the light in radius r, and initialize one As Pixel offset; if The displacement of the pixel is equal to Then you can determine the location of the intersection; otherwise set the The displacement of the pixel is Continue to iterate until the results are met Stops and returns the color of the pixel at the end of the iteration.
- the technical solution adopted by the present invention further provides a virtual reality light field dynamic refocusing display system, comprising a virtual reality display content processor, a virtual reality display, an eyeball tracker capable of acquiring an eyeball rotation direction, an eyeball tracking information processor, and having light a field dynamic refocusing processing function of the light field refocusing processor, wherein the virtual reality display content processor acquires the display content, and sequentially passes the data with the virtual reality display, the eyeball tracker, the eyeball tracking information processor, and the light field refocusing processor Wire connection.
- the virtual reality display comprises an imaging display screen corresponding to the left and right eyes, an optical lens with an amplification function, and an equal positioning sensor for displaying a binocular stereo image processed by the virtual reality display content processor.
- a micro LED lamp is disposed on an edge of the optical lens, and a camera that collects an image of the human eye is disposed above the optical lens.
- the present invention reduces the sampling of the ray information and simplifies the intersection retrieval process by random retrieval, reduces the computational complexity, and can replace the erroneous sampling by itself, thereby obtaining true and accurate refocusing; virtual reality light field dynamics
- the refocus display system can dynamically focus according to the direction of the human eye, allowing the user to focus on different objects at different distances, in line with the human eye's observation characteristics.
- Figure 1a is a schematic view of the principle of small hole imaging
- Figure 1b is a schematic view of an imaging method incorporating a virtual thin lens
- Figure 1c is a schematic diagram of image space light image tracking
- Figure 1d is a schematic view of the method of the present invention.
- Figure 2 is a schematic diagram of light shock sampling
- Figure 3a, Figure 3b are schematic diagrams of erroneous sampling analysis of light
- Figure 4a is a gamma2.2 curve and a gradient map
- Figure 4b is a schematic diagram of tone mapping corrected image brightness
- FIG. 5 is a schematic diagram of effects of different color fusion methods according to an embodiment of the present invention.
- Figure 6 is a schematic diagram of rendering effects
- Figure 7 is a comparison diagram of the depth of field effect diagram of the SLR camera and the Lytro camera
- FIG. 8 is a diagram showing a special focus effect rendered by an embodiment of the present invention.
- FIG. 9 is a schematic structural diagram of a virtual reality light field dynamic refocusing display system
- FIG. 10 is a schematic structural view of a virtual reality helmet
- FIG. 11 is a schematic structural diagram of a virtual reality display
- Figure 12 is a schematic view of a single fixed focus plane
- Figure 13 is a schematic view of a multi-fixed focus surface
- a virtual reality display content processor 100 a virtual reality display 200, an eye tracker 110, an eye direction information processor 120, a light field refocusing processor 130, a camera 1101, a micro LED light 1102, an optical lens 2201, a person The eye 310, the imaging display screen 2101, the human eye area image 2202, the left eye image 5101, and the right eye image 5102.
- This embodiment is different from the traditional ray tracing algorithm in that only one full-focus photo and its corresponding depth map can complete the same effect, which is greatly simplified in calculation; the following will show a random retrieval algorithm that displays statistical prior probability. Finally, a complete description of the virtual ray tracing method that is comparable to the SLR camera can be generated.
- the human eye 310 In the real world, when the human eye 310 is viewing the three-dimensional world, the upper and lower rectus muscles, the inner and outer rectus muscles, and the superior and inferior oblique muscles can control the rotation of the eyeball, thereby enabling the human eye 310 to see scenes in different directions or aims.
- the direction of the human head 310 can only be determined by the built-in or external sensor to determine the direction of the human eye 310, but due to the flexibility of the human eye 310.
- the gaze direction of the human eye 310 is not completely consistent with the direction of the head, which results in the inability to accurately obtain the information of the direction in which the human eye is gazing, and thus the correct picture cannot be displayed.
- the virtual reality light field dynamic refocus display system used in the implementation of the present invention mainly includes five modules, including a virtual reality display content processor 100 and a virtual reality display 200 (virtual reality display 200).
- a realistic helmet such as a VR head display such as HTC VIVE, an eye tracker 110 capable of acquiring the direction of rotation of the eyeball, an eyeball tracking information processor 120, and a light field refocusing processor 130 having a light field dynamic refocusing processing function.
- the virtual reality display content processor 100 mainly refers to projecting and transforming scenes in the display content according to the camera imaging principle in computer vision to obtain two different images with respect to the left and right eyes, and the images have parallax between them, thereby displaying Save the scene stereo information in the image.
- the virtual reality display 200 mainly includes an imaging display screen 2101 corresponding to the left and right eyes of the human eye, an optical lens 2201 having an amplification function, and a sensor function of a gyroscope, etc., for displaying a virtual reality display.
- the content processor processes the resulting binocular stereo image.
- a micro LED lamp 1102 is surrounded on the edge of the optical lens 2201, and a camera 1101 for capturing an image of the human eye is disposed above the optical lens 1102.
- the eye tracker 110 for detecting the direction of rotation of the eyeball mainly monitors the rotation of the human eye 310 by the camera 1101 in the virtual reality device, and collects the image of the eye in real time and transmits it to the eye direction information processor 120.
- the eye direction information processor 120 mainly acquires the rotation direction information of the eyeball by the eye movement tracking technology according to the human eye area image 2202, and records it as O eye , that is, the direction that the current experience person is interested in.
- the intersection position P of the human eye line of sight and the display content (ie, the displayed image) of the display system can be determined according to the direction information; and further, the position of the P can be known by the position of the P.
- the depth information which is the focus plane position. Then use the virtual ray tracing algorithm to calculate the refocusing imaging results of the human eye's region of interest or the plane of interest, and perform blurring in the non-interesting region to achieve the human eye visual effect in the real world - this part of the work is mainly in the final light.
- the field refocusing processor 130 is completed.
- the display content After acquiring the display content processed by the light field refocusing processor 130, the display content is transmitted to the virtual reality display content processor 100 for processing, and the binocular stereo image of the scene is acquired, thereby realizing the final virtual reality light field dynamic weight. Focus display.
- the device part of the embodiment can detect the rotation direction information of the human eye 310, so that the light field dynamic refocusing process can be performed on the display content according to the attention area of the human eye 310, so that the displayed content can be adapted to the viewing principle of the human eye, so that the displayed content can be adapted to the viewing principle of the human eye.
- Experiencers can get a better experience when they experience it.
- the embodiment is based on the above device, and the method part mainly comprises the following steps:
- Light information collection that is, random retrieval of pixels on a full-focus image to reduce the number of light samples; and/or simplify the intersection search by assuming that the surface of the object is smooth;
- Color fusion is performed based on the above ray information to obtain a final display result.
- the intersection point may be discretely determined along the ray direction, in order to determine whether each discrete depth z of the ray direction corresponds to a point in the space, and if there is a point, the point may be backprojected into the image, and Make sure the pixel depth is z.
- the light is "stopped" from the center r of the thin lens.
- the distance from the central axis of the Q lens is Therefore, suppose the center pixel of the image is p, and the projection point of Q in the image is the distance of q, p, q.
- the depth range of the scene is [z 0 , z 1 ].
- the ray segments of different depths are projected onto the line [q 0 , q 1 ], q 0 , q 1 are respectively Q 0 , Q 1 in the image
- the projected points on the depth are z 0 , z 1 , respectively.
- the size of z q can be extracted from the depth map.
- the left and right sides of (4) can be calculated according to the depth information, and the ray tracing of the image space avoids the reconstruction of the scene 3D model.
- This embodiment assumes that the color image and the depth image are discrete to the two-dimensional plane space, and therefore, even if the constraint of the formula (3) is satisfied, it is impossible to ensure that the light intersects with the object in the space. To alleviate this problem, this embodiment relaxes the constraints to:
- the input of the present invention has only a single color picture, the content (color and geometry) of the scene occlusion area cannot be detected.
- the solution of this embodiment is to assume that the surface (geometry and texture) is symmetrical about the center of the film lens, and that l pq is occluded, and this embodiment uses a pixel of -l pq .
- the ray tracing of the picture field of the present invention is similar to the ray tracing of a conventional geometric space, and its computational complexity is high, especially in refocusing applications.
- this embodiment adopts a method of vibration sampling:
- the first step is to divide the aperture into several grids
- the size of the mesh is the original aperture size. Times, the computational complexity is reduced to
- each grid should not be characterized by only one light.
- random sampling does not guarantee the quality of the sampled light.
- the present invention also proposes a process of simplifying the intersection search, that is, reducing K, to accelerate the intersection search by assuming that the surface of the space object is smooth.
- q ⁇ [q 0 , q 1 ] does not detect whether q satisfies the constraint (4) pixel by pixel, but directly calculates l pq according to the formula (4) and the fixed d q . See Figure 1d to initialize a sample of the light within the radius r. (t is the number of iterations) as The offset of the pixel.
- the present invention can also be accelerated by a random retrieval method.
- the most flammable random retrieval of a light does not guarantee a complete correct intersection every time, but the probability of error is very low- - Because the pixel that returned the error may be the correct sample of another ray, that is, a erroneous sample does not need to be treated as an outlier: it is still very likely to be an intersection, still valid for the final refocus rendering.
- the present embodiment distinguishes between correct and erroneous samples by different color shading.
- the wrong sampling point may be outside the fuzzy convolution kernel, and the error sampling is very small compared to the total ray sampling.
- the performance of the g p (I) response is first analyzed with the gamma 2.2 function curve.
- the gradient of the curve in the high luminance region is steeper than the low luminance region, indicating that the highlight portion is more compressed.
- the relative ratio is directly proportional to the intensity.
- the gamma 2.2 curve gives the highlighted portion a higher weight when blending highlights and highlights.
- the values of ⁇ and ⁇ can control color performance.
- Figure 5 shows the different color fusion methods.
- the linear fusion is directly averaging the integration domain.
- the gamma 2.2 can make the color blue.
- the Gaussian weight color fusion is preferred, and the fusion effect is relatively real.
- the embodiment can generate a very high depth of field effect; in particular, the effect is the same as that of the SLR and the Lytro camera, but this embodiment only needs a normal mobile camera to complete the above effect (even if It is the gray scale effect in the drawing, and can also show the comparison of the technical advantages of the present invention; the same applies hereinafter.
- the present invention is further applicable to Tilt-shift photography, which is a technique that requires special focus in a small scene. As can be seen from the comparison of the four groups in Fig. 8, it is not necessary to use a special lens to obtain the Tilt-shift effect as in the prior art. Only the user needs to modify the depth map simply, and the present invention can simulate the Tilt-shift effect very simply. The process is similar to the above, so it will not be described.
- the present embodiment reduces the sampling of the ray information and simplifies the intersection retrieval process by random retrieval, reduces the computational complexity, and can replace the erroneous sampling by itself to obtain true and accurate refocusing; the virtual reality light field dynamic refocusing display
- the system can dynamically focus according to the direction of the human eye, allowing the user to focus on different objects in the distance, in line with the human eye's observation characteristics.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Optics & Photonics (AREA)
- Computer Hardware Design (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- Processing Or Creating Images (AREA)
- Image Processing (AREA)
- Image Generation (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Claims (6)
- 一种虚拟光线跟踪方法,其特征在于包含如下步骤:获取全焦图片及其深度信息;光线信息采集,即随机检索全焦图片上的像素点,以此来减少光线采样的数量;并/或假设物体表面为光滑而简化交点检索;基于上述光线信息进行颜色融合,得到最终显示结果。
- 根据权利要求1所述的一种虚拟光线跟踪方法,其特征在于:光线采集为震动采样,即将光圈分为若干等分区域,随后在所述每个等分区域内采集一个像素点。
- 一种基于权利要求1、2或3所述内容的虚拟现实光场动态重聚焦显示系统,其特征在于:包括虚拟现实显示内容处理器、虚拟现实显示器、能够获取眼球转动方向的眼球跟踪器、眼球追踪信息处理器以及具有光场动态重聚焦处理功能的光场重聚焦处理器,其中,虚拟现实显示内容处理器获取显示内容,并且和虚拟现实显示器、眼球跟踪器、眼球追踪信息处理器和光场重聚焦处理器依次通过数据线连接。
- 一种根据权利要求4所述的虚拟现实光场动态重聚焦显示系统,其特征在于:虚拟现实显示器包括对应于左、右眼的成像显示屏、具有放大功能的光学镜片以及等定位传感器组成,用于显示虚拟现实显示内容处理器处理得到的双目立体图像。
- 一种根据权利要求4所述的虚拟现实光场动态重聚焦显示系统,其特征在于:所述虚拟现实显示器中,光学镜片的边缘上设有微型LED灯,在光学镜片的上方设有采集人眼图像的摄像头。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2019544763A JP6862569B2 (ja) | 2017-03-09 | 2018-03-06 | 仮想光線追跡方法および光照射野の動的リフォーカス表示システム |
US16/346,942 US10852821B2 (en) | 2017-03-09 | 2018-03-06 | Virtual ray tracing method and dynamic light field refocusing display system |
KR1020197009180A KR102219624B1 (ko) | 2017-03-09 | 2018-03-06 | 가상 광선 추적 방법 및 라이트 필드 동적 리포커싱 디스플레이 시스템 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710137526.8A CN107452031B (zh) | 2017-03-09 | 2017-03-09 | 虚拟光线跟踪方法及光场动态重聚焦显示系统 |
CN201710137526.8 | 2017-03-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018161883A1 true WO2018161883A1 (zh) | 2018-09-13 |
Family
ID=60486245
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/078098 WO2018161883A1 (zh) | 2017-03-09 | 2018-03-06 | 虚拟光线跟踪方法及光场动态重聚焦显示系统 |
Country Status (5)
Country | Link |
---|---|
US (1) | US10852821B2 (zh) |
JP (1) | JP6862569B2 (zh) |
KR (1) | KR102219624B1 (zh) |
CN (1) | CN107452031B (zh) |
WO (1) | WO2018161883A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113256783A (zh) * | 2021-03-29 | 2021-08-13 | 北京航空航天大学 | 基于眼动跟踪的立即辐射度渲染方法 |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107452031B (zh) * | 2017-03-09 | 2020-06-26 | 叠境数字科技(上海)有限公司 | 虚拟光线跟踪方法及光场动态重聚焦显示系统 |
CN109242900B (zh) * | 2018-08-01 | 2021-09-21 | 曜科智能科技(上海)有限公司 | 焦平面定位方法、处理装置、焦平面定位系统及存储介质 |
CN109360235B (zh) * | 2018-09-29 | 2022-07-19 | 中国航空工业集团公司上海航空测控技术研究所 | 一种基于光场数据的混合深度估计方法 |
CN110365965B (zh) * | 2019-06-26 | 2020-11-06 | 北京邮电大学 | 三维光场图像生成方法及装置 |
CN110689498B (zh) * | 2019-09-27 | 2024-03-12 | 西北大学 | 一种基于对非关注点部分分级模糊的高清视频优化方法 |
CN111650754B (zh) * | 2020-07-17 | 2022-08-12 | 北京耐德佳显示技术有限公司 | 一种平视显示设备 |
KR20220067950A (ko) | 2020-11-18 | 2022-05-25 | 삼성전자주식회사 | 디스플레이 장치 및 그의 제어 방법 |
CN112967370B (zh) * | 2021-03-03 | 2022-06-21 | 北京邮电大学 | 三维光场重建方法、装置及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110316968A1 (en) * | 2010-06-29 | 2011-12-29 | Yuichi Taguchi | Digital Refocusing for Wide-Angle Images Using Axial-Cone Cameras |
CN105144245A (zh) * | 2013-04-24 | 2015-12-09 | 高通股份有限公司 | 用于扩增现实的辐射转移取样的设备和方法 |
CN105812778A (zh) * | 2015-01-21 | 2016-07-27 | 成都理想境界科技有限公司 | 双目ar头戴显示设备及其信息显示方法 |
CN106408643A (zh) * | 2016-08-31 | 2017-02-15 | 上海交通大学 | 一种基于图像空间的图像景深模拟方法 |
CN107452031A (zh) * | 2017-03-09 | 2017-12-08 | 叠境数字科技(上海)有限公司 | 虚拟光线跟踪方法及光场动态重聚焦显示系统 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020013573A1 (en) * | 1995-10-27 | 2002-01-31 | William B. Telfair | Apparatus and method for tracking and compensating for eye movements |
US6028606A (en) * | 1996-08-02 | 2000-02-22 | The Board Of Trustees Of The Leland Stanford Junior University | Camera simulation system |
JP3629243B2 (ja) * | 2002-02-13 | 2005-03-16 | Necマイクロシステム株式会社 | モデリング時の距離成分を用いてレンダリング陰影処理を行う画像処理装置とその方法 |
US20100110069A1 (en) * | 2008-10-31 | 2010-05-06 | Sharp Laboratories Of America, Inc. | System for rendering virtual see-through scenes |
US8493383B1 (en) * | 2009-12-10 | 2013-07-23 | Pixar | Adaptive depth of field sampling |
JP6141584B2 (ja) * | 2012-01-24 | 2017-06-07 | アリゾナ ボード オブ リージェンツ オン ビハーフ オブ ザ ユニバーシティ オブ アリゾナ | 小型視線追従型ヘッドマウントディスプレイ |
US9171394B2 (en) * | 2012-07-19 | 2015-10-27 | Nvidia Corporation | Light transport consistent scene simplification within graphics display system |
CA2931776A1 (en) * | 2013-11-27 | 2015-06-04 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
US9905041B2 (en) * | 2014-11-24 | 2018-02-27 | Adobe Systems Incorporated | Depth of field synthesis using ray tracing approximation |
-
2017
- 2017-03-09 CN CN201710137526.8A patent/CN107452031B/zh active Active
-
2018
- 2018-03-06 JP JP2019544763A patent/JP6862569B2/ja active Active
- 2018-03-06 KR KR1020197009180A patent/KR102219624B1/ko active IP Right Grant
- 2018-03-06 WO PCT/CN2018/078098 patent/WO2018161883A1/zh active Application Filing
- 2018-03-06 US US16/346,942 patent/US10852821B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110316968A1 (en) * | 2010-06-29 | 2011-12-29 | Yuichi Taguchi | Digital Refocusing for Wide-Angle Images Using Axial-Cone Cameras |
CN105144245A (zh) * | 2013-04-24 | 2015-12-09 | 高通股份有限公司 | 用于扩增现实的辐射转移取样的设备和方法 |
CN105812778A (zh) * | 2015-01-21 | 2016-07-27 | 成都理想境界科技有限公司 | 双目ar头戴显示设备及其信息显示方法 |
CN106408643A (zh) * | 2016-08-31 | 2017-02-15 | 上海交通大学 | 一种基于图像空间的图像景深模拟方法 |
CN107452031A (zh) * | 2017-03-09 | 2017-12-08 | 叠境数字科技(上海)有限公司 | 虚拟光线跟踪方法及光场动态重聚焦显示系统 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113256783A (zh) * | 2021-03-29 | 2021-08-13 | 北京航空航天大学 | 基于眼动跟踪的立即辐射度渲染方法 |
Also Published As
Publication number | Publication date |
---|---|
US10852821B2 (en) | 2020-12-01 |
KR20190043590A (ko) | 2019-04-26 |
KR102219624B1 (ko) | 2021-02-23 |
JP6862569B2 (ja) | 2021-04-21 |
JP2020504984A (ja) | 2020-02-13 |
US20200057492A1 (en) | 2020-02-20 |
CN107452031B (zh) | 2020-06-26 |
CN107452031A (zh) | 2017-12-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2018161883A1 (zh) | 虚拟光线跟踪方法及光场动态重聚焦显示系统 | |
JP6246757B2 (ja) | 現実環境の視野におけるバーチャルオブジェクトを表現方法及びシステム | |
EP3101624B1 (en) | Image processing method and image processing device | |
US6618054B2 (en) | Dynamic depth-of-field emulation based on eye-tracking | |
WO2018019282A1 (zh) | 双目全景图像获取方法,装置及存储介质 | |
US20130335535A1 (en) | Digital 3d camera using periodic illumination | |
Fuhl et al. | Fast camera focus estimation for gaze-based focus control | |
WO2021197370A1 (zh) | 一种光场显示方法及系统、存储介质和显示面板 | |
JP7479729B2 (ja) | 三次元表現方法及び表現装置 | |
CN109064533B (zh) | 一种3d漫游方法及系统 | |
CN113724391A (zh) | 三维模型构建方法、装置、电子设备和计算机可读介质 | |
Zhai et al. | Image real-time augmented reality technology based on spatial color and depth consistency | |
US20230334806A1 (en) | Scaling neural representations for multi-view reconstruction of scenes | |
JP3387856B2 (ja) | 画像処理方法、画像処理装置および記憶媒体 | |
CN114157853A (zh) | 一种用于生成像素光束的数据表示的装置和方法 | |
JP3392078B2 (ja) | 画像処理方法、画像処理装置および記憶媒体 | |
CN113821107B (zh) | 一种实时、自由视点的室内外裸眼3d系统 | |
TWI831583B (zh) | 虛擬實境設備調節方法、裝置、電子設備及儲存介質 | |
Zhang et al. | [Poster] an accurate calibration method for optical see-through head-mounted displays based on actual eye-observation model | |
Du et al. | Location Estimation from an Indoor Selfie | |
Corke et al. | Image Formation | |
WO2023200936A1 (en) | Scaling neural representations for multi-view reconstruction of scenes | |
Corke et al. | Image Formation | |
Zhang et al. | Image Acquisition Modes | |
Zhang et al. | Single-Image 3-D Scene Reconstruction |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18763552 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20197009180 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2019544763 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18763552 Country of ref document: EP Kind code of ref document: A1 |