KR101818839B1 - Apparatus and method of stereo scopic 3d contents creation and display - Google Patents
Apparatus and method of stereo scopic 3d contents creation and display Download PDFInfo
- Publication number
- KR101818839B1 KR101818839B1 KR1020160022853A KR20160022853A KR101818839B1 KR 101818839 B1 KR101818839 B1 KR 101818839B1 KR 1020160022853 A KR1020160022853 A KR 1020160022853A KR 20160022853 A KR20160022853 A KR 20160022853A KR 101818839 B1 KR101818839 B1 KR 101818839B1
- Authority
- KR
- South Korea
- Prior art keywords
- distance
- stereo camera
- virtual reality
- target object
- rendering
- Prior art date
Links
Images
Classifications
-
- H04N13/0022—
-
- H04N13/0018—
-
- H04N13/0033—
-
- H04N13/0221—
Landscapes
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
TECHNICAL FIELD The present invention relates to a technique for relieving simulator fatigue in a stereoscopic three-dimensional display having a fixed screen such as a virtual reality head mounted display, and according to one aspect, The apparatus for preventing dizziness includes a calculation unit for calculating a focal length of a target object from a rendering stereo camera and an object distance from the stereo camera to the target object, and a calculation unit for calculating the calculated focal length and object distance And a processing unit for correcting the matching between the convergence angle and the focal distance between the user's eyes and the object in the 3D content.
Description
The present invention relates to a technique for mitigating user discomfort such as dizziness in a stereo 3D display for application to stereoscopic 3D image content production and display.
Virtual reality refers to a human-computer interface that makes a certain environment or situation computerized and makes the person who uses it behave as if they are interacting with the actual environment and environment. VIRTUAL REALITY is used to show and manipulate people as if they are in the environment without experiencing the environments that are difficult to experience on a daily basis. Applications of virtual reality include education, advanced programming, remote operation, remote satellite surface exploration, exploration data analysis, and scientific visualization.
Virtual reality is available through devices such as head-mounted displays.
The head mount display is a "head screen device" that allows a small display device in front of your eyes to be seen through the lens structure with a clear view of the object.
Fixed screens using head-mounted displays can cause simulator sickness or simulator fatigue, such as dizziness. Users of virtual reality experience dizziness or vomiting symptoms such as motion sickness. This simulator illness or fatigue is caused by a mismatch between what the eye sees and what the body feels or interprets.
The discrepancy of eye-convergence and the accommodation in head-mounted 3D stereoscopic display of the head are important causes of the actual eye view and the virtual view mismatch . In a 3D display using not only a head mount display type but also a stereoscopic image in which mismatch of image information and body information occurs, this simulator fatigue symptom appears.
According to one aspect, the dizziness prevention device at least temporarily implemented by a computer includes an operation unit for calculating a focal length from a stereo camera of a virtual reality to a target object and an object distance from the stereo camera to the target object, And a processor for correcting the perspective and the focus in the content for stereoscopic three-dimensional display by using the calculated focal length and object distance.
The processing unit according to an embodiment blurs the focus of another target object located outside the calculated focal length and adjusts the direction of the stereo camera in a direction corresponding to the position of the target object, And a processor for adjusting an inter-pupillary distance (IPD) of the stereo camera based on the distance.
The operation unit may calculate a first object distance from the stereo camera to the first target object and a second object distance from the stereo camera to the second target object, Pupilary distance (IPD) to be adjusted by reflecting the ratio of the first object distance and the second object distance to the IPD (inter-pupillary distance).
The processing unit according to an embodiment adjusts the direction from the direction in which the stereo camera is heading to the target object in the current content.
The processing unit may calculate an intersection between focal lengths from the stereo camera, calculate a Depth of Field (DoF) for the object of interest based on the calculated intersection, A blurring process is performed on a target object that is spaced apart from the Depth of Field (DoF) by a certain distance.
The processing unit may define a circle of interest (CoI) corresponding to the Depth of Field (DoF) and use the defined circle of interest (CoI) A bounded zone and a parameter value for blurring are defined and blurred according to the calculated Circle of Confusion (CoC).
According to an embodiment, the processing unit may use the calculated Depth of Field (DoF) and the calculated object distance,
Adjusts the direction of the stereo camera toward the focal point for the fixed object in real time.
According to one aspect, an operation method of a dizziness prevention device at least temporarily implemented by a computer includes calculating a focal length of a target object from a stereo camera of a virtual reality and an object distance from the stereo camera to the target object Blurring a focus on another target object located outside the calculated focal distance; adjusting the orientation of the stereo camera in a direction corresponding to the position of the target object; and based on the object distance, And adjusting an inter-pupillary distance (IPD) of the stereo camera.
The calculating step may include calculating a first object distance from the stereo camera to a first target object and a second object distance from the stereo camera to a second target object, The step of adjusting an inter-pupillary distance (IPD) of a stereo camera includes inter-pupillary distance (IPD) of the stereo camera to reflect the ratio of the first object distance and the second object distance to an IPD and calculating a distance.
Adjusting the orientation of the stereo camera in a direction corresponding to the position of the target object according to an embodiment includes adjusting the direction toward the target object from a direction in which the stereo camera is oriented within the current content .
The blurring processing step according to an embodiment may include calculating an intersection between focal lengths from the stereo camera, calculating a depth of field (DoF) for the object of interest based on the calculated intersection, And blurring a target object spaced by a predetermined distance from the calculated Depth of Field (DoF).
The blurring step may include defining a circle of interest (CoI) corresponding to the depth of field (DoF), defining a circle of interest (CoI) Calculating a Circle of Confusion (CoC) using CoI, calculating a bounded zone and a parameter value for blurring according to the calculated Circle of Confusion (CoC) Ring processing.
The step of adjusting the direction of the stereo camera in a direction corresponding to the position of the target object according to an embodiment may further include a step of adjusting the direction of the stereo camera in real time using the calculated depth of field (DoF) And adjusting the orientation of the stereo camera toward the focal point for the fixed object.
1 is a view for explaining an apparatus for preventing dizziness according to an embodiment.
Fig. 2 is a view for explaining an immediate matching (a) of the camera angle corresponding to the version angle of the user's eyes and an additional adjustment (b) of the IPD.
3A is a view for explaining a depth of field (DoF) in which a focal length of a target object is adjusted.
FIG. 3B is a view for explaining a specific example of adjusting the depth of field.
4A is a view for explaining the adjustment of the camera direction (camera orientation).
4B is a view for explaining an embodiment of adjusting the camera direction.
FIG. 4C is a view for explaining the procedure for adjusting the camera direction in detail.
5A is a view for explaining an embodiment for adjusting the length of the IPD.
5B is a view for explaining the change of the target object as the length of the IPD is adjusted in the stereo camera.
6 is a view for explaining an operation method of a dizziness prevention apparatus according to an embodiment.
Hereinafter, embodiments will be described in detail with reference to the accompanying drawings. However, the scope of the rights is not limited or limited by these embodiments. Like reference symbols in the drawings denote like elements.
The terms used in the following description are chosen to be generic and universal in the art to which they are related, but other terms may exist depending on the development and / or change in technology, customs, preferences of the technician, and the like. Accordingly, the terminology used in the following description should not be construed as limiting the technical thought, but should be understood in the exemplary language used to describe the embodiments.
Also, in certain cases, there may be a term chosen arbitrarily by the applicant, in which case the meaning of the detailed description in the corresponding description section. Therefore, the term used in the following description should be understood based on the meaning of the term, not the name of a simple term, and the contents throughout the specification.
1 is a view for explaining an apparatus for preventing dizziness according to an embodiment of the present invention.
2 is a view for explaining an embodiment of the present invention. In FIG. 2 (a), an area of interest, an interesting zone, and an area between these two circles of interest are set according to the position of the target object viewed by the human eye, The eye-convergence of the eye and the eye-accomodation of the eye coincide when the eye of the eye of the user looks at the target object of the region of interest. That is, FIG. 2 (a) shows the rendering angles of the two stereo cameras set by the convergence angle determined by the distance between the eyes and the object and the angle between the two focuses of the two eyes on the left and right eyes of the person An example of displaying a stereo screen is shown. In the display configuration in which the 3D image is reproduced through the stereo screen, the angle of the rendering camera of the left screen and the right screen is changed according to the agreement between the convergence angle of the two eyes and the focus point of the discussion. In FIG. 2 (a), when the distance from the object to which the eye is viewed is changed, the angle of the camera is changed while the IPD is kept constant.
FIG. 2 (b) is similar to the embodiment of FIG. 2 (a) in adjusting the rendering angles of the two stereo cameras, but the distance between the two cameras is finely adjusted to change the focal distance between the eyes We have added a method to maximize the effect that the eye-convergence and the eye-accomodation of the 3D stereo image coincide.
The
The
Specifically, the
The
On the other hand, the
The
First, the
To this end, the
The embodiment of blurring will be described in detail again with reference to FIGS. 3A and 3B.
Next, the
An embodiment of adjusting the direction of the stereo camera in the direction corresponding to the position of the target object will be described in detail again with reference to FIGS. 4A to 4C.
The
For this purpose, the
An embodiment for adjusting the inter-pupillary distance (IPD) of a stereo camera to prevent dizziness will be described in detail again in Figs. 5A and 5B.
Fig. 2 is a view for explaining an immediate matching (a) of the camera angle corresponding to the version angle of the user's eyes and an additional adjustment (b) of the IPD.
FIG. 2 illustrates a method of calculating an intersection point of a predetermined focal length in two cameras and using parameters (focal length, size of numerical aperture, etc.) for Depth of Field (DoF).
That is, the processing unit defines a circle of interest (CoI) corresponding to the depth of field (DoF) of the field, and uses a defined circle of interest (CoI) , And CoC), and define a bounded zone and a parameter value for blurring according to the calculated Circle of Confusion (CoC), thereby blurring processing.
The circle of interest represents a depth of field that is spaced a certain distance in the direction of the stereo camera and the opposite direction with respect to the fixed object, and the target object located within the circle of interest can be focused and displayed clearly. On the other hand, when the target object is positioned on the stereo camera side, the direction of the stereo camera moves in the direction toward the fixed object, thereby preventing dizziness due to mismatching. That is, in the present invention, the previously fixed focal length can be adaptively adjusted.
Meanwhile, in FIG. 2 (b), the IPD of the stereo camera can be adjusted to prevent dizziness due to mismatching. In the embodiment of FIG. 2 (b), the IPD of the stereo camera is largely adjusted in a state in which the target object is close to the stereo camera, thereby preventing dizziness due to mismatching.
3A is a view for explaining a depth of field (DoF) in which a focal length of a target object is adjusted.
According to the
The distracted image is generated by the target object at a distance S r from the stereo camera with respect to the intersection point, and the target object at a position S f near to the stereo camera with respect to the intersection point, which are located at distances d and d r or d f So that it causes confusion on the image due to d. Thus, dizziness may occur due to such confusion. In the present invention, blurring processing is performed on target objects having depth of field different from depth of field corresponding to an intersection, thereby preventing dizziness due to mismatch.
FIG. 3B is a view for explaining a specific example of adjusting the depth of field.
4A is a view for explaining the adjustment of the camera direction (camera orientation).
Referring to embodiment 410 of FIG. 4A, EyeLeft may be interpreted as a left camera in a stereo camera, and EyeRight may be interpreted as a right camera in a stereo camera.
The stereo camera can adjust the direction in the direction toward the first object corresponding to P 1 and in the direction toward the second object corresponding to P 2 . Object 1 corresponding to P 1 is displayed on the IPD axis
, And the object corresponding to P 2 is located in the direction , In other words in In the direction of the arrow.E.g,
, , And Can be calculated by the following equation (1).
[Equation 1]
= tan -1 (LeftEye 1 / d1)
L2 = tan -1 (LeftEye 2 / d2)
Here, d1 is the distance from the midpoint of the stereo camera to the first object, and d2 is the distance from the midpoint of the stereo camera to the second object.
That is, by adjusting the direction of each of the stereo cameras according to the object, dizziness due to mismatch can be prevented.
4B is a view for explaining an embodiment of adjusting the camera direction.
FIG. 4C is a view for explaining the procedure for adjusting the camera direction in detail.
When P 1 moves to P 2
And the change value of the angle in the equation (1) Therefore, it is possible to create a Camera Transformation Matrix (Homogenenous) by using it.That is, since only the direction of the camera is adjusted, the change value for the position is 0 -> x = 0.
Referring to this, the Camera Transformation Matrix can be calculated by Equation (2).
&Quot; (2) "
In the embodiment of FIG. 4C, R cam , which is a Camera Transformation Matrix, is applied to actual rendering, and L2W is calculated locally from the world (model to world) to calculate M (431). At this time, L2W corresponds to the Local to World matrix.
Next, an operation related to the camera can be performed in the world by reflecting C in M (432). At this time, C corresponds to the World to Camera matrix.
Next, the dizziness prevention apparatus according to the present invention may reflect R cam calculated through Equation (2) to M reflecting C (433). Accordingly, the direction of the stereo camera is adjusted, and dizziness due to mismatch can be mitigated.
Next, M may be projected 434 for a projection and projected 436 onto a screen via a
5A is a view for explaining an embodiment for adjusting the length of the IPD.
As shown at 510, IPD 1 for the first target object may be changed to IPD 2 for the second target object.
The IPD may be interpreted as the distance between the centers of the respective lenses of the stereo camera, and may be adjusted in consideration of the distance from the target object in the present invention.
More specifically, IPD 2 can be calculated through the following equation (3).
&Quot; (3) "
Here, Dist eye2obj1 can be interpreted as the shortest distance between the center of the stereo camera and the first object, and Dist eye2obj2 can be interpreted as the shortest distance between the center of the stereo camera and the second object.
That is, the dizziness prevention apparatus calculates the first object distance (Dist eye2obj1 ) from the stereo camera to the first target object and the second object distance (Dist eye2obj2 ) from the stereo camera to the second target object, pupillary distance (IPD) to be adjusted by reflecting the ratio of the first object distance (Dist eye 2 obj 1) and the second object distance (Dist eye 2 obj 2) to the inter-pupillary distance.
5B is a view for explaining the change of the target object as the length of the IPD is adjusted in the stereo camera.
The inter-pupillary distance (IPD) may be interpreted as the distance between the centers of the pupils of the user's eye or the distance between the centers of different lenses of the stereo camera in the virtual reality.
IPD is typically a fixed parameter during 3D rendering because it has a unique IPD value for each user.
With reference to this, the present invention adjusts the position of the stereoscopic rendering camera to mitigate binocular parallax caused by nearby objects between the eyes and the occlusion in the background.
Therefore, the present invention can adjust the IPD to maximize the effect of focus when the distance between the object and the camera is located in the foveated area. To this end, it is assumed in the present invention that the dynamic distance between the left eye and the right eye is the same.
At this time, IPD new for adjustment is calculated by the distance from the target object to the camera. When the distance from the target object to the camera is obtained, the position of the left camera is calculated by Equation (4) below.
&Quot; (4) "
Similarly, the position of the right camera is calculated by the following equation (5).
&Quot; (5) "
If two virtual camera focuses are present on the target object, the camera focus can be rendered. To render the camera focus, the rotational angles of the Cartesian x, y, z axes can be calculated. At this time, the present invention can generate a matrix of rotations for each of the x, y, and z axes as Rot x , Rot y , and Rot z . Once the matrix is calculated, it can be reflected in the final matrix to adjust the camera orientation.
You can apply this program after camera conversion to perform the results.
The confusion caused by surrounding objects can cause the mismatch between the user's eye convergence angle and the two-dimensional camera angle mismatch.
6 is a view for explaining an operation method of a dizziness prevention apparatus according to an embodiment.
The method of operation of the dizziness prevention apparatus according to an exemplary embodiment calculates the focal distance and the object distance (step 601).
Next, an operation method of the dizziness prevention apparatus according to an exemplary embodiment of the present invention is described in detail with reference to the accompanying drawings, in which: FIG. 1 is a block diagram illustrating a configuration of a wearable virtual reality device (VR HMD) Perspective and focus can be corrected.
To this end, the method of operation of the dizziness prevention device adjusts the depth of field (DoF) of the target object (step 602), adjusts the focal length of the target object (step 603), adjusts the IPD of the stereo camera (Step 604).
Recently, the virtual reality related industry has become a hot topic, among which the technology that enables users to concentrate on the contents to concentrate within the content is one of the necessary parts of all game companies, but the focal length, Camera Orientation) and IPD are fixed. Therefore, if the present invention is used, it can be solved in the content, and more efficient solution is possible.
The apparatus described above may be implemented as a hardware component, a software component, and / or a combination of hardware components and software components. For example, the apparatus and components described in the embodiments may be implemented within a computer system, such as, for example, a processor, a controller, an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable array (FPA) A programmable logic unit (PLU), a microprocessor, or any other device capable of executing and responding to instructions. The processing device may execute an operating system (OS) and one or more software applications running on the operating system. The processing device may also access, store, manipulate, process, and generate data in response to execution of the software. For ease of understanding, the processing apparatus may be described as being used singly, but those skilled in the art will recognize that the processing apparatus may have a plurality of processing elements and / As shown in FIG. For example, the processing unit may comprise a plurality of processors or one processor and one controller. Other processing configurations are also possible, such as a parallel processor.
The software may include a computer program, code, instructions, or a combination of one or more of the foregoing, and may be configured to configure the processing device to operate as desired or to process it collectively or collectively Device can be commanded. The software and / or data may be in the form of any type of machine, component, physical device, virtual equipment, computer storage media, or device , Or may be permanently or temporarily embodied in a transmitted signal wave. The software may be distributed over a networked computer system and stored or executed in a distributed manner. The software and data may be stored on one or more computer readable recording media.
The method according to an embodiment may be implemented in the form of a program command that can be executed through various computer means and recorded in a computer-readable medium. The computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination. The program instructions to be recorded on the medium may be those specially designed and configured for the embodiments or may be available to those skilled in the art of computer software. Examples of computer-readable media include magnetic media such as hard disks, floppy disks and magnetic tape; optical media such as CD-ROMs and DVDs; magnetic media such as floppy disks; Magneto-optical media, and hardware devices specifically configured to store and execute program instructions such as ROM, RAM, flash memory, and the like. Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like. The hardware devices described above may be configured to operate as one or more software modules to perform the operations of the embodiments, and vice versa.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. For example, it is to be understood that the techniques described may be performed in a different order than the described methods, and / or that components of the described systems, structures, devices, circuits, Lt; / RTI > or equivalents, even if it is replaced or replaced.
Therefore, other implementations, other embodiments, and equivalents to the claims are also within the scope of the following claims.
Claims (14)
Rendering of a virtual reality for a left screen and a right screen for reproducing a 3D image of a virtual reality A focal length for a target object of a virtual reality from a stereo camera and rendering of the virtual reality Object from a stereo camera to the target object An operation unit for calculating a distance; And
A blurring processing unit for blurring a focus of another target object located outside the calculated focal distance and adjusting a direction of the rendering stereo camera of the virtual reality in a direction corresponding to the position of the target object,
Lt; / RTI >
The operation unit,
Calculating a first object distance from the rendering stereo camera of the virtual reality to the first target object of the virtual reality and a second object distance from the rendering stereo camera of the virtual reality to the second target object of the virtual reality,
Wherein,
Calculating an inter-pupillary distance (IPD) to be adjusted by reflecting a ratio of the first object distance to the second object distance to an inter-pupillary distance (IPD) of a rendering stereo camera of the virtual reality,
The inter-pupillary distance (IPD) of the rendering stereo camera of the virtual reality is adjusted according to the calculated inter-pupillary distance (IPD) to determine the perspective and focus of the view of the virtual reality in the content for stereoscopic three- A device to prevent dizziness that reduces mismatch with the view of the actual eye.
Wherein,
And adjusts in the direction from the direction in which the rendering stereo camera is directed within the current content to the target object.
Wherein,
Calculating an intersection between focal lengths from the rendering stereo camera, calculating a Depth of Field (DoF) for the object of interest based on the calculated intersection,
And blurring a target object spaced by a predetermined distance from the calculated Depth of Field (DoF).
Wherein,
A circle of interest (CoI) corresponding to the depth of field (DoF) is defined and a Circle of Confusion (CoC) is defined using the defined circle of interest (CoI) And defines a bounded zone and a parameter value for blurring according to the calculated Circle of Confusion (CoC), and performs blurring processing.
Wherein,
Using the calculated Depth of Field (DoF) and the calculated object distance,
A dizziness prevention device that adjusts the orientation of a rendered stereo camera towards a focal point for a fixed object in real time.
Rendering of a virtual reality for a left screen and a right screen for reproducing a 3D image of a virtual reality A focal length for a target object of a virtual reality from a stereo camera and rendering of the virtual reality Object from a stereo camera to the target object Calculating a distance;
Blurring a focus of another target object located outside the calculated focal distance;
Adjusting a direction of the rendering stereo camera of the virtual reality in a direction corresponding to the position of the target object;
Calculating a first object distance from the rendering stereo camera of the virtual reality to the first target object of the virtual reality and a second object distance from the rendering stereo camera of the virtual reality to the second target object of the virtual reality;
Calculating an inter-pupillary distance (IPD) to be adjusted by reflecting a ratio of the first object distance to the second object distance to an inter-pupillary distance (IPD) of a rendering stereo camera of the virtual reality;
The inter-pupillary distance (IPD) of the rendering stereo camera of the virtual reality is adjusted according to the calculated inter-pupillary distance (IPD) to determine the perspective and focus of the view of the virtual reality in the content for stereoscopic three- Correcting a mismatch with a view of the actual eye
The method comprising the steps of:
Wherein the step of adjusting the orientation of the rendering stereo camera in a direction corresponding to the position of the target object comprises:
Adjusting in a direction from the direction of the rendering stereo camera toward the target object within the current content
The method comprising the steps of:
The blurring step includes:
Calculating an intersection between the focal lengths from the rendering stereo camera;
Calculating a depth of field (DoF) for the object of interest based on the calculated intersection; And
Blurring a target object spaced by a predetermined distance from the calculated Depth of Field (DoF)
The method comprising the steps of:
The blurring step includes:
Defining a circle of interest (CoI) corresponding to the depth of field (DoF);
Calculating a Circle of Confusion (CoC) using the defined circle of interest (CoI);
Defining a bounded zone and parameter values for blurring according to the calculated Circle of Confusion (CoC) and blurring
The method comprising the steps of:
Wherein the step of adjusting the orientation of the rendering stereo camera in a direction corresponding to the position of the target object comprises:
Adjusting a direction of a rendering stereo camera toward a focal point of an object fixed in real time using the calculated Depth of Field (DoF) and the calculated object distance,
The method comprising the steps of:
Wherein,
Pupilary distance (IPD) to be adjusted by multiplying a value obtained by dividing the second object distance by the first object distance by an inter-pupillary distance (IPD) of the rendering stereo camera.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160022853A KR101818839B1 (en) | 2016-02-25 | 2016-02-25 | Apparatus and method of stereo scopic 3d contents creation and display |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160022853A KR101818839B1 (en) | 2016-02-25 | 2016-02-25 | Apparatus and method of stereo scopic 3d contents creation and display |
Publications (2)
Publication Number | Publication Date |
---|---|
KR20170100371A KR20170100371A (en) | 2017-09-04 |
KR101818839B1 true KR101818839B1 (en) | 2018-01-15 |
Family
ID=59924335
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020160022853A KR101818839B1 (en) | 2016-02-25 | 2016-02-25 | Apparatus and method of stereo scopic 3d contents creation and display |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR101818839B1 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11877064B1 (en) * | 2019-08-20 | 2024-01-16 | Opic Technologies, Inc. | Method and apparatus for a stereoscopic smart phone |
US11927756B2 (en) | 2021-04-01 | 2024-03-12 | Samsung Electronics Co., Ltd. | Method for providing augmented reality image and head mounted display device supporting the same |
KR20220136776A (en) * | 2021-04-01 | 2022-10-11 | 삼성전자주식회사 | Method for providing augmented reality image and head mounted display device supporting the same |
KR102559913B1 (en) * | 2022-10-20 | 2023-07-26 | 주식회사 비브스튜디오스 | Method for implementing camera movement by using a virtual camera |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100688456B1 (en) * | 2005-09-13 | 2007-03-02 | 한국방송공사 | Method for controlling zoom-convergence interlock in parallel translation axis typed stereo camera system |
-
2016
- 2016-02-25 KR KR1020160022853A patent/KR101818839B1/en active IP Right Grant
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100688456B1 (en) * | 2005-09-13 | 2007-03-02 | 한국방송공사 | Method for controlling zoom-convergence interlock in parallel translation axis typed stereo camera system |
Also Published As
Publication number | Publication date |
---|---|
KR20170100371A (en) | 2017-09-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7094266B2 (en) | Single-depth tracking-accommodation-binocular accommodation solution | |
US11132056B2 (en) | Predictive eye tracking systems and methods for foveated rendering for electronic displays | |
US10241329B2 (en) | Varifocal aberration compensation for near-eye displays | |
KR101788452B1 (en) | Apparatus and method for replaying contents using eye tracking of users | |
US20160267720A1 (en) | Pleasant and Realistic Virtual/Augmented/Mixed Reality Experience | |
US20150187115A1 (en) | Dynamically adjustable 3d goggles | |
KR101818839B1 (en) | Apparatus and method of stereo scopic 3d contents creation and display | |
EP3001681B1 (en) | Device, method and computer program for 3d rendering | |
CN106293561B (en) | Display control method and device and display equipment | |
JP2022183177A (en) | Head-mounted display device | |
JP6576639B2 (en) | Electronic glasses and control method of electronic glasses | |
US11212502B2 (en) | Method of modifying an image on a computational device | |
US20220365594A1 (en) | Control method and apparatus for virtual reality device | |
JP2023515205A (en) | Display method, device, terminal device and computer program | |
US11543655B1 (en) | Rendering for multi-focus display systems | |
KR101873161B1 (en) | Method and apparatus for providing personal 3-dimensional image using convergence matching algorithm | |
CN111279293A (en) | Method for modifying image on computing device | |
US11237413B1 (en) | Multi-focal display based on polarization switches and geometric phase lenses | |
CN110197524B (en) | Stereoscopic display method, apparatus, device, and computer-readable storage medium | |
US10326976B2 (en) | Method and apparatus for providing personal 3-dimensional image using convergence matching algorithm | |
Gurrieri | Improvements in the visualization of stereoscopic 3D imagery | |
Watt et al. | 3D media and the human visual system | |
CN116466903A (en) | Image display method, device, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
A201 | Request for examination | ||
E902 | Notification of reason for refusal | ||
AMND | Amendment | ||
E601 | Decision to refuse application | ||
AMND | Amendment | ||
X701 | Decision to grant (after re-examination) | ||
GRNT | Written decision to grant |