KR101818839B1 - Apparatus and method of stereo scopic 3d contents creation and display - Google Patents

Apparatus and method of stereo scopic 3d contents creation and display Download PDF

Info

Publication number
KR101818839B1
KR101818839B1 KR1020160022853A KR20160022853A KR101818839B1 KR 101818839 B1 KR101818839 B1 KR 101818839B1 KR 1020160022853 A KR1020160022853 A KR 1020160022853A KR 20160022853 A KR20160022853 A KR 20160022853A KR 101818839 B1 KR101818839 B1 KR 101818839B1
Authority
KR
South Korea
Prior art keywords
distance
stereo camera
virtual reality
target object
rendering
Prior art date
Application number
KR1020160022853A
Other languages
Korean (ko)
Other versions
KR20170100371A (en
Inventor
김시호
이상호
Original Assignee
연세대학교 산학협력단
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 연세대학교 산학협력단 filed Critical 연세대학교 산학협력단
Priority to KR1020160022853A priority Critical patent/KR101818839B1/en
Publication of KR20170100371A publication Critical patent/KR20170100371A/en
Application granted granted Critical
Publication of KR101818839B1 publication Critical patent/KR101818839B1/en

Links

Images

Classifications

    • H04N13/0022
    • H04N13/0018
    • H04N13/0033
    • H04N13/0221

Landscapes

  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

TECHNICAL FIELD The present invention relates to a technique for relieving simulator fatigue in a stereoscopic three-dimensional display having a fixed screen such as a virtual reality head mounted display, and according to one aspect, The apparatus for preventing dizziness includes a calculation unit for calculating a focal length of a target object from a rendering stereo camera and an object distance from the stereo camera to the target object, and a calculation unit for calculating the calculated focal length and object distance And a processing unit for correcting the matching between the convergence angle and the focal distance between the user's eyes and the object in the 3D content.

Description

[0001] APPARATUS AND METHOD OF STEREO SCOPIC 3D CONTENTS [0002] BACKGROUND OF THE INVENTION [0003]

The present invention relates to a technique for mitigating user discomfort such as dizziness in a stereo 3D display for application to stereoscopic 3D image content production and display.

Virtual reality refers to a human-computer interface that makes a certain environment or situation computerized and makes the person who uses it behave as if they are interacting with the actual environment and environment. VIRTUAL REALITY is used to show and manipulate people as if they are in the environment without experiencing the environments that are difficult to experience on a daily basis. Applications of virtual reality include education, advanced programming, remote operation, remote satellite surface exploration, exploration data analysis, and scientific visualization.

Virtual reality is available through devices such as head-mounted displays.

The head mount display is a "head screen device" that allows a small display device in front of your eyes to be seen through the lens structure with a clear view of the object.

Fixed screens using head-mounted displays can cause simulator sickness or simulator fatigue, such as dizziness. Users of virtual reality experience dizziness or vomiting symptoms such as motion sickness. This simulator illness or fatigue is caused by a mismatch between what the eye sees and what the body feels or interprets.

The discrepancy of eye-convergence and the accommodation in head-mounted 3D stereoscopic display of the head are important causes of the actual eye view and the virtual view mismatch . In a 3D display using not only a head mount display type but also a stereoscopic image in which mismatch of image information and body information occurs, this simulator fatigue symptom appears.

Korean Patent No. 10-1553273 Korean Patent No. 10-0551107

According to one aspect, the dizziness prevention device at least temporarily implemented by a computer includes an operation unit for calculating a focal length from a stereo camera of a virtual reality to a target object and an object distance from the stereo camera to the target object, And a processor for correcting the perspective and the focus in the content for stereoscopic three-dimensional display by using the calculated focal length and object distance.

The processing unit according to an embodiment blurs the focus of another target object located outside the calculated focal length and adjusts the direction of the stereo camera in a direction corresponding to the position of the target object, And a processor for adjusting an inter-pupillary distance (IPD) of the stereo camera based on the distance.

The operation unit may calculate a first object distance from the stereo camera to the first target object and a second object distance from the stereo camera to the second target object, Pupilary distance (IPD) to be adjusted by reflecting the ratio of the first object distance and the second object distance to the IPD (inter-pupillary distance).

The processing unit according to an embodiment adjusts the direction from the direction in which the stereo camera is heading to the target object in the current content.

The processing unit may calculate an intersection between focal lengths from the stereo camera, calculate a Depth of Field (DoF) for the object of interest based on the calculated intersection, A blurring process is performed on a target object that is spaced apart from the Depth of Field (DoF) by a certain distance.

The processing unit may define a circle of interest (CoI) corresponding to the Depth of Field (DoF) and use the defined circle of interest (CoI) A bounded zone and a parameter value for blurring are defined and blurred according to the calculated Circle of Confusion (CoC).

According to an embodiment, the processing unit may use the calculated Depth of Field (DoF) and the calculated object distance,

Adjusts the direction of the stereo camera toward the focal point for the fixed object in real time.

According to one aspect, an operation method of a dizziness prevention device at least temporarily implemented by a computer includes calculating a focal length of a target object from a stereo camera of a virtual reality and an object distance from the stereo camera to the target object Blurring a focus on another target object located outside the calculated focal distance; adjusting the orientation of the stereo camera in a direction corresponding to the position of the target object; and based on the object distance, And adjusting an inter-pupillary distance (IPD) of the stereo camera.

The calculating step may include calculating a first object distance from the stereo camera to a first target object and a second object distance from the stereo camera to a second target object, The step of adjusting an inter-pupillary distance (IPD) of a stereo camera includes inter-pupillary distance (IPD) of the stereo camera to reflect the ratio of the first object distance and the second object distance to an IPD and calculating a distance.

Adjusting the orientation of the stereo camera in a direction corresponding to the position of the target object according to an embodiment includes adjusting the direction toward the target object from a direction in which the stereo camera is oriented within the current content .

The blurring processing step according to an embodiment may include calculating an intersection between focal lengths from the stereo camera, calculating a depth of field (DoF) for the object of interest based on the calculated intersection, And blurring a target object spaced by a predetermined distance from the calculated Depth of Field (DoF).

The blurring step may include defining a circle of interest (CoI) corresponding to the depth of field (DoF), defining a circle of interest (CoI) Calculating a Circle of Confusion (CoC) using CoI, calculating a bounded zone and a parameter value for blurring according to the calculated Circle of Confusion (CoC) Ring processing.

The step of adjusting the direction of the stereo camera in a direction corresponding to the position of the target object according to an embodiment may further include a step of adjusting the direction of the stereo camera in real time using the calculated depth of field (DoF) And adjusting the orientation of the stereo camera toward the focal point for the fixed object.

1 is a view for explaining an apparatus for preventing dizziness according to an embodiment.
Fig. 2 is a view for explaining an immediate matching (a) of the camera angle corresponding to the version angle of the user's eyes and an additional adjustment (b) of the IPD.
3A is a view for explaining a depth of field (DoF) in which a focal length of a target object is adjusted.
FIG. 3B is a view for explaining a specific example of adjusting the depth of field.
4A is a view for explaining the adjustment of the camera direction (camera orientation).
4B is a view for explaining an embodiment of adjusting the camera direction.
FIG. 4C is a view for explaining the procedure for adjusting the camera direction in detail.
5A is a view for explaining an embodiment for adjusting the length of the IPD.
5B is a view for explaining the change of the target object as the length of the IPD is adjusted in the stereo camera.
6 is a view for explaining an operation method of a dizziness prevention apparatus according to an embodiment.

Hereinafter, embodiments will be described in detail with reference to the accompanying drawings. However, the scope of the rights is not limited or limited by these embodiments. Like reference symbols in the drawings denote like elements.

The terms used in the following description are chosen to be generic and universal in the art to which they are related, but other terms may exist depending on the development and / or change in technology, customs, preferences of the technician, and the like. Accordingly, the terminology used in the following description should not be construed as limiting the technical thought, but should be understood in the exemplary language used to describe the embodiments.

Also, in certain cases, there may be a term chosen arbitrarily by the applicant, in which case the meaning of the detailed description in the corresponding description section. Therefore, the term used in the following description should be understood based on the meaning of the term, not the name of a simple term, and the contents throughout the specification.

1 is a view for explaining an apparatus for preventing dizziness according to an embodiment of the present invention.

2 is a view for explaining an embodiment of the present invention. In FIG. 2 (a), an area of interest, an interesting zone, and an area between these two circles of interest are set according to the position of the target object viewed by the human eye, The eye-convergence of the eye and the eye-accomodation of the eye coincide when the eye of the eye of the user looks at the target object of the region of interest. That is, FIG. 2 (a) shows the rendering angles of the two stereo cameras set by the convergence angle determined by the distance between the eyes and the object and the angle between the two focuses of the two eyes on the left and right eyes of the person An example of displaying a stereo screen is shown. In the display configuration in which the 3D image is reproduced through the stereo screen, the angle of the rendering camera of the left screen and the right screen is changed according to the agreement between the convergence angle of the two eyes and the focus point of the discussion. In FIG. 2 (a), when the distance from the object to which the eye is viewed is changed, the angle of the camera is changed while the IPD is kept constant.

FIG. 2 (b) is similar to the embodiment of FIG. 2 (a) in adjusting the rendering angles of the two stereo cameras, but the distance between the two cameras is finely adjusted to change the focal distance between the eyes We have added a method to maximize the effect that the eye-convergence and the eye-accomodation of the 3D stereo image coincide.

The dizziness prevention apparatus 100 performs an instantaneous matching of the convergence angle of the user's eyes and the camera angle, processes the further adjustment of the camera position according to the distance of the target object, It is possible to handle blurring for fields other than the interesting zone included in the area.

The dizziness prevention apparatus 100 reduces discrepancy of eye-convergence and accommodation of the head mounted 3D stereoscopic display to reduce the visual disturbance of the actual eye view and virtual Can be reduced to prevent dizziness.

Specifically, the dizziness prevention apparatus 100 may include a computing unit 110 and a processing unit 120. The dizziness prevention apparatus 100 may be implemented at least temporarily by the computing terminal. The computing terminal includes any type of electronic device such as a personal computer, a medical device, a smart phone, a tablet computer, and a wearable device. The computing unit 110 and the processing unit 120 may be physical and / or logical elements included in such electronic devices, respectively. For example, the computing unit 110 and the processing unit 120 may be implemented by dedicated hardware or general purpose computing resources controlled by software or an operating system. Also, the operation unit 110 and the processing unit 120 may be implemented together on one chip, and thus may not be physically distinguished, and the implementation of such hardware resources may be changed as much as possible by technological development or design changes. Accordingly, it is understood that the functions, operations, and structures of the operation unit 110 and the processing unit 120 are distinguished from each other, but there may be cases where such a division is interpreted differently according to the embodiment.

The calculation unit 110 calculates the focal length of the target object from the stereo camera of the virtual reality and the object distance from the stereo camera to the target object. The object distance can be interpreted as the shortest distance between a line segment connecting the centers of two lenses of a stereo camera and an object.

On the other hand, the processing unit 120 corrects the perspective and the focus within the content for the wearable virtual reality device (VR HMD) by using the calculated focal length and object distance. However, the application technique of the present invention is not limited to contents for a wearable virtual reality (VR HMD), but also applicable to a general 3D content display having a stereo display form as a basic constitution. For example, the present invention can be applied to the production and display of contents of 3D images using polarized glasses or electronic shutter glasses.

The processing unit 120 according to an exemplary embodiment of the present invention performs blurring on fields other than the interlaced zone, processes instantaneous convergence angles of the eyes of the user and camera angles, Additional adjustments can prevent dizziness from mismatching.

First, the processing unit 120 may perform blurring processing on a background of another object or a screen located outside the inter-resting zone calculated by the calculated focal length.

To this end, the processing unit 120 may calculate the intersection between the focal lengths from the stereo camera, and calculate the Depth of Field (DoF) for the object of interest based on the calculated intersection. In addition, the processing unit 120 may perform blurring processing on a target object or background spaced from the calculated Depth of Field (DoF) by a predetermined distance or more. That is, the Depth of Field (DoF) can be interpreted as the depth of the subject. By blurring the region beyond the depth of the subject, the image is displayed similar to the actual image, By reducing the mismatch, it is possible to reduce the dizziness felt by the person viewing the 3D image.

The embodiment of blurring will be described in detail again with reference to FIGS. 3A and 3B.

Next, the processing unit 120 may adjust the direction of the stereo camera to a direction corresponding to the position of the target object. That is, the processing unit 120 can prevent dizziness by adjusting the direction from the direction of the stereo camera toward the target object within the current content. For example, the processing unit 120 can adjust the direction of the stereo camera toward the focal point for the fixed object in real time using the calculated Depth of Field (DoF) and the calculated object distance have.

An embodiment of adjusting the direction of the stereo camera in the direction corresponding to the position of the target object will be described in detail again with reference to FIGS. 4A to 4C.

The processing unit 120 may adjust the inter-pupillary distance (IPD) of the stereo camera based on the object distance.

For this purpose, the arithmetic unit 110 calculates a first object distance from the stereo camera to the first target object and a second object distance from the stereo camera to the second target object, and the processing unit 120 calculates the inter- pupillary distance (IPD) to be adjusted by reflecting the ratio of the first object distance and the second object distance to the first object distance and the second object distance.

An embodiment for adjusting the inter-pupillary distance (IPD) of a stereo camera to prevent dizziness will be described in detail again in Figs. 5A and 5B.

Fig. 2 is a view for explaining an immediate matching (a) of the camera angle corresponding to the version angle of the user's eyes and an additional adjustment (b) of the IPD.

FIG. 2 illustrates a method of calculating an intersection point of a predetermined focal length in two cameras and using parameters (focal length, size of numerical aperture, etc.) for Depth of Field (DoF).

That is, the processing unit defines a circle of interest (CoI) corresponding to the depth of field (DoF) of the field, and uses a defined circle of interest (CoI) , And CoC), and define a bounded zone and a parameter value for blurring according to the calculated Circle of Confusion (CoC), thereby blurring processing.

The circle of interest represents a depth of field that is spaced a certain distance in the direction of the stereo camera and the opposite direction with respect to the fixed object, and the target object located within the circle of interest can be focused and displayed clearly. On the other hand, when the target object is positioned on the stereo camera side, the direction of the stereo camera moves in the direction toward the fixed object, thereby preventing dizziness due to mismatching. That is, in the present invention, the previously fixed focal length can be adaptively adjusted.

Meanwhile, in FIG. 2 (b), the IPD of the stereo camera can be adjusted to prevent dizziness due to mismatching. In the embodiment of FIG. 2 (b), the IPD of the stereo camera is largely adjusted in a state in which the target object is close to the stereo camera, thereby preventing dizziness due to mismatching.

3A is a view for explaining a depth of field (DoF) in which a focal length of a target object is adjusted.

According to the embodiment 310 of FIG. 3A, the stereo camera having the distance a as the IPD is located at the intersection of the focal lengths of the two cameras at positions corresponding to S. The position of the image on the corresponding film plane corresponding to this intersection corresponds to d. At 311, the position of d is enlarged and displayed, and at the position of c, a confusion image due to a circle of confusion may be formed.

The distracted image is generated by the target object at a distance S r from the stereo camera with respect to the intersection point, and the target object at a position S f near to the stereo camera with respect to the intersection point, which are located at distances d and d r or d f So that it causes confusion on the image due to d. Thus, dizziness may occur due to such confusion. In the present invention, blurring processing is performed on target objects having depth of field different from depth of field corresponding to an intersection, thereby preventing dizziness due to mismatch.

FIG. 3B is a view for explaining a specific example of adjusting the depth of field.

Reference numeral 321 in FIG. 3B is an image of a general virtual reality in which the depth of field is not adjusted, and dizziness may be caused by other target objects. On the other hand, the depth of field 322 is adjusted to blur the target objects other than the target object, so that dizziness due to mismatching can be significantly reduced.

4A is a view for explaining the adjustment of the camera direction (camera orientation).

Referring to embodiment 410 of FIG. 4A, EyeLeft may be interpreted as a left camera in a stereo camera, and EyeRight may be interpreted as a right camera in a stereo camera.

The stereo camera can adjust the direction in the direction toward the first object corresponding to P 1 and in the direction toward the second object corresponding to P 2 . Object 1 corresponding to P 1 is displayed on the IPD axis

Figure 112016018834353-pat00001
, And the object corresponding to P 2 is located in the direction
Figure 112016018834353-pat00002
, In other words
Figure 112016018834353-pat00003
in
Figure 112016018834353-pat00004
In the direction of the arrow.

E.g,

Figure 112016018834353-pat00005
,
Figure 112016018834353-pat00006
, And
Figure 112016018834353-pat00007
Can be calculated by the following equation (1).

[Equation 1]

Figure 112016018834353-pat00008
= tan -1 (LeftEye 1 / d1)

Figure 112016018834353-pat00009
L2 = tan -1 (LeftEye 2 / d2)

Figure 112016018834353-pat00010

Here, d1 is the distance from the midpoint of the stereo camera to the first object, and d2 is the distance from the midpoint of the stereo camera to the second object.

That is, by adjusting the direction of each of the stereo cameras according to the object, dizziness due to mismatch can be prevented.

4B is a view for explaining an embodiment of adjusting the camera direction.

Reference numeral 420 denotes a relationship between a stereo camera and an object before adjusting the direction of the camera, and reference numeral 430 denotes an embodiment in which the direction of each of the stereo cameras is adjusted toward the object direction.

FIG. 4C is a view for explaining the procedure for adjusting the camera direction in detail.

When P 1 moves to P 2

Figure 112016018834353-pat00011
Figure 112016018834353-pat00012
Figure 112016018834353-pat00013
And the change value of the angle in the equation (1)
Figure 112016018834353-pat00014
Therefore, it is possible to create a Camera Transformation Matrix (Homogenenous) by using it.

That is, since only the direction of the camera is adjusted, the change value for the position is 0 -> x = 0.

Referring to this, the Camera Transformation Matrix can be calculated by Equation (2).

&Quot; (2) "

Figure 112016018834353-pat00015

In the embodiment of FIG. 4C, R cam , which is a Camera Transformation Matrix, is applied to actual rendering, and L2W is calculated locally from the world (model to world) to calculate M (431). At this time, L2W corresponds to the Local to World matrix.

Next, an operation related to the camera can be performed in the world by reflecting C in M (432). At this time, C corresponds to the World to Camera matrix.

Next, the dizziness prevention apparatus according to the present invention may reflect R cam calculated through Equation (2) to M reflecting C (433). Accordingly, the direction of the stereo camera is adjusted, and dizziness due to mismatch can be mitigated.

Next, M may be projected 434 for a projection and projected 436 onto a screen via a viewpoint operation 435 for projection. View point operation is to adjust the orientation of the camera M reflect the three-dimensional coordinates of P obj of the target object, in order to be projected on the screen should be calculated by a 2D projection vector (M 2D).

5A is a view for explaining an embodiment for adjusting the length of the IPD.

As shown at 510, IPD 1 for the first target object may be changed to IPD 2 for the second target object.

The IPD may be interpreted as the distance between the centers of the respective lenses of the stereo camera, and may be adjusted in consideration of the distance from the target object in the present invention.

More specifically, IPD 2 can be calculated through the following equation (3).

&Quot; (3) "

Figure 112016018834353-pat00016

Here, Dist eye2obj1 can be interpreted as the shortest distance between the center of the stereo camera and the first object, and Dist eye2obj2 can be interpreted as the shortest distance between the center of the stereo camera and the second object.

That is, the dizziness prevention apparatus calculates the first object distance (Dist eye2obj1 ) from the stereo camera to the first target object and the second object distance (Dist eye2obj2 ) from the stereo camera to the second target object, pupillary distance (IPD) to be adjusted by reflecting the ratio of the first object distance (Dist eye 2 obj 1) and the second object distance (Dist eye 2 obj 2) to the inter-pupillary distance.

5B is a view for explaining the change of the target object as the length of the IPD is adjusted in the stereo camera.

Reference numeral 520 corresponds to the content before the IPD length is adjusted, and reference numeral 530 corresponds to the content before the IPD length is adjusted by the formula (3).

The inter-pupillary distance (IPD) may be interpreted as the distance between the centers of the pupils of the user's eye or the distance between the centers of different lenses of the stereo camera in the virtual reality.

IPD is typically a fixed parameter during 3D rendering because it has a unique IPD value for each user.

With reference to this, the present invention adjusts the position of the stereoscopic rendering camera to mitigate binocular parallax caused by nearby objects between the eyes and the occlusion in the background.

Therefore, the present invention can adjust the IPD to maximize the effect of focus when the distance between the object and the camera is located in the foveated area. To this end, it is assumed in the present invention that the dynamic distance between the left eye and the right eye is the same.

At this time, IPD new for adjustment is calculated by the distance from the target object to the camera. When the distance from the target object to the camera is obtained, the position of the left camera is calculated by Equation (4) below.

&Quot; (4) "

Figure 112016018834353-pat00017

Similarly, the position of the right camera is calculated by the following equation (5).

&Quot; (5) "

Figure 112016018834353-pat00018

If two virtual camera focuses are present on the target object, the camera focus can be rendered. To render the camera focus, the rotational angles of the Cartesian x, y, z axes can be calculated. At this time, the present invention can generate a matrix of rotations for each of the x, y, and z axes as Rot x , Rot y , and Rot z . Once the matrix is calculated, it can be reflected in the final matrix to adjust the camera orientation.

You can apply this program after camera conversion to perform the results.

The confusion caused by surrounding objects can cause the mismatch between the user's eye convergence angle and the two-dimensional camera angle mismatch.

6 is a view for explaining an operation method of a dizziness prevention apparatus according to an embodiment.

The method of operation of the dizziness prevention apparatus according to an exemplary embodiment calculates the focal distance and the object distance (step 601).

Next, an operation method of the dizziness prevention apparatus according to an exemplary embodiment of the present invention is described in detail with reference to the accompanying drawings, in which: FIG. 1 is a block diagram illustrating a configuration of a wearable virtual reality device (VR HMD) Perspective and focus can be corrected.

To this end, the method of operation of the dizziness prevention device adjusts the depth of field (DoF) of the target object (step 602), adjusts the focal length of the target object (step 603), adjusts the IPD of the stereo camera (Step 604).

Recently, the virtual reality related industry has become a hot topic, among which the technology that enables users to concentrate on the contents to concentrate within the content is one of the necessary parts of all game companies, but the focal length, Camera Orientation) and IPD are fixed. Therefore, if the present invention is used, it can be solved in the content, and more efficient solution is possible.

The apparatus described above may be implemented as a hardware component, a software component, and / or a combination of hardware components and software components. For example, the apparatus and components described in the embodiments may be implemented within a computer system, such as, for example, a processor, a controller, an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable array (FPA) A programmable logic unit (PLU), a microprocessor, or any other device capable of executing and responding to instructions. The processing device may execute an operating system (OS) and one or more software applications running on the operating system. The processing device may also access, store, manipulate, process, and generate data in response to execution of the software. For ease of understanding, the processing apparatus may be described as being used singly, but those skilled in the art will recognize that the processing apparatus may have a plurality of processing elements and / As shown in FIG. For example, the processing unit may comprise a plurality of processors or one processor and one controller. Other processing configurations are also possible, such as a parallel processor.

The software may include a computer program, code, instructions, or a combination of one or more of the foregoing, and may be configured to configure the processing device to operate as desired or to process it collectively or collectively Device can be commanded. The software and / or data may be in the form of any type of machine, component, physical device, virtual equipment, computer storage media, or device , Or may be permanently or temporarily embodied in a transmitted signal wave. The software may be distributed over a networked computer system and stored or executed in a distributed manner. The software and data may be stored on one or more computer readable recording media.

The method according to an embodiment may be implemented in the form of a program command that can be executed through various computer means and recorded in a computer-readable medium. The computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination. The program instructions to be recorded on the medium may be those specially designed and configured for the embodiments or may be available to those skilled in the art of computer software. Examples of computer-readable media include magnetic media such as hard disks, floppy disks and magnetic tape; optical media such as CD-ROMs and DVDs; magnetic media such as floppy disks; Magneto-optical media, and hardware devices specifically configured to store and execute program instructions such as ROM, RAM, flash memory, and the like. Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like. The hardware devices described above may be configured to operate as one or more software modules to perform the operations of the embodiments, and vice versa.

While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. For example, it is to be understood that the techniques described may be performed in a different order than the described methods, and / or that components of the described systems, structures, devices, circuits, Lt; / RTI > or equivalents, even if it is replaced or replaced.

Therefore, other implementations, other embodiments, and equivalents to the claims are also within the scope of the following claims.

Claims (14)

Implemented at least temporarily by the computer:
Rendering of a virtual reality for a left screen and a right screen for reproducing a 3D image of a virtual reality A focal length for a target object of a virtual reality from a stereo camera and rendering of the virtual reality Object from a stereo camera to the target object An operation unit for calculating a distance; And
A blurring processing unit for blurring a focus of another target object located outside the calculated focal distance and adjusting a direction of the rendering stereo camera of the virtual reality in a direction corresponding to the position of the target object,
Lt; / RTI >
The operation unit,
Calculating a first object distance from the rendering stereo camera of the virtual reality to the first target object of the virtual reality and a second object distance from the rendering stereo camera of the virtual reality to the second target object of the virtual reality,
Wherein,
Calculating an inter-pupillary distance (IPD) to be adjusted by reflecting a ratio of the first object distance to the second object distance to an inter-pupillary distance (IPD) of a rendering stereo camera of the virtual reality,
The inter-pupillary distance (IPD) of the rendering stereo camera of the virtual reality is adjusted according to the calculated inter-pupillary distance (IPD) to determine the perspective and focus of the view of the virtual reality in the content for stereoscopic three- A device to prevent dizziness that reduces mismatch with the view of the actual eye.
delete delete The method according to claim 1,
Wherein,
And adjusts in the direction from the direction in which the rendering stereo camera is directed within the current content to the target object.
The method according to claim 1,
Wherein,
Calculating an intersection between focal lengths from the rendering stereo camera, calculating a Depth of Field (DoF) for the object of interest based on the calculated intersection,
And blurring a target object spaced by a predetermined distance from the calculated Depth of Field (DoF).
6. The method of claim 5,
Wherein,
A circle of interest (CoI) corresponding to the depth of field (DoF) is defined and a Circle of Confusion (CoC) is defined using the defined circle of interest (CoI) And defines a bounded zone and a parameter value for blurring according to the calculated Circle of Confusion (CoC), and performs blurring processing.
The method according to claim 1,
Wherein,
Using the calculated Depth of Field (DoF) and the calculated object distance,
A dizziness prevention device that adjusts the orientation of a rendered stereo camera towards a focal point for a fixed object in real time.
A method of operation of a dizziness prevention device that is at least temporarily implemented by a computer,
Rendering of a virtual reality for a left screen and a right screen for reproducing a 3D image of a virtual reality A focal length for a target object of a virtual reality from a stereo camera and rendering of the virtual reality Object from a stereo camera to the target object Calculating a distance;
Blurring a focus of another target object located outside the calculated focal distance;
Adjusting a direction of the rendering stereo camera of the virtual reality in a direction corresponding to the position of the target object;
Calculating a first object distance from the rendering stereo camera of the virtual reality to the first target object of the virtual reality and a second object distance from the rendering stereo camera of the virtual reality to the second target object of the virtual reality;
Calculating an inter-pupillary distance (IPD) to be adjusted by reflecting a ratio of the first object distance to the second object distance to an inter-pupillary distance (IPD) of a rendering stereo camera of the virtual reality;
The inter-pupillary distance (IPD) of the rendering stereo camera of the virtual reality is adjusted according to the calculated inter-pupillary distance (IPD) to determine the perspective and focus of the view of the virtual reality in the content for stereoscopic three- Correcting a mismatch with a view of the actual eye
The method comprising the steps of:
delete 9. The method of claim 8,
Wherein the step of adjusting the orientation of the rendering stereo camera in a direction corresponding to the position of the target object comprises:
Adjusting in a direction from the direction of the rendering stereo camera toward the target object within the current content
The method comprising the steps of:
9. The method of claim 8,
The blurring step includes:
Calculating an intersection between the focal lengths from the rendering stereo camera;
Calculating a depth of field (DoF) for the object of interest based on the calculated intersection; And
Blurring a target object spaced by a predetermined distance from the calculated Depth of Field (DoF)
The method comprising the steps of:
12. The method of claim 11,
The blurring step includes:
Defining a circle of interest (CoI) corresponding to the depth of field (DoF);
Calculating a Circle of Confusion (CoC) using the defined circle of interest (CoI);
Defining a bounded zone and parameter values for blurring according to the calculated Circle of Confusion (CoC) and blurring
The method comprising the steps of:
9. The method of claim 8,
Wherein the step of adjusting the orientation of the rendering stereo camera in a direction corresponding to the position of the target object comprises:
Adjusting a direction of a rendering stereo camera toward a focal point of an object fixed in real time using the calculated Depth of Field (DoF) and the calculated object distance,
The method comprising the steps of:
The method according to claim 1,
Wherein,
Pupilary distance (IPD) to be adjusted by multiplying a value obtained by dividing the second object distance by the first object distance by an inter-pupillary distance (IPD) of the rendering stereo camera.
KR1020160022853A 2016-02-25 2016-02-25 Apparatus and method of stereo scopic 3d contents creation and display KR101818839B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
KR1020160022853A KR101818839B1 (en) 2016-02-25 2016-02-25 Apparatus and method of stereo scopic 3d contents creation and display

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
KR1020160022853A KR101818839B1 (en) 2016-02-25 2016-02-25 Apparatus and method of stereo scopic 3d contents creation and display

Publications (2)

Publication Number Publication Date
KR20170100371A KR20170100371A (en) 2017-09-04
KR101818839B1 true KR101818839B1 (en) 2018-01-15

Family

ID=59924335

Family Applications (1)

Application Number Title Priority Date Filing Date
KR1020160022853A KR101818839B1 (en) 2016-02-25 2016-02-25 Apparatus and method of stereo scopic 3d contents creation and display

Country Status (1)

Country Link
KR (1) KR101818839B1 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11877064B1 (en) * 2019-08-20 2024-01-16 Opic Technologies, Inc. Method and apparatus for a stereoscopic smart phone
US11927756B2 (en) 2021-04-01 2024-03-12 Samsung Electronics Co., Ltd. Method for providing augmented reality image and head mounted display device supporting the same
KR20220136776A (en) * 2021-04-01 2022-10-11 삼성전자주식회사 Method for providing augmented reality image and head mounted display device supporting the same
KR102559913B1 (en) * 2022-10-20 2023-07-26 주식회사 비브스튜디오스 Method for implementing camera movement by using a virtual camera

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100688456B1 (en) * 2005-09-13 2007-03-02 한국방송공사 Method for controlling zoom-convergence interlock in parallel translation axis typed stereo camera system

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100688456B1 (en) * 2005-09-13 2007-03-02 한국방송공사 Method for controlling zoom-convergence interlock in parallel translation axis typed stereo camera system

Also Published As

Publication number Publication date
KR20170100371A (en) 2017-09-04

Similar Documents

Publication Publication Date Title
JP7094266B2 (en) Single-depth tracking-accommodation-binocular accommodation solution
US11132056B2 (en) Predictive eye tracking systems and methods for foveated rendering for electronic displays
US10241329B2 (en) Varifocal aberration compensation for near-eye displays
KR101788452B1 (en) Apparatus and method for replaying contents using eye tracking of users
US20160267720A1 (en) Pleasant and Realistic Virtual/Augmented/Mixed Reality Experience
US20150187115A1 (en) Dynamically adjustable 3d goggles
KR101818839B1 (en) Apparatus and method of stereo scopic 3d contents creation and display
EP3001681B1 (en) Device, method and computer program for 3d rendering
CN106293561B (en) Display control method and device and display equipment
JP2022183177A (en) Head-mounted display device
JP6576639B2 (en) Electronic glasses and control method of electronic glasses
US11212502B2 (en) Method of modifying an image on a computational device
US20220365594A1 (en) Control method and apparatus for virtual reality device
JP2023515205A (en) Display method, device, terminal device and computer program
US11543655B1 (en) Rendering for multi-focus display systems
KR101873161B1 (en) Method and apparatus for providing personal 3-dimensional image using convergence matching algorithm
CN111279293A (en) Method for modifying image on computing device
US11237413B1 (en) Multi-focal display based on polarization switches and geometric phase lenses
CN110197524B (en) Stereoscopic display method, apparatus, device, and computer-readable storage medium
US10326976B2 (en) Method and apparatus for providing personal 3-dimensional image using convergence matching algorithm
Gurrieri Improvements in the visualization of stereoscopic 3D imagery
Watt et al. 3D media and the human visual system
CN116466903A (en) Image display method, device, equipment and storage medium

Legal Events

Date Code Title Description
A201 Request for examination
E902 Notification of reason for refusal
AMND Amendment
E601 Decision to refuse application
AMND Amendment
X701 Decision to grant (after re-examination)
GRNT Written decision to grant