US10298921B1 - Superstereoscopic display with enhanced off-angle separation - Google Patents

Superstereoscopic display with enhanced off-angle separation Download PDF

Info

Publication number
US10298921B1
US10298921B1 US16/044,355 US201816044355A US10298921B1 US 10298921 B1 US10298921 B1 US 10298921B1 US 201816044355 A US201816044355 A US 201816044355A US 10298921 B1 US10298921 B1 US 10298921B1
Authority
US
United States
Prior art keywords
display
light source
optical volume
superstereoscopic
viewing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US16/044,355
Inventor
Shawn Frayne
Tung Yiu Fok
Shiu Pong Lee
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Looking Glass Factory Inc
Original Assignee
Looking Glass Factory Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Looking Glass Factory Inc filed Critical Looking Glass Factory Inc
Priority to US16/044,355 priority Critical patent/US10298921B1/en
Assigned to LOOKING GLASS FACTORY, INC. reassignment LOOKING GLASS FACTORY, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FRAYNE, SHAWN, LEE, SHIU PONG, FOK, TUNG YIU
Priority to KR1020207027895A priority patent/KR102416197B1/en
Priority to JP2020567739A priority patent/JP7096371B2/en
Priority to EP19760138.8A priority patent/EP3759908A4/en
Priority to CN201980028768.0A priority patent/CN112042186B/en
Priority to PCT/US2019/015235 priority patent/WO2019168617A1/en
Priority to US16/374,955 priority patent/US11051006B2/en
Publication of US10298921B1 publication Critical patent/US10298921B1/en
Application granted granted Critical
Priority to US17/328,076 priority patent/US11683472B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking
    • H04N13/368Image reproducers using viewer tracking for two or more viewers
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0018Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for preventing ghost images
    • G02B27/2214
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B30/00Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
    • G02B30/20Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
    • G02B30/26Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type
    • G02B30/27Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the autostereoscopic type involving lenticular arrays
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B30/00Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
    • G02B30/50Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels
    • G02B30/56Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images the image being built up from image elements distributed over a 3D volume, e.g. voxels by projecting aerial or floating images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • H04N13/305Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using lenticular lenses, e.g. arrangements of cylindrical lenses
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/302Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays
    • H04N13/32Image reproducers for viewing without the aid of special glasses, i.e. using autostereoscopic displays using arrays of controllable light sources; using moving apertures or moving light sources
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/346Image reproducers using prisms or semi-transparent mirrors
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/366Image reproducers using viewer tracking
    • H04N13/383Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/398Synchronisation thereof; Control thereof

Definitions

  • This invention relates generally to the image display field, and more specifically to new and useful superstereoscopic displays with enhanced off-angle separation.
  • Image displays are an integral part of modern life. From televisions to monitors to smartphone and tablet screens, image displays provide users with the ability to view and interact with information presented in a variety of forms.
  • HMD head-mounted display
  • Free-space 3D displays remove the need for an HMD, allowing multiple users to see and manipulate content in a shared experience.
  • the few existing free-space 3D displays are hampered by a number of issues, including size, limited view angle, low resolution and brightness, scene distortion, and high cost.
  • This invention provides such new and useful displays.
  • FIG. 1 is a top-down view of a display of an invention embodiment
  • FIGS. 2A and 2B are stereoscopic views of a three-dimensional image
  • FIGS. 3A and 3B are cross-sectional views of a display of an invention embodiment
  • FIG. 4 is a cross-sectional view of a parallax generator of a display of an invention embodiment
  • FIG. 5A is a cross-sectional view of a lenticular lens in a stretched configuration of a display of an invention embodiment
  • FIG. 5B is a cross-sectional view of a lenticular lens in a compressed configuration of a display of an invention embodiment
  • FIG. 6A is a light path view of an image projected from a light source of a display of an invention embodiment
  • FIG. 6B is a light path view of an image projected from a light source and passing through a high-index optical volume of a display of an invention embodiment
  • FIG. 7 is a top-down view of a display of an invention embodiment
  • FIG. 8 is a perspective view of a high-index optical volume and perceived additional volume of an invention embodiment
  • FIGS. 9A, 9B, and 9C are various perspective views of a display of an invention embodiment.
  • FIGS. 10A and 10B are top-down views of image view dependency on viewing distance of a display of an invention embodiment.
  • a superstereoscopic display 100 with enhanced off-angle separation includes a light source 110 , a parallax generator 120 , and a high-index optical volume 140 , as shown in FIG. 1 .
  • the display 100 may additionally or alternatively include polarizers 130 , quarter waveplates 131 , optical volume guides 141 , an onboard computer 150 , and/or a contextual lighting system 160 .
  • the display 100 functions to enable viewers to see two- and/or three-dimensional image data from multiple perspectives at the same time.
  • the display 100 generates an angle-dependent viewing experience, which can be used to provide a three-dimensional viewing experience (via stereopsis) and/or to provide viewers at different angles with different images depending on viewing angle (without necessarily causing the perception of depth).
  • An example is as shown in FIGS. 2A and 2B .
  • the display 100 is capable of producing the same 3D image at multiple angles (e.g., viewer 1 has eyes at 0 and 5 degrees, which see image 1 and 2 respectively; viewer 2 has eyes at 20 and 25 degrees, which see image 1 and 2 respectively), the display 100 preferably produces a continuous unbroken scene (e.g., image 1 at 0 degrees, image 2 at 5 degrees . . . image N at 5(N ⁇ 1) degrees where N is an integer) to better provide an immersive viewing experience.
  • the display 100 may utilize any set of images for any purpose.
  • Enhanced off-angle separation is preferably enabled by the high-index optical volume 140 ; by nature of its high index of refraction and shape, the optical volume 140 may cause the primary image(s) intended for a viewer at a particular viewing angle to appear closer to the viewer than otherwise would occur (without the optical volume 140 ), increasing the perception that the image produced by the light source 110 “floats” off of the screen. Further enhancing this effect, for images off-axis to the viewer at a particular viewing angle, the images may not appear as close as on-axis images do, appearing to the viewer to separate on- and off-axis images in space. This phenomenon is discussed in greater detail in the section on the high-index optical volume 140 .
  • the light source 110 functions to generate images (i.e., light generated from image data) for display by the display 100 .
  • the light source 110 is preferably a planar two-dimensional display comprising a set of individually addressable pixels, but may additionally or alternatively be any suitable display.
  • the light source 110 may comprise one or more movable light sources; e.g., a laser that may be scanned across a set of positions to simulate the appearance of multiple light sources (i.e., display multiplexing).
  • the light source 110 is preferably an RGB color light source (e.g., each pixel includes red, green, and blue subpixels) but may additionally or alternatively be a substantially monochromatic light source or any other light source (e.g., a white light source).
  • the light source 110 is preferably a projector or projector light engine (e.g., DLP, laser, LCoS, and/or LCD projector) but may additionally or alternatively be any suitable display (e.g., an LCD monitor/TV display, an OLED display, an e-ink display, an LED array, a spinning LED display, an e-ink display, an electroluminescent display, a neon display, etc.).
  • the light source 110 includes a liquid crystal panel with a collimated backlight.
  • the display 100 preferably includes a single light source 110 , but may additionally or alternatively include multiple light sources 110 .
  • multiple light sources 110 may be placed in series and/or in parallel as shown in FIG. 3A and FIG. 3B respectively.
  • any combination of light sources 110 , polarizers 130 , quarter-waveplates 131 , and/or additional optics (e.g., mirrors, lenses, etc.) may be used in the display 100 .
  • the light sources 110 may be offset, angled, rotating, curved, or otherwise configured in any manner.
  • the light source 110 may include optical elements (e.g., lenses, mirrors, waveguides, filters) that function to couple light into the parallax generator 120 and/or the high-index optical volume 140 .
  • the light source 110 may include a collimating lens designed to increase collimating of the light source 110 output.
  • the light source 110 may include a lens designed to scale (or otherwise distort) light source 110 output (e.g., reduce in size or increase in size).
  • Such a lens may scale light source 110 output uniformly (e.g., 2 ⁇ decrease in both image dimensions) or non-uniformly (e.g., no decrease in first image dimension, 4 ⁇ decrease in other image dimension).
  • the light source 110 may include a lens that manipulates the focal plane of the viewed image; such a lens may be tunable (allowing depth of field to be swept). If such a lens is tunable at a high rate, this may provide an expanded perceived depth of field to a viewer.
  • the light source 110 may additionally or alternatively include any passive or active optical elements to prepare light for use by the display 100 for any other purpose.
  • the light source 110 may include filters or splitters.
  • the light source 110 may include a polarizing filter if the native output of the light source 110 is unpolarized.
  • the light source 110 may include microlens arrays and/or Fresnel lenses.
  • the parallax generator 120 functions to generate an angle-dependent view from the light source 110 output.
  • the parallax generator 120 may be any structure capable of generating such an angle-dependent view (e.g., lenticular lenses, micro-spherical lenses, parallax barriers, etc.).
  • the parallax generator 120 is a lenticular lens.
  • the lenticular lens generates an angle dependent view in combination with a light source 110 that displays different pixels (or image segments) based on the positioning of the lenticular lens over the light source 110 .
  • the image shown to viewers at three angles is comprised of the pixels labeled 1 , 2 , and 3 , respectively.
  • the display 100 may modify the output of the light source 110 to produce a desired result.
  • properties of the lenticular lens e.g., pitch, material, structure, orientation and position relative to the light source 110
  • desired viewing characteristics e.g., number of viewers, view distance, number of views desired, viewing mode, etc.
  • the number of different views provided by the display 100 is sufficient for superstereoscopic viewing at some viewing distance; that is, each eye of the viewer receives a different image from the display 100 , and as the viewer moves around the display 100 , the views change (with the viewing angle). For example, a viewer at angle one may see scene 1 with a right eye and scene 2 with a left eye, where scene 1 and scene 2 create a stereoscopic three-dimensional view of one or more objects. After the viewer moves from angle one to angle two, the viewer now sees scene 2 with the right eye and scene 3 with the left eye, producing a second stereoscopic three-dimensional view of the one or more objects.
  • FIG. 2A corresponding to a first view
  • FIG. 2B corresponding to a second view
  • a display that produces more than one stereoscopic three-dimensional view in this manner may be referred to as a superstereoscopic display.
  • the lenticular lens may have any suitable configuration and structure and may be made of any suitable material.
  • the lenticular lens is preferably one-dimensional (e.g., cylindrical lenses arranged in columns), but may additionally or alternatively be a two-dimensional lenticular lens, fly-eye lens array, or integral imaging lens array. Note that while there is preferably a correlation between addressable segments (e.g., pixels) of the light source 110 and the lenticular lens, the lens columns of the lenticular lens need not be at a particular orientation relative to the light source 110 .
  • columns of the lenticular lens may be aligned with pixel columns, they may also be offset at an angle (which allows the resolution loss due to image slicing to be apportioned across both image pixels columns and rows, rather than only one of these).
  • Image slicing or division may be accomplished in any manner to achieve a desired viewing result. Processing of the image is preferably performed by the onboard computer 150 but may additionally or alternatively be controlled by any computer system.
  • lenticular lenses may be reconfigurable to change the optical properties of the lenses.
  • a lenticular lens may be fabricated of a flexible or semi-flexible material so that the lens can be physically compressed or stressed to change the pitch of the lens, as shown in FIGS. 5A (stretched) and 5 B (compressed), and the pixel-to-lens mapping may be changed accordingly.
  • the optical properties of the lens may be altered by another mechanism (e.g., by modifying the index of refraction of the lenticular lens).
  • a lenticular lens may be implemented using a layered liquid crystal array (either as the entire lens or in conjunction with a fixed refractive lens), enabling dynamic configuration of the lenticular lens.
  • the display 100 may include multiple lenticular lenses and/or other lenses to produce a desired optical effect.
  • 1D lenticular lenses may be stacked at different orientations to create 2D angular view dependence.
  • the parallax generator 120 is a pinhole array or parallax barrier.
  • parallax generation may be provided by multiple light sources 110 .
  • a set of projectors at different angles and a viewing angle-dependent film or surface (together functioning as both the parallax generator 120 and the light source 110 ) may be used to produce a similar viewing effect to a single light source 110 and an overlaid parallax generator 120 .
  • Polarizers 130 and quarter waveplates 131 may be used to enhance the quality of the image output of the display 100 .
  • the polarizers 130 and quarter waveplate 131 may increase perceived image contrast or brightness (by suppressing errant reflection). This same effect may be used, for example, to reduce any image ghosting caused by the parallax generator 120 (in this case, the ‘ghost’ images may be of a different polarization than intended/on-axis images).
  • polarizers 130 and quarter waveplates 131 may be laminated onto the optical volume 140 , it may be desirable for these (and other) optical components to be separated by some material (or air/vacuum) in order to prevent the loss of light from the optical volume 140 that may result from frustrated total internal reflection (FTIR) since the polarizers 130 /quarter waveplates 131 may have an index between that of the optical volume 140 and the surrounding air.
  • FTIR frustrated total internal reflection
  • the high-index optical volume 140 functions to enhance viewing of the display 100 by enhancing perceived separation between on-angle views (i.e., primary views) and off-angle views.
  • the high-index optical volume 140 is preferably a solid rectangular prism of an optically clear material (e.g., acrylic, glass, polycarbonate), but may additionally or alternatively be any three-dimensional volume (made of any materials in any structure) capable of transmitting light and having an index of refraction greater than one.
  • the high-index optical volume may be acrylic and have a relative index of refraction of 1.49.
  • a light source 110 can create an image on the eye of a viewer.
  • the image formed on the viewer's eye is smaller, suggesting that the perceived image with the optical volume 140 in place is either closer (e.g., as shown by the perceived location of image) or larger than it would be without optical volume.
  • the optical volume 140 may have the additional benefit of increasing viewing angle of the display 100 (by bending light toward on-axis viewing).
  • this effect is less prominent at viewing angles off-axis (e.g., not perpendicular to the light source 110 and optical volume 140 ), as shown in FIG. 7 . This may result in separation between on-angle and off-angle views, further enhancing the depth effect enabled by stereoscopic images.
  • the display 100 may additionally include optical volume guides 141 , which function to enhance the ‘floating’ effect of images within the optical volume 140 .
  • the optical volume guides 141 are preferably markings visible on the optical volume 140 (or otherwise within the optical path of the light source 110 ) that draw a viewer's eye to aid in providing the sensation of depth to images of the display 100 .
  • the optical volume guides 141 may be any visible two- or three-dimensional structure present in a plane distinct from (and preferably parallel to) the perceived plane locating a primary/on-axis image viewed by a viewer.
  • the optical volume guides 141 may be etched, painted, adhered, or printed onto the surface of the optical volume 140 . Additionally or alternatively, the optical volume guides 141 may be located on a separate light guide, substrate, or on any other component.
  • the optical volume guides 141 are reflective on at least one surface.
  • the system 100 incorporates a reflective polarizer 130 (or other reflective or partially reflective surface) to create a virtual image of the optical volume guides 141 , creating the perception of a volume larger (e.g., twice as large) than the optical volume 140 's originally perceived size, as shown in FIG. 8 (the optical volume 140 's originally perceived size may be smaller than its actual size due to the use of high-index material).
  • optical volume guides 141 may also enable the three-dimensional effect of the display 100 to be better perceived in (two-dimensional) video recordings than other displays, allowing for the full impact of display 100 to be better communicated over film.
  • the display 100 may additionally or alternatively use opaque elements (e.g., portions of an opaque housing) to surround or partially surround sides of the optical volume 140 , further enhancing the perception of depth (and potentially also reducing the presence of undesired light), as shown in FIG. 9A . Additional views of this example implementation are as shown in FIG. 9B and FIG. 9C (including optional optical volume guides 141 ).
  • opaque elements e.g., portions of an opaque housing
  • the display 100 may be used with other displays 100 in any manner (e.g., in a 2 ⁇ 2 or 3 ⁇ 3 array, back to back).
  • the onboard computer 150 functions to perform image processing for image data received by the display 100 prior to display by the light source 110 .
  • the onboard computer may separate 3D model information into slices to be projected by the light source 110 .
  • the onboard computer 150 may additionally or alternatively function to prepare 3D image data for voxel representation in any manner.
  • the onboard computer 150 may generate 2D stereoscopic views based on the structure/configuration of the parallax generator 120 .
  • the onboard computer 150 may perform interpolation between pixel values to determine a new transformed pixel value.
  • the onboard computer 150 may perform dithering to simulate blurring at image edges.
  • the onboard computer may send control commands (e.g., to the contextual lighting system 160 ).
  • the onboard computer 150 may additionally or alternatively function to control general properties of the light source 110 or of other aspects of the display 100 ; for example, the onboard computer 150 may control brightness of light source 110 pixels to simulate changes of opacity in a displayed image.
  • the functions described as performed by the onboard computer 150 may additionally or alternatively be performed by another computer system (e.g., a distributed computing system in the cloud).
  • the onboard computer 150 is communicative with another electronic device (e.g., a smartphone, a tablet, a laptop computer, a desktop computer, etc.) over a wired and/or wireless communication connection.
  • another electronic device e.g., a smartphone, a tablet, a laptop computer, a desktop computer, etc.
  • data may be streamed or otherwise communicated between the onboard computer 150 and the other electronic device.
  • a smartphone may transmit video information to the onboard computer, where it is sliced into depth slices by the onboard computer 150 .
  • depth slicing may be performed by the other electronic device.
  • the task of image processing may be performed and/or split between any number of electronic devices communicative with the onboard computer 150 .
  • the contextual lighting system 160 functions to light the periphery of the display 100 (or nearby area) with a light meant to match or resemble lighting conditions programmed into digital imagery displayed by the display 100 . By doing so, the contextual lighting system 160 can ‘lock’ the imagery in the real world for some users; for example, a user's hand may be lit to match the lighting of a particular part of a digital scene near the user's hand. This may substantially increase immersiveness.
  • the contextual lighting system 160 may control lighting properties (e.g., color, duration, intensity, direction, degree of focus, collimation, etc.) based on explicit instructions in the digital imagery. Additionally or alternatively, the contextual lighting system 160 may control lighting properties in any manner. For example, the contextual lighting system 160 may (for digital images without explicit contextual lighting instructions) average the color across a subset of an image and light the display 100 with this light.
  • lighting properties e.g., color, duration, intensity, direction, degree of focus, collimation, etc.
  • the contextual lighting system 160 may include any number and/or type of lighting devices; for example, color controllable LEDs.
  • the contextual lighting system 160 is preferably controlled by the onboard computer 150 , but may additionally or alternatively be controlled by any controller or computer system.
  • the display 100 may also include means for interaction tracking.
  • the display 100 may include a depth camera that tracks user interaction with the display 100 , allowing control and/or manipulation of the image displayed based on hand gestures and/or other interaction between a viewer and the display 100 as measured by the depth camera.
  • the display 100 may include a transparent touch sensor that tracks viewer touch interactions on surfaces of the display 100 .
  • the display 100 includes an ultrasonic haptic feedback module and a head tracker (e.g., a camera or other device that tracks head position, orientation, and/or motion).
  • a head tracker e.g., a camera or other device that tracks head position, orientation, and/or motion
  • tactile feedback via the haptic feedback module may be modified according to head tracking data (or other data, e.g., hand tracking data, body tracking data, video/audio capture data, etc.).
  • Tactile feedback may also be provided by haptic gloves that are coordinated through the onboard computer 150 to provided tactile feedback that is coincident with the visual feedback of the system.
  • the display 100 includes an infrared-opaque wand for interaction with aerial display (e.g., the wand is air gap or water containing, or of an IR blocking but visible-light transparent plastic or glass).
  • This wand functions as an interaction instrument (in addition to a user's bare hands) that can be read by a depth camera, but which does not block the light of the aerial image like a visible-light interaction instrument or a hand would, in the case of interaction that extends past the plane of the aerial image.
  • the wand may feature an infrared reflector and/or light emitter to better enable tracking.
  • the wand may include internal gyros and accelerometers to provide six degree-of-freedom tracking.
  • the display 100 may additionally or alternatively include voice control (e.g., via an automated assistant such as Amazon's Alexa).
  • voice control e.g., via an automated assistant such as Amazon's Alexa.
  • Tracking and interaction are preferably controlled by the onboard computer 150 , but may additionally or alternatively be controlled by any controller or computer system.
  • the display is typically optimized for a fixed number of viewers at a set distance.
  • a viewer in a traditional stereoscopic display featuring a lenticular lens, a viewer must remain within a bounded volume (both in terms of viewing distance, along the viewing axis, and distance perpendicular to the viewing axis). This is a huge disadvantage to these systems.
  • head tracking or other mechanism to infer the location of a user's eyes as they move in space
  • some stereoscopic displays modify the projected image to extend the viewing volume.
  • the display 100 is able to address this issue in a substantially more robust manner.
  • the display 100 may compensate for viewing distance (as determined by a head tracker or other mechanism for determining the distance of a viewer from the display 100 , such as a camera) by modifying the image projected to the parallax generator 120 (and/or by modifying the optical properties of the parallax generator 120 itself) at the onboard computer 150 (or otherwise).
  • Viewing distance is important because the light projected by a lenticular lens follows constant angle (so the lateral space between rays increases with viewing distance). An example of this is as shown in FIGS. 10A and 10B .
  • a user may perceive a first image (e.g., corresponding to pixel columns 0, 10, 20, etc.) with one eye and a second image with the other eye (e.g., corresponding to pixel columns 1, 11, 21, etc.). This is similar to viewing a traditional lenticular lens-based stereoscopic display at optimal viewing distance.
  • the user may perceive several images with each eye (as the lateral distance between views is closer). For example, a user may perceive three images (e.g., corresponding to 0, 10, 20 . . . ; 1, 11, 21 . . . ; 2, 12, 22) with a first eye and three images (e.g., corresponding to 2, 12, 22 . . .
  • the display 100 is capable of adapting to this change in viewing angle in one or more ways. For instance, the display 100 may dim or turn off pixel columns that would appear brightly to both eyes of an individual viewer. Even if views do not overlap, note that in the closer viewing distance each eye is capable of seeing more pixels.
  • the display 100 may adapt to this in any of several manners; for example, the display 100 may increase the resolution of a displayed image without changing the pictured scene, either via interpolation, by selecting a higher-resolution source, or if the image is generated in real-time, changing the resolution of the generated image.
  • the lenticular lens may noticeably distort images (e.g., stretching them horizontally or otherwise changing aspect ratio) in a viewing-distance dependent manner.
  • the display 100 may adapt to this issue by rescaling the displayed image either vertically or horizontally to correct the perceived aspect ratio by a viewer at a set distance.
  • the display 100 may additionally or alternatively modify image output to preserve a given depth perception.
  • the parallax generator 120 may also be desirable to modify views displayed by the display 100 to reflect vertical movement (e.g., a user changing viewing height).
  • the display 100 may do this by detecting a y-axis viewing disparity (e.g., a viewing height relative to some reference) and changing the perspective of the displayed image accordingly. For example, a user with a viewing height above a reference might see a scene from a slightly elevated angle, whereas when that user squats the user might see the scene from an even or depressed angle.
  • the display 100 is generating the view in real time from a 3D source, this may be as simple as changing the 2D output of the light source 110 to reflect the change in angle. If the source is limited in possible perspectives, the display 100 may additionally or alternatively distort the 2D output of the light source 110 to simulate a change in elevation based on user viewing height.
  • the display 100 may additionally track multiple viewers at the same time. In some senses this is similar to the single-user case—just as “views” (distinct images projected at different angles) may be allocated to a single user's eyes dependent on viewing distance, they may in general be allocated to multiple users.
  • views may be allocated to a single user's eyes dependent on viewing distance, they may in general be allocated to multiple users.
  • cross-talk i.e., that users may see images not intended to reach their eyes (because they are off-angle). While ghost images are typically dimmer than primary images (those intended to reach a user's eyes) they may still cause blurring.
  • the display 100 may modify the light output of the display 100 to provide a continuous view even across the boundaries of “natural view” zones.
  • the display 100 may track (based on user viewing position) what views are shown to a user, and update display output as a user approaches a view boundary. For example, in a lenticular lens that has a period of ten pixel columns, as the user approaches the tenth pixel column, the display 100 may change the output of pixel column 1 to appear as “pixel column 11” (assuming the goal is to maintain a continuous view around a scene).
  • the display 100 may divide views across multiple users in any manner and may apply the various image refinement techniques discussed above to views presented to these users in any manner.
  • the display 100 may additionally or alternatively smooth transitions between applications of these image refinements. This may be particularly important in the case of “view crash”, when one user, previously shown an image in a modified fashion (e.g., projected to provide perception of a first viewing height) moves into the natural view zone of another user being shown a different perspective of that image (e.g., projected to provide perception of a second, different viewing height).
  • the display 100 may detect that a view crash is likely to occur, and may change the output presented to one or more users to (ideally) bring the views presented to different users to unity as the users' views “crash” (i.e., begin to intersect).
  • the display 100 may damp transitions. This may be particularly useful in scenarios where head tracking produces erroneous values: transition damping may prevent views from rapidly changing if erroneous values occur. Also, if the display 100 loses head tracking lock on a viewer, the display 100 may attempt to perform dead reckoning until lock is regained.
  • aspects of displayed images may be modified based on environmental factors (e.g., temperature, humidity, altitude, etc.). Aspects of displayed images may also be modified to maintain compatibility between images in implementations of the display 100 featuring multiple light sources (e.g., as in an array of displays 100 ).
  • off-axis views in a high-index volume may appear at different perceived depths (due to the different distance traveled by light within the high-index volume with angle, at least for a non-radially-symmetric volume), so the image refinement techniques above (and display in general) may be additionally modified to account for the geometry and optical properties (particularly, index of refraction) of the high-index optical volume 140 .

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • General Physics & Mathematics (AREA)
  • Optics & Photonics (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

A superstereoscopic display with enhanced off-angle separation includes a first light source; a lenticular lens optically coupled to the first light source that, with the first light source, generates a first light output having viewing angle dependency; and a high-index optical volume optically coupled to the lenticular lens.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This invention claims the benefit of U.S. Provisional Application No. 62/635,728, filed on 27 Feb. 2018, and of U.S. Provisional Application No. 62/661,605, filed on 23 Apr. 2018, all of which are incorporated in their entireties by this reference.
TECHNICAL FIELD
This invention relates generally to the image display field, and more specifically to new and useful superstereoscopic displays with enhanced off-angle separation.
BACKGROUND
Image displays are an integral part of modern life. From televisions to monitors to smartphone and tablet screens, image displays provide users with the ability to view and interact with information presented in a variety of forms.
The advent of three-dimensional displays has enabled users to experience images with higher realism than would be possible with their two-dimensional counterparts; however, the vast majority of 3D displays require the use of a head-mounted display (HMD) or other cumbersome peripheral.
Free-space 3D displays remove the need for an HMD, allowing multiple users to see and manipulate content in a shared experience. Unfortunately, the few existing free-space 3D displays are hampered by a number of issues, including size, limited view angle, low resolution and brightness, scene distortion, and high cost. Thus, there exists a need in the image display field to create new and useful superstereoscopic displays with enhanced off-angle separation. This invention provides such new and useful displays.
BRIEF DESCRIPTION OF THE FIGURES
FIG. 1 is a top-down view of a display of an invention embodiment;
FIGS. 2A and 2B are stereoscopic views of a three-dimensional image;
FIGS. 3A and 3B are cross-sectional views of a display of an invention embodiment;
FIG. 4 is a cross-sectional view of a parallax generator of a display of an invention embodiment;
FIG. 5A is a cross-sectional view of a lenticular lens in a stretched configuration of a display of an invention embodiment;
FIG. 5B is a cross-sectional view of a lenticular lens in a compressed configuration of a display of an invention embodiment;
FIG. 6A is a light path view of an image projected from a light source of a display of an invention embodiment;
FIG. 6B is a light path view of an image projected from a light source and passing through a high-index optical volume of a display of an invention embodiment;
FIG. 7 is a top-down view of a display of an invention embodiment;
FIG. 8 is a perspective view of a high-index optical volume and perceived additional volume of an invention embodiment;
FIGS. 9A, 9B, and 9C are various perspective views of a display of an invention embodiment; and
FIGS. 10A and 10B are top-down views of image view dependency on viewing distance of a display of an invention embodiment.
DESCRIPTION OF THE INVENTION EMBODIMENTS
The following description of the invention embodiments of the invention is not intended to limit the invention to these invention embodiments, but rather to enable any person skilled in the art to make and use this invention.
1. Superstereoscopic Display with Enhanced Off-Angle Separation
A superstereoscopic display 100 with enhanced off-angle separation includes a light source 110, a parallax generator 120, and a high-index optical volume 140, as shown in FIG. 1. The display 100 may additionally or alternatively include polarizers 130, quarter waveplates 131, optical volume guides 141, an onboard computer 150, and/or a contextual lighting system 160.
As shown in FIG. 1, the display 100 functions to enable viewers to see two- and/or three-dimensional image data from multiple perspectives at the same time. The display 100 generates an angle-dependent viewing experience, which can be used to provide a three-dimensional viewing experience (via stereopsis) and/or to provide viewers at different angles with different images depending on viewing angle (without necessarily causing the perception of depth). An example is as shown in FIGS. 2A and 2B. While the display 100 is capable of producing the same 3D image at multiple angles (e.g., viewer 1 has eyes at 0 and 5 degrees, which see image 1 and 2 respectively; viewer 2 has eyes at 20 and 25 degrees, which see image 1 and 2 respectively), the display 100 preferably produces a continuous unbroken scene (e.g., image 1 at 0 degrees, image 2 at 5 degrees . . . image N at 5(N−1) degrees where N is an integer) to better provide an immersive viewing experience. Alternatively, the display 100 may utilize any set of images for any purpose.
Enhanced off-angle separation is preferably enabled by the high-index optical volume 140; by nature of its high index of refraction and shape, the optical volume 140 may cause the primary image(s) intended for a viewer at a particular viewing angle to appear closer to the viewer than otherwise would occur (without the optical volume 140), increasing the perception that the image produced by the light source 110 “floats” off of the screen. Further enhancing this effect, for images off-axis to the viewer at a particular viewing angle, the images may not appear as close as on-axis images do, appearing to the viewer to separate on- and off-axis images in space. This phenomenon is discussed in greater detail in the section on the high-index optical volume 140.
The light source 110 functions to generate images (i.e., light generated from image data) for display by the display 100. The light source 110 is preferably a planar two-dimensional display comprising a set of individually addressable pixels, but may additionally or alternatively be any suitable display. For example, the light source 110 may comprise one or more movable light sources; e.g., a laser that may be scanned across a set of positions to simulate the appearance of multiple light sources (i.e., display multiplexing).
The light source 110 is preferably an RGB color light source (e.g., each pixel includes red, green, and blue subpixels) but may additionally or alternatively be a substantially monochromatic light source or any other light source (e.g., a white light source).
The light source 110 is preferably a projector or projector light engine (e.g., DLP, laser, LCoS, and/or LCD projector) but may additionally or alternatively be any suitable display (e.g., an LCD monitor/TV display, an OLED display, an e-ink display, an LED array, a spinning LED display, an e-ink display, an electroluminescent display, a neon display, etc.). In one variation of a preferred embodiment, the light source 110 includes a liquid crystal panel with a collimated backlight.
The display 100 preferably includes a single light source 110, but may additionally or alternatively include multiple light sources 110. For example, multiple light sources 110 may be placed in series and/or in parallel as shown in FIG. 3A and FIG. 3B respectively. Note that any combination of light sources 110, polarizers 130, quarter-waveplates 131, and/or additional optics (e.g., mirrors, lenses, etc.) may be used in the display 100. In configurations utilizing multiple light sources 110, the light sources 110 may be offset, angled, rotating, curved, or otherwise configured in any manner.
The light source 110 may include optical elements (e.g., lenses, mirrors, waveguides, filters) that function to couple light into the parallax generator 120 and/or the high-index optical volume 140. For example, the light source 110 may include a collimating lens designed to increase collimating of the light source 110 output. As a second example, the light source 110 may include a lens designed to scale (or otherwise distort) light source 110 output (e.g., reduce in size or increase in size). Such a lens may scale light source 110 output uniformly (e.g., 2× decrease in both image dimensions) or non-uniformly (e.g., no decrease in first image dimension, 4× decrease in other image dimension). As a third example, the light source 110 may include a lens that manipulates the focal plane of the viewed image; such a lens may be tunable (allowing depth of field to be swept). If such a lens is tunable at a high rate, this may provide an expanded perceived depth of field to a viewer.
The light source 110 may additionally or alternatively include any passive or active optical elements to prepare light for use by the display 100 for any other purpose. For example, the light source 110 may include filters or splitters. As a more specific example, the light source 110 may include a polarizing filter if the native output of the light source 110 is unpolarized. As another example, the light source 110 may include microlens arrays and/or Fresnel lenses.
The parallax generator 120 functions to generate an angle-dependent view from the light source 110 output. The parallax generator 120 may be any structure capable of generating such an angle-dependent view (e.g., lenticular lenses, micro-spherical lenses, parallax barriers, etc.).
In a first example, the parallax generator 120 is a lenticular lens. The lenticular lens generates an angle dependent view in combination with a light source 110 that displays different pixels (or image segments) based on the positioning of the lenticular lens over the light source 110. For example, as shown in FIG. 4, the image shown to viewers at three angles is comprised of the pixels labeled 1, 2, and 3, respectively.
Based on the properties of the lenticular lens (e.g., pitch, material, structure, orientation and position relative to the light source 110) and desired viewing characteristics (e.g., number of viewers, view distance, number of views desired, viewing mode, etc.), the display 100 may modify the output of the light source 110 to produce a desired result.
In one example embodiment, the number of different views provided by the display 100 is sufficient for superstereoscopic viewing at some viewing distance; that is, each eye of the viewer receives a different image from the display 100, and as the viewer moves around the display 100, the views change (with the viewing angle). For example, a viewer at angle one may see scene 1 with a right eye and scene 2 with a left eye, where scene 1 and scene 2 create a stereoscopic three-dimensional view of one or more objects. After the viewer moves from angle one to angle two, the viewer now sees scene 2 with the right eye and scene 3 with the left eye, producing a second stereoscopic three-dimensional view of the one or more objects. In this manner, the viewer perceives a three-dimensional image (thanks to the stereoscopic effect) at a given viewing angle, and that perception is preserved (thanks to the changing views, which correspond to a rotated view of the one or more objects) as the viewer moves around the display 100, as shown in FIG. 2A (corresponding to a first view) and FIG. 2B (corresponding to a second view). A display that produces more than one stereoscopic three-dimensional view in this manner may be referred to as a superstereoscopic display.
The lenticular lens may have any suitable configuration and structure and may be made of any suitable material. The lenticular lens is preferably one-dimensional (e.g., cylindrical lenses arranged in columns), but may additionally or alternatively be a two-dimensional lenticular lens, fly-eye lens array, or integral imaging lens array. Note that while there is preferably a correlation between addressable segments (e.g., pixels) of the light source 110 and the lenticular lens, the lens columns of the lenticular lens need not be at a particular orientation relative to the light source 110. For example, while columns of the lenticular lens may be aligned with pixel columns, they may also be offset at an angle (which allows the resolution loss due to image slicing to be apportioned across both image pixels columns and rows, rather than only one of these). This technique is described further in U.S. Pat. No. 6,064,424. Image slicing or division (of light source 110 output) may be accomplished in any manner to achieve a desired viewing result. Processing of the image is preferably performed by the onboard computer 150 but may additionally or alternatively be controlled by any computer system.
Note that lenticular lenses may be reconfigurable to change the optical properties of the lenses. For example, a lenticular lens may be fabricated of a flexible or semi-flexible material so that the lens can be physically compressed or stressed to change the pitch of the lens, as shown in FIGS. 5A (stretched) and 5B (compressed), and the pixel-to-lens mapping may be changed accordingly. Additionally or alternatively, the optical properties of the lens may be altered by another mechanism (e.g., by modifying the index of refraction of the lenticular lens). As another example, a lenticular lens may be implemented using a layered liquid crystal array (either as the entire lens or in conjunction with a fixed refractive lens), enabling dynamic configuration of the lenticular lens.
The display 100 may include multiple lenticular lenses and/or other lenses to produce a desired optical effect. For example, 1D lenticular lenses may be stacked at different orientations to create 2D angular view dependence.
In a second example, the parallax generator 120 is a pinhole array or parallax barrier. In a third example, parallax generation may be provided by multiple light sources 110. For example, a set of projectors at different angles and a viewing angle-dependent film or surface (together functioning as both the parallax generator 120 and the light source 110) may be used to produce a similar viewing effect to a single light source 110 and an overlaid parallax generator 120.
Polarizers 130 and quarter waveplates 131 may be used to enhance the quality of the image output of the display 100. For example, in the configuration as shown in FIG. 1, the polarizers 130 and quarter waveplate 131 may increase perceived image contrast or brightness (by suppressing errant reflection). This same effect may be used, for example, to reduce any image ghosting caused by the parallax generator 120 (in this case, the ‘ghost’ images may be of a different polarization than intended/on-axis images).
While polarizers 130 and quarter waveplates 131 may be laminated onto the optical volume 140, it may be desirable for these (and other) optical components to be separated by some material (or air/vacuum) in order to prevent the loss of light from the optical volume 140 that may result from frustrated total internal reflection (FTIR) since the polarizers 130/quarter waveplates 131 may have an index between that of the optical volume 140 and the surrounding air.
The high-index optical volume 140 functions to enhance viewing of the display 100 by enhancing perceived separation between on-angle views (i.e., primary views) and off-angle views.
The high-index optical volume 140 is preferably a solid rectangular prism of an optically clear material (e.g., acrylic, glass, polycarbonate), but may additionally or alternatively be any three-dimensional volume (made of any materials in any structure) capable of transmitting light and having an index of refraction greater than one. For example, the high-index optical volume may be acrylic and have a relative index of refraction of 1.49.
As shown in FIG. 6A, by itself, a light source 110 can create an image on the eye of a viewer. After incorporating the high-index optical volume 140, the image formed on the viewer's eye is smaller, suggesting that the perceived image with the optical volume 140 in place is either closer (e.g., as shown by the perceived location of image) or larger than it would be without optical volume. Given appropriate relative size clues (e.g., physical boundaries of the display) it is likely that a viewer may perceive the image as closer (e.g., ‘floating’ within the optical volume 140), as shown in FIG. 6B. The optical volume 140 may have the additional benefit of increasing viewing angle of the display 100 (by bending light toward on-axis viewing).
In some configurations of the optical volume 140 (e.g., a rectangular prism of uniform index of refraction), this effect is less prominent at viewing angles off-axis (e.g., not perpendicular to the light source 110 and optical volume 140), as shown in FIG. 7. This may result in separation between on-angle and off-angle views, further enhancing the depth effect enabled by stereoscopic images.
The display 100 may additionally include optical volume guides 141, which function to enhance the ‘floating’ effect of images within the optical volume 140. The optical volume guides 141 are preferably markings visible on the optical volume 140 (or otherwise within the optical path of the light source 110) that draw a viewer's eye to aid in providing the sensation of depth to images of the display 100.
The optical volume guides 141 may be any visible two- or three-dimensional structure present in a plane distinct from (and preferably parallel to) the perceived plane locating a primary/on-axis image viewed by a viewer. For example, the optical volume guides 141 may be etched, painted, adhered, or printed onto the surface of the optical volume 140. Additionally or alternatively, the optical volume guides 141 may be located on a separate light guide, substrate, or on any other component.
In one implementation of an invention embodiment, the optical volume guides 141 are reflective on at least one surface. In this implementation, the system 100 incorporates a reflective polarizer 130 (or other reflective or partially reflective surface) to create a virtual image of the optical volume guides 141, creating the perception of a volume larger (e.g., twice as large) than the optical volume 140's originally perceived size, as shown in FIG. 8 (the optical volume 140's originally perceived size may be smaller than its actual size due to the use of high-index material).
Note that the use of optical volume guides 141 may also enable the three-dimensional effect of the display 100 to be better perceived in (two-dimensional) video recordings than other displays, allowing for the full impact of display 100 to be better communicated over film.
The display 100 may additionally or alternatively use opaque elements (e.g., portions of an opaque housing) to surround or partially surround sides of the optical volume 140, further enhancing the perception of depth (and potentially also reducing the presence of undesired light), as shown in FIG. 9A. Additional views of this example implementation are as shown in FIG. 9B and FIG. 9C (including optional optical volume guides 141).
Likewise, the display 100 may be used with other displays 100 in any manner (e.g., in a 2×2 or 3×3 array, back to back).
The onboard computer 150 functions to perform image processing for image data received by the display 100 prior to display by the light source 110. For example, the onboard computer may separate 3D model information into slices to be projected by the light source 110. The onboard computer 150 may additionally or alternatively function to prepare 3D image data for voxel representation in any manner. For example, the onboard computer 150 may generate 2D stereoscopic views based on the structure/configuration of the parallax generator 120. As another example, if light folding is performed by the display 100 (i.e., images are sliced and anisotropically scaled), the onboard computer 150 may perform interpolation between pixel values to determine a new transformed pixel value. As another example, the onboard computer 150 may perform dithering to simulate blurring at image edges. As a third example, the onboard computer may send control commands (e.g., to the contextual lighting system 160).
The onboard computer 150 may additionally or alternatively function to control general properties of the light source 110 or of other aspects of the display 100; for example, the onboard computer 150 may control brightness of light source 110 pixels to simulate changes of opacity in a displayed image.
Note that the functions described as performed by the onboard computer 150 may additionally or alternatively be performed by another computer system (e.g., a distributed computing system in the cloud).
In one implementation of an invention embodiment, the onboard computer 150 is communicative with another electronic device (e.g., a smartphone, a tablet, a laptop computer, a desktop computer, etc.) over a wired and/or wireless communication connection. In this implementation, data may be streamed or otherwise communicated between the onboard computer 150 and the other electronic device. For example, a smartphone may transmit video information to the onboard computer, where it is sliced into depth slices by the onboard computer 150. Additionally or alternatively, depth slicing may be performed by the other electronic device. In general, the task of image processing may be performed and/or split between any number of electronic devices communicative with the onboard computer 150.
The contextual lighting system 160 functions to light the periphery of the display 100 (or nearby area) with a light meant to match or resemble lighting conditions programmed into digital imagery displayed by the display 100. By doing so, the contextual lighting system 160 can ‘lock’ the imagery in the real world for some users; for example, a user's hand may be lit to match the lighting of a particular part of a digital scene near the user's hand. This may substantially increase immersiveness.
The contextual lighting system 160 may control lighting properties (e.g., color, duration, intensity, direction, degree of focus, collimation, etc.) based on explicit instructions in the digital imagery. Additionally or alternatively, the contextual lighting system 160 may control lighting properties in any manner. For example, the contextual lighting system 160 may (for digital images without explicit contextual lighting instructions) average the color across a subset of an image and light the display 100 with this light.
The contextual lighting system 160 may include any number and/or type of lighting devices; for example, color controllable LEDs.
The contextual lighting system 160 is preferably controlled by the onboard computer 150, but may additionally or alternatively be controlled by any controller or computer system.
The display 100 may also include means for interaction tracking. For example, the display 100 may include a depth camera that tracks user interaction with the display 100, allowing control and/or manipulation of the image displayed based on hand gestures and/or other interaction between a viewer and the display 100 as measured by the depth camera. As another example, the display 100 may include a transparent touch sensor that tracks viewer touch interactions on surfaces of the display 100.
In one implementation of a preferred embodiment, the display 100 includes an ultrasonic haptic feedback module and a head tracker (e.g., a camera or other device that tracks head position, orientation, and/or motion). In this implementation, tactile feedback via the haptic feedback module may be modified according to head tracking data (or other data, e.g., hand tracking data, body tracking data, video/audio capture data, etc.). Tactile feedback may also be provided by haptic gloves that are coordinated through the onboard computer 150 to provided tactile feedback that is coincident with the visual feedback of the system.
In another implementation of a preferred embodiment, the display 100 includes an infrared-opaque wand for interaction with aerial display (e.g., the wand is air gap or water containing, or of an IR blocking but visible-light transparent plastic or glass). This wand functions as an interaction instrument (in addition to a user's bare hands) that can be read by a depth camera, but which does not block the light of the aerial image like a visible-light interaction instrument or a hand would, in the case of interaction that extends past the plane of the aerial image. Additionally or alternatively, the wand may feature an infrared reflector and/or light emitter to better enable tracking. In some example, the wand may include internal gyros and accelerometers to provide six degree-of-freedom tracking.
The display 100 may additionally or alternatively include voice control (e.g., via an automated assistant such as Amazon's Alexa).
Tracking and interaction are preferably controlled by the onboard computer 150, but may additionally or alternatively be controlled by any controller or computer system.
Note that while the components of the display 100 are shown in particular configurations, it is understood the that components of the display 100 may be coupled/ordered in any manner to produce the effects as described in the present application.
2. Superstereoscopic Display Image Processing
In traditional stereoscopic displays, the display is typically optimized for a fixed number of viewers at a set distance. For example, in a traditional stereoscopic display featuring a lenticular lens, a viewer must remain within a bounded volume (both in terms of viewing distance, along the viewing axis, and distance perpendicular to the viewing axis). This is a huge disadvantage to these systems. With the advent of head tracking (or other mechanism to infer the location of a user's eyes as they move in space), some stereoscopic displays modify the projected image to extend the viewing volume.
The display 100 is able to address this issue in a substantially more robust manner. For example, the display 100 may compensate for viewing distance (as determined by a head tracker or other mechanism for determining the distance of a viewer from the display 100, such as a camera) by modifying the image projected to the parallax generator 120 (and/or by modifying the optical properties of the parallax generator 120 itself) at the onboard computer 150 (or otherwise). Viewing distance is important because the light projected by a lenticular lens follows constant angle (so the lateral space between rays increases with viewing distance). An example of this is as shown in FIGS. 10A and 10B. At a first viewing distance, a user may perceive a first image (e.g., corresponding to pixel columns 0, 10, 20, etc.) with one eye and a second image with the other eye (e.g., corresponding to pixel columns 1, 11, 21, etc.). This is similar to viewing a traditional lenticular lens-based stereoscopic display at optimal viewing distance. At a second, closer, distance, the user may perceive several images with each eye (as the lateral distance between views is closer). For example, a user may perceive three images (e.g., corresponding to 0, 10, 20 . . . ; 1, 11, 21 . . . ; 2, 12, 22) with a first eye and three images (e.g., corresponding to 2, 12, 22 . . . ; 3, 13, 23 . . . ; 4, 14, 22) with a second eye-note that the images may overlap at close distances. The display 100 is capable of adapting to this change in viewing angle in one or more ways. For instance, the display 100 may dim or turn off pixel columns that would appear brightly to both eyes of an individual viewer. Even if views do not overlap, note that in the closer viewing distance each eye is capable of seeing more pixels. The display 100 may adapt to this in any of several manners; for example, the display 100 may increase the resolution of a displayed image without changing the pictured scene, either via interpolation, by selecting a higher-resolution source, or if the image is generated in real-time, changing the resolution of the generated image. Another issue with changing viewing distance is that the lenticular lens may noticeably distort images (e.g., stretching them horizontally or otherwise changing aspect ratio) in a viewing-distance dependent manner. The display 100 may adapt to this issue by rescaling the displayed image either vertically or horizontally to correct the perceived aspect ratio by a viewer at a set distance.
Note also that viewing distance affects perceived depth (due to differences in separation of the stereoscopic images); the display 100 may additionally or alternatively modify image output to preserve a given depth perception.
To the extent that the parallax generator 120 varies views horizontally (e.g., as with a columnar lenticular lens), it may also be desirable to modify views displayed by the display 100 to reflect vertical movement (e.g., a user changing viewing height). The display 100 may do this by detecting a y-axis viewing disparity (e.g., a viewing height relative to some reference) and changing the perspective of the displayed image accordingly. For example, a user with a viewing height above a reference might see a scene from a slightly elevated angle, whereas when that user squats the user might see the scene from an even or depressed angle. If the display 100 is generating the view in real time from a 3D source, this may be as simple as changing the 2D output of the light source 110 to reflect the change in angle. If the source is limited in possible perspectives, the display 100 may additionally or alternatively distort the 2D output of the light source 110 to simulate a change in elevation based on user viewing height.
These are examples of accommodations that the display 100 can make when tracking a user's head. The display 100 may additionally track multiple viewers at the same time. In some senses this is similar to the single-user case—just as “views” (distinct images projected at different angles) may be allocated to a single user's eyes dependent on viewing distance, they may in general be allocated to multiple users. In the single user case, a primary concern with views generally outside of perception is cross-talk—i.e., that users may see images not intended to reach their eyes (because they are off-angle). While ghost images are typically dimmer than primary images (those intended to reach a user's eyes) they may still cause blurring. In the single user case, it may be possible simply to turn off views that cause ghosting (or to time multiplex views that are reaching a viewer's eyes). This may still be possible in the multi-user case as well, but the “ghost” views for one person may be the primary views for another, and thus this is a higher complexity issue. Another primary concern with views is the distinction between “natural” and “artificial” views. As an individual moves his or head laterally, the user sees sequentially different views (even though these views may be identical)—this is a function of the lenticular lens—but eventually the views shown to the user repeat unless the light output of the display 100 changes (due to an effect known as ray crossover). These views (those that exist without changing the light output of the display 100) are “natural” views. Additionally, the display 100 may modify the light output of the display 100 to provide a continuous view even across the boundaries of “natural view” zones. In other words, the display 100 may track (based on user viewing position) what views are shown to a user, and update display output as a user approaches a view boundary. For example, in a lenticular lens that has a period of ten pixel columns, as the user approaches the tenth pixel column, the display 100 may change the output of pixel column 1 to appear as “pixel column 11” (assuming the goal is to maintain a continuous view around a scene).
This principle applies in the multi-user scenario as well, except that views are now divided across users (to the extent that different images need to be shown to different users, such as if they are at different viewing distances, viewing heights, or are intended to see different content). The display 100 may divide views across multiple users in any manner and may apply the various image refinement techniques discussed above to views presented to these users in any manner.
The display 100 may additionally or alternatively smooth transitions between applications of these image refinements. This may be particularly important in the case of “view crash”, when one user, previously shown an image in a modified fashion (e.g., projected to provide perception of a first viewing height) moves into the natural view zone of another user being shown a different perspective of that image (e.g., projected to provide perception of a second, different viewing height). In such a scenario, the display 100 may detect that a view crash is likely to occur, and may change the output presented to one or more users to (ideally) bring the views presented to different users to unity as the users' views “crash” (i.e., begin to intersect).
As another example of transition smoothing, the display 100 may damp transitions. This may be particularly useful in scenarios where head tracking produces erroneous values: transition damping may prevent views from rapidly changing if erroneous values occur. Also, if the display 100 loses head tracking lock on a viewer, the display 100 may attempt to perform dead reckoning until lock is regained.
Additionally or alternatively, aspects of displayed images may be modified based on environmental factors (e.g., temperature, humidity, altitude, etc.). Aspects of displayed images may also be modified to maintain compatibility between images in implementations of the display 100 featuring multiple light sources (e.g., as in an array of displays 100).
Note that, as shown in FIG. 7, off-axis views in a high-index volume may appear at different perceived depths (due to the different distance traveled by light within the high-index volume with angle, at least for a non-radially-symmetric volume), so the image refinement techniques above (and display in general) may be additionally modified to account for the geometry and optical properties (particularly, index of refraction) of the high-index optical volume 140.
As a person skilled in the art will recognize from the previous detailed description and from the figures and claims, modifications and changes can be made to the preferred embodiments of the invention without departing from the scope of this invention defined in the following claims.

Claims (25)

We claim:
1. A superstereoscopic display with enhanced off-angle separation comprising:
a first light source;
a lenticular lens optically coupled to the first light source that, with the first light source, generates a first light output having viewing angle dependency;
a high-index optical volume optically coupled to the lenticular lens; wherein the optical volume has an index of refraction greater than one; and
a set of optical volume guides located at a first surface of the high-index optical volume; wherein the set of optical volume guides enhances depth perception of images displayed by the display;
wherein the first light source transmits the first light output to the high-index optical volume; wherein the high-index optical volume transmits the first light output to free-space; wherein, after transmission by the high-index optical volume, the first light output comprises a first visible image at a first viewing angle and a second visible image, non-identical to the first visible image, at a second viewing angle separated by a first non-zero angle from the first viewing angle, and a third visible image non-identical to the first and second visible images, at a third viewing angle separated by a second non-zero angle from the second viewing angle; wherein the first, second, and third visible images correspond to views of a first three-dimensional image and enable superstereoscopic viewing of the first three-dimensional image; wherein the high-index optical volume decreases the perceived viewing distance of the first, second, and third visible images such that the first, second, and third visible images appear to be located within the high-index optical volume.
2. The superstereoscopic display of claim 1, wherein the high-index optical volume is a rectangular prism comprised of a transparent polymer having an index of refraction greater than 1.4.
3. The superstereoscopic display of claim 1, wherein the optical volume guides are etched into a surface of the high-index optical volume.
4. The superstereoscopic display of claim 1, further comprising a reflector located at a second surface, opposite the first surface, of the high-index optical volume; wherein the reflector creates a virtual image of the optical volume guides, creating a perception that the high-index optical volume is larger than it otherwise would appear.
5. The superstereoscopic display of claim 1, wherein the lenticular lens is a one-dimensional lenticular lens parallel to addressable columns of the first light source; wherein the one-dimensional lenticular lens is perpendicular to addressable rows of the first light source.
6. The superstereoscopic display of claim 5, wherein the lenticular lens is a one-dimensional lenticular lens oriented at an angle of more than zero but less than ninety degrees relative to addressable columns of the first light source, resulting in apportionment of resolution loss across both the addressable columns and addressable rows of the first light source.
7. The superstereoscopic display of claim 5, wherein the lenticular lens is a two dimensional lenticular lens resulting from stacking two one-dimensional lenticular lenses.
8. The superstereoscopic display of claim 5, further comprising a tracking sensor, wherein the first light source modifies the first light output to enable superstereoscopic viewing of the first three-dimensional image for a first viewer based on head or eye tracking data corresponding to the viewer and captured by the tracking sensor.
9. The superstereoscopic display of claim 8, wherein the first light source further modifies the first light output to reduce view ghosting in response to the first viewer moving closer to the display.
10. The superstereoscopic display of claim 8, wherein the first light source further modifies the first light output to correct for aspect ratio distortion in response to the first viewer moving closer to the display.
11. The superstereoscopic display of claim 8, wherein the first light source further modifies the first light output to correct for depth perception change in response to the first viewer moving closer to the display.
12. The superstereoscopic display of claim 8, wherein the first light source damps modifications of the first light output in response to changes in head or eye tracking data.
13. The superstereoscopic display of claim 5, wherein the first light source modifies the first light output to enable superstereoscopic viewing for a first viewer and a second viewer based on head or eye tracking data corresponding to the viewers and captured by the tracking sensor.
14. The superstereoscopic display of claim 13, wherein the first light source modifies the first light output to enable superstereoscopic viewing of the first three-dimensional image for the first viewer; wherein the first light source modifies the first light output to enable superstereoscopic viewing of a second three-dimensional image for the second viewer.
15. The superstereoscopic display of claim 14, wherein the first and second three-dimensional images are identical.
16. The superstereoscopic display of claim 15, wherein the first light source modifies the first light output to provide a first perspective of the first three-dimensional image to the first viewer based upon a viewing angle, viewing distance, and viewing height of the first viewer.
17. The superstereoscopic display of claim 15, wherein the first light source modifies the first light output to provide a second perspective of the first three-dimensional image to the second viewer based upon a viewing angle, viewing distance, and viewing height of the second viewer; wherein the first and second perspectives are non-identical.
18. The superstereoscopic display of claim 17, wherein the first light source initially modifies the first light output such that the first perspective is modified to reflect a first viewing height of the first viewer and such that the second perspective is modified to reflect a second viewing height of the second viewer; wherein the first and second viewing heights are nonidentical.
19. The superstereoscopic display of claim 18, wherein the first light source later modifies the first light output such that both of the first and second perspectives reflect the first viewing height of the first viewer in response to the second viewer moving toward the first viewer.
20. The superstereoscopic display of claim 1, wherein the optical volume guides are painted, adhered, or printed on a surface of the high-index optical volume.
21. The superstereoscopic display of claim 20, further comprising a reflector located at a second surface, opposite the first surface, of the high-index optical volume; wherein the reflector creates a virtual image of the optical volume guides.
22. The superstereoscopic display of claim 1, wherein the optical volume guides are reflective of visible light.
23. The superstereoscopic display of claim 22, further comprising a reflector located at a second surface, opposite the first surface, of the high-index optical volume; wherein the reflector creates a virtual image of the optical volume guides.
24. The superstereoscopic display of claim 1, wherein the optical volume guides are etched, painted, adhered, or printed on a surface of a light guide optically coupled to the high-index optical volume.
25. The superstereoscopic display of claim 20, further comprising a reflector located at a second surface, opposite the first surface, of the high-index optical volume; wherein the reflector creates a virtual image of the optical volume guides.
US16/044,355 2018-02-27 2018-07-24 Superstereoscopic display with enhanced off-angle separation Active US10298921B1 (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
US16/044,355 US10298921B1 (en) 2018-02-27 2018-07-24 Superstereoscopic display with enhanced off-angle separation
CN201980028768.0A CN112042186B (en) 2018-02-27 2019-01-25 Super-stereoscopic display with enhanced off-angle separation
JP2020567739A JP7096371B2 (en) 2018-02-27 2019-01-25 Super stereoscopic display with enhanced off-angle separation
EP19760138.8A EP3759908A4 (en) 2018-02-27 2019-01-25 Superstereoscopic display with enhanced off-angle separation
KR1020207027895A KR102416197B1 (en) 2018-02-27 2019-01-25 Hyperstereoscopic Display with Enhanced Off-Angle Separation
PCT/US2019/015235 WO2019168617A1 (en) 2018-02-27 2019-01-25 Superstereoscopic display with enhanced off-angle separation
US16/374,955 US11051006B2 (en) 2018-02-27 2019-04-04 Superstereoscopic display with enhanced off-angle separation
US17/328,076 US11683472B2 (en) 2018-02-27 2021-05-24 Superstereoscopic display with enhanced off-angle separation

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201862635728P 2018-02-27 2018-02-27
US201862661605P 2018-04-23 2018-04-23
US16/044,355 US10298921B1 (en) 2018-02-27 2018-07-24 Superstereoscopic display with enhanced off-angle separation

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/374,955 Continuation US11051006B2 (en) 2018-02-27 2019-04-04 Superstereoscopic display with enhanced off-angle separation

Publications (1)

Publication Number Publication Date
US10298921B1 true US10298921B1 (en) 2019-05-21

Family

ID=66541248

Family Applications (3)

Application Number Title Priority Date Filing Date
US16/044,355 Active US10298921B1 (en) 2018-02-27 2018-07-24 Superstereoscopic display with enhanced off-angle separation
US16/374,955 Active 2039-04-15 US11051006B2 (en) 2018-02-27 2019-04-04 Superstereoscopic display with enhanced off-angle separation
US17/328,076 Active 2039-04-21 US11683472B2 (en) 2018-02-27 2021-05-24 Superstereoscopic display with enhanced off-angle separation

Family Applications After (2)

Application Number Title Priority Date Filing Date
US16/374,955 Active 2039-04-15 US11051006B2 (en) 2018-02-27 2019-04-04 Superstereoscopic display with enhanced off-angle separation
US17/328,076 Active 2039-04-21 US11683472B2 (en) 2018-02-27 2021-05-24 Superstereoscopic display with enhanced off-angle separation

Country Status (6)

Country Link
US (3) US10298921B1 (en)
EP (1) EP3759908A4 (en)
JP (1) JP7096371B2 (en)
KR (1) KR102416197B1 (en)
CN (1) CN112042186B (en)
WO (1) WO2019168617A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220166963A1 (en) * 2020-09-30 2022-05-26 Magnetic Media Holdings, Inc. dba Magnetic 3D Method for optimized viewing experience and reduced rendering for autostereoscopic 3d, multiview and volumetric displays.
EP4124033A1 (en) 2021-07-23 2023-01-25 VitreaLab GmbH Stereoscopic display device

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10298921B1 (en) 2018-02-27 2019-05-21 Looking Glass Factory, Inc. Superstereoscopic display with enhanced off-angle separation
WO2021237065A1 (en) 2020-05-21 2021-11-25 Looking Glass Factory, Inc. System and method for holographic image display
WO2021243037A1 (en) * 2020-05-27 2021-12-02 Looking Glass Factory, Inc. System and method for holographic displays
US11936844B1 (en) 2020-08-11 2024-03-19 Apple Inc. Pre-processing in a display pipeline
WO2022119396A1 (en) * 2020-12-04 2022-06-09 Samsung Electronics Co., Ltd. Mid-air image display device and method of operating the same
CN113031299A (en) * 2021-03-29 2021-06-25 四川虹微技术有限公司 Desktop true three-dimensional display method
CN113552549B (en) * 2021-07-28 2023-05-12 北京环境特性研究所 Method and device for calibrating airborne downward-looking measurement by using active calibration equipment

Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6064424A (en) 1996-02-23 2000-05-16 U.S. Philips Corporation Autostereoscopic display apparatus
US6462871B1 (en) * 1997-02-18 2002-10-08 Canon Kabushiki Kaisha Stereoscopic image display apparatus using specific mask pattern
US20040165262A1 (en) * 2001-06-29 2004-08-26 Jose Javier Alejo Trevijano Autostereoscopic vision system
US6798390B1 (en) * 1997-08-29 2004-09-28 Canon Kabushiki Kaisha 3D image reconstructing apparatus and 3D object inputting apparatus
US20050007837A1 (en) * 2003-03-25 2005-01-13 Lsi Logic Corporation Clock generator for pseudo dual port memory
US20050078370A1 (en) * 2002-04-05 2005-04-14 Hiroshi Nishihara Stereoscopic image display apparatus and stereoscopic image display system
US20070016514A1 (en) * 2005-07-15 2007-01-18 Al-Abdulqader Hisham A System, program product, and methods for managing contract procurement
US20070164950A1 (en) * 2006-01-19 2007-07-19 Sanyo Epson Imaging Devices Corporation Image display and electronic device
US20070165145A1 (en) * 2006-01-19 2007-07-19 Sanyo Epson Imaging Devices Corporation Electro-optical apparatus and electronic equipment
US20120031389A1 (en) * 2010-08-05 2012-02-09 Mcmaster William J Oven door
US20120313896A1 (en) * 2011-06-07 2012-12-13 Sony Corporation Information processing apparatus, information processing method, and program
US20140011851A1 (en) * 2007-07-09 2014-01-09 Furiex Pharmaceuticals, Inc. Novel crystals and process of making 5-(-METHYL)-2-METHOXY-BENZOIC ACID
US20140118511A1 (en) * 2012-10-31 2014-05-01 Elwha Llc Systems and methods to confirm that an autostereoscopic display is accurately aimed
US20140320614A1 (en) * 2011-09-19 2014-10-30 Ecrans Polaires Inc. /Polar Screens Inc. Method and display for showing a stereoscopic image
US20160021365A1 (en) * 2014-07-18 2016-01-21 Au Optronics Corp. Image displaying method and image displaying device
US20160023448A1 (en) * 2013-03-14 2016-01-28 Corning Incorporated Methods and apparatus for fabricating and cutting flexible glass and polymer composite structures
US20160234487A1 (en) * 2013-09-30 2016-08-11 Koninklijke Philips N.V. Autostereoscopic display device
US20170013921A1 (en) * 2014-06-26 2017-01-19 Stephan Hugo Frei Quick lock for attaching one end of a band to an object
US20170078650A1 (en) * 2015-09-15 2017-03-16 Looking Glass Factory, Inc. Enhanced 3d volumetric display
US20170139213A1 (en) * 2015-10-25 2017-05-18 Oculus Vr, Llc Combination Prism Array for Focusing Light
US9686535B2 (en) * 2012-07-27 2017-06-20 Samsung Display Co., Ltd. Method of displaying three dimensional image and three dimensional display apparatus for performing the method
US20180035096A1 (en) * 2014-03-20 2018-02-01 Ipc Hk Limited Automated 3d photo booth

Family Cites Families (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4140370A (en) 1977-05-12 1979-02-20 Snaper Alvin A Planar optical viewing screens and lenses
US4340275A (en) 1980-06-09 1982-07-20 General Electric Company Rear projection screen with patterned lenticular prismatic structure
US4964695A (en) 1987-10-06 1990-10-23 North American Philips Corporation Lenticular arrays for front projection screens and contrast improving method and device
US5359454A (en) 1992-08-18 1994-10-25 Applied Physics Research, L.P. Apparatus for providing autostereoscopic and dynamic images
US5521724A (en) 1993-11-15 1996-05-28 Shires; Mark R. Real-time automultiscopic 3D video display using holographic optical elements (HOEs)
US5886675A (en) 1995-07-05 1999-03-23 Physical Optics Corporation Autostereoscopic display system with fan-out multiplexer
GB2307058A (en) 1995-11-13 1997-05-14 Thomson Multimedia Sa Stereoscopic display with lens,prism and barrier arrays
US6097394A (en) 1997-04-28 2000-08-01 Board Of Trustees, Leland Stanford, Jr. University Method and system for light field rendering
US6195184B1 (en) 1999-06-19 2001-02-27 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration High-resolution large-field-of-view three-dimensional hologram display system and method thereof
JP3770006B2 (en) 1999-11-01 2006-04-26 松下電器産業株式会社 Rear projection image display device
GB0119176D0 (en) 2001-08-06 2001-09-26 Ocuity Ltd Optical switching apparatus
GB2387664B (en) 2002-04-17 2005-08-24 Philip Anthony Surman Autostereoscopic display
US7364314B2 (en) 2002-05-15 2008-04-29 Reflexite Corporation Optical structures
US7000091B2 (en) 2002-08-08 2006-02-14 Hewlett-Packard Development Company, L.P. System and method for independent branching in systems with plural processing elements
NZ521505A (en) 2002-09-20 2005-05-27 Deep Video Imaging Ltd Multi-view display
GB2405519A (en) 2003-08-30 2005-03-02 Sharp Kk A multiple-view directional display
EP1776837A1 (en) * 2004-08-10 2007-04-25 SeeReal Technologies GmbH Sweet-spot image separation device for autostereoscopic multi-user displays
US8675125B2 (en) 2005-04-27 2014-03-18 Parellel Consulting Limited Liability Company Minimized-thickness angular scanner of electromagnetic radiation
CN101568888B (en) 2006-10-26 2013-01-30 视瑞尔技术公司 Mobile tlephony system comprising holographic display
US7715103B2 (en) 2007-09-10 2010-05-11 Microvision, Inc. Buried numerical aperture expander having transparent properties
US8068278B2 (en) 2007-09-19 2011-11-29 The Aerospace Corporation Photostructured imaging display panels
CN101903809B (en) 2007-12-21 2012-02-29 3M创新有限公司 Light control film
JP2010068202A (en) 2008-09-10 2010-03-25 Michiyoshi Nagashima Image display device
WO2010049868A1 (en) 2008-10-28 2010-05-06 Koninklijke Philips Electronics N.V. A three dimensional display system
US20120092339A1 (en) * 2009-06-26 2012-04-19 Koninklijke Philips Electronics N.V. Multi-view autostereoscopic display device
JP5494284B2 (en) 2010-06-24 2014-05-14 ソニー株式会社 3D display device and 3D display device control method
US8502816B2 (en) 2010-12-02 2013-08-06 Microsoft Corporation Tabletop display providing multiple views to users
US8953242B2 (en) * 2011-03-31 2015-02-10 Honeywell International Inc. Varible focus stereoscopic display system and method
TWI561857B (en) 2011-05-25 2016-12-11 3M Innovative Properties Co Light control film
KR101269631B1 (en) 2011-10-06 2013-05-30 한국과학기술연구원 3-dimensional image display apparatus and method of performing the same
WO2013079607A1 (en) 2011-11-30 2013-06-06 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. View rendering for the provision of virtual eye contact using special geometric constraints in combination with eye-tracking
KR20140105479A (en) * 2011-12-23 2014-09-01 톰슨 라이센싱 Computer device with power-consumption management and method for managing power-consumption of computer device
JP2013178469A (en) 2012-02-03 2013-09-09 Nlt Technologies Ltd Optical element
GB2499426A (en) 2012-02-16 2013-08-21 Dimenco B V Autostereoscopic display device with viewer tracking system
US9704220B1 (en) 2012-02-29 2017-07-11 Google Inc. Systems, methods, and media for adjusting one or more images displayed to a viewer
KR101958447B1 (en) 2012-05-16 2019-03-15 삼성디스플레이 주식회사 3 dimensional image display device and driving method thereof
US9216133B2 (en) * 2013-01-16 2015-12-22 Elwha Llc Using a 3D display to train a weak eye
KR102070800B1 (en) * 2013-03-11 2020-01-29 삼성전자주식회사 Stereoscopic display apparatus, and display method thereof
CN105917277B (en) 2014-01-07 2020-04-17 视瑞尔技术公司 Display device for holographic reconstruction
CN103852819B (en) * 2014-03-28 2016-01-20 福州大学 A kind of grating strengthening bore hole stereo display space continuity
EP3041231A1 (en) 2014-12-30 2016-07-06 SeeFront GmbH Autostereoscopic multi-view system
US10152154B2 (en) 2015-01-22 2018-12-11 Superd Technology Co., Ltd. 3D interaction method and display device
US10853625B2 (en) 2015-03-21 2020-12-01 Mine One Gmbh Facial signature methods, systems and software
WO2016154123A2 (en) 2015-03-21 2016-09-29 Mine One Gmbh Virtual 3d methods, systems and software
US9983717B2 (en) 2015-04-21 2018-05-29 Dell Products L.P. Disambiguation of false touch inputs at an information handling system projected user interface
KR102415502B1 (en) 2015-08-07 2022-07-01 삼성전자주식회사 Method and apparatus of light filed rendering for plurality of user
US10511895B2 (en) * 2015-10-09 2019-12-17 Warner Bros. Entertainment Inc. Cinematic mastering for virtual reality and augmented reality
US9813673B2 (en) 2016-01-20 2017-11-07 Gerard Dirk Smits Holographic video capture and telepresence system
WO2017127897A1 (en) 2016-01-27 2017-08-03 Paul Lapstun Shuttered waveguide light field display
WO2017145154A1 (en) 2016-02-22 2017-08-31 Real View Imaging Ltd. Wide field of view hybrid holographic display
JP6911269B2 (en) 2016-05-12 2021-07-28 国立研究開発法人情報通信研究機構 Stereoscopic image observation system
CN106125322A (en) * 2016-09-05 2016-11-16 万维云视(上海)数码科技有限公司 Bore hole 3D display floater and bore hole 3D display device
CN114967217B (en) 2017-01-04 2024-07-26 3M创新有限公司 Light control film and method of forming the same
US10191295B2 (en) 2017-01-05 2019-01-29 Looking Glass Factory, Inc. Advanced retroreflecting aerial displays
WO2018227098A1 (en) 2017-06-09 2018-12-13 Vid Scale, Inc. External camera assisted virtual reality
US10269159B2 (en) 2017-07-27 2019-04-23 Rockwell Collins, Inc. Neural network foreground separation for mixed reality
US10948648B2 (en) 2017-09-29 2021-03-16 Reald Spark, Llc Backlights having stacked waveguide and optical components with different coefficients of friction
US10778962B2 (en) 2017-11-10 2020-09-15 Misapplied Sciences, Inc. Precision multi-view display
US20190196539A1 (en) 2017-12-27 2019-06-27 Vincent Gerard Dominique Cassar Privacy Screen Attached by Micro-suction Tape
US10298921B1 (en) 2018-02-27 2019-05-21 Looking Glass Factory, Inc. Superstereoscopic display with enhanced off-angle separation
US10898818B2 (en) 2018-07-25 2021-01-26 Light Field Lab, Inc. Light field display system based amusement park attraction
US10867430B2 (en) 2018-11-19 2020-12-15 Intel Corporation Method and system of 3D reconstruction with volume-based filtering for image processing
US10904479B2 (en) 2019-03-12 2021-01-26 Light Field Lab, Inc. Video communication including holographic content
US11212514B2 (en) 2019-03-25 2021-12-28 Light Field Lab, Inc. Light field display system for cinemas
KR20220045166A (en) 2019-08-09 2022-04-12 라이트 필드 랩 인코포레이티드 Digital signage system based on light field display system
US10981046B2 (en) 2019-08-26 2021-04-20 Light Field Lab, Inc. Light field display system for sporting events
WO2021087450A1 (en) 2019-11-01 2021-05-06 Raxium, Inc. Light field displays incorporating eye trackers and methods for generating views for a light field display using eye tracking information
GB2589575B (en) 2019-12-02 2022-01-12 Envisics Ltd Pupil expander
US11211433B2 (en) 2020-05-04 2021-12-28 Intel Corporation In-display sensors and viewing angle adjustment microassemblies

Patent Citations (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6064424A (en) 1996-02-23 2000-05-16 U.S. Philips Corporation Autostereoscopic display apparatus
US6462871B1 (en) * 1997-02-18 2002-10-08 Canon Kabushiki Kaisha Stereoscopic image display apparatus using specific mask pattern
US6798390B1 (en) * 1997-08-29 2004-09-28 Canon Kabushiki Kaisha 3D image reconstructing apparatus and 3D object inputting apparatus
US20040165262A1 (en) * 2001-06-29 2004-08-26 Jose Javier Alejo Trevijano Autostereoscopic vision system
US20050078370A1 (en) * 2002-04-05 2005-04-14 Hiroshi Nishihara Stereoscopic image display apparatus and stereoscopic image display system
US20050007837A1 (en) * 2003-03-25 2005-01-13 Lsi Logic Corporation Clock generator for pseudo dual port memory
US20070016514A1 (en) * 2005-07-15 2007-01-18 Al-Abdulqader Hisham A System, program product, and methods for managing contract procurement
US8248694B2 (en) * 2006-01-19 2012-08-21 Epson Imaging Devices Corporation Electro-optical apparatus and electronic equipment
US20070165145A1 (en) * 2006-01-19 2007-07-19 Sanyo Epson Imaging Devices Corporation Electro-optical apparatus and electronic equipment
US20070164950A1 (en) * 2006-01-19 2007-07-19 Sanyo Epson Imaging Devices Corporation Image display and electronic device
US20140011851A1 (en) * 2007-07-09 2014-01-09 Furiex Pharmaceuticals, Inc. Novel crystals and process of making 5-(-METHYL)-2-METHOXY-BENZOIC ACID
US20120031389A1 (en) * 2010-08-05 2012-02-09 Mcmaster William J Oven door
US20120313896A1 (en) * 2011-06-07 2012-12-13 Sony Corporation Information processing apparatus, information processing method, and program
US20140320614A1 (en) * 2011-09-19 2014-10-30 Ecrans Polaires Inc. /Polar Screens Inc. Method and display for showing a stereoscopic image
US9686535B2 (en) * 2012-07-27 2017-06-20 Samsung Display Co., Ltd. Method of displaying three dimensional image and three dimensional display apparatus for performing the method
US20140118511A1 (en) * 2012-10-31 2014-05-01 Elwha Llc Systems and methods to confirm that an autostereoscopic display is accurately aimed
US20160023448A1 (en) * 2013-03-14 2016-01-28 Corning Incorporated Methods and apparatus for fabricating and cutting flexible glass and polymer composite structures
US20160234487A1 (en) * 2013-09-30 2016-08-11 Koninklijke Philips N.V. Autostereoscopic display device
US20180035096A1 (en) * 2014-03-20 2018-02-01 Ipc Hk Limited Automated 3d photo booth
US20170013921A1 (en) * 2014-06-26 2017-01-19 Stephan Hugo Frei Quick lock for attaching one end of a band to an object
US20160021365A1 (en) * 2014-07-18 2016-01-21 Au Optronics Corp. Image displaying method and image displaying device
US20170078650A1 (en) * 2015-09-15 2017-03-16 Looking Glass Factory, Inc. Enhanced 3d volumetric display
US20170139213A1 (en) * 2015-10-25 2017-05-18 Oculus Vr, Llc Combination Prism Array for Focusing Light

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220166963A1 (en) * 2020-09-30 2022-05-26 Magnetic Media Holdings, Inc. dba Magnetic 3D Method for optimized viewing experience and reduced rendering for autostereoscopic 3d, multiview and volumetric displays.
US11601633B2 (en) * 2020-09-30 2023-03-07 Magnetic Media Holdings, Inc. Method for optimized viewing experience and reduced rendering for autostereoscopic 3D, multiview and volumetric displays
EP4124033A1 (en) 2021-07-23 2023-01-25 VitreaLab GmbH Stereoscopic display device
WO2023002003A1 (en) 2021-07-23 2023-01-26 Vitrealab Gmbh Stereoscopic display device

Also Published As

Publication number Publication date
KR20200139680A (en) 2020-12-14
CN112042186A (en) 2020-12-04
EP3759908A1 (en) 2021-01-06
EP3759908A4 (en) 2021-12-15
CN112042186B (en) 2023-05-09
WO2019168617A1 (en) 2019-09-06
US11051006B2 (en) 2021-06-29
US11683472B2 (en) 2023-06-20
US20210281824A1 (en) 2021-09-09
JP2021516517A (en) 2021-07-01
US20190268588A1 (en) 2019-08-29
KR102416197B1 (en) 2022-07-05
JP7096371B2 (en) 2022-07-05

Similar Documents

Publication Publication Date Title
US11683472B2 (en) Superstereoscopic display with enhanced off-angle separation
US10241344B1 (en) Advanced retroreflecting aerial displays
US9936193B2 (en) Device for generation of colored virtual three-dimensional images
US10554960B2 (en) Unassisted stereoscopic display device using directional backlight structure
US9191661B2 (en) Virtual image display device
US20210021804A1 (en) Method for displaying stereoscopic image and stereoscopic image display apparatus
JP3453086B2 (en) Three-dimensional display method and head-mounted display device
US20220163816A1 (en) Display apparatus for rendering three-dimensional image and method therefor
CN113272710A (en) Extending field of view by color separation
WO2023143505A1 (en) Image generation apparatus, display device and image generation method
US11822086B2 (en) Lightguide based illuminator for reflective display panel
Travis et al. Flat projection for 3-D
CN207625713U (en) Vision display system and head-wearing display device
US10605968B2 (en) Imaging system
KR101978790B1 (en) Multi View Display Device And Method Of Driving The Same
TWI608255B (en) Stereoscopic floating image display apparatus
KR20190085567A (en) Multi-Layer Based Three-Dimension Image Forming Apparatus
CN218524969U (en) Image display device and vehicle
US20230124178A1 (en) System and Method for Generating Compact Light-Field Displays through Varying Optical Depths
US20230314716A1 (en) Emission of particular wavelength bands utilizing directed wavelength emission components in a display system
US20230176274A1 (en) Adjustable focal length illuminator for a display panel

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment: 4