US8717392B2 - Apparatus for enabling users to view images, methods and computer readable storage mediums - Google Patents
Apparatus for enabling users to view images, methods and computer readable storage mediums Download PDFInfo
- Publication number
- US8717392B2 US8717392B2 US12/455,531 US45553109A US8717392B2 US 8717392 B2 US8717392 B2 US 8717392B2 US 45553109 A US45553109 A US 45553109A US 8717392 B2 US8717392 B2 US 8717392B2
- Authority
- US
- United States
- Prior art keywords
- image
- expander
- calibration
- calibration image
- combination
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 title claims description 22
- 238000010168 coupling process Methods 0.000 claims description 28
- 238000005859 coupling reaction Methods 0.000 claims description 28
- 210000001747 pupil Anatomy 0.000 claims description 3
- 238000004590 computer program Methods 0.000 description 17
- 238000010586 diagram Methods 0.000 description 10
- 239000000758 substrate Substances 0.000 description 7
- 230000005670 electromagnetic radiation Effects 0.000 description 5
- 238000004519 manufacturing process Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 230000004438 eyesight Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 241001025261 Neoraja caerulea Species 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 208000003464 asthenopia Diseases 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1423—Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display
- G06F3/1431—Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display using a single graphics controller
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/06—Adjustment of display parameters
- G09G2320/0693—Calibration of display systems
Definitions
- Embodiments of the present invention relate to apparatus for enabling users to view images, methods and computer readable storage mediums.
- they relate to apparatus for enabling users to view images, methods and computer readable storage mediums in portable devices.
- Apparatus such as portable devices, may include displays that are configured to be positioned in close proximity to a user's eyes. Such displays are usually referred to as ‘near to eye displays’. However, the images provided by the displays may not be aligned for various reasons. For example, the displays may not have been calibrated during manufacturing or the calibration may have been altered due to an impact received during use of the apparatus. A user may develop eye strain using such an apparatus.
- an apparatus comprising: a first expander configured to provide a first image and including a first portion configured to provide a first calibration image; a second expander configured to provide a second image and including a second portion configured to provide a second calibration image; and a controller configured to control alignment of the first image and the second image, a combination of the first calibration image and the second calibration image being useable to control alignment of the first image and the second image.
- an apparatus comprising means for controlling alignment of a first image from a first expander and a second image from a second expander using a combination of a first calibration image from a first portion of the first expander and a second calibration image from a second potion of the second expander.
- an apparatus comprising: a first expander configured to provide a first image and including a first portion configured to provide a first calibration image; a second expander configured to provide a second image and including a second portion configured to provide a second calibration image; the first image and the second image being non-overlapping images and the first calibration image and the second calibration image being overlapping images, the first calibration image and the second calibration image having a predetermined relationship with the first image and the second image, and a controller configured to control an angle between, and/or the content of, the first image and the second image for alignment, the first calibration image and the second calibration image being useable to control the angle between, and/or the content of, the first image and the second image.
- the apparatus may be for enabling a user to view images.
- a portable device comprising an apparatus as described in any of the preceding paragraphs.
- a module comprising an apparatus as described in any of the preceding paragraphs.
- a method comprising: providing a first expander configured to provide a first image and including a first portion configured to provide a first calibration image; providing a second expander configured to provide a second image and including a second portion configured to provide a second calibration image; and providing a controller configured to control alignment of the first image and the second image, a combination of the first calibration image and the second calibration image being useable to control alignment of the first image and the second image.
- a method comprising: controlling alignment of a first image from a first expander and a second image from a second expander using a combination of a first calibration image from a first portion of the first expander and a second calibration image from a second potion of the second expander.
- a computer readable storage medium encoded with instructions that, when executed by a processor, perform: controlling alignment of a first image from a first expander and a second image from a second expander using a combination of a first calibration image from a first portion of the first expander and a second calibration image from a second potion of the second expander.
- a computer program that, when executed by a processor, performs controlling alignment of a first image from a first expander and a second image from a second expander using a combination of a first calibration image from a first portion of the first expander and a second calibration image from a second potion of the second expander.
- FIG. 1 illustrates a schematic diagram of an apparatus according to various embodiments of the invention
- FIG. 2A illustrates a first diagram of a combination of a first calibration image and a second calibration image according to various embodiments of the invention
- FIG. 2B illustrates a second diagram of a combination of a first calibration image and a second calibration image according to various embodiments of the invention
- FIG. 2C illustrates a third diagram of a combination of a first calibration image and a second calibration image according to various embodiments of the invention
- FIG. 3 illustrates a schematic diagram of another apparatus according to various embodiments of the invention.
- FIG. 4 illustrates a flow diagram of a method of controlling alignment of a first image and a second image according to various embodiments of the invention.
- FIG. 5 illustrates a flow diagram of a method of manufacturing an apparatus according to various embodiments of the invention.
- FIGS. 1 and 3 illustrate an apparatus 10 comprising: a first expander 22 configured to provide a first image 44 and including a first portion 54 configured to provide a first calibration image 56 ; a second expander 28 configured to provide a second image 60 and including a second portion 70 configured to provide a second calibration image 72 ; and a controller 32 configured to control alignment of the first image 44 and the second image 60 , a combination of the first calibration image 56 and the second calibration image 72 being useable to control alignment of the first image 44 and the second image 60 .
- connection and ‘couple’ and their derivatives mean operationally connected/coupled. It should be appreciated that any number or combination of intervening components can exist (including no intervening components).
- FIG. 1 illustrates a schematic diagram of an apparatus 10 according to various embodiments of the invention.
- the apparatus 10 includes a processor 12 , a memory 14 , a user input device 16 , a first display 18 , first optics 20 , a first expander 22 , a second display 24 , second optics 26 , a second expander 28 and a housing 30 .
- the processor 12 , the memory 14 and the user input device 16 together form a controller 32 .
- the apparatus 10 may be any device and may be a portable device.
- the apparatus 10 may be, for example, a mobile cellular telephone, a personal digital assistant (PDA), a portable internet appliance, a hand-held computer, digital video or still camera, wearable computer, computer game device, specialized bring-to-the-eye product for viewing or a module for such devices.
- PDA personal digital assistant
- module refers to a unit or apparatus that excludes certain parts/components that would be added by an end manufacturer or a user.
- the processor 12 may be any suitable processor and may be a microprocessor for example.
- the implementation of the processor 12 can be in hardware alone (a circuit for example), have certain aspects in software including firmware alone or can be a combination of hardware and software (including firmware).
- the processor 12 may be implemented using instructions that enable hardware functionality, for example, by using executable computer program instructions in a general-purpose or special-purpose processor that may be stored on a computer readable storage medium (disk, memory etc) to be executed by such a processor.
- a general-purpose or special-purpose processor that may be stored on a computer readable storage medium (disk, memory etc) to be executed by such a processor.
- the processor 12 is configured to read from and write to the memory 14 .
- the processor 12 may also comprise an output interface 34 via which data and/or commands are output by the processor 12 and an input interface 36 via which data and/or commands are input to the processor 12 .
- the memory 14 may be any suitable memory and may, for example be permanent built-in memory such as flash memory or it may be a removable memory such as a hard disk, secure digital (SD) card or a micro-drive.
- the memory 14 stores a computer program 38 comprising computer program instructions that control the operation of the apparatus 10 when loaded into the processor 12 .
- the computer program instructions 38 provide the logic and routines that enables the apparatus 10 to perform the method described below in relation to FIGS. 1 , 2 A, 2 B, 2 C.
- the computer program instructions 38 enable the apparatus 10 to perform the method illustrated in FIG. 4 .
- the processor 12 by reading the memory 14 is able to load and execute the computer program 38 .
- the computer program instructions 38 provide means for controlling alignment of a first image from the first expander 22 and a second image from a second expander 28 using a combination of a first calibration image from a first portion of the first expander 22 and a second calibration image from a second potion of the second expander 28 .
- the computer program 38 may arrive at the apparatus 10 via any suitable delivery mechanism 40 .
- the delivery mechanism 40 may be, for example, a computer-readable storage medium, a computer program product, a memory device, a record medium such as a Blue-ray disk, CD-ROM or DVD, an article of manufacture that tangibly embodies the computer program 38 .
- the delivery mechanism 40 may be a signal configured to reliably transfer the computer program 38 .
- the apparatus 10 may propagate or transmit the computer program 38 as a computer data signal.
- memory 14 is illustrated as a single component it may be implemented as one or more separate components some or all of which may be integrated/removable and/or may provide permanent/semi-permanent/dynamic/cached storage.
- references to ‘computer-readable storage medium’, ‘computer program product’, ‘tangibly embodied computer program’ etc. or a ‘controller’, ‘computer’, ‘processor’ etc. should be understood to encompass not only computers having different architectures such as single/multi-processor architectures and sequential (Von Neumann)/parallel architectures but also specialized circuits such as field-programmable gate arrays (FPGA), application specific circuits (ASIC), signal processing devices and other devices.
- References to computer program, instructions, code etc. should be understood to encompass software for a programmable processor or firmware such as, for example, the programmable content of a hardware device whether instructions for a processor, or configuration settings for a fixed-function device, gate array or programmable logic device etc.
- the user input device 16 is configured to receive an input from a user and provide the input to the processor 12 for processing.
- the user input device 16 may be any suitable user input device and may be, for example, a keypad, a joystick, a pointing stick, a trackball or a touchpad.
- the first display 18 may be any suitable display device and may be a micro-display that may provide full color pixels at more than fifty lines per millimeter.
- the first display 18 may be, for example, a liquid crystal display (LCD) or an organic light emitting diode (OLED) display.
- the first display 18 is configured to provide a first image 42 to the first optics 20 .
- the first image 42 comprises electromagnetic radiation having a wavelength that is visible to the human eye (approximately 400 nm to 700 nm).
- the first optics 20 may be any suitable optical device for guiding the first image 42 to the first expander 22 .
- the first optics 20 may comprise, for example, one or more lenses.
- the apparatus 10 may not include the first optics 20 and the first image 42 may be provided directly to the first expander 22 .
- the first expander 22 is configured to receive the first image 42 and provide a first image 44 to a first eye 46 of a user (the left eye of the user for example).
- the width of the light beam corresponding to the first image 44 is larger than the one corresponding to the image 42 .
- the first expander 22 may be any suitable device for expanding the beam of the first image 42 and may be, for example, a diffractive exit pupil expander (EPE).
- EPE diffractive exit pupil expander
- the first expander 22 includes a light transmissive substrate 48 , a first in-coupling element, 50 , that may be diffractive in nature and a first out-coupling element 52 that may also be diffractive in nature.
- the first out-coupling element 52 includes a first portion 54 that is configured to provide a first calibration image 56 . In other embodiments, the first portion 54 may be separated from the first out-coupling element 52 .
- the first portion 54 is provided at one end of the first expander 22 and the first in-coupling element 50 is provided at the opposite end of the first expander 22 .
- the first in-coupling element 50 , the first out-coupling element 52 and the first portion 54 may comprise, for example, diffraction gratings that are configured to diffract electromagnetic radiation having visible wavelengths.
- the first image 42 is received and coupled in to the transmissive substrate by the first in-coupling element 50 .
- the light from the first image 42 is thus guided from the first in-coupling element 50 towards the first out-coupling element 52 by total internal reflection within the light transmissive substrate 48 .
- the light from the first image 42 is further coupled out at the first out-coupling element 52 and is directed out of the substrate 48 as the first image 44 and the first calibration image 56 .
- the second display 24 may be any suitable display device and may be a micro-display that may provide full color pixels at more than fifty lines per millimeter.
- the second display 24 may be, for example, a liquid crystal display (LCD) or an organic light emitting diode (OLED) display.
- the second display 24 is configured to provide a second image 58 to the second optics 26 .
- the second image 58 comprises electromagnetic radiation having a wavelength that is visible to the human eye (approximately 400 nm to 700 nm) and includes substantially the same content as the first image 42 .
- the second optics 26 may be any suitable optical device for guiding the second image 58 to the second expander 28 .
- the second optics 26 may comprise, for example, one or more lenses.
- the apparatus 10 may not include the second optics 26 and the second image 58 may be provided directly to the second expander 28 .
- the second expander 28 is configured to receive the second image 58 and provide a second image 60 to a second eye 62 of the user (the right eye of the user for example).
- the width of the light beam corresponding to the second image 60 is larger than the width of the light beam corresponding to the second image 58 and is consequently larger in the user's vision.
- the second expander 28 may be any suitable device for expanding the light beam of the second image 58 and may be, for example, a diffractive exit pupil expander (EPE).
- the second expander 28 includes a light transmissive substrate 64 , a second in-coupling element 66 , that may be diffractive in nature and a second out-coupling element 68 , that may also be diffractive in nature.
- the second out-coupling element 68 includes a second portion 70 that is configured to provide a second calibration image 72 . In other embodiments, the second portion 70 may be separated from the out-coupling element 68 .
- the second portion 70 is provided at one end of the second expander 28 and the second in-coupling element 66 is provided at the opposite end of the second expander 28 .
- the second in-coupling element 66 , the second out-coupling element 68 and the second portion 70 may comprise, for example, diffraction gratings that are configured to diffract electromagnetic radiation having visible wavelengths.
- the second image 58 is received and coupled in by the second in-coupling element 66 .
- the second image 58 is guided from the first in-coupling element 66 towards the second out-coupling element 68 by total internal reflection within the light transmissive substrate 64 .
- the second image 58 is further coupled out at the second out-coupling element 68 and is directed out of the substrate 64 as the second image 60 and the second calibration image 72 .
- the first image 44 and the second image 60 are provided through a first and second aperture 74 , 76 respectively in the housing 30 of the apparatus 10 .
- a user is able to place his eyes at the first and second apertures 74 , 76 and view the first image 44 with his first eye 46 and the second image 60 with his second eye 62 . It should be appreciated that at the first and second apertures 74 , 76 , the first image 44 and the second image 60 are non-overlapping.
- the first calibration image 56 and the second calibration image 72 are provided through a third aperture 78 in the housing 30 of the apparatus 10 .
- the viewing direction of the first and second calibration images 56 , 72 are opposite to the first and second expanded images 44 , 60 (one hundred and eighty degrees difference in orientation in this embodiment). Since the first portion 54 and the second portion 70 are positioned adjacent one another with little or no gap therebetween, the first calibration image 56 and the second calibration image 72 substantially overlap one another and form a single image. Consequently, a user is able to view the combination of the first calibration image 56 and the second calibration image 72 through the third aperture 78 with either his first eye 46 or his second eye 62 .
- the first expander 22 and the second expander 28 are configured so that when the first calibration image 56 and the second calibration image 72 are aligned, the first image 44 and the second image 60 are also aligned in the user's vision.
- the first and second calibration images 56 , 72 overlay one another at the position of observation so that they appear to be a single image to the user.
- a user is able to view the combination of the first calibration image 56 and the second calibration image 72 through the third aperture 78 to determine whether the first image 44 and the second image 60 are aligned. If the combination of the first calibration image 56 and the second calibration image 72 are not aligned (that is, the user views two images through the third aperture 78 ), the user may operate the user input device 16 to control the alignment of the first and second calibration images 56 , 72 (and hence the alignment of the first and second images 44 , 60 ) so that they are aligned.
- the processor 12 receives an input signal from the user input device 16 that indicates how the first calibration image 56 and the second calibration image 72 should move relative to one another so that they overlap when viewed through the third aperture 78 .
- the processor 12 may control the position and/or orientation of the first display 18 and the second display 24 to control alignment of the first calibration image 56 and the second calibration image 72 .
- the processor 12 may alternatively/additionally control the content of the first display 18 and the second display 24 (that is, the image 42 , 58 may be shifted in the first display 18 and/or the second display 24 ) to control alignment.
- the processor 12 may alternatively/additionally control the position and/or orientation of the first optics 20 and/or the second optics 26 to control alignment.
- the user input device 16 may be operated by a user to mechanically control the alignment of the first calibration image 56 and the second calibration image 72 .
- the user input device 16 may be mechanically coupled to the first display 18 and/or the second display 24 and/or the first optics 20 and/or the second optics 26 to enable a user to mechanically control the position and/or orientation of the first display 18 and/or the second display 24 and/or the first optics and/or the second optics 26 .
- FIG. 2A illustrates a user's view of the combination of the first calibration image 56 and the second calibration image 72 through the third aperture 78 .
- the first calibration image 56 includes a first cross 80 and the second calibration image 72 includes a second cross 82 that may assist a user with calibration.
- the first cross 80 is positioned in the top left of the user's view and the second cross 82 is positioned in the bottom right of the user's view. Since alignment is achieved when the user views a single cross in the centre of the third aperture 78 , the first calibration image 56 and the second calibration image 72 are not aligned in FIG. 2A .
- the user may then control the user input device 16 so that the first cross 80 and the second cross 82 are brought towards one another but do not overlap (as illustrated in FIG. 2B ).
- the user may provide further input to the user input device 16 so that the first cross 80 and the second cross 82 substantially overlap in the centre of the third aperture (as illustrated in FIG. 2C ), indicating that the first and second calibration images 56 , 72 (and hence first and second images 44 , 60 ) are aligned.
- Embodiments of the present invention provide an advantage in that they enable a user to control the alignment of the first and second images 44 , 60 using the combination of the first and second calibration images 56 , 72 . Since the apparatus 10 may receive impacts during use (from the user for example) which may result in the first and second images 44 , 60 not being aligned, embodiments of the invention may enable the user to calibrate the apparatus 10 after each such impact.
- FIG. 3 illustrates a schematic diagram of another apparatus 10 according to various embodiments of the invention.
- the apparatus 10 illustrated in FIG. 3 is similar to the apparatus illustrated in FIG. 1 and where the features are similar, the same reference numerals are used.
- the apparatus 10 illustrated in FIG. 3 is different to the apparatus illustrated in FIG. 1 in that the apparatus 10 further comprises a sensor 84 and may not include the user input device 16 or the third aperture 78 .
- the sensor 84 may be any suitable device for detecting electromagnetic radiation and may be, for example, a charge coupled device (CCD) camera or a complementary metal oxide semiconductor (CMOS) camera.
- CCD charge coupled device
- CMOS complementary metal oxide semiconductor
- the sensor 84 receives and detects the first calibration image 56 and the second calibration image 72 and provides the combination of the first calibration image 56 and the second calibration image 72 to the processor 12 for processing.
- the processor 12 receives the combination of the first calibration image 56 and the second calibration image 72 .
- the processor 12 controls the alignment of the first image 44 and the second image 60 using the combination of the first calibration image 56 and the second calibration image 72 .
- the processor 12 may automatically control the alignment of the first image 44 and the second image 60 in response to receiving the combination of the first calibration image 56 and the second calibration image 72 .
- the processor 12 may determine that the first and second images 44 , 60 are not aligned and request instructions from the user to confirm that alignment should be carried out.
- the first and second calibration images 56 , 72 may comprise crosses 80 , 82 as illustrated in FIGS. 2A , 2 B, 2 C and the processor 12 may use image recognition to determine the positions of the crosses 80 , 82 .
- the processor 12 may then determine a vector for each cross 80 , 82 that is indicative of the movement required for each cross 80 , 82 to reach the centre for example.
- the processor 12 may then use a look-up table (stored in the memory 14 ) to determine how to control the displays 18 , 24 and/or optics 20 , 26 and/or expanders 22 , 28 .
- the processor 12 may use the result of the interrogation of the look-up table to control the displays 18 , 24 and/or optics 20 , 26 to achieve alignment of the first and second images 44 , 60 .
- Embodiments of the invention as illustrated in FIG. 3 may provide an advantage in that the alignment of the first and second images 44 , 60 may be carried out by the processor 12 without requiring significant input from the user of the apparatus 10 . This may result in the apparatus 10 being simpler to operate for the user.
- first cross 80 and the second cross 82 of the first and second calibration images 56 , 72 may have a wavelength that is not visible to human eyesight (infra-red for example).
- the sensor 84 may be configured to detect the wavelength of the first and second calibration images 56 , 72 (infra-red) and provide the combination of the first calibration image 56 and the second calibration image 72 to the processor 12 .
- This embodiment may provide an advantage in that the crosses 80 , 82 are substantially invisible to the user and the calibration process may be transparent to the user.
- FIG. 5 illustrates a flow diagram of a method of manufacturing an apparatus 10 according to various embodiments of the invention.
- the method includes providing, positioning and orienting the first expander 22 .
- the method includes providing, positioning and orienting the second expander 28 .
- the method includes providing a controller 32 and configuring the controller 12 to be able to control the alignment of the first image 44 and the second image 60 using the combination of the first calibration image 56 and the second calibration image 72 .
- the method then includes providing a housing 30 defining a third aperture 78 for housing the first expander 22 , the second expander 28 and the controller 32 .
- the method then includes providing a sensor 84 for receiving the first calibration image 56 and the second calibration image 72 .
- the blocks illustrated in the FIGS. 4 and 5 may represent steps in a method and/or sections of code in the computer program 38 .
- the illustration of a particular order to the blocks does not necessarily imply that there is a required or preferred order for the blocks and the order and arrangement of the block may be varied. Furthermore, it may be possible for some steps to be omitted.
- the apparatus 10 may include the features of the apparatus illustrated in FIG. 1 and features of the apparatus illustrated in FIG. 3 .
- the apparatus 10 may enable a user to align the first and second images 44 , 60 using the user input device 16 and may include a sensor 84 for enabling the processor 12 to perform the alignment without user output.
- first expander 22 and the second expander 28 may not be diffractive as described above, but may instead be refractive or reflective.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Graphics (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
Abstract
Description
Claims (22)
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/455,531 US8717392B2 (en) | 2009-06-02 | 2009-06-02 | Apparatus for enabling users to view images, methods and computer readable storage mediums |
DE112010002211.9T DE112010002211B4 (en) | 2009-06-02 | 2010-04-15 | Apparatus for enabling users to view images, methods, and computer readable storage media |
CN201080024424.1A CN102804025B (en) | 2009-06-02 | 2010-04-15 | Support that user watches the device of image, method and computer-readable recording medium |
PCT/FI2010/050307 WO2010139848A1 (en) | 2009-06-02 | 2010-04-15 | Apparatus for enabling users to view images, methods and computer readable storage mediums |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/455,531 US8717392B2 (en) | 2009-06-02 | 2009-06-02 | Apparatus for enabling users to view images, methods and computer readable storage mediums |
Publications (2)
Publication Number | Publication Date |
---|---|
US20100302276A1 US20100302276A1 (en) | 2010-12-02 |
US8717392B2 true US8717392B2 (en) | 2014-05-06 |
Family
ID=43219726
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/455,531 Active 2032-05-22 US8717392B2 (en) | 2009-06-02 | 2009-06-02 | Apparatus for enabling users to view images, methods and computer readable storage mediums |
Country Status (4)
Country | Link |
---|---|
US (1) | US8717392B2 (en) |
CN (1) | CN102804025B (en) |
DE (1) | DE112010002211B4 (en) |
WO (1) | WO2010139848A1 (en) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140139517A1 (en) * | 2012-11-20 | 2014-05-22 | Innolux Corporation | Three dimensional image display system and adjusting method thereof |
US10645348B2 (en) | 2018-07-07 | 2020-05-05 | Sensors Unlimited, Inc. | Data communication between image sensors and image displays |
US10742913B2 (en) | 2018-08-08 | 2020-08-11 | N2 Imaging Systems, LLC | Shutterless calibration |
US10753709B2 (en) | 2018-05-17 | 2020-08-25 | Sensors Unlimited, Inc. | Tactical rails, tactical rail systems, and firearm assemblies having tactical rails |
US10796860B2 (en) | 2018-12-12 | 2020-10-06 | N2 Imaging Systems, LLC | Hermetically sealed over-molded button assembly |
US10801813B2 (en) | 2018-11-07 | 2020-10-13 | N2 Imaging Systems, LLC | Adjustable-power data rail on a digital weapon sight |
US10921578B2 (en) | 2018-09-07 | 2021-02-16 | Sensors Unlimited, Inc. | Eyecups for optics |
US11079202B2 (en) | 2018-07-07 | 2021-08-03 | Sensors Unlimited, Inc. | Boresighting peripherals to digital weapon sights |
US11103763B2 (en) | 2018-09-11 | 2021-08-31 | Real Shot Inc. | Basketball shooting game using smart glasses |
US11122698B2 (en) | 2018-11-06 | 2021-09-14 | N2 Imaging Systems, LLC | Low stress electronic board retainers and assemblies |
US11141645B2 (en) | 2018-09-11 | 2021-10-12 | Real Shot Inc. | Athletic ball game using smart glasses |
US11143838B2 (en) | 2019-01-08 | 2021-10-12 | N2 Imaging Systems, LLC | Optical element retainers |
US11162763B2 (en) | 2015-11-03 | 2021-11-02 | N2 Imaging Systems, LLC | Non-contact optical connections for firearm accessories |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
IL166799A (en) | 2005-02-10 | 2014-09-30 | Lumus Ltd | Substrate-guided optical device utilizing beam splitters |
US10073264B2 (en) | 2007-08-03 | 2018-09-11 | Lumus Ltd. | Substrate-guide optical device |
CN104471463B (en) * | 2012-05-03 | 2018-02-13 | 诺基亚技术有限公司 | Image providing device, method and computer program |
IL219907A (en) | 2012-05-21 | 2017-08-31 | Lumus Ltd | Head-mounted display eyeball tracker integrated system |
CN104704423A (en) * | 2012-10-05 | 2015-06-10 | 诺基亚技术有限公司 | An apparatus and method for capturing images |
CN103841392B (en) * | 2012-11-20 | 2016-08-03 | 群康科技(深圳)有限公司 | 3 D image display system and bearing calibration thereof |
TW201515433A (en) * | 2013-10-14 | 2015-04-16 | Etron Technology Inc | Image calibration system and calibration method of a stereo camera |
US20170353714A1 (en) * | 2016-06-06 | 2017-12-07 | Navid Poulad | Self-calibrating display system |
US10212414B2 (en) * | 2016-08-01 | 2019-02-19 | Microsoft Technology Licensing, Llc | Dynamic realignment of stereoscopic digital consent |
US10216263B2 (en) | 2016-09-12 | 2019-02-26 | Microsoft Technology Licensing, Llc | Display active alignment systems utilizing test patterns for calibrating signals in waveguide displays |
US10324291B2 (en) * | 2016-09-12 | 2019-06-18 | Microsoft Technology Licensing, Llc | Display active alignment system for waveguide displays |
WO2018138714A1 (en) | 2017-01-28 | 2018-08-02 | Lumus Ltd. | Augmented reality imaging system |
US11086059B2 (en) | 2017-06-13 | 2021-08-10 | Vuzix Corporation | Image light guide with expanded light distribution overlapping gratings |
JP7174929B2 (en) | 2017-07-19 | 2022-11-18 | ルムス エルティーディー. | LCOS illumination via LOE |
IL275615B (en) * | 2018-01-02 | 2022-08-01 | Lumus Ltd | Augmented reality displays with active alignment and corresponding methods |
US10506220B2 (en) | 2018-01-02 | 2019-12-10 | Lumus Ltd. | Augmented reality displays with active alignment and corresponding methods |
IL259518B2 (en) | 2018-05-22 | 2023-04-01 | Lumus Ltd | Optical system and method for improvement of light field uniformity |
WO2020014700A1 (en) | 2018-07-13 | 2020-01-16 | Magic Leap, Inc. | Systems and methods for display binocular deformation compensation |
JP7398131B2 (en) | 2019-03-12 | 2023-12-14 | ルムス エルティーディー. | image projector |
Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5142357A (en) * | 1990-10-11 | 1992-08-25 | Stereographics Corp. | Stereoscopic video camera with image sensors having variable effective position |
US6191809B1 (en) | 1998-01-15 | 2001-02-20 | Vista Medical Technologies, Inc. | Method and apparatus for aligning stereo images |
US6507359B1 (en) * | 1993-09-20 | 2003-01-14 | Canon Kabushiki Kaisha | Image display system |
US6552332B2 (en) * | 2000-01-06 | 2003-04-22 | Minolta Co., Ltd. | Image-sensing apparatus |
US20030184860A1 (en) | 2002-03-28 | 2003-10-02 | Nokia Corporation | Method to detect misalignment and distortion in near-eye displays |
WO2006058188A2 (en) | 2004-11-24 | 2006-06-01 | Kopin Corporation | Binocular display system with two alignment fixtures |
US20060123182A1 (en) * | 2004-12-07 | 2006-06-08 | Francisc Sandulescu | Distributed KVM and peripheral switch |
US20060228073A1 (en) | 2004-03-29 | 2006-10-12 | Hiroshi Mukawa | Optical device and virtual image display device |
US7206107B2 (en) | 2004-12-13 | 2007-04-17 | Nokia Corporation | Method and system for beam expansion in a display device |
WO2007102144A1 (en) | 2006-03-08 | 2007-09-13 | Lumus Ltd. | Device and method for alignment of binocular personal display |
WO2007125257A1 (en) | 2006-04-26 | 2007-11-08 | Essilor International (Compagnie Générale d'Optique) | Driver for display comprising a pair of binocular-type spectacles |
US20080024391A1 (en) | 2006-07-31 | 2008-01-31 | Manuel Oliver | Image alignment method for binocular eyewear displays |
US20080100633A1 (en) * | 2003-04-24 | 2008-05-01 | Dallas James M | Microdisplay and interface on a single chip |
US20080218822A1 (en) * | 2004-02-04 | 2008-09-11 | Microvision, Inc. | Scanned-Beam Heads-Up Display and Related Systems and Methods |
US20090097122A1 (en) * | 2005-09-14 | 2009-04-16 | Mirage Innovations Ltd | Diffractive Optical Device and System |
US7764413B2 (en) * | 2004-12-13 | 2010-07-27 | Nokia Corporation | General diffractive optics method for expanding an exit pupil |
US20100232016A1 (en) * | 2005-09-28 | 2010-09-16 | Mirage Innovations Ltd. | Stereoscopic Binocular System, Device and Method |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101452117B (en) * | 2004-11-24 | 2011-01-26 | 寇平公司 | Binocular display system and method thereof |
-
2009
- 2009-06-02 US US12/455,531 patent/US8717392B2/en active Active
-
2010
- 2010-04-15 CN CN201080024424.1A patent/CN102804025B/en active Active
- 2010-04-15 WO PCT/FI2010/050307 patent/WO2010139848A1/en active Application Filing
- 2010-04-15 DE DE112010002211.9T patent/DE112010002211B4/en active Active
Patent Citations (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5142357A (en) * | 1990-10-11 | 1992-08-25 | Stereographics Corp. | Stereoscopic video camera with image sensors having variable effective position |
US6507359B1 (en) * | 1993-09-20 | 2003-01-14 | Canon Kabushiki Kaisha | Image display system |
US6191809B1 (en) | 1998-01-15 | 2001-02-20 | Vista Medical Technologies, Inc. | Method and apparatus for aligning stereo images |
US6552332B2 (en) * | 2000-01-06 | 2003-04-22 | Minolta Co., Ltd. | Image-sensing apparatus |
US20030184860A1 (en) | 2002-03-28 | 2003-10-02 | Nokia Corporation | Method to detect misalignment and distortion in near-eye displays |
US7229174B2 (en) | 2002-03-28 | 2007-06-12 | Nokia Corporation | Method to detect misalignment and distortion in near-eye displays |
US20080100633A1 (en) * | 2003-04-24 | 2008-05-01 | Dallas James M | Microdisplay and interface on a single chip |
US7460305B2 (en) * | 2004-02-04 | 2008-12-02 | Microvision, Inc. | Scanned-beam heads-up display and related systems and methods |
US20080218822A1 (en) * | 2004-02-04 | 2008-09-11 | Microvision, Inc. | Scanned-Beam Heads-Up Display and Related Systems and Methods |
US20060228073A1 (en) | 2004-03-29 | 2006-10-12 | Hiroshi Mukawa | Optical device and virtual image display device |
WO2006058188A2 (en) | 2004-11-24 | 2006-06-01 | Kopin Corporation | Binocular display system with two alignment fixtures |
US20060123182A1 (en) * | 2004-12-07 | 2006-06-08 | Francisc Sandulescu | Distributed KVM and peripheral switch |
US7206107B2 (en) | 2004-12-13 | 2007-04-17 | Nokia Corporation | Method and system for beam expansion in a display device |
US7764413B2 (en) * | 2004-12-13 | 2010-07-27 | Nokia Corporation | General diffractive optics method for expanding an exit pupil |
US20090097122A1 (en) * | 2005-09-14 | 2009-04-16 | Mirage Innovations Ltd | Diffractive Optical Device and System |
US20090128911A1 (en) * | 2005-09-14 | 2009-05-21 | Moti Itzkovitch | Diffraction Grating With a Spatially Varying Duty-Cycle |
US20100232016A1 (en) * | 2005-09-28 | 2010-09-16 | Mirage Innovations Ltd. | Stereoscopic Binocular System, Device and Method |
WO2007102144A1 (en) | 2006-03-08 | 2007-09-13 | Lumus Ltd. | Device and method for alignment of binocular personal display |
WO2007125257A1 (en) | 2006-04-26 | 2007-11-08 | Essilor International (Compagnie Générale d'Optique) | Driver for display comprising a pair of binocular-type spectacles |
US20080024391A1 (en) | 2006-07-31 | 2008-01-31 | Manuel Oliver | Image alignment method for binocular eyewear displays |
Non-Patent Citations (1)
Title |
---|
Shechter et al., Compact beam expander with linear gratings, Applied Optics, vol. 41, No. 7, pp. 1236-1240. * |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140139517A1 (en) * | 2012-11-20 | 2014-05-22 | Innolux Corporation | Three dimensional image display system and adjusting method thereof |
US11162763B2 (en) | 2015-11-03 | 2021-11-02 | N2 Imaging Systems, LLC | Non-contact optical connections for firearm accessories |
US10753709B2 (en) | 2018-05-17 | 2020-08-25 | Sensors Unlimited, Inc. | Tactical rails, tactical rail systems, and firearm assemblies having tactical rails |
US10645348B2 (en) | 2018-07-07 | 2020-05-05 | Sensors Unlimited, Inc. | Data communication between image sensors and image displays |
US11079202B2 (en) | 2018-07-07 | 2021-08-03 | Sensors Unlimited, Inc. | Boresighting peripherals to digital weapon sights |
US10742913B2 (en) | 2018-08-08 | 2020-08-11 | N2 Imaging Systems, LLC | Shutterless calibration |
US10921578B2 (en) | 2018-09-07 | 2021-02-16 | Sensors Unlimited, Inc. | Eyecups for optics |
US11103763B2 (en) | 2018-09-11 | 2021-08-31 | Real Shot Inc. | Basketball shooting game using smart glasses |
US11141645B2 (en) | 2018-09-11 | 2021-10-12 | Real Shot Inc. | Athletic ball game using smart glasses |
US11122698B2 (en) | 2018-11-06 | 2021-09-14 | N2 Imaging Systems, LLC | Low stress electronic board retainers and assemblies |
US10801813B2 (en) | 2018-11-07 | 2020-10-13 | N2 Imaging Systems, LLC | Adjustable-power data rail on a digital weapon sight |
US10796860B2 (en) | 2018-12-12 | 2020-10-06 | N2 Imaging Systems, LLC | Hermetically sealed over-molded button assembly |
US11143838B2 (en) | 2019-01-08 | 2021-10-12 | N2 Imaging Systems, LLC | Optical element retainers |
Also Published As
Publication number | Publication date |
---|---|
CN102804025B (en) | 2016-08-03 |
DE112010002211T5 (en) | 2013-10-02 |
CN102804025A (en) | 2012-11-28 |
US20100302276A1 (en) | 2010-12-02 |
WO2010139848A1 (en) | 2010-12-09 |
DE112010002211B4 (en) | 2020-09-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8717392B2 (en) | Apparatus for enabling users to view images, methods and computer readable storage mediums | |
CN112470163B (en) | In-field illumination and imaging for eye tracking | |
US10969585B2 (en) | Waveguide display with increased uniformity and reduced cross-coupling between colors | |
CN108292048B (en) | Extended field of view in near-eye displays using wide spectrum imagers | |
US10698204B1 (en) | Immersed hot mirrors for illumination in eye tracking | |
JP5608242B2 (en) | Infrared vision using liquid crystal display devices | |
US10108013B2 (en) | Indirect-view augmented reality display system | |
US10725302B1 (en) | Stereo imaging with Fresnel facets and Fresnel reflections | |
US10712576B1 (en) | Pupil steering head-mounted display | |
US20180321736A1 (en) | Beam guiding device | |
EP2852859A1 (en) | Head -mounted display with an eyeball - tracker integrated system | |
US11675192B2 (en) | Hybrid coupling diffractive optical element | |
US11852830B2 (en) | Augmented reality glass and operating method therefor | |
US11455031B1 (en) | In-field illumination for eye tracking | |
US20240004202A1 (en) | Wearable electronic device including lens assembly | |
US11809622B2 (en) | Electronic device and method for eye-tracking of user and providing augmented reality service thereof | |
US11733952B2 (en) | Wearable electronic device including display, method for controlling display, and system including wearable electronic device and case | |
US11237628B1 (en) | Efficient eye illumination using reflection of structured light pattern for eye tracking | |
TW202343080A (en) | Suppression of first-order diffraction in a two-dimensional grating of an output coupler for a head-mounted display | |
US20230083745A1 (en) | Head-mounted display | |
US20230319428A1 (en) | Camera comprising lens array | |
US20240179284A1 (en) | Dual-path disparity sensor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: NOKIA CORPORATION, FINLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LEVOLA, TAPANI;REEL/FRAME:023113/0154 Effective date: 20090804 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: NOKIA TECHNOLOGIES OY, FINLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NOKIA CORPORATION;REEL/FRAME:035280/0093 Effective date: 20150116 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551) Year of fee payment: 4 |
|
AS | Assignment |
Owner name: MAGIC LEAP, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NOKIA TECHNOLGIES LTD;REEL/FRAME:049399/0158 Effective date: 20190313 |
|
AS | Assignment |
Owner name: MAGIC LEAP, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:NOKIA TECHNOLOGIES OY;REEL/FRAME:049719/0250 Effective date: 20190627 |
|
AS | Assignment |
Owner name: JP MORGAN CHASE BANK, N.A., NEW YORK Free format text: PATENT SECURITY AGREEMENT;ASSIGNORS:MAGIC LEAP, INC.;MOLECULAR IMPRINTS, INC.;MENTOR ACQUISITION ONE, LLC;REEL/FRAME:050138/0287 Effective date: 20190820 |
|
AS | Assignment |
Owner name: CITIBANK, N.A., NEW YORK Free format text: ASSIGNMENT OF SECURITY INTEREST IN PATENTS;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:050967/0138 Effective date: 20191106 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |