NZ758961B2 - Virtual and augmented reality systems and methods - Google Patents
Virtual and augmented reality systems and methodsInfo
- Publication number
- NZ758961B2 NZ758961B2 NZ758961A NZ75896114A NZ758961B2 NZ 758961 B2 NZ758961 B2 NZ 758961B2 NZ 758961 A NZ758961 A NZ 758961A NZ 75896114 A NZ75896114 A NZ 75896114A NZ 758961 B2 NZ758961 B2 NZ 758961B2
- Authority
- NZ
- New Zealand
- Prior art keywords
- light
- user
- waveguide
- focus
- lens
- Prior art date
Links
- 230000003190 augmentative effect Effects 0.000 title abstract description 33
- 238000000034 method Methods 0.000 title description 74
- 230000003287 optical effect Effects 0.000 claims abstract description 249
- 210000001747 pupil Anatomy 0.000 claims abstract description 177
- 210000001508 eye Anatomy 0.000 claims description 347
- 230000004308 accommodation Effects 0.000 claims description 101
- 230000008859 change Effects 0.000 claims description 39
- 230000003068 static effect Effects 0.000 claims description 21
- 230000004048 modification Effects 0.000 claims description 3
- 238000012986 modification Methods 0.000 claims description 3
- 230000001186 cumulative effect Effects 0.000 claims 1
- 210000000695 crystalline len Anatomy 0.000 description 353
- 241000219739 Lens Species 0.000 description 336
- 239000000835 fiber Substances 0.000 description 261
- 239000013307 optical fiber Substances 0.000 description 100
- 239000000758 substrate Substances 0.000 description 78
- 230000000007 visual effect Effects 0.000 description 46
- 239000011295 pitch Substances 0.000 description 40
- 230000008447 perception Effects 0.000 description 39
- 238000003491 array Methods 0.000 description 34
- 230000033001 locomotion Effects 0.000 description 33
- 210000003128 head Anatomy 0.000 description 32
- 239000004973 liquid crystal related substance Substances 0.000 description 32
- 125000001475 halogen functional group Chemical group 0.000 description 31
- 238000009877 rendering Methods 0.000 description 30
- 210000001525 retina Anatomy 0.000 description 29
- 230000010287 polarization Effects 0.000 description 28
- 230000000875 corresponding effect Effects 0.000 description 26
- 230000000694 effects Effects 0.000 description 25
- 239000000463 material Substances 0.000 description 23
- 230000006870 function Effects 0.000 description 22
- 238000013459 approach Methods 0.000 description 19
- 210000004556 brain Anatomy 0.000 description 19
- 238000000576 coating method Methods 0.000 description 19
- 238000005286 illumination Methods 0.000 description 19
- 230000002238 attenuated effect Effects 0.000 description 16
- 239000011521 glass Substances 0.000 description 16
- 239000011248 coating agent Substances 0.000 description 15
- 230000005540 biological transmission Effects 0.000 description 14
- 150000002500 ions Chemical class 0.000 description 13
- GQYHUHYESMUTHG-UHFFFAOYSA-N lithium niobate Chemical compound [Li+].[O-][Nb](=O)=O GQYHUHYESMUTHG-UHFFFAOYSA-N 0.000 description 13
- 238000012545 processing Methods 0.000 description 13
- 238000009826 distribution Methods 0.000 description 12
- 230000001965 increasing effect Effects 0.000 description 12
- 210000004087 cornea Anatomy 0.000 description 11
- 238000010168 coupling process Methods 0.000 description 11
- 238000005859 coupling reaction Methods 0.000 description 11
- 239000003086 colorant Substances 0.000 description 10
- 230000008878 coupling Effects 0.000 description 10
- 239000007788 liquid Substances 0.000 description 9
- 230000004044 response Effects 0.000 description 9
- 210000000887 face Anatomy 0.000 description 8
- 238000002347 injection Methods 0.000 description 8
- 239000007924 injection Substances 0.000 description 8
- 239000012528 membrane Substances 0.000 description 8
- 238000012800 visualization Methods 0.000 description 8
- 239000004983 Polymer Dispersed Liquid Crystal Substances 0.000 description 7
- 230000004075 alteration Effects 0.000 description 7
- 230000008901 benefit Effects 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 7
- 239000011159 matrix material Substances 0.000 description 7
- 238000011282 treatment Methods 0.000 description 7
- 230000000903 blocking effect Effects 0.000 description 6
- 238000005253 cladding Methods 0.000 description 6
- 230000001427 coherent effect Effects 0.000 description 6
- 238000013461 design Methods 0.000 description 6
- 239000010408 film Substances 0.000 description 6
- 238000001914 filtration Methods 0.000 description 6
- AMGQUBHHOARCQH-UHFFFAOYSA-N indium;oxotin Chemical compound [In].[Sn]=O AMGQUBHHOARCQH-UHFFFAOYSA-N 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 241000282412 Homo Species 0.000 description 5
- 230000001886 ciliary effect Effects 0.000 description 5
- 239000002131 composite material Substances 0.000 description 5
- 230000001934 delay Effects 0.000 description 5
- 239000011263 electroactive material Substances 0.000 description 5
- 238000013507 mapping Methods 0.000 description 5
- 230000000873 masking effect Effects 0.000 description 5
- 108091008695 photoreceptors Proteins 0.000 description 5
- 230000007704 transition Effects 0.000 description 5
- 238000005259 measurement Methods 0.000 description 4
- 229920000642 polymer Polymers 0.000 description 4
- 230000001902 propagating effect Effects 0.000 description 4
- 230000005855 radiation Effects 0.000 description 4
- 230000002123 temporal effect Effects 0.000 description 4
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 3
- 206010025421 Macule Diseases 0.000 description 3
- 239000000654 additive Substances 0.000 description 3
- 230000000996 additive effect Effects 0.000 description 3
- 230000003098 cholesteric effect Effects 0.000 description 3
- 230000003247 decreasing effect Effects 0.000 description 3
- 239000013013 elastic material Substances 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000002156 mixing Methods 0.000 description 3
- 230000003534 oscillatory effect Effects 0.000 description 3
- 229940125730 polarisation modulator Drugs 0.000 description 3
- 238000004064 recycling Methods 0.000 description 3
- 239000010409 thin film Substances 0.000 description 3
- KRQUFUKTQHISJB-YYADALCUSA-N 2-[(E)-N-[2-(4-chlorophenoxy)propoxy]-C-propylcarbonimidoyl]-3-hydroxy-5-(thian-3-yl)cyclohex-2-en-1-one Chemical compound CCC\C(=N/OCC(C)OC1=CC=C(Cl)C=C1)C1=C(O)CC(CC1=O)C1CCCSC1 KRQUFUKTQHISJB-YYADALCUSA-N 0.000 description 2
- XUKUURHRXDUEBC-KAYWLYCHSA-N Atorvastatin Chemical compound C=1C=CC=CC=1C1=C(C=2C=CC(F)=CC=2)N(CC[C@@H](O)C[C@@H](O)CC(O)=O)C(C(C)C)=C1C(=O)NC1=CC=CC=C1 XUKUURHRXDUEBC-KAYWLYCHSA-N 0.000 description 2
- 201000004569 Blindness Diseases 0.000 description 2
- WHXSMMKQMYFTQS-UHFFFAOYSA-N Lithium Chemical compound [Li] WHXSMMKQMYFTQS-UHFFFAOYSA-N 0.000 description 2
- 101100400378 Mus musculus Marveld2 gene Proteins 0.000 description 2
- 230000004931 aggregating effect Effects 0.000 description 2
- 230000004888 barrier function Effects 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 238000005520 cutting process Methods 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 210000000613 ear canal Anatomy 0.000 description 2
- 238000005530 etching Methods 0.000 description 2
- 239000012530 fluid Substances 0.000 description 2
- 229910021389 graphene Inorganic materials 0.000 description 2
- 229910052744 lithium Inorganic materials 0.000 description 2
- FOXFZRUHNHCZPX-UHFFFAOYSA-N metribuzin Chemical compound CSC1=NN=C(C(C)(C)C)C(=O)N1N FOXFZRUHNHCZPX-UHFFFAOYSA-N 0.000 description 2
- 230000005405 multipole Effects 0.000 description 2
- 210000003205 muscle Anatomy 0.000 description 2
- 230000002688 persistence Effects 0.000 description 2
- 238000005498 polishing Methods 0.000 description 2
- 238000004886 process control Methods 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 230000000638 stimulation Effects 0.000 description 2
- 238000010408 sweeping Methods 0.000 description 2
- 230000032258 transport Effects 0.000 description 2
- 238000011144 upstream manufacturing Methods 0.000 description 2
- 230000004462 vestibulo-ocular reflex Effects 0.000 description 2
- 239000004986 Cholesteric liquid crystals (ChLC) Substances 0.000 description 1
- 239000004593 Epoxy Substances 0.000 description 1
- 244000228957 Ferula foetida Species 0.000 description 1
- 206010019233 Headaches Diseases 0.000 description 1
- 241001508691 Martes zibellina Species 0.000 description 1
- 229920001410 Microfiber Polymers 0.000 description 1
- RAQQRQCODVNJCK-JLHYYAGUSA-N N-[(4-amino-2-methylpyrimidin-5-yl)methyl]-N-[(E)-5-hydroxy-3-(2-hydroxyethyldisulfanyl)pent-2-en-2-yl]formamide Chemical compound C\C(N(Cc1cnc(C)nc1N)C=O)=C(\CCO)SSCCO RAQQRQCODVNJCK-JLHYYAGUSA-N 0.000 description 1
- 239000004988 Nematic liquid crystal Substances 0.000 description 1
- 206010033799 Paralysis Diseases 0.000 description 1
- 241000270295 Serpentes Species 0.000 description 1
- 239000011149 active material Substances 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 229910052782 aluminium Inorganic materials 0.000 description 1
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 1
- 238000000429 assembly Methods 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 208000003464 asthenopia Diseases 0.000 description 1
- 239000011324 bead Substances 0.000 description 1
- 230000035559 beat frequency Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000002041 carbon nanotube Substances 0.000 description 1
- 229910021393 carbon nanotube Inorganic materials 0.000 description 1
- 150000001768 cations Chemical class 0.000 description 1
- 210000003161 choroid Anatomy 0.000 description 1
- 239000002642 cobra venom Substances 0.000 description 1
- 238000002591 computed tomography Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000003989 dielectric material Substances 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000004424 eye movement Effects 0.000 description 1
- 230000004438 eyesight Effects 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 230000004886 head movement Effects 0.000 description 1
- 231100000869 headache Toxicity 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 101150085091 lat-2 gene Proteins 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000003658 microfiber Substances 0.000 description 1
- 230000004459 microsaccades Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000008450 motivation Effects 0.000 description 1
- 230000007433 nerve pathway Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 238000010428 oil painting Methods 0.000 description 1
- 210000001328 optic nerve Anatomy 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 238000012856 packing Methods 0.000 description 1
- 239000003973 paint Substances 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 230000000737 periodic effect Effects 0.000 description 1
- 230000002085 persistent effect Effects 0.000 description 1
- 239000000049 pigment Substances 0.000 description 1
- 239000004033 plastic Substances 0.000 description 1
- 238000004321 preservation Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000002829 reductive effect Effects 0.000 description 1
- 238000002310 reflectometry Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 230000002207 retinal effect Effects 0.000 description 1
- 230000004256 retinal image Effects 0.000 description 1
- 238000005096 rolling process Methods 0.000 description 1
- 238000005488 sandblasting Methods 0.000 description 1
- 210000003786 sclera Anatomy 0.000 description 1
- 238000012163 sequencing technique Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 239000007921 spray Substances 0.000 description 1
- 238000005507 spraying Methods 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 239000002435 venom Substances 0.000 description 1
- 210000001048 venom Anatomy 0.000 description 1
- 231100000611 venom Toxicity 0.000 description 1
- 230000004470 vergence movement Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
- NWONKYPBYAMBJT-UHFFFAOYSA-L zinc sulfate Chemical compound [Zn+2].[O-]S([O-])(=O)=O NWONKYPBYAMBJT-UHFFFAOYSA-L 0.000 description 1
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B17/00—Systems with reflecting surfaces, with or without refracting elements
- G02B17/08—Catadioptric systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B2006/0098—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings for scanning
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
- G02B2027/0125—Field-of-view increase by wavefront division
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0132—Head-up displays characterised by optical features comprising binocular systems
- G02B2027/0134—Head-up displays characterised by optical features comprising binocular systems of stereoscopic type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0149—Head-up displays characterised by mechanical features
- G02B2027/0161—Head-up displays characterised by mechanical features characterised by the relative positioning of the constitutive elements
- G02B2027/0163—Electric or electronic control thereof
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B26/00—Optical devices or arrangements for the control of light using movable or deformable optical elements
- G02B26/08—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
- G02B26/0808—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light by means of one or more diffracting elements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0087—Phased arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0176—Head mounted characterised by mechanical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/42—Diffraction optics, i.e. systems including a diffractive element being designed for providing a diffractive effect
- G02B27/4205—Diffraction optics, i.e. systems including a diffractive element being designed for providing a diffractive effect having a diffractive optical element [DOE] contributing to image formation, e.g. whereby modulation transfer function MTF or optical aberrations are relevant
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B3/00—Simple or compound lenses
- G02B3/0006—Arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B3/00—Simple or compound lenses
- G02B3/0006—Arrays
- G02B3/0037—Arrays characterized by the distribution or form of lenses
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/22—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the stereoscopic type
- G02B30/24—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the stereoscopic type involving temporal multiplexing, e.g. using sequentially activated left and right shutters
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/34—Stereoscopes providing a stereoscopic pair of separated images corresponding to parallactically displaced views of the same object, e.g. 3D slide viewers
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/005—Diaphragms
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/18—Diffraction gratings
- G02B5/1814—Diffraction gratings structurally combined with one or more further optical elements, e.g. lenses, mirrors, prisms or other diffraction gratings
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/18—Diffraction gratings
- G02B5/1828—Diffraction gratings having means for producing variable diffraction
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/20—Filters
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/02—Optical fibres with cladding with or without a coating
- G02B6/02042—Multicore optical fibres
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/04—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres
- G02B6/06—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres the relative position of the fibres being the same at both ends, e.g. for transporting images
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/24—Coupling light guides
- G02B6/26—Optical coupling means
- G02B6/32—Optical coupling means having lens focusing means positioned between opposed fibre ends
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/24—Coupling light guides
- G02B6/26—Optical coupling means
- G02B6/34—Optical coupling means utilising prism or grating
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/24—Coupling light guides
- G02B6/36—Mechanical coupling means
- G02B6/40—Mechanical coupling means having fibre bundle mating means
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C2202/00—Generic optical aspects applicable to one or more of the subgroups of G02C7/00
- G02C2202/16—Laminated or compound lenses
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C7/00—Optical parts
- G02C7/02—Lenses; Lens systems ; Methods of designing lenses
- G02C7/04—Contact lenses for the eyes
- G02C7/049—Contact lenses having special fitting or structural features achieved by special materials or material structures
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C7/00—Optical parts
- G02C7/10—Filters, e.g. for facilitating adaptation of the eyes to the dark; Sunglasses
- G02C7/104—Filters, e.g. for facilitating adaptation of the eyes to the dark; Sunglasses having spectral characteristics for purposes other than sun-protection
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/0102—Constructional details, not otherwise provided for in this subclass
- G02F1/0105—Illuminating devices
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/13—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on liquid crystals, e.g. single liquid crystal display cells
- G02F1/133—Constructional arrangements; Operation of liquid crystal cells; Circuit arrangements
- G02F1/1333—Constructional arrangements; Manufacturing methods
- G02F1/1334—Constructional arrangements; Manufacturing methods based on polymer dispersed liquid crystals, e.g. microencapsulated liquid crystals
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/17—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on variable-absorption elements not provided for in groups G02F1/015 - G02F1/169
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
- G02F1/292—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection by controlled diffraction or phased-array beam steering
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
- G02F1/294—Variable focal length devices
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
- G02F1/31—Digital deflection, i.e. optical switching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1423—Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/506—Illumination models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/10—Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/003—Navigation within 3D models or images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
- G06T2207/10021—Stereoscopic video; Stereoscopic image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2213/00—Indexing scheme for animation
- G06T2213/08—Animation software package
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/001—Image restoration
- G06T5/002—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/001—Image restoration
- G06T5/003—Deblurring; Sharpening
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/02—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/10—Intensity circuits
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/18—Timing circuits for raster scan displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
- H04N13/117—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/286—Image signal generators having separate monoscopic and stereoscopic modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/327—Calibration thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/341—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] using temporal multiplexing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/344—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/361—Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/363—Image reproducers using image projection screens
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
- H04N13/383—Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0085—Motion estimation from stereoscopic image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/14—Picture signal circuitry for video frequency region
- H04N5/21—Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
-
- H04N5/225—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/74—Projection arrangements for image reproduction, e.g. using eidophor
Abstract
Configurations are disclosed for presenting virtual reality and augmented reality experiences to users. Disclosed is a system for displaying virtual content to a user. At least one light source multiplexes a plurality of light beams to display a respective plurality of light patterns associated with one or more frames of image data. A plurality of waveguides receives the plurality of light beams and directs the plurality of light beams toward an exit pupil. The plurality of waveguides are stacked along an optical axis of the user. At least one optical element modifies a focus of a light beam of the plurality of light beams directed by the plurality of waveguides. one or more frames of image data. A plurality of waveguides receives the plurality of light beams and directs the plurality of light beams toward an exit pupil. The plurality of waveguides are stacked along an optical axis of the user. At least one optical element modifies a focus of a light beam of the plurality of light beams directed by the plurality of waveguides.
Description
L AND AUGMENTED REALITY SYSTEMS AND METHODS
FIELD OF THE INVENTION
The present disclosure relates to virtual reality and augmented reality g and
visualization systems.
BACKGROUND
Modern computing and display technologies have facilitated the development of
systems for so called “virtual y” or “augmented reality” experiences, wherein digitally
reproduced images or portions thereof are presented to a user in a manner wherein they seem
to be, or may be perceived as, real. A virtual reality, or “VR”, scenario typically involves
presentation of digital or l image information without transparency to other actual real-
world visual input; an augmented reality, or “AR”, scenario typically involves presentation of
digital or virtual image information as an tation to visualization of the actual world
around the user. For example, referring to Figure 1, an augmented reality scene (4) is depicted
wherein a user of an AR technology sees a real-world ike setting (6) featuring people,
trees, buildings in the background, and a concrete platform (1120). In addition to these items,
the user of the AR technology also perceives that he “sees” a robot statue (1110) standing
upon the real-world platform (1120), and a cartoon-like avatar character (2) flying by which
seems to be a personification of a bumble bee, even though these elements (2, 1110) do not
exist in the real world. As it turns out, the human visual perception system is very complex, and
producing a VR or AR technology that facilitates a comfortable, natural-feeling, rich
presentation of virtual image elements amongst other l or orld imagery elements is
nging.
Referring to Figure 2A, stereoscopic wearable glasses (8) type configurations have
been developed which generally e two displays (10, 12) that are configured to y
images with slightly different element presentation such that a three-dimensional perspective is
perceived by the human visual system. Such configurations have been found to be
uncomfortable for many users due to a ch between vergence and accommodation which
must be overcome to perceive the images in three dimensions; indeed, some users are not
able to tolerate stereoscopic configurations. Figure 2B shows another stereoscopic wearable
glasses (14) type configuration featuring two forward-oriented s (16, 18) configured to
capture images for an augmented reality presentation to the user through scopic
displays. The on of the cameras (16, 18) and displays generally blocks the natural field of
view of the user when the glasses (14) are mounted on the user’s head.
Referring to Figure 2C, an augmented reality configuration (20) is shown which
features a visualization module (26) coupled to a glasses frame (24) which also holds
conventional glasses lenses (22). The user is able to see an at least partially unobstructed
view of the real world with such a system, and has a small display (28) with which digital
imagery may be presented in an AR configuration to one eye — for a monocular AR
presentation. Figure 2D features a configuration wherein a ization module (32) may be
coupled to a hat or helmet (30) and ured to present monocular augmented digital imagery
to a user through a small display (34). Figure 2E illustrates another similar configuration
wherein a frame (36) couple-able to a user’s head in a manner similar to an eyeglasses
ng so that a visualization module (38) may be utilized to capture images and also present
monocular augmented digital imagery to a user through a small display (40). Such a
configuration is available, for example, from , Inc, of Mountain View, CA under the trade
name GoogleGlass (RTM). None of these configurations is optimally suited for presenting a
rich, binocular, three-dimensional augmented reality experience in a manner that will be
comfortable and lly useful to the user, in part because prior systems fail to address
some of the ental aspects of the human perception system, including the
photoreceptors of the retina and their interoperation with the brain to produce the perception of
visualization to the user.
Referring to Figure 3, a simplified cross-sectional view of a human eye is depicted
ing a cornea (42), iris (44), lens — or “crystalline lens” (46), sclera (48), choroid layer (50),
macula (52), retina (54), and optic nerve pathway (56) to the brain. The macula is the center of
the retina, which is utilized to see moderate detail; at the center of the macula is a n of
the retina that is referred to as the “fovea”, which is utilized for seeing the finest details, and
which contains more eceptors (approximately 120 cones per visual degree) than any
other portion of the . The human visual system is not a passive sensor type of system; it
is configured to actively scan the environment. In a manner somewhat akin to use of a flatbed
scanner to capture an image, or use of a finger to read Braille from a paper, the photoreceptors
of the eye fire in response to changes in stimulation, rather than constantly responding to a
constant state of ation. Thus motion is required to present photoreceptor ation to
the brain (as is motion of the linear scanner array across a piece of paper in a flatbed scanner,
or motion of a finger across a word of Braille ted into a paper). Indeed, experiments with
substances such as cobra venom, which has been utilized to paralyze the muscles of the eye,
have shown that a human subject will experience blindness if positioned with his eyes open,
viewing a static scene with venom-induced paralysis of the eyes. In other words, without
changes in stimulation, the photoreceptors do not provide input to the brain and blindness is
experienced. It is believed that this is at least one reason that the eyes of normal humans have
been observed to move back and forth, or , in side-to-side motion in what are called
“microsaccades”.
As noted above, the fovea of the retina contains the st density of
photoreceptors, and while humans lly have the perception that they have high-resolution
visualization capabilities throughout their field of view, they generally actually have only a small
high-resolution center that they are mechanically sweeping around a lot, along with a persistent
memory of the high-resolution information recently captured with the fovea. In a somewhat
similar manner, the focal distance control mechanism of the eye (ciliary muscles operatively
coupled to the crystalline lens in a manner wherein ciliary relaxation causes taut ciliary
connective fibers to flatten out the lens for more distant focal lengths; ciliary ction causes
loose ciliary connective fibers, which allow the lens to assume a more rounded geometry for
more close-in focal lengths) dithers back and forth by imately 1A to 1/2 diopter to cyclically
induce a small amount of what is called “dioptric blur” on both the close side and far side of the
targeted focal ; this is utilized by the accommodation control circuits of the brain as
cyclical negative feedback that helps to constantly correct course and keep the retinal image of
a fixated object approximately in focus.
The visualization center of the brain also gains valuable perception information from
the motion of both eyes and components thereof relative to each other. Vergence movements
(i.e., rolling nts of the pupils toward or away from each other to converge the lines of
sight of the eyes to fixate upon an object) of the two eyes relative to each other are closely
associated with ng (or “accommodation”) of the lenses of the eyes. Under normal
conditions, changing the focus of the lenses of the eyes, or accommodating the eyes, to focus
upon an object at a different distance will tically cause a matching change in ce
to the same distance, under a relationship known as the “accommodation-vergence .”
Likewise, a change in ce will trigger a matching change in accommodation, under normal
conditions. Working against this , as do most conventional stereoscopic AR or VR
configurations, is known to produce eye fatigue, headaches, or other forms of discomfort in
users.
Movement of the head, which houses the eyes, also has a key impact upon
visualization of objects. Humans move their heads to visualize the world around them; they
often are in a fairly constant state of repositioning and reorienting the head relative to an object
of interest. Further, most people prefer to move their heads when their eye gaze needs to
move more than about 20 degrees off center to focus on a particular object (i.e., people do not
typically like to look at things “from the corner of the eye”). Humans also typically scan or move
their heads in relation to sounds — to improve audio signal capture and utilize the geometry of
the ears relative to the head. The human visual system gains powerful depth cues from what is
called “head motion parallax”, which is related to the relative motion of s at different
distances as a function of head motion and eye vergence ce (i.e., if a person moves his
head from side to side and maintains fixation on an , items farther out from that object will
move in the same direction as the head; items in front of that object will move opposite the
head motion; these are very salient cues for where things are spatially in the environment
relative to the person — perhaps as ul as stereopsis). Head motion also is utilized to look
around objects, of course.
Further, head and eye motion are nated with something called the “vestibulo-
ocular reflex”, which stabilizes image information relative to the retina during head rotations,
thus keeping the object image information approximately centered on the retina. In response to
a head rotation, the eyes are reflexively and proportionately rotated in the te direction to
maintain stable fixation on an object. As a result of this satory relationship, many
humans can read a book while shaking their head back and forth (interestingly, if the book is
panned back and forth at the same speed with the head imately stationary, the same
generally is not true — the person is not likely to be able to read the moving book; the vestibuloocular
reflex is one of head and eye motion coordination, generally not developed for hand
motion). This paradigm may be important for augmented reality systems, because head
motions of the user may be associated relatively directly with eye s, and the system
preferably will be ready to work with this onship.
Indeed, given these various relationships, when placing digital content (e.g., 3-D
content such as a virtual chandelier object presented to t a orld view of a room;
or 2-D content such as a planar/flat virtual oil painting object presented to augment a real-world
view of a room), design choices may be made to control behavior of the objects. For example,
the 2-D oil painting object may be head-centric, in which case the object moves around along
with the user’s head (e.g., as in a GoogleGlass approach); or the object may be world-centric,
in which case it may be presented as though it is part of the real world coordinate system, so
that the user may move his head or eyes without moving the position of the object relative to
the real world.
Thus when placing virtual content into the augmented reality world presented with
an augmented reality system, whether the object should be presented as world centric (i.e., the
virtual object stays in position in the real world so that the user may move his body, head, eyes
around it without changing its position relative to the real world s surrounding it, such as a
real world wall); body, or torso, centric, in which case a virtual element may be fixed relative to
the user’s torso, so that the user can move his head or eyes without moving the object, but that
is slaved to torso movements; head c, in which case the yed object (and/or display
itself) may be moved along with head movements, as described above in reference to
GoogleGlass; or eye centric, as in a “foveated display” configuration, as is described below,
wherein t is slewed around as a function of what the eye position is.
With world-centric configurations, it may be desirable to have inputs such as
accurate head pose measurement, accurate representation and/or measurement of real world
objects and geometries around the user, low-latency dynamic rendering in the augmented
reality display as a function of head pose, and a generally tency display.
The systems and techniques described herein are configured to work with the visual
configuration of the l human to address these challenges.
Embodiments of the present ion are directed to devices, systems and s
for facilitating l reality and/or augmented reality interaction for one or more users. In one
aspect, a system for displaying virtual content is disclosed.
In one or more embodiments, the system comprises a light source to lex one
or more light patterns associated with one or more frames of image data in a time-sequential
manner; and an array of reflectors to receive the one or more light ns and variably
converge light on an exit pupil.
In one or more embodiment, the system comprises an image-generating source to
provide one or more frames of image data in a time-sequential manner, a light modulator
configured to transmit light associated with the one or more frames of image data, a substrate
to direct image information to a user’s eye, wherein the substrate houses a plurality of
tors, a first reflector of the plurality of reflectors to reflect light associated with a first frame
of image data at a first angle to the user’s eye, and a second reflector of the plurality of
reflectors to t light associated with a second frame of image data at a second angle to the
user’s eye.
The angle of reflection of the plurality of reflectors may be variable in one or more
embodiments. The reflectors may be switchable in one or more embodiments. The plurality of
reflectors may be electro-optically active in one or more embodiments. A refractive index of the
plurality of reflectors may be varied to match a refractive index of the substrate in one or more
embodiments. In an optional embodiment, the system may also comprise a high-frequency
gating layer configurable to be placed in between the substrate and the user’s eye, the highfrequency
gating layer having an aperture that is controllably movable. The aperture of the
high-frequency gating layer may be moved in a manner such that image data is selectively
transmitted only through the light that is reflected h the aperture in one or more
embodiments. One or more reflectors of the transmissive beamsplitter substrate may be
d by the high-frequency gating layer. The aperture may be an LCD aperture in one or
more embodiments. The re may be a MEMs array in one or more embodiments. The
first angle may be the same as the second angle in one or more embodiments. The first angle
may be different than the second angle in one or more embodiments.
In one or more embodiments, the system may further comprise a first lens to steer
a set of light rays through a nodal point and to the user’s eye. The first lens may be
configurable to be placed on the substrate and in front of the first reflector such that the set of
light rays exiting the reflector pass through the first lens before reaching the user’s eye in one
or more embodiments.
The system may further comprise a second lens to compensate for the first lens,
the second lens configurable to be placed on the substrate and on a side opposite to the side
on which the first lens is placed, thereby resulting in zero magnification, in one or more
embodiments.
The first reflector of the plurality of reflectors may be a curved reflective surface to
collect a set of light rays associated with the image data into a single output point before being
red to the user’s eye in one or more ments. The curved reflector may a parabolic
tor in one or more embodiments. The curved tor may be an elliptical reflector in one
or more embodiments.
In another embodiment, a method for displaying virtual content comprises providing
one or more light patterns associated with one or more frames of image data in a time-
tial manner and reflecting the one or more light patterns associated with the one or
more frames of image data via a transmissive beamsplitter to an exit pupil, the transmissive
beamsplitter having a ity of reflectors to variably converge on the exit pupil.
The angle of reflection of the plurality of reflectors may be variable in one or more
embodiments. The tors may be switchable in one or more ments. The plurality of
reflectors may be electro-optically active in one or more embodiments. A refractive index of the
plurality of tors may be varied to match a refractive index of the substrate in one or more
embodiments. In an al embodiment, the system may also se a high-frequency
gating layer configurable to be placed in between the substrate and the user’s eye, the highfrequency
gating layer having an re that is controllably movable. The aperture of the
high-frequency gating layer may be moved in a manner such that image data is selectively
transmitted only through the light that is reflected through the aperture in one or more
ments. One or more reflectors of the transmissive litter substrate may be
blocked by the high-frequency gating layer. The aperture may be an LCD aperture in one or
more embodiments. The aperture may be a MEMs array in one or more embodiments. The
first angle may be the same as the second angle in one or more embodiments. The first angle
may be different than the second angle in one or more ments.
In one or more embodiments, the system may further comprise a first lens to steer
a set of light rays through a nodal point and to the user’s eye. The first lens may be
configurable to be placed on the substrate and in front of the first reflector such that the set of
light rays exiting the tor pass through the first lens before reaching the user’s eye in one
or more embodiments.
The system may further comprise a second lens to compensate for the first lens,
the second lens configurable to be placed on the substrate and on a side opposite to the side
on which the first lens is placed, thereby resulting in zero magnification, in one or more
embodiments.
The first reflector of the plurality of reflectors may be a curved reflective surface to
collect a set of light rays associated with the image data into a single output point before being
delivered to the user’s eye in one or more embodiments. The curved reflector may a parabolic
reflector in one or more embodiments. The curved reflector may be an elliptical reflector in one
or more embodiments.
In one or more embodiments, the ont may be collimated. In one or more
embodiments, the wavefront may be curved. The collimated wavefront may be perceived as an
infinity depth plane, in some embodiments. The curved wavefront may be perceived as a depth
plane closer than optical infinity, in some embodiments.
In another embodiment, a system for displaying virtual content to a user, comprises
a light source to lex one or more light patterns associated with one or more frames of
image data in a time-sequential manner, an array of reflectors to receive the one or more light
patterns, the array of tors oriented at a particular angle, and a plurality of optical elements
coupled to the array of reflectors to variably converge the light patterns on an exit pupil.
In one or more embodiments, the array of reflectors may be separate from the
l elements, in one or more embodiments. The array of reflectors may comprise flat
mirrors, in one or more embodiments. The optical elements may be lenslets coupled to the
array of reflectors, in one or more embodiments. The one or more tors of the array of
reflectors may be curved, in one or more embodiments. The optical elements may be
integrated into the array of reflectors. The plurality of optical elements may expand an exit
pupil, in one or more embodiments.
The system may further comprise a first lens to steer a set of light rays h a
nodal point and to the user’s eye, wherein the first lens is configurable to be placed on the
ate and in front of the first reflector such that the set of light rays exiting the reflector
pass through the first lens before reaching the user’s eye, in one or more embodiments.
The system may further comprise a second lens to compensate for the first lens,
the second lens configurable to be placed on the substrate and on a side opposite to the side
on which the first lens is placed, thereby ing in zero magnification, in one or more
embodiments. The plurality of reflectors may comprise wavelength-selective reflectors, in one
or more embodiments. The plurality of reflectors may comprise half-silvered mirrors, in one or
more embodiments. The plurality of optical elements may comprise refractive lenses. The
plurality of optical elements may comprise diffractive lenses, in one or more embodiments. The
curved reflectors may comprise wave-length selective notch filters, in one or more
embodiments.
In another embodiment, a method for displaying virtual content to a user comprises
providing one or more light patterns associated with one or more frames of image data in a
time-sequential manner, reflecting the one or more light ns associated with the one or
more frames of image data via a transmissive beamsplitter to an exit pupil, the issive
beamsplitter having a plurality of reflectors to variably converge on the exit pupil, and
expanding an exit pupil through a plurality of optical elements coupled to the plurality of
reflectors of the transmissive beamsplitter.
In one or more embodiments, the array of tors may be separate from the
optical elements. In one or more embodiments the array of reflectors se flat mirrors.
The optical elements may be lenslets coupled to the array of tors, in one or more
ments.
In another embodiment, a system for displaying virtual content to a user, ses
a light source to multiplex one or more light ns associated with one or more frames of
image data in a time-sequential manner, and a ide to receive the one or more light
patterns and converge the light patterns to a first focus, and a variable focus element (VFE)
coupled to the waveguide to converge at least some of the light patterns to a second focus.
In one or more embodiments, the VFE is telecentric, in one or more embodiments.
The VFE is non-telecentric, in one or more embodiments. The system further comprises a
compensating lens such that the user’s view of the outside world is undistorted, in one or more
embodiments. The plurality of frames are presented to the user at a high ncy such that
the user perceives the frames as part of a single nt scene, wherein the VFE varies the
focus from a first frame to a second frame, in one or more embodiments. The light source is a
scanned light display, and wherein the VFE varies the focus in a line-by-line manner, in one or
more embodiments. The light source is a scanned light display, and wherein the VFE varies
the focus in a pixel-by—pixel manner, in one or more embodiments.
The VFE is a diffractive lens, in one or more ments. The VFE is a refractive
lens, in one or more embodiments. The VFE is a reflective , in one or more
embodiments. The reflective mirror is opaque, in one or more embodiments. The reflective
mirror is partially reflective, in one or more embodiments. The system further ses an
accommodation module to track an accommodation of a user’s eyes, wherein the VFE varies
the focus of the light patterns based at least in part on the accommodation of the user’s eyes,
in one or more embodiments.
In yet another embodiment, system for displaying virtual content to a user,
comprises a light source to multiplex one or more light ns associated with one or more
frames of image data in a time-sequential manner, a waveguide to receive the one or more
light patterns and converge the light ns to a first focus, and a variable focus element
(VFE) coupled to the waveguide to converge at least some of the light ns to a second
focus, wherein the VFE is integrated into the waveguide.
In another embodiment, a system for displaying virtual content to a user ses
a light source to multiplex one or more light patterns associated with one or more frames of
image data in a time-sequential manner, a waveguide to receive the one or more light patterns
and converge the light patterns to a first focus, and a variable focus t (VFE) coupled to
the waveguide to converge at least some of the light patterns to a second focus, wherein the
VFE is separate from the waveguide.
In another aspect, a method for ying virtual content to a user, comprises
providing one or more light patterns associated with one or more frames of image data,
converging the one or more light patterns associated with the one or more frames of image
data to a first focus h a waveguide, and modifying, through a variable focus element
(VFE), the first focus of the light to produce a wavefront at a second focus.
The VFE is separate from the waveguide, in one or more embodiments. The VFE
is ated into the waveguide, in one or more embodiments. The one or more frames of
image data are provided in a time-sequential manner, in one or more embodiments. The VFE
modifies the focus of the one or more frames of image data on a frame-by-frame basis, in one
or more embodiments. The VFE modifies the focus of the one or more frames of image data
on a pixel-by—pixel basis, in one or more embodiments. The VFE modifies the first focus to
produce a wavefront at a third focus, wherein the second focus is different than the third focus,
in one or more embodiments. The wavefront at the second focus is perceived by the user as
coming from a particular depth plane, in one or more embodiments.
In some embodiments, the plurality of frames are presented to the user at a high
frequency such that the user ves the frames as part of a single coherent scene, wherein
the VFE varies the focus from a first frame to a second frame. The light source is a scanned
light display, and wherein the VFE varies the focus in a line-by-line , in one or more
embodiments.
In another embodiment, a system for displaying virtual content to a user, comprises
a ity of waveguides to receive light rays associated with image data and to transmit the
light rays toward the user’s eyes, wherein the plurality of waveguides are stacked in a direction
facing the user’s eye, and a first lens coupled to a first waveguide of the plurality of
waveguides to modify light rays transmitted from the first waveguide, thereby ring light
rays having a first wavefront curvature, and a second lens coupled to a second waveguide of
the plurality of waveguides to modify light rays transmitted from the second waveguide,
thereby delivering light rays having a second wavefront curvature, n the first lens
coupled to the first waveguide and the second lens coupled to the second ide are
stacked horizontally in a ion facing the user’s eye.
In one or more embodiments, the first ont curvature is different from the
second ont ure. The system further comprises a third waveguide of the plurality of
waveguides to deliver co||imated light to the user’s eye, such that the user perceives the image
data as coming from an optical infinity plane, in one or more embodiments. The waveguide is
configured to transmit co||imated light to the lens, in one or more embodiments.
The system further comprises a compensating lens layer to compensate for an
aggregate power of the lenses stacked in the direction facing the user’s eyes, wherein the
compensating lens layer is stacked st from the user’s eye, in one or more embodiments.
The waveguide comprises a plurality of reflectors configurable to t the light rays injected
into the waveguide toward the user’s eye, in one or more embodiments.
The waveguide is electro-active, in one or more embodiments. The waveguide is
switchable, in one or more embodiments. The light rays having the first wavefront curvature
and the light rays having the second wavefront curvature are delivered simultaneously, in one
or more embodiments. The light rays having the first wavefront curvature and the light rays
having the second wavefront curvature are delivered sequentially, in one or more
embodiments. The second wavefront curvature corresponds to a margin of the first wavefront
curvature, y providing a focal range in which the user can accommodate, in one or more
ments. The system further comprises an accommodation module to track an
accommodation of a user’s eyes, and wherein the VFE varies the focus of the light patterns
based at least in part on the odation of the user’s eyes, in one or more embodiments.
In yet another embodiments, a system for displaying virtual content to a user,
comprises a light source to multiplex one or more light patterns associated with one or more
frames of image data in a time-sequential manner, a ity of waveguides to e the one
or more light patterns and to converge light to an exit pupil, n the plurality of waveguides
are stacked along a z-axis and away from the user’s line of vision, and at least one optical
element d to the stacked waveguides to modify a focus of the light transmitted by the
plurality of ides.
The waveguide of the plurality of waveguides may se a waveguide to
distribute the projected light across the length of the waveguide, and a lens to modify the light
in a manner such that a wavefront curvature is created, wherein the created wavefront
curvature corresponds to a focal plane when viewed by the user in one or more embodiments.
The ide of the plurality of waveguides comprises a diffractive optical
element (DOE) in one or more embodiments. The DOE is switchable between an on and off
state, in one or more embodiments. The waveguide of the plurality of waveguides comprises a
refractive lens, in one or more embodiments. The waveguide of the plurality of waveguides
comprises a Fresnel zone plate, in one or more embodiments. The waveguide of the plurality
of waveguides comprises a substrate guided optics (SGO) element, in one or more
embodiments. The waveguide is switchable between an on and off state, in one or more
embodiments. The ide is static, in one or more embodiments. The first frame of
image data and second frame of image data are delivered to the user’s eye simultaneously, in
one or more embodiments. The first frame of image data and second frame of image data are
delivered to the user’s eye sequentially, in one or more embodiments.
The system r comprises a plurality of angled reflectors to deliver light to the
user’s eye, wherein the first waveguide component and the second waveguide component
direct light to the one or more angled reflectors, in one or more embodiments. The system
further comprises a beam distribution waveguide optic, the beam distribution waveguide
coupled to the waveguide assembly, wherein the beam distribution waveguide optic is
configurable to spread the projected light across the ide assembly, such that a light ray
injected into the beam distributed waveguide optic is cloned and injected into waveguide
components of the waveguide ly, in one or more embodiments.
In another embodiment, a system for displaying virtual content to a user comprises
an image-generating source to provide one or more frames of image data in a time-sequential
manner, a light modulator to project light associated with the one or more frames of image
data, a waveguide assembly to receiving the projected light and deliver the light towards the
user’s eye, wherein the waveguide ly comprises at least a first waveguide component
configurable to modify light associated with a first frame of the image data such the light is
perceived as coming from a first focal plane, and a second waveguide ent configurable
to modify light associated with a second frame of the image data such that the light is
perceived as coming from a second focal plane, and n the first waveguide component
and second waveguide component are stacked along a z-axis in front of the user’s eye.
In some embodiments, the waveguide ent of the ide assembly
ses a ide to distribute the projected light across the length of the waveguide,
and a lens to modify the light in a manner such that a wavefront curvature is created, wherein
the created wavefront curvature corresponds to a focal plane when viewed by the user. The
waveguide component of the waveguide ly comprises a diffractive optical element
(DOE) in one or more ments.
The DOE is switchable between an on and off state in one or more embodiments. .
The waveguide component of the waveguide assembly comprises a tive lens in one or
more ments. The ide component of the waveguide assembly comprises a
Fresnel zone plate in one or more embodiments. The first frame of image data and second
frame of image data are delivered to the user’s eye simultaneously in one or more
embodiments. The first frame of image data and second frame of image data are delivered to
the user’s eye sequentially, in one or more ments.
The system further comprises a ity of angled reflectors to deliver light to the
user’s eye, wherein the first waveguide component and the second waveguide component
direct light to the one or more angled reflectors in one or more embodiments. The system
further comprises a beam distribution waveguide optic, the beam distribution waveguide
coupled to the waveguide assembly, wherein the beam distribution waveguide optic is
configurable to spread the projected light across the waveguide assembly, such that a light ray
injected into the beam distributed waveguide optic is cloned and injected into waveguide
components of the waveguide assembly in one or more embodiments.
The waveguide component of the waveguide assembly comprises a reflector
configurable to reflect the projected light at a desired angle toward the user’s eye. The first
waveguide component comprises a first reflector configured to reflect the projected light at a
first angle, and wherein the second waveguide component comprises a second reflector to
reflect the projected light at a second angle in one or more ments. The first reflector is
staggered in relation to the second reflector, thereby expanding a field of view of the image as
viewed by the user in one or more embodiments.
The reflectors of the ide components are positioned in a manner to form a
continuous curved tion surface across the waveguide assembly in one or more
embodiments. The continuous curved reflection surface comprises a parabolic curve in one or
more embodiments. The continuous curved reflection e comprises an elliptical curve in
one or more embodiments.
In yet another embodiment, a method for displaying virtual content to a user
comprises delivering, through a first waveguide, light rays associated with a first frame of
image data to the user, the light rays having a first wavefront ure, and delivering, through
a second waveguide, light rays associated with a second frame of image data to the user, the
light rays having a second wavefront curvature, wherein the first waveguide and second
waveguide are stacked along a z axis facing the user’s eyes.
The first ont curvature and the second wavefront curvature are delivered
simultaneously, in one or more embodiments. The first ont curvature and the second
wavefront curvature are delivered sequentially, in one or more ments. The first and
second wavefront curvatures are perceived as a first and second depth plane by the user, in
one or more embodiments. The first and second ides are coupled to one or more
optical elements, in one or more embodiments. The method may further comprise
compensating for an effect of the one or more optical elements through a compensation lens,
in one or more embodiments.
The method may further comprise determining an accommodation of the user’s
eyes, and delivering light rays through at least one of the first and second waveguides based
at least in part on the determined accommodation, in one or more embodiments.
In r embodiment, a method for displaying virtual content to a user comprises
determining an accommodation of the user’s eyes, delivering, through a first ide of a
stack of waveguides, light rays having a first wavefront curvature based at least in part on the
ined accommodation, wherein the first wavefront curvature corresponds to a focal
distance of the determined accommodation, and delivering, through a second waveguide of
the stack of waveguides, light rays having a second wavefront curvature, the second wavefront
curvature associated with a predetermined margin of the focal distance of the ined
accommodation.
The margin is a positive margin, in one or more embodiments. The margin is a
negative margin, in one or more embodiments. The second waveguide increases a focal range
in which the user can accommodate, in one or more embodiments. The first waveguide is
coupled to a variable focus element (VFE), wherein the VFE varies a focus at which the
waveguide focuses the light rays, in one or more embodiments. The focus is varied based at
least in part on the determined accommodation of the users’ eyes, in one or more
embodiments. The first ont curvature and the second wavefront curvature are delivered
simultaneously, in one or more embodiments.
The first and second wavefront curvatures are ved as a first and second
depth plane by the user, in one or more embodiments. The waveguide is a diffractive optical
element (DOE) in one or more embodiments. The waveguide is a substrate guided
, optic
(SGO), in one or more embodiments. The first and second waveguides are switchable, in one
or more embodiments. The waveguide comprises one or more switchable elements, in one or
more embodiments.
In yet another embodiment, a system for displaying virtual content to a user
comprises an image-generating source to provide one or more frames of image data in a time-
sequential manner, a y assembly to project light rays associated with the one or more
frames of image data, the display ly comprises a first display element corresponding to
a first frame-rate and a first bit depth, and a second y element corresponding to a second
frame-rate and a second bit depth, and a variable focus element (VFE) configurable to vary a
focus of the projected light and transmit the light to the user’s eye.
The first frame-rate is higher than the second frame-rate, and the first bit depth is
lower than the second bit depth, in one or more embodiments. The first display element is a
DLP projection system, in one or more embodiments. The second display element is a
liquid l display (LCD), in one or more embodiments. The first display element projects
light to a subset of the second display element such that a ery of the LCD has nt
illumination, in one or more embodiments. Only light transmitted from the first display element
is focused through the VFE, in one or more embodiments.
The VFE is optically conjugate to an exit pupil, such that a focus of the projected
light is varied without affecting a magnification of the image data, in one or more embodiments.
The first display element is a DLP, and the second display element is an LCD, and wherein the
DLP is of low resolution, and n the LCD is of high resolution, in one or more
embodiments. The intensity of backlight is varied over time to equalize a brightness of sub-
images ted by the first display t, thereby increase a frame rate of the first display
element, in one or more ments.
The VFE is configurable to vary the focus of the projected light on a frame-by-frame
basis, in one or more embodiments. The system further comprises software to compensate for
optical magnification associated with an operation of the VFE, in one or more embodiments.
The generating source produces slices of a particular image that when projected
together or sequentially produces a three-dimensional volume of an object, in one or more
embodiments. The DLP is operated in a binary mode, in one or more embodiments. The
DLP is operated in a grayscale mode, in one or more embodiments.
The VFE varies the projected light such that a first frame is perceived as coming
from a first focal plane, and a second frame is perceived as coming from a second focal plane,
wherein the first focal plane is different from the second focal plane, in one or more
ments. The focal distance associated with the focal plane is fixed, in one or more
embodiments. The focal distance associated with the focal plane is variable, in one or more
embodiments.
In r embodiment, a method for displaying virtual content to a user comprises
providing one or more image slices, wherein a first and second image slice of the one or more
image slices ents a three-dimensional volume, projecting light ated with the first
image slice through a spatial light modulator, focusing, through a variable focus element
(VFE), the first image slice to a first focus, delivering the first image slice having the first focus
to the user, providing light associated with the second image slice, focusing, through the VFE,
the second image slice to a second focus, wherein the first focus is different from the second
focus, and delivering the second image slice having the second focus to the user.
The method may further comprise determining an accommodation of the user’s
eyes, wherein the VFE focuses the projected light based at least in part on the determined
accommodation, in one or more embodiments. The image slices are provided in a frame-
sequential fashion, in one or more embodiments. The first image slice and the second image
slice are delivered aneously, in one or more embodiments. The first image slice and the
second image slice are delivered sequentially, in one or more embodiments.
In yet another embodiment, a method for displaying virtual content to a user
comprises combining a first display element with a second display element, the first display
element corresponding to a high frame rate and a low bit depth, and the second display
element corresponding to a low frame rate and a high bit depth, such that the combined
y elements pond to a high frame rate and high bit depth, projecting light associated
with one or more frames of image data h the combined display elements, and switching
a focus of the projected light, through a variable focus element (VFE), on a frame-by-frame
basis, such that a first image slice is projected at a first focus, and a second image slice is
projected at a second focus.
In another embodiment, a system for displaying l content to a user, ses
a ity of lightguides to receive coherent light associated with one or more frames of image
data and to produce an aggregate wavefront, a phase modulator coupled to one or more
lightguides of the plurality of light guides to induce a phase delay in the light projected by the
one or more lightguides, and a processor to control the phase modulator in a manner such that
the ate wavefront ed by the plurality of lightguides is varied.
The wavefront produced a lightguide of the plurality of light guides is a spherical
wavefront, in one or more embodiments. The spherical onts produced by at least two
lightguides constructively interfere with each other, in one or more embodiments. The
spherical wavefronts produced by the at least two lightguides destructively interfere with each
other, in one or more embodiments. The aggregate wavefront is an approximately planar
wavefront, in one or more ments.
The planar wavefront corresponds to an optical infinity depth plane. The
aggregate wavefront is cal, in one or more embodiments. The spherical wavefront
corresponds to a depth plane closer than optical infinity, in one or more embodiments. An
inverse Fourier transform of a desired beam is injected into the multicore , such that a
desired aggregate wavefront is produced, in one or more embodiments.
In another aspect, a system of displaying virtual content to a user comprises an
image-generating source to provide one or more frames of image data, a multicore assembly
comprising a plurality of multicore fibers to project light associated with the one or more frames
of image data, a multicore fiber of the plurality of multicore fibers emitting light in a wavefront,
such that the multicore assembly produces an aggregate wavefront of the projected light, and
a phase tor to induce phase delays between the multicore fibers in a manner such that
the ate wavefront emitted by the multicore assembly is varied, thereby varying a focal
ce at which the user perceives the one or more frames of image data.
In yet another aspect, a method for displaying virtual t to a user comprises
emitting light through a multicore fiber, the ore fiber comprising a plurality of single core
fibers, wherein the singlecore fibers emit a spherical ont, providing an aggregate
ont from light emitted from the plurality of single core fibers, and inducing a phase delay
between the singlecore fibers of the multicore fiber such the aggregate ont ed by
the ore fiber is varied based at least in part on the induced phase delay.
The aggregate wavefront is a planar wavefront, in one or more embodiments. The
planar wavefront corresponds to optical infinity, in one or more embodiments. The aggregate
wavefront is spherical, in one or more embodiments. The spherical wavefront corresponds to
a depth plane closer than optical infinity, in one or more embodiments. The method further
comprises injecting an e Fourier transform of a desired wavefront into the multicore fiber
such that the aggregate wavefront corresponds to the desired wavefront, in one or more
embodiments.
In yet another embodiment, a system for displaying l content to a user
comprises an image-generating source to provide one or more frames of image data, a
multicore ly comprising a plurality of multicore fibers to project light associated with the
one or more frames of image data, an image injector to input images into the multicore
assembly, wherein the input injector is further configurable to input an inverse Fourier
transform of a desired wavefront into the multicore assembly such that the multicore assembly
outputs the Fourier transform by producing light ated with the image data in the desired
wavefront, thereby allowing the user to perceive the image data at a desired focal distance.
The d wavefront is associated with a hologram, in one or more
embodiments. The inverse Fourier transform is input to modulate a focus of the one or more
light beams, in one or more embodiments. A multicore fiber of the plurality of multicore fibers
is a multi-mode fiber, in one or more embodiments. The multicore fiber of the ity of
multicore fibers is configured to propagate light along a plurality of paths along the fiber, in one
or more embodiments. The multicore fiber is a single core fiber, in one or more embodiments.
The multicore fiber is a concentric core fiber, in one or more embodiments.
The image or is configured to input a wavelet pattern into the multicore
assembly, in one or more embodiments. The image injector is configured to input a Zernike
coefficient into the multicore assembly, in one or more embodiments. The system further
ses an odation tracking module to determine an accommodation of the user’s
eye, wherein the image injector is configured to input an inverse Fourier transform of a
wavefront corresponding to the determined accommodation of the user’s eyes, in one or more
embodiments.
In yet another embodiment, a method of displaying l content to a user
comprises determining an accommodation of the user’s eyes, wherein the determined
odation is associated with a focal ce corresponding to a user’s t state of
focus, projecting, through a waveguide, light associated with one or more frames of image
data, varying a focus of the projected light based at least in part on the determined
accommodation, and delivering the projected light to the user’s eyes, such that the light is
perceived by the user as coming from the focal distance corresponding to the user’s current
state of focus.
The accommodation is measured directly, in one or more embodiments. The
accommodation is measured indirectly, in one or more embodiments. The accommodation is
measured through an infrared autorefractor. The odation is measured through
eccentric photorefraction, in one or more embodiments. The method further comprises
measuring a convergence level of two eyes of the user to te the accommodation, in one
or more embodiments. The method further ses blurring one or more portions of the one
or more frames of image data based at least in part on the determined accommodation, in one
or more embodiments. The focus is varied between fixed depth planes, in one or more
embodiments. The method further comprises a compensating lens to compensate for an
optical effect of the waveguide, in one or more embodiments.
In one or more embodiments, a method of displaying virtual content to a user
comprises ining an accommodation of the user’s eyes, wherein the determined
accommodation is associated with a focal distance corresponding to a user’s current state of
focus, projecting, h a diffractive optics element (DOE), light associated with one or more
frames of image data, varying a focus of the projected light based at least in part on the
determined odation, and delivering the projected light to the user’s eyes, such that the
light is perceived by the user as coming from the focal distance corresponding to the user’s
current state of focus.
In another embodiment, a method of displaying virtual content to a user comprises
determining an accommodation of the user’s eyes, wherein the determined accommodation is
associated with a focal ce corresponding to a user’s t state of focus, projecting,
through a rm optic, light associated with one or more frames of image data, varying a
focus of the projected light based at least in part on the determined accommodation, and
delivering the projected light to the user’s eyes, such that the light is perceived by the user as
coming from the focal distance corresponding to the user’s current state of focus.
In another , a method of displaying l content to a user comprises
ining an accommodation of the user’s eyes, wherein the determined accommodation is
associated with a focal distance corresponding to a user’s t state of focus, projecting light
associated with one or more frames of image data, varying a focus of the ted light based
at least in part on the determined accommodation, and delivering the projected light to the
user’s eyes, such that the light is perceived by the user as coming from the focal distance
corresponding to the user’s current state of focus.
The light is delivered to the user through a substrate guided optics assembly, in one
or more embodiments. The light is delivered to the user h a freeform optical element, in
one or more embodiments. The light is delivered to the user through a ctive optical
element (DOE) in one or more embodiments. The light is projected through a stack of
waveguides, a first of the stack of waveguide configured to output light at a particular wavefront,
a second waveguide to output a positive margin wavefront relative to the particular wavefront, a
third waveguide to output a ve margin wavefront relative to the particular wavefront, in
one or more embodiments. The method further comprises blurring a portion of the one or
more frames of image data in a manner such that the portion is out of focus when the projected
light is delivered to the user’s eyes, in one or more embodiments.
In yet another embodiment, a system for displaying l content to a user,
comprises an image-generating source to provide one or more frames of image data in a timesequential
manner, a light generator to provide light associated with the one or more frames of
image data, an accommodation tracking module to track an accommodation of the user’s eye,
and a waveguide assembly to vary a focus of the light associated with the one or more frames
of image data, wherein different frames of image data are focused differently based at least in
part on the tracked accommodation.
In another aspect, a system for displaying virtual content to a user, comprises an
odation tracking module to ine an odation of the user’s eyes, an
image-generating source to provide one or more frames of image data in a time-sequential
manner, a light generator to project light associated with the one or more frames of image
data, a plurality of waveguides to e light rays ated with image data and to it
the light rays toward the user’s eyes, wherein the plurality of waveguides are stacked in a
direction facing the user’s eye, and a variable focus element (VFE) to vary a focus of the
itted light based at least in part on the determined accommodation of the user’s eyes.
The waveguide of the plurality of waveguides is a waveguide element, wherein the
focus of a first frame of image data transmitted from a first waveguide of the plurality of
waveguides is different from the focus of a second frame of image data transmitted from a
second waveguide of the plurality of waveguides, in one or more embodiments. The first
frame is a first layer of a 3D scene, and a second frame is a second layer of the 3D scene, in
one or more embodiments. The system further comprises a ng module to blur a portion
of the one or more frames of image data in a manner such that the portion is out of focus when
viewed by the user, in one or more embodiments.
The VFE is common to the plurality of waveguides, in one or more ments.
The VFE is ated with a waveguide of the plurality of waveguides, in one or more
embodiments. The VFE is coupled to a waveguide of the plurality of waveguides, such that
the VFE is interleaved between two waveguides of the plurality of ides, in one or more
embodiments. The VFE is embedded into a waveguide of the plurality of waveguides, in one
or more embodiments. The VFE is a diffractive optical element, in one or more embodiments.
The VFE is a refractive element, in one or more embodiments.
The waveguide is electro-active, in one or more embodiments. One or more
waveguides of the plurality of waveguides is switched off, in one or more embodiments. A
waveguide of the plurality of waveguides corresponds to a fixed focal plane, in one or more
embodiments. The system further comprises an exit pupil, wherein a diameter of the exit pupil
is no greater than 0.5 mm, in one or more embodiments. The light generator is a scanning
fiber display. The system of claim 302, r sing an array of exit pupils, in one or
more embodiments.
The system further comprises a plurality of light generators, a light generator
coupled to an exit pupil, in one or more embodiments. The system further comprises an exit
pupil expander, in one or more embodiments. The exit pupil is switchable based at least in
part on the determined accommodation of the user’s eyes, in one or more embodiments.
In another aspect, a system comprises an accommodation tracking module to
determine an accommodation of a user’s eyes, a fiber scanning display to scan a plurality of
light beams associated with one or more frames of image data, wherein a light beam of the
plurality of light beams is movable, a blur software to render a simulated dioptric blur in the one
or more frames of image data based at least in part on the determined accommodation of the
user’s eyes.
The diameter of light beams is no greater than 2 mm, in one or more ments.
The diameter of light beams is no greater than 0.5 mm, in one or more embodiments. The
scanned light beam is duplicated to create a plurality of exit pupils, in one or more
embodiments. The scanned light beam is ated to create a larger eye box, in one or more
ments. The exit pupils are switchable, in one or more embodiments.
In another embodiment, a method for displaying virtual content comprises
determining an odation of a user’s eyes, scanning a plurality of light beams associated
with one or more frames of image data, h a fiber scanning y, wherein a diameter of
the light beam is no greater than 0.5 mm, such that the frames of image data appear in focus
when viewed by the user, and blurring, using blur software, one or more portions of the frame
based at least in part on the ined accommodation of the user’s eyes.
A plurality of exit pupils are created, in one or more embodiments. The light beam
is generated by a singlecore fiber, in one or more embodiments. The light beam is duplicated
to create a plurality of exit pupils, in one or more embodiments. The exit pupils are switchable,
in one or more embodiments.
In another embodiment, a method for displaying virtual content to a user comprises
determining a position of the user’s pupil relative to a bundle of light projectors, wherein the
bundle of light projectors corresponds to a sub-image of an image to be presented to the user,
and g, based on the determined position of the user’s pupil, light corresponding to the
sub-image into a portion of the user’s pupil.
The method further comprises driving light corresponding to another sub-image of
the image to be presented to another portion of the user’s pupil h another bundle of light
projectors, in one or more embodiments. The method further ses mapping one or more
bundles of light projectors of the fiber scanning display with one or more portions of the user’s
pupil, in one or more ments. The mapping is a 1:1 mapping, in one or more
ments.
The diameter of the light is no greater than 0.5 mm, in one or more embodiments.
The bundle of light projectors produces an aggregate wavefront, in one or more embodiments.
The beamlets produced by the light projectors form a discretized ate wavefront, in one
or more ments. The beamlets approach the user’s eye in parallel, the eye deflects the
beamlets to ge upon a same spot on the retina, in one or more embodiments. The
user’s eye receives a superset of beamlets, the beamlets corresponding to a plurality of angles
at which they interest the pupil, in one or more embodiments.
In r embodiment, a system for displaying l content to a user comprises
a light source to provide light associated with one or more frames of image data, and a light
display assembly to receive the light associated with the one or more frames of image data,
wherein the light display assembly corresponds to multiple exit pupils spaced together, and
wherein the multiple exit pupils transmit light into a pupil of the user.
The le exit pupils are arranged in a hexagonal lattice, in one or more
embodiments. The multiple exit pupils are ed in a square lattice, in one or more
embodiments. The multiple exit pupils are arranged in a two-dimensional array, in one or more
embodiments. The multiple exit pupils are arranged in a three-dimensional array, in one or
more embodiments. The multiple exit pupils are arranged in a time-varying array, in one or
more embodiments.
In one or more embodiments, a method for displaying virtual content to a user,
comprising grouping a plurality of light projectors to form an exit pupil, driving a first light
pattern, h a first exit pupil, into a first portion of the user’s pupil, and driving a second
light pattern, through a second exit pupil, into a second portion of the user’s pupil, wherein the
first light pattern and second light pattern corresponds to sub-images of an image to be
presented to the user, and wherein the first light pattern is different than the second light
pattern. The method further comprises creating a discretized aggregate wavefront, in one or
more ments.
In yet another embodiment, a method for displaying virtual t to a user,
comprises ining a location of a pupil of the user relative to a light display assembly, and
calculating a focus at which to converge light to the pupil based at least in part on a limited eye
box around the determined location of the pupil.
The diameter of light is no greater than 0.5 mm, in one or more embodiments. The
method further comprises creating a discretized aggregate wavefront, in one or more
embodiments. The method further comprises aggregating a plurality of discrete neighboring
collimated light beams based at least in part on a center of a radius of ure of a desired
aggregate wavefront, in one or more embodiments. The method further comprises
determining an accommodation of the user’s eyes, wherein the focus is calculated based at
least in part on the determined accommodation, in one or more embodiments.
The method r comprising selecting an angular trajectory of light of a ity
of beamlets to create an out-of focus light beam, in one or more embodiments. A plurality of
beamlets represent a pixel of image data to the presented to the user, in one or more
embodiments. The beamlets hit the eye at a plurality of incident , in one or more
ments.
In yet another embodiment, a system for displaying virtual content to a user
comprises an image generating source to provide one or more ns of an image to be
presented to the user, and a plurality of microprojectors to t light associated with the one
or more portions of the image, the microprojectors positioned in a manner facing the user’s
pupil, and wherein a microprojector of the plurality of microprojectors is configured to t a
set of light rays representing a portion of the sub-image, the set of light rays projected to a
portion of the user’s pupil.
A first portion of the user’s pupil receives light rays from a plurality of
rojectors, in one or more embodiments. The system further comprises a reflective
surface to reflect the light from the plurality of microprojectors to one or more portions of the
user’s pupil, in one or more embodiments. The reflective surface is positioned in a manner
such that the user is able to view the real world through the reflective surface, in one or more
embodiments. A diameter of light is no greater than 0.5 mm, in one or more embodiments.
The system further comprises a discretized aggregate wavefront, in one or more
embodiments.
In another embodiment, a system comprises a processor to determine a location of
a user’s pupil, and an array of spatial light modulators (SLMs) to project light associated with
one or more frames of image data, n the array of SLMs are positioned based at least in
part on the determined on of the user’s pupil, and wherein the array of SLMs generate a
lightfield when viewed by the user.
In another aspect, a system for displaying l t to a user, comprises an
image generating source to provide one or more frames of image data, a first spatial light
modulator (SLM) configured to selectively transmit light rays associated with the one or more
frames of image data, a second SLM positioned in relation to the first SLM, the second SLM
also configured to selectively transmit light rays associated with the one or more frames of
image data, and a processor to control the first and second SLMs in a manner such that a
lightfield is created when the transmitted light rays are viewed by the user.
The system further comprises an accommodation tracking module to determine an
accommodation of the user’s eyes, in one or more embodiments. The SLM is an LCD, in one
or more embodiments. The LCD is attenuated, in one or more embodiments. The LCD
rotates a polarization of the transmitted light, in one or more embodiments. The SLM is a
DMD, in one or more embodiments. The DMD is coupled to one or more lenses, in one or
more ments. The SLM is a MEMs array, in one or more embodiments. The MEMs
array comprises an array of sliding MEMs shutters, in one or more embodiments. The MEMs
array is a Pixtronics ® MEMs array, in one or more ments.
In another embodiment, a system for displaying virtual content to a user comprises
a plurality of optical fibers to project light associated with one or more frames of image data to
be presented to the user, n an optical fiber of the plurality of optical fibers is coupled to a
lens, the lens configured to alter a diameter of the light beam projected by the scanning fiber,
wherein the lens comprises a nt refractive index.
The lens is a GRIN lens, in one or more embodiments. The lens ates the light
beams, in one or more embodiments. The system further ses an actuator coupled to the
optical fiber of the plurality of optical fibers to scan the fiber, in one or more embodiments.
The actuator is a piezo-electric actuator, in one or more embodiments. An end of the optical
fiber is polished at an angle to create a lensing effect, in one or more embodiments. An end of
the optical fiber is melted to create a lensing , in one or more embodiments.
A method for displaying virtual content to a user comprises projecting light
ated with one or more frames of image data, wherein the light is projected through a
plurality of optical fibers, modifying the light projected through the plurality of optical fibers
through a lens, wherein the lens is coupled to a tip of the plurality of l fibers, and
delivering the modified light to the user, in one or more embodiments.
In one or more embodiments, a system for displaying l content, comprises a
multicore ly comprising a plurality of fibers to multiplex light associated with one or
more frames of image data, and a ide to receive the light patterns, and transmit the
light patterns such that a first viewing zone only receives light associated with a first portion of
an image, and a second viewing zone only receives light associated with a second portion of
the image, wherein the first and second viewing zone is no greater than 0.5 mm. The system
further ses blurring software to blur out one or more portions of the frames of image
data, in one or more embodiments. The system r comprises an accommodation module
to determine an accommodation of the user’s eyes, in one or more embodiments. The
waveguide projects light to a user’s eye directly without an intermediate viewing optic, in one or
more embodiments.
A system comprises a multicore assembly sing a ity of fibers to
multiplex light associated with one or more frames of image data, a waveguide to e the
light patterns, and transmit the light patterns such that a first viewing zone only receives light
associated with a first portion of an image, and a second viewing zone only receives light
associated with a second portion of the image, wherein the first and second viewing zone is no
greater than 0.5 mm, and an optical assembly coupled to the ide to modify the
transmitted light beams to the first and second viewing zones, in one or more embodiments.
The plurality of fibers project light into a single waveguide array. The multicore
assembly is scanned, in one or more embodiments. A time-varying light field is generated, in
one or more embodiments. The l assembly is a DOE t. The optical assembly is
an LC layer, in one or more embodiments.
A method comprising projecting light associated with one or more frames of image
data through a ore assembly, the multicore assembly comprising a plurality of optical
fibers, and delivering the projected light, through a ide, such that a first portion of the
user’s pupil receives light associated with a first portion of an image, and a second portion of
the user’s pupil receives light associated with a second portion of the image, in one or more
embodiments.
A diameter of the first and second portions is no greater than 0.5 mm, in one or
more embodiments. The plurality of l fibers project light into a single waveguide array, in
one or more embodiments. The ore assembly is scanned, in one or more embodiments.
The waveguide comprises a plurality of reflectors, in one or more embodiments. An angle of
the reflectors is variable, in one or more embodiments. A set of optics to modify light being
delivered to the first and second viewing zones, in one or more embodiments. The set of
optics is a DOE t. The set of optics is a freeform optic. The set of optics is an LC layer,
in one or more embodiments.
In one aspect, a system comprises an array of microprojectors to project light
associated with one or more frames of image data to be presented to a user, wherein the array
of microprojectors is positioned relative to a location of the user’s pupil, and wherein the light
is projected into the user’s pupil, in one or more embodiments. The fiber scanning display of
claim 407, wherein the first and second light beams are mposed, in one or more
embodiments. The fiber scanning display of claim 407, wherein the first and second light
beam are deflected based at least in part on the critical angle of the polished bundled fiber, in
one or more embodiments. The fiber scanning display of claim 407, wherein the polished
d fibers is used to increase a resolution of the display. The polished bundled fibers is
used to create a lightfield, in one or more embodiments.
In another embodiment, a system comprises an array of microprojectors to project
light associated with one or more frames of image data to be presented to a user, wherein the
array of microprojectors is positioned relative to a on of the user’s pupil, and wherein the
light is projected into the user’s pupil, and an optical element d to the array of
microprojectors to modify the light projected into the user’s pupil.
In yet another embodiment, a system comprises a plurality of multicore fibers to
transmit light beams, the plurality of beams coupled together, and a coupling element to
bundle the ity of multicore fibers together, wherein the bundle of multicore fibers is
ed at a critical angle relative to a longitudinal axis of the fiber such that a first light beam
transmitted from a first fiber of the bundled fibers has a first path , and a second light
beam transmitted from a second fiber of the bundled fibers has a second path length, and
wherein the first path length is different from the second path length such that the first light
beam is out of phase relative to the second light beam.
The first and second light beams are superimposed, in one or more ments.
The first and second light beam are deflected based at least in part on the al angle of the
polished bundled fiber, in one or more embodiments. The ed bundled fibers is used to
increase a resolution of the display, in one or more embodiments. The polished bundled fibers
is used to create a lightfield, in one or more embodiments.
In another embodiment, a system for displaying virtual content to a user comprises
an image-generating source to provide one or more frames of image data, a plurality of optical
fibers to transmit light beams ated with the one or more frames of image data, and an
optical t coupled to the plurality of optical fibers to receive collimated light from the
optical fibers and deliver the light beams to the user’s eye, wherein the light beams are
delivered at a plurality of angles to the user’s eye such that a first light beam is red to a
portion of a user’s eye at a first angle, and a second light beam is delivered to the same
portion of the user’s eye at a second angle, wherein the first angle is different from the second
angle. The optical element is a waveguide, in one or more embodiments. The system further
comprises a phase modulator to modulate the transmission of light h the optical fibers, in
one or more embodiments.
In yet another embodiment, a method comprises providing one or more frames of
image data, transmitting light beams associated with the one or more frames of image data
through a plurality of optical fibers, and delivering the light beams to the user’s eyes at a
plurality of angles.
The method further comprises modulating a phase delay of the plurality of l
fibers, in one or more embodiments. The method further comprises coupling an optical
element to the plurality of optical fibers, in one or more embodiments. The optical element is a
waveguide, in one or more ments. The l element is a freeform optic. The optical
element is a DOE, in one or more embodiments. The optical element is a waveguide, in one
or more embodiments.
In one or more embodiments, a virtual y display system comprises a plurality
of optical fibers to te light beams associated with one or more images to be presented
to a user, and a plurality of phase modulators coupled to the plurality of optical fibers to
modulate the light beams, n the plurality of phase modulators modulate the light in a
manner that affects a wavefront generated as a result of the plurality of light beams.
The one or more optical fibers are deflected, at one or more , in one or more
embodiments. An optical fiber of the ity of optical fibers is coupled to a GRIN lens, in one
or more embodiments. The plurality of l fibers is physically actuated to scan the optical
fibers, in one or more embodiments.
In yet another aspect, a method comprises providing one or more frames of image
data to be presented to a user, projecting, through a plurality of optical fibers, light associated
with the one or more frames of image data, and modulating the light, h a plurality of
phase modulators, projects by the ity of optical fibers in a manner that affects an
ate wavefront produced by the plurality of optical fibers.
The light projected by the one or more optical fibers is deflected at one or more
angles, in one or more embodiments. The one or more optical fibers is coupled to a GRIN
lens, in one or more embodiments. The method further comprises scanning the optical light
beams, wherein the plurality of optical fibers is physically actuated to scan the optical , in
one or more embodiments.
In r aspect, a system for displaying virtual content, comprises an array of
optical fibers to transmit light beams ated with an image to be presented to a user, and a
lens coupled to the array of the optical fibers to deflect a plurality of light beams output by the
array of optical fibers through a single nodal point, wherein the lens is physically attached to
the optical fibers such that a movement of the optical fiber causes the lens to move, and
wherein the single nodal point is scanned.
The light beams output by the array of optical fibers ents a pixel of the image
to be presented to the user, in one or more embodiments. The lens is a GRIN lens, in one or
more embodiments. The array of optical fibers is used to display a lightfield, in one or more
embodiments. Another set of light beams output by another array of optical fibers represents
another pixel of the image to be presented to the user, in one or more embodiments. Multiple
arrays of optical fibers are combined to represent a pixel of the image to be presented to the
user, in one or more embodiments. The array of optical fibers is configured to deliver the light
beams to a predetermined portion of the user’s pupil, in one or more embodiments. The
output light beams are diverging, in one or more embodiments. The output light beams are
converging, in one or more embodiments.
A numerical aperture of the output light beams is increased relative to the light
beams transmitted by the individual optical fibers, in one or more embodiments. The
increased numerical re allows for higher resolution, in one or more embodiments. The
array of optical fibers is beveled in a manner such that a path length of a first light beam
traveling through a first optical fiber is different than a second light beam traveling through a
second optical fiber, thereby allowing for a plurality of focal lengths of the light beams delivered
to the user’s eye, in one or more embodiments.
In another aspect, a system for displaying virtual content to a user comprises an
array of microprojectors to t light associated with one or more frames of image data,
n one or more microprojectors of the array of microprojectors is polished at an angle
such that the projected light is deflected, and wherein the polished angle causes path length
differences between a first and second microprojectors of the array of microprojectors relative
to an optical element, and a light r to e the deflected light beams and to scan
them in at least one axis.
In yet another aspect, a system to provide at least one of a virtual or an augment
reality experience to a user comprises a frame, an array of micro-projectors carried by the
frame and positionable in front of at least one eye of the user when the frame is worn by the
user, and a local controller communicatively coupled to the array of micro-projectors to provide
image information to the micro-projectors, the local controller comprising at least one
processor, and at least one nontransitory processor readable media communicatively coupled
to the at least one processor, the at least one nsitory processor readable media which
stores at last one of processor-executable instructions or data, which when executed by the at
least one processor causes the at least one sor to at least one of process, cache, and
store data and provide the image information to the projectors to produce at least one of
a virtual or an t reality visual experience to the user.
] The system further comprises at least one reflector ted by the frame and
positioned and oriented to direct light from the micro-projectors toward at least one eye of the
user when the frame is worn by the user, in one or more embodiments. The micro-projectors
comprise tive ones of a plurality of scanning fiber displays, in one or more
embodiments. Each of the scanning fiber displays has a respective collimating lens at a distal
tip thereof, in one or more embodiments. The respective collimating lens is a gradient
refractive index (GRIN) lens, in one or more embodiments.
The tive collimating lens is a curved lens, in one or more ments. The
respective collimating lens is fused to the distal tip of the respective scanning fiber display, in
one or more embodiments. The scanning fiber displays has a respective diffractive lens at a
distal tip thereof, in one or more embodiments. Each of the scanning fiber displays has a
respective diffuser at a distal tip thereof, in one or more embodiments.
The diffuser is etched into the respective distal tip, in one or more embodiments.
Each of the scanning fiber displays has a respective lens at a distal tip f, the lens which
extends from the distal tip by a sufficient ce as to freely vibrate in response to a stimulus,
in one or more embodiments. Each of the ng fiber displays has a respective reflector at
a distal tip thereof, the reflector which extends from the distal tip by a sufficient distance as to
freely vibrate in response to a stimulus, in one or more embodiments. The scanning fiber
displays each includes a respective single mode optical fiber, in one or more embodiments.
The scanning fiber displays each include a respective mechanical transducer
coupled to move at least a distal tip of the single mode optical fiber, in one or more
embodiments. The respective mechanical transducers are each piezoelectric actuators, in one
or more embodiments. Each the single mode optical fibers has a distal tip, the distal tips
having a hemispherical lens shape, in one or more embodiments. Each the single mode
l fibers has a distal tip, the distal tips having a refractive lens affixed thereto, in one or
more embodiments.
The system further comprises a arent holder substrate which retains the
plurality of single mode optical fibers er, in one or more embodiments. The transparent
holder substrate has a refractive index that at least approximately matches a refractive index of
a cladding of the single mode optical fibers, in one or more embodiments. The transparent
holder substrate retains the ity of single mode optical fibers each angled toward a
common spot, in one or more embodiments.
The system further ses at least one mechanical ucer coupled to move
the plurality of single mode optical fibers in unison, in one or more embodiments. The at least
one mechanical transducer vibrates the ity of single mode l fibers at a mechanical
resonant frequency of the single mode optical fibers a portion of which are cantilevered out
from the transparent holder substrate, in one or more ments. The micro-projectors
comprise respective ones of a ity of planar waveguides, a portion of each of the planar
waveguides which extends cantilevered from a holder substrate, in one or more embodiments.
The system further comprises at least one mechanical transducer coupled to move the plurality
of planar waveguides in , in one or more embodiments.
The at least one ical transducer vibrates the holder substrate at a
mechanical resonant frequency of the planar waveguides, in one or more embodiments. The
micro-projectors comprise respective ones of a plurality of piezoelectric actuators coupled to
move respective ones of the planar waveguides with respect to the holder substrate, in one or
more embodiments. The planar waveguides each define an totally internally reflective path
along a respective length of the planar waveguide, and the planar waveguides comprise
respective ones of a plurality of electronically switchable diffractive optical elements (DOEs)
le to propagate light outward of the respective totally internally reflective path, in one or
more ments. The array of micro-projectors comprises an array of optical fibers, each
having a distal tip and at least one bevel edge, in one or more embodiments. The at least one
bevel edge is at the distal tip, and the distal tip is a polished distal tip, in one or more
embodiments.
Each of the optical fibers has a reflective surface at the respective distal tip thereof,
in one or more embodiments. The distal tip has an output edge at the distal tip at a defined
critical angle to a longitudinal axis of the respective optical fiber, in one or more embodiments.
The defined critical angle is an approximately forty-five (45) degree to the longitudinal axis of
the tive optical fiber, in one or more embodiments. The system further comprises a
focusing lens in an l path of light exiting the distal ends of the optical fibers, to receive a
plurality of beams of the light, the beams out of phase with one another, in one or more
embodiments. The system further comprises at least one transducer coupled to move at least
one of the optical fibers in an X-Y Cartesian coordinate system, to move light emitted by the at
least one optical fiber in an X-Z Cartesian coordinate system, in one or more embodiments.
The at least one transducer is a first piezoelectric actuator that resonates a cantilevered
portion of the optical fibers in a direction perpendicular to a direction at which the cantilevered
portions extend, in one or more embodiments.
The optical fibers comprise a thin ribbon of l , in one or more
ments. The at least one transducer is a second piezoelectric actuator that moves at
least the evered portion of the optical fibers in a direction longitudinal to the ion at
which the cantilevered portions extend, in one or more embodiments. The microprojectors
include at least one a single axis mirror le to provide a slow scan along a longitudinal
axis of at least one of the optical fibers, in one or more embodiments. The array of optical
fibers comprises a multicore fiber, in one or more embodiments. The multicore fiber includes a
plurality of approximately seven sparsely positioned clusters within a single conduit, each
cluster comprising three optical fibers, each optical fiber to carry a respective one of three
different colors of light, in one or more embodiments.
The multicore fiber es a plurality of imately nineteen sparsely
positioned clusters within a single conduit, each cluster comprising three optical fibers, each
optical fiber to carry a respective one of three different colors of light to e a triad of
overlapped spots of three different colors, in one or more embodiments. The ore fiber
includes at least one cluster within a single conduit, the cluster comprising at least three optical
fibers, each, each of the optical fibers to carry at least two different colors of light, in one or
more embodiments.
The multicore fiber includes at least one cluster within a single conduit, the at least
one cluster comprising four optical fibers, each optical fiber to carry a respective one of four
different colors of light, where one of the four colors is ed or nfrared, in one or more
embodiments. The multicore fiber includes a plurality of cores in a tight bundle, and r
comprises at least one transducer coupled to move the cores in a sparse spiral pattern, in one
or more embodiments. The at least one bevel edge is spaced inwardly from the distal tip, in
one or more embodiments. The at least one bevel edge is ed, in one or more
embodiments.
] The system further ses at least one transducer coupled to move at least one
of the optical fibers in an X—Y Cartesian coordinate system, to move light emitted by the at
least one optical fiber in an X-Z Cartesian coordinate system, in one or more embodiments.
The system further comprises a focusing lens in an optical path of light exiting the
bevel edges of the optical fibers, to receive a plurality of beams of the light, the beams out of
phase with one another, in one or more embodiments. The system further comprises a laser,
and at least one phase modulator that optically couples an output of the laser to a number of
cores of the multicore fiber to achieve mutual coherence, in one or more embodiments.
The system further ses a lenslet array optically d upstream of an input
end of respective ones of a number of cores of the multicore fiber, and a prism array lly
coupled between the plurality of collimation lenses and the input end of the cores of the
multicore fiber to deflect light from the lenslet array to the cores of the multicore fiber, in one or
more embodiments.
The system further comprises a lenslet array optically d upstream of an input
end of respective ones of a number of cores of the multicore fiber, and a shared focusing lens
optically coupled between the lenslet array and the input end of the cores of the multicore fiber
to deflect light from the lenslet array to the cores of the multicore fiber, in one or more
embodiments.
The array of micro-projectors further comprises at least one reflector, the at least
one reflector operable to produce scan pattern and optically coupled to the array of l
fibers, in one or more embodiments. The at least one reflector is operable to e at least
one of a raster scan pattern, a Lissajous scan pattern, or a spiral scan pattern of a multifocal
beam, in one or more ments. The each core of the ore fiber addresses a
respective part of an image plane without overlap, in one or more embodiments. The each
core of the multicore fiber addresses a respective part of an image plane with substantial
overlap, in one or more embodiments.
In another embodiment, a system for displaying virtual content, comprises an
image-source to provide one or more frames of image data to be presented to a user, a fiber
scanning y, the fiber scanning display comprising a plurality of fibers to project light
associated with the one or more frames of image data, wherein the ity of fibers are
d using an actuator and a processor to control the fiber ng display in a manner
such that a light field is presented to the user.
The actuator is shared among all the fibers of the fiber scanning display, in one or
more embodiments. The each fiber has its individual actuator, in one or more embodiments.
The plurality of fibers are mechanically coupled by a lattice, such that the plurality of fibers
move together, in one or more embodiments. The lattice is a graphene plane, in one or more
ments. The lattice is a lightweight strut, in one or more embodiments.
] In another embodiment, a system to provide at least one of a virtual or an augment
reality ence to a user, comprises a frame, a display system carried by the frame and
positionable in front of at least one eye of the user when the frame is worn by the user, and a
local controller communicatively coupled to the display system to e image information to
the display system, the local ller comprising at least one sor, and at least one
nontransitory processor readable media communicatively coupled to the at least one
sor, the at least one nontransitory processor readable media which stores at last one of
processor-executable instructions or data, which when executed by the at least one processor
causes the at least one processor to at least one of process, cache, and store data and
provide the image information to the display to produce at least one of a virtual or an augment
reality visual experience to the user.
The y comprises at least one wedge-shaped waveguide, the wedge-shaped
waveguide having at least two flat surfaces d from one another across a thickness of
the first wedge-shaped waveguide and having a length along which light ng the wedge-
shaped waveguide at defined angles via an entrance portion of the wedge-shaped waveguide
propagates via total internal reflection, the thickness of the wedge-shaped waveguide which
varies linearly along the length of the wedge-shaped waveguide, in one or more embodiments.
The wedge-shaped waveguide provides a al total internal reflection, in one or more
embodiments.
The system further comprises at least two projectors optically coupled to the
wedge-shaped waveguide at respective different locations along the entrance portion of the
wedge-shaped waveguide, in one or more embodiments. The system further comprises a first
linear array of a plurality of projectors optically coupled to the wedge-shaped waveguide at
respective different locations along the entrance portion of the wedge-shaped waveguide, in
one or more embodiments.
The tors of the first linear array of a plurality of projectors are scanning fiber
displays, in one or more embodiments. The system r comprises a stack of a plurality of
spatial light modulators optically coupled to the wedge-shaped waveguide along the entrance
n of the wedge-shaped waveguide, in one or more embodiments. The system further
comprises a multicore optical fiber optically coupled to the wedge-shaped waveguide at one or
more locations along the entrance portion of the wedge-shaped waveguide, in one or more
ments.
The projectors of the first linear array of tors are optically coupled to the
wedge-shaped waveguide to inject light into the wedge-shaped waveguide at a first angle,
further comprising a second linear array of a plurality of projectors optically coupled to the
wedge-shaped waveguide at tive different locations along the entrance portion of the
wedge-shaped waveguide, wherein the projectors of the second linear array of projectors are
optically coupled to the shaped waveguide to inject light into the wedge-shaped
waveguide at a second angle, the second angle different from the first angle, in one or more
embodiments.
The entrance portion is ude end of the wedged-shaped waveguide, in one or
more embodiments. The entrance portion is a lateral edge of the wedged-shaped waveguide,
in one or more ments. The entrance portion is a one of the flat es of the wedged-
shaped waveguide, in one or more embodiments. The system further comprises at least one
optical component optically coupled to a projector, and which changes an angle of light
received from the projector to optically couple the light to the wedge-shaped waveguide at
angles that achieve total internal reflection of the light within the wedge-shaped waveguide, in
one or more ments.
In another aspect, a system for displaying virtual content to a user, comprises an
array of rojectors to project light beams ated with one or more frames of image
data to be presented to the user, wherein the microprojector is configurable to be movable
relative to one or more microprojectors of the array of the microprojectors, a frame to house
the array of microprojectors, a processor operatively coupled to the one or more
microprojectors of the array of microprojectors to control one or more light beams transmitted
from the one or more projectors in a manner such that the one or more light beams are
modulated as a on of a position of the one or more microprojectors relative to the array of
microprojectors, thereby enabling delivery of a lightfield image to the user.
The microprojector of the array of microprojectors is d to a lens, in one or
more embodiments. The array of microprojectors is arranged in a manner based on a desired
tion of the image to be presented to the user, in one or more embodiments. The array of
microprojectors is arranged based on a desired field of view, in one or more embodiments.
The light beams of a plurality of microprojectors overlap, in one or more embodiments. The
system further comprises an or, wherein the actuator is coupled to one or more
microprojectors, and wherein the actuator is configurable to move the one or more
microprojectors, in one or more embodiments.
The actuator is coupled to a plurality of rojectors, in one or more
embodiments. The or is coupled to a single microprojector, in one or more
embodiments. The microprojector of the array of ojectors is mechanically coupled to a
lattice, in one or more embodiments.
In yet another embodiment, a t lens to interface with a cornea of an eye of a
user of a virtual or augmented reality y comprises a partially hemispherical substrate and
a selective filter. The selective filter is configured to ively pass light beams to a user’s
eye, in one or more embodiments. The selective filter is a notch filter, in one or more
embodiments. The notch filter substantially blocks wavelengths at approximately 450nm (peak
blue) and substantially passes other wavelengths in a visible portion of the electromagnetic
spectrum, in one or more embodiments. The notch filter substantially blocks wavelengths at
approximately 530nm (green) and ntially passes other wavelengths in a visible portion
of the electromagnetic spectrum, in one or more embodiments. The notch filter substantially
blocks wavelengths at approximately 650nm and ntially passes other ngths in a
visible portion of the electromagnetic um, in one or more embodiments.
The notch filter comprises a plurality of layers of dielectric materials carried by the
substrate, in one or more embodiments. The filter has a pinhole opening of less than 1.5mm
diameter, in one or more embodiments. The pinhole opening allows light beams of a plurality
of wavelengths to pass through, in one or more embodiments. A size of the pinhole is varied
based at least in part on a desired depth of focus of the y, in one or more embodiments.
The contact lens further comprises a plurality of modes of operation, in one or more
embodiments. The contact lens further comprises a multi-depth of focus display configuration
of the virtual content, in one or more embodiments.
The contact lens r comprises an odation tracking module to
determine an accommodation of the user’s eye, in one or more embodiments. A depth of
focus of a particular display object is varied based at least in part on the determined
accommodation, in one or more embodiments. The image is relayed through a waveguide, the
d image associated with a particular depth of focus, in one or more embodiments.
In another embodiment, a method for displaying virtual content to a user, comprises
providing one or more frames of image data to be presented to a user, projecting light
associated with the one or more frames of image data, and receiving, through a partially
hemispherical substrate coupled to the user’s pupil, the projected light and selectively filtering
out the light beams to the user’s pupil.
In another embodiment, a system for displaying virtual content to a user, comprises
a light projection system to project light associated with one or more frames of image data to a
user’s eyes, the light project system configured to project light corresponding to a plurality of
pixels associated with the image data, and a processor to modulate a depth of focus of the
plurality of pixels displayed to the user.
The depth of focus is modulated spatially, in one or more embodiments. The depth
of focus is modulated over time, in one or more ments. The system r comprises
an image-generating source to provide the one or more frames of image data in a time-
sequential manner, in one or more embodiments. The depth of focus is modulated on a frame-
to-frame basis, in one or more ments. The light projection system comprises a plurality
of optical fibers, and wherein the depth of focus is ted across the plurality of optical
fibers such that a portion of the optical fibers is associated with a first depth of focus, and
another portion of the optical fibers is ated with a second depth of focus, wherein the
first depth of focus is different from the second depth of focus in one or more embodiments.
A first display object of a particular frame is displayed through a first depth of focus,
and a second display object of the particular frame is displayed through a second depth of
focus, wherein the first depth of focus is ent from the second depth of focus, in one or
more embodiments. A first pixel of a particular frame is associated with a first depth of focus,
and a second pixel of the particular frame is associated with a second depth of focus, wherein
the first depth of focus is ent from the second depth of focus, in one or more
embodiments. The system further ses an accommodation tracking module to determine
an accommodation of the user’s eyes, wherein the depth of focus is modulated based at least
in part on the determined accommodation, in one or more embodiments.
A pattern of light generation associated with the light generation system is
dynamically slaved to the determined odation, in one or more embodiments. The
pattern is scanning pattern of a plurality of optical fibers, in one or more embodiments. The
system further ses a blurring module to blur one or more ns of the image data,
wherein the blurring is created to smooth a transition between a first scan pattern and a
second scan pattern or a first resolution scan pitch to a second resolution scan pitch, in one or
more embodiments.
In another embodiment, a system for displaying virtual content to a user, comprises
a light projection system to t light associated with one or more frames of image data to a
user’s eyes, the light project system configured to project light corresponding to a plurality of
pixels associated with the image data, a sor to modulate a size of the plurality of pixels
displayed to the user.
] The light projection system is a fiber scanning display, in one or more
embodiments. The projected light is displayed through a scanning pattern, in one or more
embodiments. The processor modulates the size of a particular pixel based at least in part on
a type of scanning pattern, in one or more embodiments. The size of the one or more pixels
may be ted based at least in part on a distance between scan lines of the scanning
pattern, in one or more embodiments. The size of a first pixel is different from the size of a
second pixel in the same frame, in one or more embodiments.
In another aspect, a method for displaying virtual content to a user, comprises
projecting light associated with one or more frames of image data, wherein one or more light
beams of the projected light correspond to one or more pixels, wherein light is projected
through a fiber scanning display, and ting a size of the one or more pixels displayed to
the user, in one or more embodiments.
] The size of a particular pixel is varied based at least in part on a scanning pattern
of the fiber scanning display, in one or more ments. The size of the one or more pixels
is modulated based at least in part on a distance between scan lines of the scanning pattern,
in one or more embodiments. The size of the one or more pixels is variable, in one or more
embodiments.
In yet another embodiment, a system for displaying virtual content to a user,
comprises a display system that delivers light associated with one or more frames of image
data, n the display system comprises a plurality of pixels, wherein the display system
scans light having variable line pitch, a blurring module to variably blur one or more pixels of
the plurality of pixels to modify a size of the one or more pixels, and a processor to control the
blurring module in a manner such pixel size is varied based at least in part on the line pitch of
the display system, in one or more embodiments. The display system is a fiber scanning
system, in one or more embodiments. The pixel size is enlarged, in one or more
embodiments. The pixel size is reduced, in one or more embodiments. The pitch line is
sparse, in one or more embodiments. The pitch line is dense, in one or more embodiments.
] In r aspect, a method of displaying virtual content to a user, the method
comprises projecting light associated with one or more frames of image data to be ted
to the user, selectively attenuating at least a portion of the ted light beams based at least
in part on a characteristic of the image data, and delivering the attenuated light beams to the
user’s eyes.
The light beam is selectively attenuated based at least in part on angle of nce
of the light beam, in one or more embodiments. Different portions of the frame are attenuated
to different amounts, in one or more embodiments. A depth of focus of the attenuated light
beams is varied, in one or more embodiments.
In one or more embodiments, a system for displaying virtual content to a user,
comprises an image generating source to provide one or more frames of image data, a stack
of two or more spatial light tors (SLMs) positioned such that the stack delivers light
associated with the one or more frames of image data to the user, wherein the SLM spatially
attenuates light from an outside environment, and a sor to control the stack of SLMs in
a manner such that an angle at which light beams pass through one or more cells of the SLM
is modulated.
The system further comprises a set of display optics, wherein the set of display
optics is positioned between the user’s eye and the outside environment, in one or more
embodiments. The SLMs of the stack of SLMs are cholesteric LCDs. The at least one of the
SLMs is a cholesteric LCD, in one or more embodiments. The stack of SLMs is positioned
such that the user views an outside world h the stack of SLMs, wherein the SLMs is at
least semi-transparent, in one or more embodiments.
The spatial light modulator arrays comprise at least one of a number or liquid
crystal arrays, a number of digital mirror device ts of digital light processing s, a
number of electro-mechanical system (MEMS) arrays, or a number of MEMS shutters, in
one or more embodiments. The system further comprises an occluder sing at least one
optical component, and wherein the processor controls the at least one optical component of
the occluder to produce a darkfield representation of a dark virtual object. in one or more
embodiments
In another aspect, a system for displaying virtual content, the system comprises an
array of spatial light modulators, the array of spatial light modulators configured to generate
light patterns, and wherein the array of spatial light modulators ses at least two
modulators, and a processor to control the array of spatial modulators in a manner such that
the at least two spatial modulators form a Moire n, n the Moire pattern is a periodic
l pattern attenuates light at a different period than a period of the light patterns forms on
the at least two spatial light modulators.
The spatial light modulator arrays se at least two spatial light modulator
arrays optically coupled to one another, and which are l passage of light via moire
s, in one or more embodiments. The at least two spatial light modulator arrays each bear
a respective attenuation pattern, in one or more embodiments. The at least two spatial light
modulator arrays each bear a respective fine-pitch sine wave pattern printed, etched, or other
inscribed thereon or therein, in one or more embodiments. The at least two spatial light
modulator arrays are in registration with one another, in one or more embodiments. The at
least two spatial light modulator arrays each bear a respective attenuation pattern, in one or
more embodiments.
In yet another ment, a system for display virtual content to a user,
comprises a light generating source to provide light associated with one or more frames of
image data, wherein the light ting source is a spatial light modulator, a e array
positioned in a manner relative to the spatial light modulator such that a pinhole of the pinhole
array receives light from a plurality of cells of the spatial light modulator, and n a first
light beam passing through the pinhole corresponds to a different angle than a second light
beam passing through the e, and wherein the cell of spatial light modulator selectively
attenuate light.
An outside nment is viewed through the pinhole array and the SLMs, and
wherein light beams are selectively attenuated based at least in part on the angle of incidence
of the light beams, in one or more embodiments. The light from different portions of a visual
field is selectively attenuated, in one or more embodiments. The system further comprises a
ive ation layer selectively operable to attenuation transmission of light
therethrough, the selective attenuation layer optically in series with the pinhole layer, in one or
more embodiments.
The selective attenuation layer ses a liquid crystal array, digital light projector
system, or spatial light modulator arrays which bear tive attenuation patterns, in one or
more embodiments. The pinhole array placed at a distance of approximately 30mm from a
cornea of an eye of the user, and the selective ation panel is located opposite the
pinhole array from the eye, in one or more embodiments. The pinhole array comprises a
plurality of es, and wherein the process controls the SLMs in a manner such light is
attenuated as a function of the angles at which light beams pass through the plurality of
pinholes, thereby producing an aggregate light field, in one or more embodiments. The
aggregate light field causes occlusion at a desired focal distance, in one or more
embodiments.
In another embodiment, a system comprises a light generating source to provide
light associated with one or more frames of image data, wherein the light generating source is
a spatial light tor, a lens array positioned in a manner relative to the spatial light
modulator such that a lens of the lens array receives light from a plurality of cells of the spatial
light modulator, and wherein a first light beam received at the lens corresponds to a different
angle than a second light beam ed at the lens, and n the cells of spatial light
modulator selectively attenuate light, in one or more embodiments.
The outside environment is viewed through the lens array and the SLMs, and
wherein light beams are selectively attenuated based at least in part on the angle of incidence
of the light beams, in one or more embodiments. The light from different portions of a visual
field is selectively attenuated, in one or more embodiments. The lens array comprises a
plurality of lenses, and wherein the process controls the SLMs in a manner such light is
ated as a function of the angles at which light beams are received at the plurality of
lenses, thereby producing an aggregate light field, in one or more ments. The
aggregate light field causes occlusion at a desired focal distance, in one or more
embodiments.
In another embodiment, a system for displaying virtual content to a user,
ses a light projector to project light associated with one or more frames of an image
data, at least one polarization sensitive layer to receive the light and rotate a polarization of the
light, and an array of polarization modulators to modulate the polarization of the polarization
sensitive layer, and wherein a state of the cell in the array determines how much light passes
through the polarization sensitive layer. The system is placed in a near-to-eye configuration, in
one or more embodiments. The polarization modulator is a liquid crystal array, in one or more
embodiments.
] The system further comprises a parallax barrier to offset the polarizer such that
ent exit pupils have different paths through the polarizer, in one or more embodiments.
The polarizer is an xpol zer, in one or more embodiments. The polarizer is a multiPol
polarizer, in one or more embodiments. The polarizer is a patterned zer, in one or more
embodiments. The light interacts with one or more MEMs arrays, in one or more
embodiments.
The system further comprises SLMs to project light, wherein the SLMs are
positioned between one or more optical elements, wherein the optical elements correspond to
a zero magnification ope, in one or more ments. The user views an outside
environment through the agnification telescope, in one or more embodiments. The at
least one SLM is positioned at an image plane within the zero-magnification telescope, in one
or more embodiments. The system further comprises a DMD, wherein the DMD corresponds
to a transparent ate, in one or more embodiments.
The system further ses an occluder comprising at least one l
component, and wherein the processor controls the at least one optical component of the
occluder to produce a darkfield representation of a dark virtual object, in one or more
embodiments. The system further comprises one or more LCDs, wherein the one or more
LCDs selectively attenuate light beams, in one or more embodiments. The system further
comprises one or more LCDs, wherein the one or more LCDs serve as polarization rotators, in
one or more embodiments. The occluder is a louver MEMs device, in one or more
embodiments.
The louver MEMs device is opaque, and wherein the louver MEMs device s
an angle of nce on a pixel-by-pixel basis, in one or more embodiments. The occluder is
a sliding panel MEMs device, wherein the sliding panel MEMs device slides back and forth to
modify a region of occlusion, in one or more embodiments.
In another embodiment, a method for displaying virtual content comprises
ting light ated with one or more frames of image data, rotating a polarization of
light through a polarization sensitive layer at a substrate which receives the projected light,
and modulating a polarization of light to selectively attenuate light passing through the
polarization layer.
The polarization modulator is a liquid crystal array, in one or more embodiments.
The method further comprises creating a parallax barrier to offset the polarizer such that
different exit pupils have different paths through the polarizer, in one or more embodiments.
The polarizer is an xpol polarizer, in one or more embodiments. The polarizer is a multiPol
polarizer, in one or more embodiments. The polarizer is a patterned zer, in one or more
embodiments.
In another embodiment, a system for displaying virtual content, comprises a light
generating source to provide light associated with one or more frames of image data, wherein
the light generating source is a spatial light modulator, an array of electro-mechanical
(MEMs) louvers, n the MEMs louvers are housed in a substantially transparent
substrate, and wherein the MEMs louvers are configurable to change an angle at which light is
delivered to a pixel, and wherein the angle of a first pixel delivered to the user is different from
a second pixel delivered to the user.
The at least one l component comprises a first array of micro-electro-
mechanical system (MEMS) louvers, in one or more embodiments. The array of MEMS louvers
comprises a plurality of substantially opaque s carried by an lly transparent
substrate, in one or more embodiments. The array of micro-electro-mechanical system
(MEMS) louvers has a louver pitch iently fine to selectably occlude light on a pixel-by-
pixel basis, in one or more embodiments. The system further comprises at least one optical
component of the occluder comprises a second array of MEMS louvers, the second array of
MEMS louvers in a stack configuration with the first array of MEMS louvers, in one or more
embodiments.
The array of MEMS louvers comprises a ity of polarizing louvers carried by an
optically transparent substrate, a respective polarization state of each of the louvers selectively
controllable, in one or more embodiments. The louvers of the first and the second arrays of
MEMS panels are polarizers, in one or more ments. The at least one optical
ent of the occluder comprises a first array of micro-electro-mechanical system (MEMS)
panels mounted for movement in a frame, in one or more ments.
The panels of the first array of MEMS panels are slidably mounted for movement in
the frame, in one or more embodiments. The panels of the first array of MEMS panels are
pivotably mounted for movement in the frame, in one or more ments. The panels of the
first array of MEMS panels are both translationally and pivotably mounted for movement in the
frame, in one or more embodiments. The panels of moveably to produce a moire n, in
one or more embodiments. The at least one optical component of the occluder further
ses a second array of MEMS panels mounted for movement in a frame, the second
array in a stack uration with the first array, in one or more embodiments. The panels of
the first and the second arrays of MEMS panels are polarizers. The at least one optical
component of the occluder comprises a reflector array, in one or more embodiments.
] In another embodiment, a system comprises at least one waveguide to receive light
from an outside environment and direct the light to one or more spatial light tors,
wherein the one or more spatial light modulators selectively attenuate the received light in
different portions of a visual field of the user. The at least one waveguide comprises a first and
second ides, and wherein the second waveguide is configured to deliver light exiting
the SLMs to the user’s eye, in one or more embodiments.
In another embodiment, a method comprises receiving light from an outside
environment, directing the light to a selective attenuator, and selectively attenuating, through
the selective attenuator, the received light in different portions of a visual field of the user.
The at least one waveguide ses a first and second waveguides, and wherein
the second waveguide is configured to deliver light exiting the SLMs to the user’s eye, in one
or more embodiments. The selective attenuator is a spatial light modulator, in one or more
embodiments. The spatial light modulator is a DMD array, in one or more embodiments. The
light is directed to the one or more spatial light modulators through one or more ides, in
one or more embodiments. The method further comprises recoupling light back to the
waveguide, causing light to partially exit toward the user’s eye, in one or more embodiments.
The waveguide is oriented substantially perpendicular to the selective attenuator, in one or
more embodiments.
In r embodiment, a system for ying virtual content to a user,
comprises a light generating source to e light associated with one or more frames of
image data, wherein the light generating source comprises a plurality of microprojectors, and
a waveguide configured to receive light from the plurality of microprojectors and transmit light
to a user’s eye.
The microprojectors are placed in a linear array, in one or more embodiments. The
microprojectors are placed in one edge of the waveguide, in one or more embodiments. The
microprojectors are placed in multiple edges of the waveguide. The microprojectors are
placed in a two dimensional array, in one or more embodiments. The microprojectors are
placed in a three-dimensional array, in one or more ments. The rojectors are
placed at multiple edges of the substrate, in one or more embodiments. The microprojectors
are placed at le angles, in one or more embodiments.
In another embodiment, a system for displaying virtual content, comprises an image
generating source to provide one or more frames of image data, wherein the image data
comprises one or more virtual objects to be presented to a user, and a rendering engine to
render the one or more virtual objects in a manner such that a halo is ved by the user
around the one or more virtual objects.
The system further comprises a light attenuator, wherein the light attenuator
balances a light intensity of the halo across the visual field of the user, in one or more
embodiments.
In another embodiment, a method for displaying l content, comprises
providing one or more frames of image data, n the image data comprises one or more
virtual objects to be presented to a user, and rendering the one or more virtual objects in a
manner such that a halo is perceived by the user around the one or more virtual objects,
thereby making it easier for the user to view the l object, n the virtual object is a
dark virtual object.
The method further comprises selectively attenuating light receiving from an outside
nment, through a light ator, wherein the light attenuator balances a light intensity
of the halo across a visual field of the user, in one or more embodiments.
In another embodiment, a system for displaying virtual content, comprises a
camera system to capture a view of a real environment, an optical see-through system that
displays one or more virtual objects superimposed over the view of the real environment,
wherein the captured view is used to render the one or more virtual objects presented to the
user, and a light intensity module to modulate a light intensity of the view of the real
environment, based at least on a ation between one or more real objects, and the one or
more virtual objects, such that a dark virtual object is visible in st with the one or more
real s, in one or more embodiments.
The captured view is used to generate a halo around one or more virtual objects,
wherein the halo fades across space, in one or more embodiments. The system further
comprises a light attenuator, wherein the light attenuator balances a light ity of the halo
across the visual field of the user, in one or more embodiments.
In yet another embodiment, a method of driving an augmented reality display
system, the method comprises rendering a first virtual object at a location on a field of view of
a user, and rendering a visual emphasis at least spatially proximate the rendered first virtual
object in the field of view of the user ntially concurrently with the rendering of the first
l object.
The rendering a visual emphasis includes rendering the visual emphasis with an
intensity gradient, in one or more embodiments. The rendering a visual emphasis includes
ing the visual emphasis with ng proximate a perimeter of the visual is, in
one or more embodiments.
The rendering a visual emphasis at least spatially proximate the rendered first
virtual object es rendering a halo visual effect spatially proximate the rendered first virtual
object, in one or more embodiments. The rendering a halo visual effect spatially proximate the
rendered first virtual object includes rendering the halo visual effect to be brighter than the
rendered first virtual object, in one or more embodiments.
The rendering the halo visual effect to be brighter than the ed first virtual
object is in response to a determination that the rendered first virtual object is darker than a
threshold value of darkness, in one or more embodiments. The rendering a halo visual effect
es ing the halo visual effect in a separate focal plane from the rendered first l
object in a perceived dimensional space, in one or more embodiments. The rendering a
halo visual effect includes rendering the halo visual effect with an intensity gradient, in one or
more embodiments. The rendering a halo visual effect includes rendering the halo visual
effect with an intensity gradient that matches a dark halo resulting from occlusion applied to
the rendering of the first virtual object to compensate for a darkfield effect of the occlusion, in
one or more embodiments.
The rendering a halo visual effect es rendering the halo visual effect with
blurring proximate a perimeter of the halo visual effect, in one or more embodiments. The
rendered first visual object has a non-circular perimeter and the rendered halo visual effect
conforms to the non-circular perimeter, in one or more embodiments. The ing a visual
emphasis at least spatially ate the rendered first virtual object includes rendering the
visual effect in a separate focal plane from the rendered first virtual object in a ved three-
dimensional space, in one or more embodiments. The rendering the visual effect in a separate
focal plane from the rendered first virtual object in a perceived three-dimensional space
includes rendering the visual effect in a focal plane spaced relative away from the user with
respect to a focal plane in which the rendered first virtual object is rendered, in one or more
embodiments.
In another embodiment, a system for displaying virtual content, comprises an
image generating source to provide one or more frames of image data to be presented to a
user, wherein the one or more frames of image data comprises at least one black virtual
object, and a rendering image to render the one or more frames of image data, and wherein
the rendering engine renders the black l object as a blue virtual object, such that the
black virtual object is e to the user.
The rendering a first virtual object at a location on a field of view of a user includes,
first changing any black tions of the first virtual object to a dark blue color, in one or more
embodiments.
In yet another embodiment, a system for transmitting light beams for display of
virtual content, ses at least one waveguide, the at least one waveguide having a first
end, a second end spaced from the first end across a length of the at least one waveguide, the
length along which light entering the respective waveguide at defined angles propagates via
total al reflection, at least one edge reflector positioned at least proximate a first end of
the at least one waveguide to optically reflectively couple light back to the first end of the at
least one waveguide, and at least one edge reflector positioned at least proximate a second
end of the at least one waveguide to optically reflectively couple light back to the second end
of the at least one ide.
The at least one waveguide has a number of transverse tive and/or diffractive
surfaces al to the waveguide that redirect at least a portion of the light transversely
outward of the ide, in one or more embodiments. The transverse reflective and/or
ctive surfaces are low diffraction efficiency diffractive optical elements (DOEs) in one or
more embodiments. The at least one edge reflector positioned at least proximate a first end of
the at least one waveguide comprises a plurality of reflectors positioned at least proximate the
first end of the at least one waveguide, in one or more embodiments.
The at least one edge reflector positioned at least proximate a second end of the at
least one waveguide comprises a plurality of reflectors positioned at least proximate the
second end of the at least one waveguide, in one or more embodiments. The at least one
waveguide is a single waveguide, in one or more embodiments.
In yet r embodiment, a system for transmitting light beams for display of
virtual content, comprises a waveguide assembly comprising a plurality of planar waveguides,
each of the planar waveguides respectively having at least two flat parallel major faces
opposed from one another across a thickness of the planar waveguide, a first end, and a
second opposed to the first end across a length of the waveguide, the length along which light
entering the respective waveguide at defined angles propagates via total internal reflection,
and two flat major edges opposed to one another across a width of the waveguide, the plurality
of planar waveguides in a stacked configuration along a first axis that is parallel with a direction
of the thicknesses of planar waveguides and along a second axis that is el with the
widths of the planar waveguides to form a three-dimensional array of planar waveguides.
There are at least three planar waveguides stacked in the direction of the first axis,
in one or more embodiments. There are at least three planar waveguides stacked in the
direction of the second axis, in one or more embodiments. There are at least three planar
waveguides stacked in the direction of the second axis, in one or more embodiments. The
successive planar ides in the stack along the first axis are immediately nt one
another, and successive planar waveguides in the stack along the second axis are
immediately adjacent one another, in one or more embodiments. The waveguide assembly
further ses a plurality of reflective layers carried on at least one surface of at least one
of the planar waveguides, in one or more embodiments.
The reflective layers include a completely reflective metalized coating. The
reflective layers include a wavelength specific reflector, in one or more embodiments. The
reflective layers separate the planar waveguides in each successive pair of the planar
waveguides along at least one of the first or the second axes, in one or more embodiments.
The reflective layers separate the planar waveguides in each successive pair of the planar
ides along both the first and the second axes, in one or more embodiments.
Each of a number of the planar waveguides respectively includes a number of
transverse reflective and/or diffractive surfaces that redirect at least a portion of the light
received by the respective planar waveguide transversely d of the planar waveguide, in
one or more embodiments. The transverse reflective and/or diffractive es comprise
diffractive optical elements sandwiched in the respective planar waveguides between the
major faces of the respective planar waveguide, in one or more embodiments. The diffractive
optical ts are ively operable to vary a focal distance, in one or more
embodiments.
The first axis is a curved axis, and at least one of the major edges of each of the
planar waveguides in at least one set in the waveguide assembly is oriented to focus on a
single line, the single line parallel to the lengths of the planar waveguides, in one or more
embodiments.
In one or more embodiments, a system for displaying l content to a user, the
system comprises a light projector to project light associated with one or more frames of image
data, wherein the light projector is a fiber scanning display, a waveguide assembly to ly
t light to a user’s eye, wherein the ide is curved concavely toward the eye, in one
or more embodiments.
The curved waveguide expands a field of view, in one or more embodiments. The
curved waveguide efficiently directs light to the user’s eye, in one or more embodiments. The
curved ide comprises a time-varying grating, thereby ng an axis for scanning the
light for the fiber scanning display, in one or more embodiments.
In another embodiment, a system for displaying virtual content to a user, comprises
a transmissive beam splitter substrate having an entrance to receive light and a number of
internal reflective or diffractive surfaces angled with respect to the entrance to redirect at least
a portion of the light ed at the entrance transversely outward of the transmissive beam
splitter substrate toward an eye of the user, wherein the number of internal reflective or
diffractive surfaces es a plurality of transverse reflective and/or ctive surfaces
spaced along a udinal axis of the transmissive beam splitter substrate, each of the
transverse reflective and/or diffractive surfaces angled or angleable with respect to the
entrance to redirect at least a portion of the light received at the entrance transversely outward
of the transmissive beam splitter substrate along an optical path toward an eye of the user, a
light generating system to transmit light to the transmissive beam splitter, and a local controller
icatively coupled to the y system to provide image information to the display
system, the local controller comprising at least one processor, and at least one nontransitory
processor readable media communicatively coupled to the at least one processor, the at least
one nontransitory processor readable media which stores at last one of processor-executable
instructions or data, which when executed by the at least one processor causes the at least
one processor to at least one of process, cache, and store data and provide the image
information to the display to produce at least one of a virtual or an augment y visual
experience to the user, in one or more embodiments.
The erse reflective and/or diffractive surfaces comprise at least one diffractive
optical element (DOE), where a collimated beam that enters the beam splitter at a number of
defined angles is totally internally reflected along the length of thereof, and intersects the DOE
at one or more locations, in one or more ments. The at least one diffractive optical
element (DOE) comprises a first grating, in one or more embodiments. The first grating is a
first Bragg grating, in one or more embodiments.
The DOE ses a second grating, the first grating on a first plane and the
second grating on a second plane, the second plane spaced from the first plane such that the
first and the second gs interact to produce a moire beat pattern, in one or more
embodiments. The first grating has a first pitch and the second grating has a second pitch, the
first pitch the same as the second pitch, in one or more embodiments. The first grating has a
first pitch and the second grating has a second pitch, the first pitch the different from the
second pitch, in one or more embodiments. The first grating pitch is controllable to change the
first grating pitch over time, in one or more embodiments. The first grating comprises an elastic
material and is subject to mechanical deformation in one or more embodiments.
The first g is d by an elastic material which is subject to mechanical
deformation, in one or more embodiments. The first grating pitch is controllable to change the
first grating pitch over time, in one or more embodiments. The second g pitch is
controllable to change the second grating pitch over time, in one or more embodiments. The
first grating is an electro-active grating, having at least one ON state and an OFF state, in one
or more embodiments. The first grating ses a polymer dispersed liquid l, a plurality
of liquid crystal droplets of the polymer dispersed liquid crystal controllably activated to change
a refractive index of the first grating, in one or more embodiments.
The first grating is a time-varying grating wherein the first g is a time-varying
grating, and the local controller controls at least the first grating to expand a field of view of the
display, in one or more embodiments. The first grating is a time-varying grating, and the local
controller employs time-varying control of at least the first grating to correction for a chromatic
aberration, in one or more embodiments. The local controller drives at least the first grating to
vary a placement of a red xel of a pixel of an image with respect to at least one of a blue
or a green sub-pixel of corresponding pixel of the image, in one or more embodiments. The
local controller drives at least the first grating to laterally shift an exit pattern to fill a gap in an
outbound image pattern, in one or more embodiments.
The at least one DOE element has a first circularly—symmetric term, in one or more
embodiments. The at least one DOE element has a first linear term, the first linear term
summed with the first circularly—symmetric term, in one or more embodiments. The circularly—
symmetric term is llable, in one or more embodiments. The at least one DOE t
has a second first circularly—symmetric term, in one or more embodiments. The at least one
diffractive optical (DOE) element comprises a first DOE, in one or more embodiments. The
first DOE is a circular DOE, in one or more embodiments.
The circular DOE is a time-varying DOE, in one or more embodiments. The circular
DOE is layered in relation to a waveguide for focus modulation, in one or more embodiments.
A diffraction pattern of the circular DOE is static, in one or more embodiments. A diffraction
n of the circular DOE is dynamic, in one or more ments. The system comprises
onal circular DOEs, wherein the additional circular DOEs are positioned in relation to the
circular DOE, such that many focus levels are achieved through a small number of switchable
DOEs, in one or more embodiments.
The system further comprises a matrix of able DOE elements, in one or
more embodiments. The matrix is utilized to expand a field of view, in one or more
embodiments. The matrix is utilized to expand a size of an exit pupil, in one or more
embodiments.
In one or more embodiments, a system for displaying virtual content to a user,
comprises a light projecting system to project light beams associated with one or more frames
of image data, a ctive optical element (DOE) to receive the projected light beams and
deliver the light beams at a d focus, n the DOE is a circular DOE, in one or more
embodiments.
The DOE is stretchable along a single axis to adjust an angle of a linear DOE term,
in one or more ments. The DOE ses a membrane, and at least one transducer
operable to selectively vibrate the membrane with an oscillatory motion in a Z-axis to provide
Z-axis control and change of focus over time, in one or more embodiments. The DOE is
embedded in a stretchable medium, such that a pitch of the DOE can be adjusted by physically
stretching the medium, in one or more embodiments. The DOE is stretched biaxially, and
wherein the stretching of the DOE affects a focal length of the DOE, in one or more
embodiments. The system of claim 762, further comprising a plurality of circular DOEs,
wherein the DOEs are d along a Z axis, in one or more embodiments. A circular DOE is
d in front of an ide for focus modulation. The system of claim 768, wherein the
DOE is static, in one or more embodiments.
In one or more ments, a system for displaying virtual content to a user,
comprises a light projecting system to project light beams associated with one or more frames
of image data, a first ide without any diffractive optical elements (DOEs), the first
waveguide which propagates light received by the first waveguide at a number of defined
angles along at least a portion of a length of the first waveguide via total internal reflection and
which provides the light ally from the first waveguide as collimated light, a second
waveguide with at least a first circularly-symmetric diffractive optical element (DOE), the
second waveguide optically coupled to receive the collimated light from the first waveguide,
and a processor to control the gratings of the DOE, in one or more embodiments.
The first DOE is selectively llable, in one or more embodiments. The display
ses a plurality of additional DOEs in addition to the first DOE, the DOEs arranged in a
stack configuration, in one or more embodiments. Each of the DOEs of the plurality of
additional DOEs is selectively controllable, in one or more embodiments. The local controller
controls the first DOE and the plurality of additional DOEs to dynamically te a focus of
light passing through the display, in one or more embodiments. The processor selectively
respectively switches the first DOE and the plurality of additional DOEs to realize a number of
focus levels, the number of realizable focus levels greater than a total number of the DOEs in
the stack, in one or more embodiments.
Each of the DOES in the stack has a tive optical power, the optical power of
the DOES in the static controllable additive to one another, in one or more embodiments. The
respective optical power of at least one of the DOEs in the stack is twice the respective optical
power of at least one other of the DOEs in the stack, in one or more embodiments. The
processor selectively respectively switches the first DOE and the plurality of additional DOEs
to modulate tive linear and radial terms of the DOEs over time, in one or more
embodiments. The processor selectively respectively switches the first DOE and the plurality of
additional DOEs on a frame sequential basis, in one or more embodiments.
The stack of DOEs comprises a stack of polymer dispersed liquid crystal elements.
In absence of an applied voltage, a host medium refraction index matches that of a set of
dispersed molecules of the polymer dispersed liquid crystal elements, in one or more
embodiments. The polymer dispersed liquid crystal elements comprise molecules of lithium
niobate, and a number of transparent indium tin oxide layer ode on either side of a host
, wherein the sed molecules of lithium niobate controllably change index of
refraction and functionally form a diffraction pattern within the host medium, in one or more
ments.
In another embodiment, a method for displaying virtual content, comprises
projecting light associated with one or more frames of image data to a user, receiving light at a
first waveguide, the first ide without any diffractive optical elements, and propagating
the light through internal reflection, receiving collimated light at a second waveguide with at
least a first circularly-symmetric diffractive optical element (DOE), the second waveguide
optically coupled to receive the collimated light from the first waveguide, wherein a g of
the circularly symmetric DOE is varied, and wherein the first waveguide and second
waveguide are assembled in a stack of DOEs, in one or more ments.
In one or more ments, an optical element for displaying virtual content to a
user, comprises at least one diffractive optical t (DOE) positioned to receive light, the at
least one DOE sing a first array of a plurality of separately addressable sections, with at
least one ode for each of the separately addressable subsection, each of the separately
addressable subsections responsive to at least one respective single received via the
respective at least one electrode to selectively switch between at least a first state and a
section state, the second state different from the first state, in one or more embodiments.
A field of view is expanded by multiplexing ntly addressable subsections, in
one or more embodiments. The first state is an ON state and the second state is an OFF
state, in one or more embodiments. The each of the tely addressable subsections has a
respective set of at least two indium tin oxide odes, in one or more embodiments. The
first array of a plurality of separately addressable sections of the at least one DOE is a one-
dimensional array, in one or more embodiments. The first array of a plurality of separately
addressable sections of the at least one DOE is a two-dimensional array, in one or more
embodiments. The first array of separately addressable ns are sections of a first DOE
that resides on a first planar layer, in one or more embodiments.
The at least one DOE comprises at least second DOE, the second DOE
comprising a second array of a plurality of separately addressable sections, with at least one
electrode for each of the separately addressable subsection, each of the separately
addressable subsections responsive to at least one respective single received via the
respective at least one electrode to selectively switch between at least a first state and a
n state, the second state different from the first state, the second array of DOEs ng
on a second planar layer, the second planar layer in a stacked configuration with the first
planar layer, in one or more embodiments.
The at least one DOE comprises at least third DOE, the third DOE sing a
third array of a plurality of separately addressable sections, with at least one electrode for each
of the separately addressable subsection, each of the separately addressable subsections
responsive to at least one respective single received via the respective at least one electrode
to ively switch between at least a first state and a section state, the second state different
from the first state, the third array of DOEs residing on a third planar layer, the third planar
layer in a stacked configuration with the first and the second planar layers, in one or more
embodiments.
The first array of separately sable sections are embedded in a single planar
waveguide, in one or more embodiments. The local controller controls the separately
addressable tions to selectively emit collimated light from the planar waveguide at a first
time and to emit a diverging light from the planar waveguide at a second time, the second time
different from the first time, in one or more embodiments. The local control controls the
separately addressable subsections to selectively emit light in a first ion from the planar
waveguide at a first time and to emit light in a second direction from the planar waveguide at
the first time, the second direction different from the first ion, in one or more
embodiments.
The local control controls the separately addressable tions to selectively
scan light across a direction over time, in one or more embodiments. The local control controls
the separately addressable subsections to ively focus light over time, in one or more
embodiments. The local control controls the separately addressable subsections to selectively
vary a field of view of an exit pupil over time, in one or more embodiments.
In one or more embodiments, a system comprises a first freeform tive and
lens optical ent to increase a size of a field-of—view for a defined set of optical
parameters, the first freeform reflective and lens optical component comprising: a first curved
surface, a second curved surface, and a third curved surface, the first curved surface at least
partially optically issive and refractive and which imparts a focal change to the light
received by the first freeform reflective and lens optical ent via the first curved surface,
the second curved surface which at least partially ts light received by the second curved
surface from the first curved surface toward the third curved surface and which passes light
received by the second curved surface from the third curved surface, the third curved surface
which at least partially reflects light out of the first freeform tive and lens optical
ent via the second curved surface, in one or more embodiments.
The first curved surface of the first freeform reflective and lens optical component is
a tive freeform curved surface, in one or more embodiments. The first curved surface of
the first freeform tive and lens optical ent adds a stigmatism to the light, in one or
more embodiments. The third curved surface of the first rm reflective and lens optical
component adds an opposite stigmatism to cancel the stigmatism added by the first curved
surface of the first freeform reflective and lens optical component, in one or more
embodiments. The second curved surface of the first freeform reflective and lens optical
component is a respective freeform curved surface, in one or more embodiments. The second
curved surface of the first freeform reflective and lens l component reflects defined
angles of light to be reflected by total internal reflection toward the third curved surface, in one
or more embodiments.
In one or more embodiments, a system comprises a fiber scanning display to
project light associated with one or more frames of image data, wherein the fiber scanning
y is configured to deliver the light to a first free form optical element, and a first freeform
reflective and lens optical component to increase a size of a field-of-view for a defined set of
optical parameters, the first freeform reflective and lens optical component comprising: a first
curved surface, a second curved surface, and a third curved surface, the first curved surface at
least partially optically transmissive and refractive and which imparts a focal change to the light
received by the first freeform reflective and lens l component via the first curved surface,
the second curved surface which at least partially reflects light received by the second curved
surface from the first curved surface toward the third curved surface and which passes light
received by the second curved surface from the third curved surface, the third curved surface
which at least lly reflects light out of the first freeform reflective and lens l
ent via the second curved surface, in one or more embodiments.
The freeform optic is a TIR rm optic, in one or more embodiments. The
freeform optic has non-uniform thickness, in one or more embodiments. The freeform optic is a
wedge optic, in one or more embodiments. The freeform optic is a conic, in one or more
embodiments. The freeform optic corresponds to arbitrary curves, in one or more
embodiments.
In one or more embodiments, a system comprises an image generating source to
e one or more frames of image data to be presented to a user, a display system to
provide light associated with the one or more frames of image data; and a free form optical
element to modify the provided light and deliver the light to the user, wherein the freeform optic
includes reflective coating, wherein the y system is configured to illuminate the freeform
optical element with light such that a wavelength of the light matches a corresponding
wavelength of the reflective coating, in one or more embodiments.
The one or more freeform l elements are tiled in relation to one another. The
one or more freeform optical elements are tiled along a z axis, in one or more embodiments.
In one or more embodiments, a system comprises an image generating source to
provide one or more frames of image data to be presented to a user; a y system to
provide light associated with the one or more frames of image data, wherein the display
system comprises a plurality of microdisplays; and a free form optical element to modify the
provided light and r the light to the user, in one or more embodiments.
The one or more freeform optics are tiled in relation to one another. The light
projected by the plurality of microdisplays increases a field of view, in one or more
embodiments. The freeform optical elements are configured such that only one color is
delivered by a particular freeform optical t, in one or more embodiments. The tiled
freeform is a star shape, in one or more embodiments. The tiled freeform optical elements
increase a size of an exit pupil, in one or more ments. The system further comprises
another free form l element, wherein the freeform optical element and stacked together
in a manner to create a uniform material ess, in one or more embodiments. The system
r comprises r free form optical element, wherein the other l element is
configured to capture light corresponding to an outside environment, in one or more
embodiments.
The system further comprises a DMD, wherein the DMD is configured to occlude
one or more pixels, in one or more embodiments. The system further comprises one or more
LCDs. The system further comprises a t lens substrate, wherein the freeform optics is
coupled to the contact lens substrate, in one or more embodiments. The plurality of
isplays provides an array of small exit pupils that in an aggregate form a functional
equivalent of a large exit pupil, in one or more embodiments.
The at least one image source includes at least a first monochromatic image source
that es light of a first color, at least a second monochromatic image source that provides
light of a second color, the second color different from the first color, and at least a third
monochromatic image source that provides light of a third color, the third color different from
the first and the second colors, in one or more embodiments. The at least a first
monochromatic image source comprises a first up of scanning fibers, the at least a
second monochromatic image source comprises a second subgroup of ng fibers, and
the at least a third monochromatic image source comprises a third subgroup of scanning
fibers, in one or more embodiments.
The system further comprises an occluder positioned in an optical path between the
first rm reflective and lens optical component and the at least one reflector, the occluder
operable to ive occlude light on a pixel-by-pixel basis. The first freeform reflective and
lens optical component forms at least a portion of a contact lens. The system further comprises
a compensator lens optically coupled to a portion of the first rm reflective and lens optical
component, in one or more embodiments.
In one or more ments, a system comprises a first freeform reflective and
lens optical component to increase a size of a field-of—view for a defined set of optical
parameters, the first freeform reflective and lens optical component comprising: a first surface,
a second surface, and a third surface, the first surface at least lly optically transmissive to
light ed by the first freeform reflective and lens optical component via the first surface,
the second surface which is curved and at least lly reflects light received by the second
surface from the first surface toward the third surface and which passes light received by the
second surface from the curved surface, the third surface which is curved and at least partially
reflects light out of the first freeform reflective and lens optical component via the second
surface, and a second freeform reflective and lens optical component, the second freeform
reflective and lens optical component comprising: a first surface, a second surface, and a third
surface, the first surface of the second freeform reflective and lens optical component at least
partially optically transmissive to light received by the second rm reflective and lens
optical component via the first surface, the second surface of the second freeform reflective
and lens optical component which is curved and at least partially reflects light received by the
second surface from the first e of the second freeform reflective and lens optical
component toward the third e of the second freeform reflective and lens optical
component and which passes light received by the second e from the third surface of the
second freeform reflective and lens optical component, the third surface of the second freeform
tive and lens optical component which is curved and at least partially reflects light out of
the second freeform reflective and lens optical component via the second surface, n the
first and the second freeform reflective and lens optical components are in an oppositely
ed stacked configuration along a Z-axis, in one or more embodiments.
The second surface of the second freeform reflective and lens optical component is
adjacent the third surface of the first freeform reflective and lens optical component, in one or
more embodiments. The second surface of the second freeform reflective and lens optical
ent is concave and the third surface of the first freeform reflective and lens optical
component is convex, that the third surface of the first freeform reflective and lens optical
component closely receives the second surface of the second freeform reflective and lens
optical component, in one or more embodiments. The first surface of the first freeform
tive and lens optical component is flat and the first surface of the second freeform
reflective and lens optical component is flat, and further ses at least a first projector
optically coupled to the first rm reflective and lens optical component via the first surface
of the first freeform reflective and lens optical component; and at least a second projector
optically coupled to the second rm reflective and lens optical ent via the first
surface of the second freeform reflective and lens optical component, in one or more
ments.
The system further comprises at least one wavelength selective material carried by
at least one of the first or the second freeform reflective and lens optical components, in one or
more embodiments. The system further comprises at least a first wavelength selective material
carried by the first freeform reflective and lens optical components, at least a second
ngth selective material carried by the second freeform reflective and lens optical
components, the first wavelength selective material selective of a first set of wavelengths and
the second wavelength ive al selective of a second set of wavelengths, the second
set of wavelengths different from the first set of wavelengths, in one or more embodiments.
] The system further comprises at least a first polarizer carried by the first freeform
reflective and lens optical ents, at least a second polarizer carried by the second
freeform reflective and lens optical components, the first polarizer having a different
polarization orientation than the second polarizer, in one or more embodiments.
The optical fiber cores are in the same fiber ng, in one or more embodiments.
The optical fiber cores are in te fiber ngs, in one or more embodiments. The
accommodation module tracks accommodation indirectly, by ng the vergence or gaze of
the user’s eyes, in one or more embodiments. The partially reflective mirror has vely high
reflectance for the polarization of light provided by the light source, and relative low reflectance
for the other polarization states of light provided by the e world, in one or more
embodiments. The plurality of partially reflective mirrors comprises a dielectric coating, in one
or more embodiments. The plurality of reflective mirrors has relatively high reflectance for the
waveguides for the wavelengths of light provided by the light source, and relatively low
reflectance for the other waveguides of light provided by the outside world, in one or more
embodiments. The VFE is a deformable mirror, the e shape of which can be varied over
time, in one or more embodiments. The VFE is an electrostatically actuated membrane mirror,
and wherein the waveguide or an additional transparent layer comprises one or more
substantially transparent electrodes, and wherein a voltage applied to the one or more
odes electrostatically deforms the membrane mirror, in one or more embodiments, in one
or more ments. The light source is a scanned light display, and wherein the VFE varies
the focus on a line segment basis, in one or more embodiments. The waveguide comprises an
exit pupil expansion function, n an input ray of light is split and outcoupled as multiple
rays of light exiting the waveguide at multiple locations, in one or more embodiments. The
image data is scaled by a processor in accordance with and to compensate for changing
optical image magnification, before the waveguide receives the one or more light patterns,
such that the image magnification appears to remain ntially fixed while adjusting focus
level, in one or more embodiments.
In another embodiment, a system for displaying virtual content comprises an
image-generating source to provide one or more frames of image data in a time-sequential
manner, a display assembly to project light rays associated with the one or more frames of
image data, the display assembly comprises a first display element corresponding to a first
frame-rate and a first bit depth, and a second display element corresponding to a second
rate and a second bit depth, and a le focus element (VFE) configurable to vary a
focus of the ted light and transmit the light to the user’s eye.
In yet another ment, a system for displaying virtual content comprises an
array of optical fibers to transmit light beams associated with an image to be presented to a
user, and a lens coupled to the array of the optical fibers to deflect a plurality of light beams
output by the array of optical fibers through a single nodal point, n the lens is physically
attached to the optical fibers such that a movement of the optical fiber causes the lens to move,
and n the single nodal point is scanned.
In another embodiment, a virtual reality display system comprises a plurality of
l fibers to generate light beams associated with one or more images to be presented to a
user, and a plurality of phase tors coupled to the plurality of optical fibers to modulate
the light beams, wherein the plurality of phase modulators modulate the light in a manner that
affects a wavefront generated as a result of the plurality of light beams.
In one embodiment, a system for displaying virtual content to a user comprises a
light projection system to t light associated with one or more frames of image data to a
user’s eyes, the light project system configured to project light corresponding to a plurality of
pixels associated with the image data and a processor to modulate a size of the plurality of
pixels displayed to the user.
In one embodiment, a system of displaying virtual content to a user,
comprises an image-generating source to provide one or more frames of image
data, a multicore assembly comprising a plurality of multicore fibers to project light
associated with the one or more frames of image data, a multicore fiber of the
plurality of multicore fibers ng light in a wavefront, such that the multicore
assembly produces an aggregate wavefront of the projected light, and a phase
tor to induce phase delays between the multicore fibers in a manner such
that the ate ont emitted by the multicore ly is varied, thereby
varying a focal distance at which the user perceives the one or more frames of
image data.
In another embodiment, a system for displaying virtual content to a user
comprises an array of microprojectors to t light beams associated with one or
more frames of image data to be presented to the user, wherein the microprojector
is configurable to be movable relative to one or more microprojectors of the array of
the microprojectors, a frame to house the array of microprojectors, a processor
operatively coupled to the one or more microprojectors of the array of
microprojectors to control one or more light beams transmitted from the one or more
projectors in a manner such that the one or more light beams are modulated as a
function of a position of the one or more microprojectors ve to the array of
microprojectors, thereby enabling ry of a lightfield image to the user.
Additional and other objects, features, and advantages of the invention are
bed in the detail description, figures and claims.
[00266A] In another embodiment, a system for displaying virtual content to a
user, comprising: at least one light source to multiplex a plurality of light beams to
display a respective plurality of light patterns associated with one or more frames of
image data; first, second and third planar waveguides to receive the plurality of light
beams and to direct the plurality of light beams toward an exit pupil, n the
first, second and third planar waveguides are d along an optical axis of the
user; and at least one optical element to modify a focus of a light beam of the
plurality of light beams directed by the first, second and third planar waveguides
wherein the first, second and third planar waveguides have first, second and third
wavefront curvatures corresponding to respective first, second and third focal
distances; wherein the first focal ce is a focal distance of a determined
accommodation of the user’s eyes; wherein the second focal ce is based at
least in part on the determined accommodation and is located at a first
predetermined error margin farther from the user’s eyes relative to the first focal
distance; n the third focal distance is based at least in part on the determined
odation and is located at a second predetermined error margin closer to the
user’s eyes relative to the first focal distance; and, wherein the first, second and
third planar waveguides are configured to allow a range of accommodation before a
physical adjustment of the focus level is necessary due to a change in the
accommodation of the user’s eyes.
[00266B] The system, where the at least one optical element modifies the
plurality of light beams in a manner such that a wavefront curvature is created, and
wherein the created wavefront curvature corresponds to a focal plane when the
plurality of light beams are viewed by the user.
[00266C] The , wherein the first, second and third planar waveguides are
each switchable between on and off states.
D] The system, wherein first, second and third planar ides are
static planar waveguides.
[00266E] The system, wherein an optical element of the at least one optical
elements is switchable n on and off states.
[00266F] The system, n an optical element of the at least one optical
elements is a static optical element.
[00266G] The system, wherein an optical element of the at least one optical
elements is a diffractive optical element (DOE) in one of the first, second or third
planar waveguides, wherein the DOE is switchable between on and off states.
[00266H] The system, wherein the at least one optical element comprises a
weak lens.
[00266I] The system, wherein the at least one optical element comprises a
Fresnel zone plate.
[00266J] The system, n first, second and third frames of the one or more
frames of image data are delivered through the first, second and third planar
ides to the user's eye simultaneously.
[00266K] The system, wherein first, second and third frames of the one or more
frames of image data are red through the first, second and third planar
waveguides to the user's eye sequentially.
[00266L] The system, wherein the first planar waveguide is disposed between
the second and third planar waveguides, and wherein the first wavefront curvature
corresponds to the focal distance of the determined accommodation of the user’s
eyes.
[00266M] The system, wherein the second planar waveguide is closer than the
first planar waveguide to the user, and n the second ont curvature
corresponds to the first predetermined error margin from the user’s eyes relative to
the first focal distance.
[00266N] The system, wherein the third planar waveguide is farther than the first
planar waveguide from the user, wherein the third wavefront curvature corresponds
to the second predetermined error margin closer to the user’s eyes relative to the
first focal distance; wherein the third wavefront curvature is greater than the first
wavefront ure; and, wherein the first wavefront curvature is greater than the
second wavefront curvature.
[00266O] The system, further comprising a compensating lens to compensate
for a tive focus modification of the at least one optical elements on a light
beam g through the first, second and third planar waveguides along the
optical axis of the user.
[00266P] The system, wherein the compensating lens is disposed on an
opposite side of the plurality of planar waveguides from the user.
[00266Q] The system, wherein the at least one light source comprises a plurality
of light sources corresponding to respective ones of the plurality of waveguides.
[00266R] The system, n the at least one light source is a single
multiplexed light source.
[00266S] The system, wherein the first predetermined error margin is an
accommodation error margin, and wherein second predetermined error margin is an
error accommodation margin.
[00266T] The system, wherein the at least one light source comprises first,
second and third light sources respectively addressing the first, second and third
planar waveguides along respective longitudinal axes of the first, second and third
planar waveguides orthogonal to the optical axis of the user.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1 illustrates a user's view of augmented y (AR) through a
wearable AR user device, in one illustrated embodiment.
Figs. 2A-2E rates various embodiments of wearable AR devices.
] Fig. 3 illustrates a cross-sectional view of the human eye, in one
illustrated embodiment.
Figs. 4A -4D illustrate one or more embodiments of various internal
sing ents of the wearable AR device.
Figs. 5A-5H illustrate embodiments of transmitting focused light to a
user through a transmissive beamsplitter substrate.
Figs. 6A and 6B illustrate embodiments of coupling a lens element with
the transmissive litter substrate of Figs. 5A-5H.
Figs. 7A and 7B illustrate embodiments of using one or more
waveguides to transmit light to a user.
] Figs. 8A-8Q illustrate embodiments of a diffractive optical element
(DOE).
Figs 9A and 9B illustrate a wavefront produced from a light projector, according to
one illustrated embodiment.
Fig. 10 illustrates an embodiment of a stacked configuration of le transmissive
beamsplitter substrate d with optical ts, according to one illustrated embodiment.
Figs 11A-11C illustrate a set of beamlets projected into a user’s pupil, according to
the illustrated embodiments.
Figs. 12A and 12B illustrate configurations of an array of microprojectors, according
to the illustrated embodiments.
Figs. 13A—13M illustrate embodiments of coupling microprojectors with l
ts, according to the illustrated embodiments.
Figs. 14A— 14F illustrate embodiments of spatial light modulators coupled with
l elements, according to the illustrated embodiments.
Figs. 15A—15C illustrate the use of a wedge type waveguides along with a plurality of
light sources, according to the illustrated embodiments.
Figs. 16A—16O illustrate embodiments of coupling optical ts to optical ,
according to the illustrated embodiments.
Fig. 17 illustrates a notch filter, according to one illustrated embodiment.
Fig. 18 illustrates a spiral pattern of a fiber scanning display, according to one
illustrated embodiment.
Figs. N illustrate occlusion effects in presenting a darkfield to a user,
according to the illustrated embodiments.
Figs. 20A—200 illustrate embodiments of various waveguide assemblies, according
to the illustrated embodiments.
Figs. 21A—21 N illustrate various configurations of DOEs d to other optical
elements, ing to the illustrated embodiments.
Figs. 22A—22Y illustrate various configurations of freeform optics, according to the
illustrated embodiments.
ED PTION
Referring to Figures 4A—4D, some general componentry options are illustrated. In
the portions of the ed description which follow the discussion of Figures 4A—4D, various
systems, subsystems, and components are presented for sing the objectives of
providing a high-quality, comfortably-perceived display system for human VR and/or AR.
As shown in Figure 4A, an AR system user (60) is depicted wearing a frame (64)
ure coupled to a display system (62) positioned in front of the eyes of the user. A speaker
(66) is coupled to the frame (64) in the depicted configuration and positioned adjacent the ear
canal of the user (in one embodiment, another speaker, not shown, is positioned adjacent the
other ear canal of the user to provide for stereo / shapeable sound control). The display (62) is
operatively coupled (68), such as by a wired lead or wireless connectivity, to a local processing
and data module (70) which may be mounted in a variety of configurations, such as fixedly
ed to the frame (64), fixedly attached to a helmet or hat (80) as shown in the embodiment
of Figure 4B, embedded in headphones, removably attached to the torso (82) of the user (60) in
a backpack-style configuration as shown in the embodiment of Figure 4C, or removably
attached to the hip (84) of the user (60) in a belt-coupling style configuration as shown in the
embodiment of Figure 4D.
] The local processing and data module (70) may comprise a power-efficient
processor or ller, as well as digital memory, such as flash memory, both of which may be
utilized to assist in the processing, g, and storage of data a) captured from sensors which
may be operatively coupled to the frame (64), such as image capture devices (such as
s), microphones, inertial measurement units, accelerometers, compasses, GPS units,
radio s, and/or gyros; and/or b) acquired and/or processed using the remote processing
module (72) and/or remote data repository (74), possibly for passage to the display (62) after
such processing or retrieval. The local processing and data module (70) may be operatively
coupled (76, 78), such as via a wired or wireless ication links, to the remote processing
module (72) and remote data repository (74) such that these remote modules (72, 74) are
operatively coupled to each other and available as resources to the local processing and data
module (70). In one embodiment, the remote processing module (72) may se one or
more relatively powerful processors or controllers configured to analyze and process data
and/or image information. In one embodiment, the remote data repository (74) may se a
relatively large-scale digital data storage facility, which may be available through the internet or
other networking configuration in a “cloud” resource configuration. In one ment, all data
is stored and all ation is performed in the local processing and data module, allowing
fully autonomous use from any remote modules.
Referring to Figures 5A through 22Y, various display urations are presented
that are designed to present the human eyes with photon-based radiation patterns that can be
comfortably perceived as tations to physical reality, with high-levels of image quality
and three-dimensional perception, as well as being capable of presenting two-dimensional
content.
Referring to Figure 5A, in a simplified e, a transmissive beamsplitter
substrate (104) with a 45-degree ting surface (102) directs ng radiation (106), which
may be output from a lens (not shown), through the pupil (45) of the eye (58) and to the retina
(54). The field of view for such a system is limited by the geometry of the beamsplitter (104).
To accommodate the desire to have comfortable viewing with minimal hardware, in one
embodiment, a larger field of view can be created by ating the outputs/reflections of
various different reflective and/or diffractive surfaces and using, e.g., a sequential
configuration wherein eye (58) is presented with a sequence of frames at high frequency that
provides the tion of a single coherent scene. As an alternative to, or in addition to,
presenting different image data via different reflectors in a time-sequential fashion, the
reflectors may separate content by other means, such as zation selectivity or wavelength
selectivity. In addition to being capable of relaying two-dimensional images, the reflectors can
relay the three-dimensional wavefronts associated with true-three-dimensional g of actual
physical objects.
Referring to Figure 5B, a substrate (108) comprising a plurality of tors at a
plurality of angles (110) is shown, with each reflector actively reflecting in the ed
configuration for illustrative purposes. The reflectors may be switchable elements to tate
temporal selectivity. In one embodiment, the reflective surfaces would intentionally be
tially activated with frame-sequential input ation (106), in which each reflective
surface presents a narrow field of view sub-image which is tiled with other narrow field of view
sub-images presented by the other reflective surfaces to form a composite wide field of view
image. For example, referring to Figures 5C, 5D, and 5E, surface (110), about in the middle of
substrate (108), is switched “on” to a reflecting state, such that it reflects incoming image
information (106) to present a relatively narrow field of view sub-image in the middle of a larger
field of view, while the other potential reflective surfaces are in a transmissive state.
Referring to Figure 5C, incoming image information (106) coming from the right of
the narrow field of view sub-image (as shown by the angle of incoming beams 106 relative to
the substrate 108 input interface 112, and the resultant angle at which they exit the substrate
108) is reflected toward the eye (58) from reflective surface (110). Figure 5D illustrates the
same reflector (110) active, with image information coming from the middle of the narrow field
of view sub-image, as shown by the angle of the input information (106) at the input interface
(112) and its angle as it exits substrate (108). Figure 5E illustrates the same reflector (110)
active, with image ation coming from the left of the field of view, as shown by the angle of
the input information (106) at the input interface (112) and the resultant exit angle at the surface
of the substrate (108). Figure 5F rates a configuration wherein the bottom reflector (110) is
active, with image information (106) coming in from the far right of the overall field of view. For
example, Figures 5C, 5D, and 5E can illustrate one frame representing the center of a framesequential
tiled image, and Figure 5F can illustrate a second frame representing the far right of
that tiled image.
In one embodiment, the light carrying the image information (106) may strike the
reflective surface (110) directly after entering substrate (108) at input interface (112), t
first reflecting from the surfaces of substrate (108). In one embodiment, the light carrying the
image information (106) may reflect from one or more surfaces of ate (108) after entering
at input interface (112) and before striking the reflective surface (110); for instance, substrate
(108) may act as a planar waveguide, propagating the light carrying image information (106) by
total internal reflection. Light may also reflect from one or more surfaces of the substrate (108)
from a partially reflective coating, a wavelength-selective coating, an angle-selective g,
and/or a polarization-selective coating.
In one embodiment, the angled reflectors may be constructed using an electro-
active material, such that upon application of a voltage and/or current to a ular reflector,
the tive index of the material sing such reflector s from an index
substantially d to the rest of the substrate (108), in which case the reflector is in a
transmissive configuration, to a reflective uration wherein the refractive index of the
reflector mismatches the refractive index of the substrate (108) such that a reflection effect is
created. Example electro-active al includes lithium niobate and electro-active polymers.
Suitable ntially arent electrodes for controlling a plurality of such reflectors may
comprise materials such as indium tin oxide, which is utilized in liquid crystal displays.
In one embodiment, the electro-active reflectors (110) may comprise liquid crystal,
embedded in a ate (108) host medium such as glass or plastic. In some variations, liquid
crystal may be selected that changes refractive index as a function of an applied electric signal,
so that more analog changes may be accomplished as opposed to binary (from one
transmissive state to one tive state). In an embodiment wherein 6 sub-images are to be
presented to the eye frame-sequential to form a large tiled image with an overall refresh rate of
60 frames per second, it is desirable to have an input y that can refresh at the rate of
about 360 Hz, with an electro-active reflector array that can keep up with such frequency. In
one embodiment, lithium e may be utilized as an electro-active reflective material as
opposed to liquid crystal; lithium niobate is utilized in the photonics industry for high-speed
switches and fiber optic networks and has the capability to switch refractive index in response
to an applied voltage at a very high frequency; this high frequency may be used to steer line-
sequential or pixel-sequential sub-image information, especially if the input display is a scanned
light display, such as a fiber-scanned display or scanning mirror-based display.
In another embodiment, a variable switchable angled mirror configuration may
comprise one or more high-speed ically repositionable reflective surfaces, such as a
MEMS (micro-electro-mechanical system) device. A MEMS device may include what is known
as a “digital mirror device”, or “DMD”, (often part of a “digital light processing”, or “DLP” ,
such as those available from Texas Instruments, Inc). In another electromechanical
ment, a ity of air-gapped (or in vacuum) reflective surfaces could be mechanically
moved in and out of place at high frequency. In r electromechanical embodiment, a
single reflective e may be moved up and down and re-pitched at very high frequency.
Referring to Figure 5G, it is e that the switchable variable angle reflector
configurations described herein are e of passing not only collimated or flat wavefront
information to the retina (54) of the eye (58), but also curved ont (122) image
ation, as shown in the illustration of Figure 5G. This generally is not the case with other
waveguide-based configurations, wherein total internal reflection of curved wavefront
information causes undesirable complications, and therefore the inputs generally must be
collimated. The ability to pass curved wavefront information facilitates the ability of
configurations such as those shown in Figures 5B-5H to provide the retina (54) with input
perceived as focused at s distances from the eye (58), notjust optical infinity (which
would be the interpretation of collimated light absent other cues).
Referring to Figure 5H, in another embodiment, an array of static partially reflective
surfaces (116) (Le, always in a reflective mode; in another embodiment, they may be electro-
active, as above) may be embedded in a substrate (1 14) with a high-frequency gating layer
(118) controlling outputs to the eye (58) by only allowing transmission through an aperture
(120) which is controllably movable. In other words, everything may be selectively blocked
except for transmissions through the aperture (120). The gating layer (1 18) may comprise a
liquid crystal array, a lithium niobate array, an array of MEMS shutter elements, an array of DLP
DMD elements, or an array of other MEMS devices configured to pass or transmit with
relatively high-frequency switching and high issibility upon being switched to
transmission mode.
Referring to Figures 6A—6B, other embodiments are depicted wherein d optical
ts may be combined with exit pupil expansion configurations to assist with the comfort of
the virtual or augmented reality experience of the user. With a larger “exit pupil” for the optics
uration, the user’s eye positioning relative to the display (which, as in Figures 4A—4D, may
be mounted on the user’s head in an eyeglasses sort of configuration) is not as likely to disrupt
his experience — because due to the larger exit pupil of the system, there is a larger acceptable
area wherein the user’s ical pupil may be d to still receive the information from the
display system as desired. In other words, with a larger exit pupil, the system is less likely to be
sensitive to slight misalignments of the display relative to the user’s anatomical pupil, and
greater comfort for the user may be achieved through less geometric constraint on his or her
relationship with the display/glasses.
As shown in Figure 6A, the display (140) on the left feeds a set of parallel rays into
the substrate (124). In one embodiment, the display may be a scanned fiber display scanning a
narrow beam of light back and forth at an angle as shown to project an image through the lens
or other optical element (142), which may be utilized to collect the angularly-scanned light and
convert it to a parallel bundle of rays. The rays may be reflected from a series of reflective
surfaces (126, 128, 130, 132, 134, 136) which may be configured to partially t and
partially transmit incoming light so that the light may be shared across the group of reflective
surfaces (126, 128, 130, 132, 134, 136) approximately y. With a small lens (138) placed
at each exit point from the waveguide (124), the exiting light rays may be steered through a
nodal point and scanned out toward the eye (58) to provide an array of exit , or the
functional equivalent of one large exit pupil that is usable by the user as he or she gazes toward
the display system.
For virtual reality configurations wherein it is ble to also be able to see through
the waveguide to the real world (144), a similar set of lenses (139) may be presented on the
opposite side of the waveguide (124) to compensate for the lower set of lenses; thus ng a
the equivalent of a zero-magnification telescope. The reflective surfaces (126, 128, 130, 132,
134, 136) each may be aligned at approximately 45 s as shown, or may be configured to
have different alignments, akin to the configurations of Figures 5B-5H, for e). The
reflective es (126, 128, 130, 132, 134, 136) may comprise wavelength-selective
reflectors, band pass reflectors, half silvered mirrors, or other tive configurations. The
lenses (138, 139) shown are refractive lenses, but diffractive lens elements may also be
utilized.
Referring to Figure 6B, a somewhat similar configuration is depicted wherein a
plurality of curved reflective surfaces (148, 150, 152, 154, 156, 158) may be utilized to
effectively combine the lens (element 138 of Figure 6A) and reflector (elements 126, 128, 130,
132, 134, 136 of Figure 6A) functionality of the embodiment of Figure 6A, thereby obviating the
need for the two groups of lenses (element 138 of Figure 6A). The curved tive surfaces
(148, 150, 152, 154, 156, 158) may be various curved configurations selected to both t
and impart angular change, such as parabolic or elliptical curved surfaces. With a parabolic
shape, a parallel set of incoming rays will be collected into a single output point; with an
elliptical configuration, a set of rays diverging from a single point of origin are collected to a
single output point. As with the configuration of Figure 6A, the curved reflective surfaces (148,
150, 152, 154, 156, 158) preferably are configured to partially reflect and partially transmit so
that the incoming light is shared across the length of the waveguide (146). The curved
reflective surfaces (148, 150, 152, 154, 156, 158) may comprise wavelength-selective notch
reflectors, half silvered mirrors, or other reflective configurations. In another embodiment, the
curved reflective surfaces (148, 150, 152, 154, 156, 158) may be replaced with ctive
reflectors configured to reflect and also deflect.
ing to Figure 7A, perceptions of Z-axis difference (i.e., ce straight out
from the eye along the optical axis) may be facilitated by using a waveguide in ction with
a variable focus optical element configuration. As shown in Figure 7A, image information from
a display (160) may be collimated and injected into a ide (164) and distributed in a large
exit pupil manner using, e.g., configurations such as those described in nce to Figures
6A and 6B, or other substrate-guided optics methods known to those skilled in the art — and
then variable focus optical t capability may be utilized to change the focus of the
wavefront of light ng from the waveguide and provide the eye with the perception that
the light coming from the waveguide (164) is from a particular focal distance. In other words,
since the incoming light has been collimated to avoid challenges in total internal reflection
waveguide configurations, it will exit in collimated fashion, requiring a viewer’s eye to
accommodate to the far point to bring it into focus on the retina, and lly be interpreted as
being from optical infinity — unless some other intervention causes the light to be refocused and
perceived as from a different g distance; one suitable such intervention is a variable
focus lens.
In the ment of Figure 7A, collimated image information is injected into a
piece of glass (162) or other material at an angle such that it totally internally reflects and is
passed into the adjacent waveguide (164). The waveguide (164) may be configured akin to the
waveguides of s 6A or 6B (124, 146, respectively) so that the collimated light from the
display is distributed to exit somewhat uniformly across the distribution of reflectors or
diffractive features along the length of the waveguide. Upon exit toward the eye (58), in the
depicted configuration the exiting light is passed through a variable focus lens element (166)
wherein, depending upon the controlled focus of the variable focus lens element (166), the light
exiting the variable focus lens element (166) and entering the eye (58) will have various levels
of focus (a collimated flat wavefront to represent optical infinity, more and more beam
divergence / wavefront curvature to represent closer viewing distance relative to the eye 58).
To compensate for the variable focus lens element (166) n the eye (58) and
the waveguide (164), another similar variable focus lens element (167) is placed on the
opposite side of the waveguide (164) to cancel out the optical effects of the lenses (166) for
light coming from the world (144) for ted reality (i.e., as described above, one lens
compensates for the other, producing the functional equivalent of a zero-magnification
telescope).
The variable focus lens element (166) may be a refractive element, such as a liquid
crystal lens, an electro-active lens, a conventional refractive lens with moving ts, a
mechanical-deformation-based lens (such as a fluid-filled membrane lens, or a lens akin to the
human crystalline lens wherein a flexible t is flexed and relaxed by actuators), an
electrowetting lens, or a plurality of fluids with different refractive indices. The variable focus
lens element (166) may also comprise a switchable diffractive optical element (such as one
ing a polymer dispersed liquid crystal approach n a host medium, such as a
polymeric material, has microdroplets of liquid l dispersed within the material; when a
voltage is applied, the molecules reorient so that their refractive indices no longer match that of
the host medium, thereby creating a high-frequency switchable diffraction n).
One embodiment es a host medium in which microdroplets of a Kerr effect-
based electro-active material, such as lithium niobate, is dispersed within the host medium,
enabling refocusing of image information on a pixel-by—pixel or line-by-line basis, when coupled
with a scanning light display, such as a fiber-scanned display or scanning-mirror-based display.
In a variable focus lens t (166) uration wherein liquid crystal, lithium niobate, or
other technology is ed to present a n, the pattern spacing may be modulated to not
only change the focal power of the le focus lens element (166), but also to change the
focal power of the overall optical system — for a zoom lens type of functionality.
In one embodiment, the lenses (166) could be telecentric, in that focus of the display
y can be altered while keeping magnification constant — in the same way that a
photography zoom lens may be configured to decouple focus from zoom position. In another
embodiment, the lenses (166) may be non-telecentric, so that focus changes will also slave
zoom changes. With such a uration, such magnification changes may be compensated
for in software with dynamic scaling of the output from the graphics system in sync with focus
changes).
Referring back to the projector or other video y unit (160) and the issue of how
to feed images into the optical display system, in a “frame sequential” configuration, a stack of
sequential two-dimensional images may be fed to the display sequentially to produce three-
dimensional perception over time; in a manner akin to the manner in which a computed
tomography system uses stacked image slices to represent a three-dimensional structure. A
series of mensional image slices may be presented to the eye, each at a ent focal
distance to the eye, and the ain would integrate such a stack into a perception of a
coherent three-dimensional volume. Depending upon the display type, line-by-line, or even
pixel-by-pixel sequencing may be ted to produce the perception of three-dimensional
viewing. For example, with a d light display (such as a scanning fiber display or
scanning mirror display), then the display is presenting the waveguide (164) with one line or
one pixel at a time in a sequential fashion.
If the variable focus lens t (166) is able to keep up with the high-frequency of
pixel-by-pixel or line-by-line presentation, then each line or pixel may be presented and
cally focused through the le focus lens element (166) to be perceived at a different
focal distance from the eye (58). Pixel-by—pixel focus modulation generally requires an
extremely fast/ high-frequency variable focus lens element (166). For example, a 1080P
resolution display with an overall frame rate of 60 frames per second typically presents around
125 million pixels per second. Such a configuration also may be constructed using a solid state
switchable lens, such as one using an electro-active material, e.g., lithium e or an electro-
active polymer. In addition to its compatibility with the system illustrated in Figure 7A, a frame
sequential multi-focal display driving approach may be used in conjunction with a number of the
display system and optics ments described in this disclosure.
] Referring to Figure 7B, with an electro-active layer (172) (such as one comprising
liquid crystal or lithium niobate) surrounded by functional electrodes (170, 174) which may be
made of indium tin oxide, a waveguide (168) with a conventional transmissive substrate (176,
such as one made from glass or c with known total internal reflection characteristics and
an index of refraction that matches the on or off state of the electro-active layer 172) may be
controlled such that the paths of entering beams may be dynamically altered to essentially
create a time-varying light field.
Referring to Figure 8A, a stacked waveguide ly (178) may be utilized to
provide three-dimensional perception to the eye/brain by having a plurality of waveguides (182,
184, 186, 188, 190) and a plurality of weak lenses (198, 196, 194, 192) configured togetherto
send image information to the eye with various levels of wavefront curvature for each
waveguide level indicative of focal distance to be perceived for that waveguide level. A ity
of displays (200, 202, 204, 206, 208), or in another embodiment a single multiplexed display,
may be utilized to inject collimated image information into the waveguides (182, 184, 186, 188,
190), each of which may be ured, as described above, to distribute incoming light
substantially equally across the length of each waveguide, for exit down toward the eye.
The ide (182) nearest the eye is configured to deliver collimated light, as
injected into such waveguide (182), to the eye, which may be representative of the optical
infinity focal plane. The next waveguide up (184) is configured to send out collimated light
which passes through the first weak lens (192; e.g., a weak negative lens) before it can reach
the eye (58); such first weak lens (192) may be ured to create a slight convex wavefront
curvature so that the eye/brain interprets light coming from that next waveguide up (184) as
coming from a first focal plane closer inward toward the person from optical infinity. Similarly,
the third up waveguide (186) passes its output light through both the first (192) and second
(194) lenses before reaching the eye (58); the combined optical power of the first (192) and
second (194) lenses may be configured to create another incremental amount of wavefront
divergence so that the eye/brain interprets light coming from that third waveguide up (186) as
coming from a second focal plane even closer inward toward the person from optical infinity
than was light from the next waveguide up (184).
] The other waveguide layers (188, 190) and weak lenses (196, 198) are similarly
configured, with the highest waveguide (190) in the stack sending its output through all of the
weak lenses between it and the eye for an aggregate focal power representative of the closest
focal plane to the person. To compensate forthe stack of lenses (198, 196, 194, 192) when
viewing/interpreting light coming from the world (144) on the other side of the stacked
ide assembly (178), a sating lens layer (180) is disposed at the top of the stack
to compensate for the aggregate power of the lens stack (198, 196, 194, 192) below. Such a
uration provides as many perceived focal planes as there are available waveguide/lens
pairings, again with a vely large exit pupil configuration as described above. Both the
reflective aspects of the waveguides and the focusing aspects of the lenses may be static (i.e.,
not dynamic or o-active). In an alternative embodiment they may be dynamic using
electro-active features as described above, enabling a small number of waveguides to be
lexed in a time sequential n to produce a larger number of effective focal planes.
Referring to Figures 8B-8N, various aspects of diffraction configurations for focusing
and/or redirecting collimated beams are depicted. Other aspects of diffraction systems for such
es are disclosed in US. Patent Application Serial No. 61/845,907 (US Patent
Application No. 14/331,218), which is incorporated by reference herein in its entirety. Referring
to Figure 8B, passing a collimated beam through a linear diffraction pattern (210), such as a
Bragg grating, will deflect, or “steer”, the beam. Passing a collimated beam through a radially
symmetric diffraction pattern (212), or el zone plate”, will change the focal point of the
beam. Figure 8C illustrates the deflection effect of passing a ated beam through a linear
diffraction pattern (210); Figure 8D rates the focusing effect of passing a collimated beam
through a radially symmetric diffraction pattern (212).
Referring to Figures 8E and 8F, a combination diffraction pattern that has both linear
and radial elements (214) produces both deflection and focusing of a collimated input beam.
These tion and focusing effects can be produced in a reflective as well as issive
mode. These principles may be applied with waveguide urations to allow for additional
optical system control, as shown in s 8G-8N, for example. As shown in Figures 8G-8N, a
diffraction pattern (220), or “diffractive optical element” (or “DOE”) has been embedded within a
planar waveguide (216) such that as a collimated beam is totally ally reflected along the
planar waveguide (216), it intersects the diffraction pattern (220) at a multiplicity of locations.
Preferably, the DOE (220) has a relatively low diffraction efficiency so that only a
portion of the light of the beam is deflected away toward the eye (58) with each intersection of
the DOE (220) while the rest continues to move through the planar waveguide (216) via total
internal reflection; the light carrying the image information is thus divided into a number of
related light beams that exit the waveguide at a multiplicity of locations and the result is a fairly
uniform pattern of exit on toward the eye (58) for this particular collimated beam
ng around within the planar waveguide (216), as shown in Figure 8H. The exit beams
toward the eye (58) are shown in Figure 8H as substantially parallel, e, in this case, the
DOE (220) has only a linear diffraction pattern. As shown in the comparison n Figures
8L, 8M, and 8N, changes to this linear diffraction pattern pitch may be utilized to controllably
deflect the exiting parallel beams, thereby producing a scanning or tiling functionality.
Referring back to Figure 8|, with changes in the radially symmetric ction
pattern component of the embedded DOE (220), the exit beam n is more divergent, which
would require the eye to accommodation to a closer ce to bring it into focus on the retina
and would be interpreted by the brain as light from a viewing distance closer to the eye than
optical infinity. Referring to Figure 8J, with the addition of another waveguide (218) into which
the beam may be injected (by a projector or display, for example), a DOE (221) embedded in
this other waveguide (218), such as a linear diffraction pattern, may function to spread the light
across the entire larger planar waveguide (216), which functions to provide the eye (58) with a
very large incoming field of ng light that exits from the larger planar waveguide (216), Le,
a large eye box, in accordance with the particular DOE configurations at work.
The DOES (220, 221) are depicted bisecting the associated waveguides (216, 218)
but this need not be the case; they could be placed closer to, or upon, either side of either of
the waveguides (216, 218) to have the same onality. Thus, as shown in Figure 8K, with
the injection of a single collimated beam, an entire field of cloned collimated beams may be
directed toward the eye (58). In addition, with a combined linear diffraction pattern / radially
symmetric diffraction pattern scenario such as that depicted in Figures 8F (214) and 8| (220), a
beam distribution waveguide optic (for functionality such as exit pupil functional expansion;
with a configuration such as that of Figure 8K, the exit pupil can be as large as the optical
element itself, which can be a very significant advantage for user comfort and ergonomics) with
Z-axis focusing capability is presented, in which both the divergence angle of the cloned beams
and the wavefront curvature of each beam represent light coming from a point closer than
optical infinity.
In one ment, one or more DOEs are switchable between “on” states in
which they actively diffract, and “off” states in which they do not significantly diffract. For
instance, a switchable DOE may comprise a layer of polymer dispersed liquid crystal, in which
roplets se a diffraction n in a host medium, and the refractive index of the
microdroplets can be switched to substantially match the refractive index of the host material (in
which case the pattern does not appreciably diffract nt light) or the microdroplet can be
switched to an index that does not match that of the host medium (in which case the pattern
actively diffracts incident . Further, with dynamic changes to the diffraction terms, such as
the linear diffraction pitch term as in Figures 8L-8N, a beam scanning or tiling functionality may
be achieved. As noted above, it is ble to have a relatively low diffraction grating
efficiency in each of the DOEs (220, 221) because it facilitates distribution of the light, and also
because light coming through the waveguides that is desirably itted (for example, light
coming from the world 144 toward the eye 58 in an augmented y configuration) is less
affected when the diffraction efficiency of the DOE that it crosses (220) is lower — so a better
view of the real world through such a configuration is achieved.
Configurations such as those illustrated in Figure 8K preferably are driven with
injection of image information in a time sequential approach, with frame sequential driving being
the most straightforward to implement. For example, an image of the sky at optical infinity may
be injected at time1 and the diffraction g retaining ation of light may be ed; then
an image of a closer tree branch may be injected at time2 while a DOE controllably imparts a
focal change, say one diopter or 1 meter away, to provide the eye/brain with the tion that
the branch light information is coming from the closer focal range. This kind of gm can
be repeated in rapid time sequential fashion such that the eye/brain perceives the input to be all
part of the same image. This isjust a two focal plane example; preferably the system will be
configured to have more focal planes to e a smoother transition between objects and
their focal distances. This kind of configuration generally assumes that the DOE is switched at
a relatively low speed (i.e., in sync with the frame-rate of the display that is injecting the images
— in the range of tens to hundreds of cycles/second).
The opposite extreme may be a configuration wherein DOE ts can shift focus
at tens to hundreds of MHz or greater, which facilitates switching of the focus state of the DOE
elements on a pixel-by—pixel basis as the pixels are scanned into the eye (58) using a scanned
light display type of approach. This is desirable because it means that the overall display
frame-rate can be kept quite low; just low enough to make sure that “flicker” is not a problem
(in the range of about 60-120 frames/sec).
In between these ranges, if the DOEs can be switched at KHz rates, then on a line-
e basis the focus on each scan line may be adjusted, which may afford the user with a
visible benefit in terms of temporal artifacts during an eye motion ve to the display, for
example. For instance, the different focal planes in a scene may, in this manner, be
interleaved, to minimize e artifacts in response to a head motion (as is discussed in
greater detail later in this disclosure). A line-by-line focus tor may be operatively coupled
to a line scan display, such as a grating light valve y, in which a linear array of pixels is
swept to form an image; and may be operatively coupled to scanned light displays, such as
fiber-scanned displays and -scanned light displays.
A stacked configuration, similar to those of Figure 8A, may use c DOEs
(rather than the static waveguides and lenses of the embodiment of Figure 8A) to provide multi-
planar focusing simultaneously. For example, with three aneous focal planes, a primary
focus plane (based upon measured eye accommodation, for example) could be presented to
the user, and a + margin and — margin (i.e., one focal plane closer, one farther out) could be
utilized to provide a large focal range in which the user can accommodate before the planes
need be updated. This increased focal range can provide a temporal advantage if the user
switches to a closer orfarther focus (i.e., as determined by accommodation measurement);
then the new plane of focus could be made to be the middle depth of focus, with the + and —
margins again ready for a fast switchover to either one while the system catches up.
] Referring to Figure 80, a stack (222) of planar ides (244, 246, 248, 250,
252) is shown, each having a reflector (254, 256, 258, 260, 262) at the end and being
configured such that collimated image information injected in one end by a display (224, 226,
228, 230, 232) bounces by total internal tion down to the reflector, at which point some or
all of the light is reflected out toward an eye or other target. Each of the reflectors may have
slightly different angles so that they all reflect exiting light toward a common destination such as
a pupil. Such a configuration is somewhat similar to that of Figure 5B, with the exception that
each different angled reflector in the embodiment of Figure 80 has its own waveguide for less
interference when projected light is travelling to the targeted reflector. Lenses (234, 236, 238,
240, 242) may be interposed n the displays and waveguides for beam steering and/or
Figure 8P illustrates a geometrically staggered version wherein reflectors (276, 278,
280, 282, 284) are positioned at staggered lengths in the waveguides (266, 268, 270, 272, 274)
so that g beams may be relatively easily aligned with objects such as an anatomical pupil.
With knowledge of how far the stack (264) is going to be from the eye (such as 28mm between
the cornea of the eye and an eyeglasses lens, a typical comfortable geometry), the geometries
of the reflectors (276, 278, 280, 282, 284) and ides (266, 268, 270, 272, 274) may be
set up to fill the eye pupil (typically about 8mm across or less) with exiting light. By directing
light to an eye box larger than the diameter of the eye pupil, the viewer may make eye
movements while retaining the ability to see the displayed imagery. Referring back to the
discussion related to Figure 5A and 5B about field of view expansion and tor size, an
expanded field of view is presented by the configuration of Figure 8P as well, and it does not
e the complexity of the switchable reflective elements of the embodiment of Figure 5B.
Figure 8Q rates a version wherein many reflectors (298) form a vely
continuous curved reflection surface in the aggregate or discrete flat facets that are oriented to
align with an overall curve. The curve could a parabolic or elliptical curve and is shown cutting
across a plurality of waveguides (288, 290, 292, 294, 296) to minimize any crosstalk issues,
although it also could be utilized with a monolithic waveguide configuration.
In one implementation, a high-frame-rate and lower persistence display may be
ed with a lower-frame-rate and higher persistence y and a variable focus element
to comprise a relatively high-frequency frame sequential volumetric display. In one
embodiment, the high-frame-rate display has a lower bit depth and the lower-frame-rate display
has a higher bit depth, and are ed to comprise an effective high-frame-rate and high bit
depth display, that is well suited to presenting image slices in a frame sequential fashion. With
such an approach, a three-dimensional volume that is desirably represented is functionally
divided into a series of two-dimensional slices. Each of those two-dimensional slices is
projected to the eye frame sequentially, and in sync with this presentation, the focus of a
variable focus element is changed.
In one embodiment, to get enough frame rate to support such a uration, two
display elements may be integrated: a olor, high-resolution liquid crystal display (“LCD”; a
backlighted ferroelectric panel display also may be utilized in another embodiment; in a further
embodiment a scanning fiber display may be utilized) ing at 60 frames per second, and
aspects of a higher-frequency DLP system. Instead of illuminating the back of the LCD panel in
a conventional manner (i.e., with a full size fluorescent lamp or LED array), the conventional
lighting configuration may be removed to accommodate using the DLP projector to project a
mask pattern on the back of the LCD (in one embodiment, the mask pattern may be binary in
that the DLP either projects illumination, or not-illumination; in another embodiment described
below, the DLP may be utilized to project a grayscale mask image).
DLP tion systems can operate at very high frame rates; in one embodiment
for 6 depth planes at 60 frames per , a DLP tion system is operated against the
back of the LCD display at 360 /second. Then the DLP projector is utilized to ively
illuminate portions of the LCD panel in sync with a high-frequency variable focus element (such
as a deformable membrane ) that is disposed between the viewing side of the LCD panel
and the eye of the user, the variable focus element being used to change the global display
focus on a frame by frame basis at 360 frames/second. In one embodiment, the variable focus
element is positioned to be optically conjugate to the exit pupil, to enable adjustments of focus
without simultaneously affecting image magnification or “zoom.” In another embodiment, the
variable focus element is not ate to the exit pupil, such that image magnification s
accompany focus adjustments, and software is used to compensate for these optical
magnification changes and any distortions by pre-scaling or warping the images to be
presented.
Operationally, it’s useful to consider an example again wherein a three-dimensional
scene is to be presented to a user n the sky in the background is to be at a viewing
distance of optical infinity, and wherein a branch coupled to a tree d at a certain location
closer to the user than optical infinity extends from the tree trunk in a direction toward the user,
so that the tip of the branch is closer to the user than is the proximal portion of the branch that
joins the tree trunk.
In one embodiment, for a given global frame, the system may be configured to
present on an LCD a full-color, all in-focus image of the tree branch in front the sky. Then at
subframe1, within the global frame, the DLP tor in a binary masking configuration (i.e.,
illumination or absence of illumination) may be used to only illuminate the portion of the LCD
that represents the cloudy sky while functionally black-masking (i.e., failing to illuminate) the
portion of the LCD that represents the tree branch and other elements that are not to be
perceived at the same focal distance as the sky, and the variable focus element (such as a
deformable membrane mirror) may be utilized to on the focal plane at optical infinity so
that the eye sees a sub-image at me1 as being clouds that are infinitely far away.
Then at subframe2, the variable focus element may be switched to focusing on a
point about 1 meter away from the user’s eyes (or whatever distance is required; here 1 meter
for the branch location is used for illustrative es), the pattern of illumination from the DLP
can be switched so that the system only illuminates the portion of the LCD that ents the
tree branch while functionally black-masking (i.e., g to illuminate) the portion of the LCD
that represents the sky and other elements that are not to be perceived at the same focal
distance as the tree branch. Thus the eye gets a quick flash of cloud at optical infinity followed
by a quick flash of tree at 1 meter, and the sequence is integrated by the eye/brain to form a
three-dimensional perception. The branch may be oned diagonally relative to the viewer,
such that it extends through a range of g distances, e.g., it may join with the trunk at
around 2 meters viewing ce while the tips of the branch are at the closer position of 1
meter.
] In this case, the display system can divide the 3-D volume of the tree branch into
multiple slices, rather than a single slice at 1 meter. For instance, one focus slice may be used
to represent the sky (using the DLP to mask all areas of the tree during presentation of this
slice), while the tree branch is divided across 5 focus slices (using the DLP to mask the sky and
all portions of the tree except one, for each part of the tree branch to be presented). Preferably,
the depth slices are positioned with a spacing equal to or smaller than the depth of focus of the
eye, such that the viewer will be unlikely to notice the transition between slices, and instead
perceive a smooth and uous flow of the branch through the focus range.
In another embodiment, rather than utilizing the DLP in a binary (illumination or
darkfield only) mode, it may be utilized to project a grayscale (for example, 256 shades of
grayscale) mask onto the back of the LCD panel to enhance three-dimensional tion. The
grayscale shades may be utilized to impart to the eye/brain a perception that something resides
in between adjacent depth orfocal planes. Back to the branch and clouds scenario, if the
leading edge of the branch closest to the user is to be in focalplane1, then at subframe1, that
n branch on the LCD may be lit up with full intensity white from the DLP system with the
variable focus element at focalplane1.
Then at subframe2, with the variable focus element at focalplane2 right behind the
part that was lit up, there would be no illumination. These are similar steps to the binary DLP
masking configuration above. r, if there is a portion of the branch that is to be
perceived at a position between focalplane1 and lane1, e.g., halfway, grayscale g
can be utilized. The DLP can project an illumination mask to that portion during both
subframe1 and subframe2, but at half-illumination (such as at level 128 out of 256 grayscale)
for each subframe. This es the perception of a blending of depth of focus layers, with the
perceived focal ce being proportional to the illuminance ratio between subframe1 and
subframe2. For instance, for a n of the tree branch that should lie 3/4ths of the way
between lane1 and focalplane2, an about 25% intensity grayscale mask can be used to
illuminate that portion of the LCD at subframe1 and an about 75% grayscale mask can be used
to illuminate the same portion of the LCD at subframe2.
In one embodiment, the bit depths of both the low-frame-rate display and the high-
frame-rate display can be combined for image modulation, to create a high dynamic range
display. The high dynamic range driving may be conducted in tandem with the focus plane
addressing function described above, to comprise a high dynamic range focal 3-D display.
In another embodiment that may be more efficient on computation resources, only a
certain portion of the display (i.e., LCD) output may be mask-illuminated by the DMD and
variably focused en route to the user’s eye. For e, the middle portion of the display may
be mask nated, with the periphery of the display not providing varying odation
cues to the user (i.e. the periphery could be uniformly illuminated by the DLP DMD, while a
central n is actively masked and variably focused en route to the eye).
In the above described embodiment, a refresh rate of about 360 Hz allows for 6
depth planes at about 60 frames/second each. In another embodiment, even higher refresh
rates may be achieved by increasing the operating frequency of the DLP. A standard DLP
uration uses a MEMS device and an array of micro-mirrors that toggle between a mode of
reflecting light toward the display or user to a mode of reflecting light away from the display or
user, such as into a light trap—thus they are inherently binary. DLPs typically create grayscale
images using a pulse width modulation schema wherein the mirror is left in the “on” state for a
variable amount of time for a variable duty cycle in order to create a brighter pixel, or pixel of
interim brightness. Thus, to create grayscale images at moderate frame rate, they are running
at a much higher binary rate.
In the above bed configurations, such setup works well for creating grayscale
masking. However, if the DLP drive scheme is adapted so that it is flashing subimages in a
binary pattern, then the frame rate may be sed icantly — by thousands of frames per
second, which allows for ds to thousands of depth planes being refreshed at 60
frames/second, which may be utilized to obviate the between-depth-plane ale
interpolating as described above. A typical pulse width modulation scheme for a Texas
Instruments DLP system has an 8—bit command signal (first bit is the first long pulse of the
mirror; second bit is a pulse that is half as long as the first; third bit is half as long again; and
so on) — so that the configuration can create 2 to the 8th power different illumination . In
one ment, the backlighting from the DLP may have its intensity varied in sync with the
different pulses of the DMD to equalize the brightness of the subimages that are created, which
is a practical workaround to get existing DMD drive electronics to produce significantly higher
frame rates.
In another embodiment, direct l s to the DMD drive electronics and
software may be utilized to have the mirrors always have an equal on-time instead of the
variable on-time configuration that is conventional, which would facilitate higher frame rates. In
another embodiment, the DMD drive electronics may be configured to t low bit depth
images at a frame rate above that of high bit depth images but lower than the binary frame rate,
enabling some grayscale blending between focus planes, while moderately increasing the
number of focus .
In another embodiment, when limited to a finite number of depth planes, such as 6 in
the example above, it is desirable to functionally move these 6 depth planes around to be
maximally useful in the scene that is being presented to the user. For example, if a user is
standing in a room and a virtual monster is to be placed into his augmented reality view, the
virtual monster being about 2 feet deep in the Z axis straight away from the user’s eyes, then it
makes sense to cluster all 6 depth planes around the center of the monster’s t location
(and dynamically move them with him as he moves relative to the user) — so that more rich
accommodation cues may be provided for the user, with all six depth planes in the direct region
of the monster (for example, 3 in front of the center of the monster, 3 in back of the center of
the monster). Such allocation of depth planes is content dependent.
For example, in the scene above the same monster is to be presented in the same
room, but also to be presented to the user is a virtual window frame element, and then a virtual
view to optical ty out of the l window frame, it will be useful to spend at least one
depth plane on optical infinity, one on the depth of the wall that is to house the virtual window
frame, and then perhaps the remaining four depth planes on the monster in the room. If the
content causes the virtual window to disappear, then the two depth planes may be dynamically
reallocated to the region around the monster, and so on — content-based dynamic allocation of
focal plane resources to provide the most rich ence to the user given the computing and
presentation resources.
In another embodiment, phase delays in a multicore fiber or an array of single-core
fibers may be utilized to create variable focus light wavefronts. Referring to Figure 9A, a
multicore fiber (300) may comprise the ation of multiple individual fibers (302); Figure
QB shows a up view of a multicore assembly, which emits light from each core in the form
of a spherical wavefront (304) from each. If the cores are transmitting coherent light, e.g., from
a shared laser light source, these small cal onts ultimately constructively and
ctively interfere with each other, and if they were emitted from the multicore fiber in
phase, they will develop an approximately planar wavefront (306) in the aggregate, as shown.
However, if phase delays are induced between the cores (using a conventional phase
modulator such as one using lithium niobate, for example, to slow the path of some cores
relative to others), then a curved or spherical wavefront may be created in the aggregate, to
represent at the eyes/brain an object coming from a point closer than optical infinity, which
presents another option that may be used in place of the variable focus elements described
above. In other words, such a phased multicore configuration, or phased array, may be utilized
to create le optical focus levels from a light source.
In another embodiment related to the use of optical fibers, a known Fourier
orm aspect of multi-mode optical fiber or light guiding rods or pipes may be utilized for
control of the wavefronts that are output from such fiber. Optical fibers typically are available in
two categories: single mode and multi-mode. Multi-mode optical fiber typically has larger core
ers and allows light to ate along multiple angular paths, rather than just the one of
single mode optical fiber. It is known that if an image is injected into one end of a multi-mode
fiber, that angular differences that are encoded into that image will be retained to some degree
as it propagates through the multi-mode fiber, and for some configurations the output from the
fiber will be significantly similar to a Fourier transform of the image that was input.
Thus in one embodiment, the inverse Fourier transform of a wavefront (such as a
diverging spherical wavefront to represent a focal plane nearer to the user than l infinity)
may be input so that, after passing through the fiber that optically imparts a Fourier transform,
the output is the desired shaped, or focused, wavefront. Such output end may be scanned
about to be used as a scanned fiber display, or may be used as a light source for a scanning
mirror to form an image, for instance. Thus such a configuration may be utilized as yet another
focus tion subsystem. Other kinds of light patterns and wavefronts may be injected into
a multi-mode fiber, such that on the output end, a certain spatial pattern is emitted. This may
be utilized to have the lent of a wavelet pattern (in optics, an optical system may be
analyzed in terms of what are called the ke coefficients; images may be similarly
characterized and decomposed into smaller principal ents, or a weighted combination
of comparatively simpler image components). Thus if light is scanned into the eye using the
principal components on the input side, a higher resolution image may be recovered at the
output end of the multi-mode fiber.
In another embodiment, the Fourier transform of a hologram may be injected into the
input end of a multi-mode fiber to output a wavefront that may be used for three-dimensional
focus modulation and/or resolution ement. Certain single fiber core, multi-core fibers, or
tric core + cladding configurations also may be utilized in the entioned inverse
r transform urations.
In r ment, rather than physically manipulating the wavefronts
approaching the eye of the user at a high frame rate without regard to the user’s particular state
of accommodation or eye gaze, a system may be configured to monitor the user’s
accommodation and rather than presenting a set of multiple different light onts, present a
single wavefront at a time that corresponds to the accommodation state of the eye.
Accommodation may be measured directly (such as by infrared autorefractor or eccentric
photorefraction) or indirectly (such as by measuring the convergence level of the two eyes of
the user; as described above, vergence and accommodation are strongly linked neurologically,
so an estimate of accommodation can be made based upon vergence geometry). Thus with a
determined accommodation of, say, 1 meter from the user, then the wavefront presentations at
the eye may be configured for a 1 meter focal distance using any of the above variable focus
configurations. If an accommodation change to focus at 2 meters is detected, the wavefront
presentation at the eye may be reconfigured for a 2 meter focal ce, and so on.
Thus in one embodiment incorporating accommodation tracking, a variable focus
element may be placed in the l path between an outputting er (e.g., a waveguide
or beamsplitter) and the eye of the user, so that the focus may be changed along with (Le,
preferably at the same rate as) accommodation changes of the eye. Software effects may be
utilized to produce variable amounts blur (e.g., Gaussian) to objects which should not be in
focus to simulate the dioptric blur expected at the retina if an object were at that viewing
ce and enhance the three-dimensional tion by the eyes/brain.
A simple ment is a single plane whose focus level is slaved to the viewer’s
accommodation level, however the performance demands on the accommodation tracking
system can be relaxed if even a low number of multiple planes are used. Referring to Figure 10,
in another embodiment, a stack (328) of about 3 waveguides (318, 320, 322) may be utilized to
create three focal planes worth of wavefronts simultaneously. In one embodiment, the weak
lenses (324, 326) may have static focal distances, and a variable focal lens (316) may be
slaved to the accommodation tracking of the eyes such that one of the three waveguides (say
the middle waveguide 320) s what is deemed to be the in-focus wavefront, while the
other two ides (322, 318) output a + margin wavefront and a — margin ont (Le, a
little farther than detected focal distance, a little closer than detected focal distance) which may
improve the three-dimensional perception and also provide enough difference for the brain/eye
accommodation control system to sense some blur as negative feedback, which enhances the
perception of reality, and allows a range of accommodation before an physical adjustment of
the focus levels is necessary.
] A variable focus compensating lens (314) is also shown to ensure that light coming
in from the real world (144) in an ted reality configuration is not refocused or magnified
by the assembly of the stack (328) and output lens (316). The variable focus in the lenses
(316, 314) may be achieved, as discussed above, with refractive, ctive, or reflective
techniques.
In another embodiment, each of the waveguides in a stack may contain their own
capability for changing focus (such as by having an included electronically switchable DOE) so
that the variable focus t need not be centralized as in the stack (328) of the
configuration of Figure 10.
] In another embodiment, variable focus elements may be interleaved between the
waveguides of a stack (i.e., rather than fixed focus weak lenses as in the embodiment of Figure
) to obviate the need for a combination of fixed focus weak lenses plus whole-stack-
refocusing variable focus element.
Such stacking configurations may be used in accommodation tracked variations as
described herein, and also in a frame-sequential multi-focal display approach.
In a configuration wherein light enters the pupil with a small exit pupil, such as 1/2
mm diameter or less, one has the equivalent of a pinhole lens configuration wherein the beam
is always interpreted as in-focus by the eyes/brain—e.g., a scanned light display using a 0.5
mm diameter beam to scan images to the eye. Such a configuration is known as a Maxwellian
view configuration, and in one embodiment, accommodation tracking input may be utilized to
induce blur using software to image ation that is to be perceived as at a focal plane
behind or in front of the focal plane ined from the odation tracking. In other
words, if one starts with a display presenting a Maxwellian view, then everything theoretically
can be in focus, and to e a rich and natural three-dimensional perception, simulated
dioptric blur may be induced with software, and may be slaved to the accommodation tracking
status.
In one embodiment a scanning fiber y is well suited to such configuration
because it may be configured to only output small-diameter beams in a Maxwellian form. In
r embodiment, an array of small exit pupils may be created to increase the functional eye
box of the system (and also to reduce the impact of a light-blocking particle which may reside in
the vitreous or cornea of the eye), such as by one or more scanning fiber displays, or by a DOE
uration such as that described in reference to Figure 8K, with a pitch in the array of
presented exit pupils that ensure that only one will hit the anatomical pupil of the user at any
given time (for e, if the average anatomical pupil diameter is 4mm, one configuration
may comprise 1/2 mm exit pupils spaced at intervals of imate 4mm apart). Such exit
pupils may also be switchable in response to eye position, such that only the eye always
receives one, and only one, active small exit pupil at a time; allowing a denser array of exit
pupils. Such user will have a large depth of focus to which software-based blur techniques may
be added to enhance perceived depth perception.
As discussed above, an object at optical infinity creates a substantially planar
wavefront; an object closer, such as 1m away from the eye, creates a curved wavefront (with
about 1m convex radius of curvature). The eye’s optical system needs to have enough optical
power to bend the incoming rays of light so that they end up focused on the retina (convex
wavefront gets turned into concave, and then down to a focal point on the retina). These are
basic functions of the eye.
In many of the embodiments described above, light directed to the eye has been
treated as being part of one continuous ont, some subset of which would hit the pupil of
the particular eye. In another approach, light ed to the eye may be effectively discretized
or broken down into a plurality of beamlets or individual rays, each of which has a diameter less
than about 0.5mm and a unique propagation pathway as part of a greater aggregated
wavefront that may be functionally created with the an aggregation of the beamlets or rays. For
example, a curved ont may be approximated by aggregating a plurality of discrete
neighboring collimated beams, each of which is ching the eye from an appropriate angle
to represent a point of origin that matches the center of the radius of curvature of the d
aggregate wavefront.
When the beamlets have a diameter of about 0.5mm or less, it is as though it is
coming through a pinhole lens configuration, which means that each individual beamlet is
always in relative focus on the retina, independent of the accommodation state of the eye—
however the trajectory of each beamlet will be affected by the accommodation state. For
instance, if the beamlets approach the eye in parallel, representing a discretized collimated
aggregate wavefront, then an eye that is correctly accommodated to ty will deflect the
beamlets to all converge upon the same shared spot on the retina, and will appear in focus. If
the eye accommodates to, say, 1 m, the beams will be converged to a spot in front of the ,
cross paths, and fall on multiple neighboring or partially overlapping spots on the retina—
appearing blurred.
If the beamlets approach the eye in a diverging uration, with a shared point of
origin 1 meter from the viewer, then an accommodation of 1 m will steer the beams to a single
spot on the retina, and will appear in focus; if the viewer odates to infinity, the beamlets
will converge to a spot behind the , and produce multiple neighboring or partially
overlapping spots on the retina, producing a d image. Stated more generally, the
accommodation of the eye ines the degree of overlap of the spots on the , and a
given pixel is “in focus” when all of the spots are directed to the same spot on the retina and
“defocused” when the spots are offset from one another. This notion that all of the 0.5mm
diameter or less beamlets are always in focus, and that they may be aggregated to be
perceived by the eyes/brain as though they are substantially the same as coherent wavefronts,
may be utilized in producing configurations for comfortable three-dimensional virtual or
augmented reality perception.
In other words, a set of multiple narrow beams may be used to emulate what is
going on with a larger diameter variable focus beam, and if the beamlet diameters are kept to a
maximum of about 0.5mm, then they maintain a relatively static focus level, and to produce the
perception of out-of-focus when desired, the beamlet r trajectories may be selected to
create an effect much like a larger out-of-focus beam (such a defocussing treatment may not be
the same as a Gaussian blur treatment as for the larger beam, but will create a multimodal
point spread function that may be interpreted in a similar fashion to a Gaussian blur).
In a preferred embodiment, the beamlets are not mechanically deflected to form this
aggregate focus effect, but rather the eye receives a superset of many beamlets that includes
both a multiplicity of incident angles and a multiplicity of ons at which the beamlets
intersect the pupil; to ent a given pixel from a particular viewing ce, a subset of
beamlets from the superset that comprise the appropriate angles of incidence and points of
intersection with the pupil (as if they were being emitted from the same shared point of origin in
space) are turned on with matching color and intensity, to represent that aggregate wavefront,
while beamlets in the superset that are inconsistent with the shared point of origin are not
turned on with that color and intensity (but some of them may be turned on with some other
color and intensity level to represent, e.g., a different pixel).
Referring to Figure 11A, each of a multiplicity of incoming beamlets (332) is passing
through a small exit pupil (330) relative to the eye (58) in a discretized wavefront display
configuration. ing to Figure 118, a subset (334) of the group of ts (332) may be
driven with matching color and intensity levels to be perceived as though they are part of the
same larger-sized ray (the bolded subgroup 334 may be deemed an “aggregated beam”). In
this case, the subset of beamlets are parallel to one another, representing a collimated
aggregate beam from optical infinity (such as light coming from a distant mountain). The eye is
odated to infinity, so the subset of beamlets are deflected by the eye’s cornea and lens
to all fall substantially upon the same location of the retina and are ved to comprise a
single in focus pixel.
Figure 11C shows another subset of beamlets representing an aggregated
collimated beam (336) coming in from the right side of the field of view of the user’s eye (58) if
the eye (58) is viewed in a l-style planar view from above. Again, the eye is shown
accommodated to infinity, so the beamlets fall on the same spot of the retina, and the pixel is
perceived to be in focus. If, in contrast, a different subset of beamlets were chosen that were
ng the eye as a diverging fan of rays, those beamlets would not fall on the same location
of the retina (and be perceived as in focus) until the eye were to shift odation to a near
point that matches the geometrical point of origin of that fan of rays.
As regards patterns of points of ection of beamlets with the anatomical pupil of
the eye (i.e., the pattern of exit pupils), they may be organized in urations such as a
cross-sectionally efficient hex-lattice (for example, as shown in Figure 12A) or a square lattice
or other two-dimensional array. Further, a three-dimensional array of exit pupils could be
created, as well as time-varying arrays of exit .
Discretized aggregate wavefronts may be created using several configurations, such
as an array of microdisplays or microprojectors placed optically conjugate with the exit pupil of
viewing optics, microdisplay or microprojector arrays coupled to a direct field of view substrate
(such as an eyeglasses lens) such that they project light to the eye directly, t additional
intermediate viewing optics, successive l light modulation array techniques, or ide
techniques such as those described in relation to Figure 8K.
Referring to Figure 12A, in one embodiment, a lightfield may be created by bundling
a group of small projectors or display units (such as scanned fiber displays). Figure 12A
depicts a hexagonal lattice tion bundle (338) which may, for example, create a 7mm-
diameter hex array with each fiber display outputting a sub-image (340). If such an array has
an optical , such as a lens, placed in front of it such that the array is placed optically
conjugate with the eye’s entrance pupil, this will create an image of the array at the eye’s pupil,
as shown in Figure 12B, which essentially provides the same optical ement as the
embodiment of Figure 11A.
Each of the small exit pupils of the configuration is created by a dedicated small
display in the bundle (338), such as a ng fiber display. Optically, it’s as though the entire
hex array (338) is positioned right into the anatomical pupil (45). Such embodiments are
means for g different subimages to different small exit pupils within the larger anatomical
ce pupil (45) of the eye, comprising a superset of beamlets with a multiplicity of incident
angles and points of intersection with the eye pupil. Each of the separate projectors or displays
may be driven with a slightly different image, such that subimages may be created that pull out
different sets of rays to be driven at different light intensities and colors.
In one variation, a strict image conjugate may be created, as in the embodiment of
Figure 12B, n there is direct 1-to-1 mapping of the array (338) with the pupil (45). In
another variation, the spacing may be changed between displays in the array and the optical
system (lens 342, in Figure 12B) so that instead of getting a conjugate mapping of the array to
the eye pupil, the eye pupil may be catching the rays from the array at some other ce.
With such a configuration, one would still get an angular diversity of beams through which one
could create a discretized aggregate wavefront representation, but the mathematics regarding
how to drive which ray and at which power and intensity may become more complex (although,
on the other hand, such a configuration may be considered simpler from a viewing optics
perspective). The mathematics involved with light field image capture may be ged for
these calculations.
Referring to Figure 13A, another lightfield ng embodiment is depicted wherein
an array of microdisplays or microprojectors (346) may be coupled to a frame (344; such as an
eyeglasses frame) to be positioned in front of the eye (58). The depicted configuration is a
nonconjugate arrangement wherein there are no large-scale optical elements interposed
between the displays (for example, scanning fiber displays) of the array (346) and the eye (58).
One can imagine a pair of glasses, and d to those glasses are a ity of ys,
such as scanning fiber engines, positioned orthogonal to the eyeglasses surface, and all angled
inward so they are pointing at the pupil of the user. Each display may be configured to create a
set of rays representing different elements of the beamlet superset.
With such a configuration, at the ical pupil (45) the user is going to receive a
similar result as received in the embodiments discussed in reference to Figure 11A, in which
every point at the user’s pupil is ing rays with a multiplicity of angles of incidence and
points of intersection that are being contributed from the ent displays. Figure 13B
illustrates a nonconjugate configuration similar to that of Figure 13A, with the ion that the
embodiment of Figure 13B es a reflecting surface (348) to facilitate moving the display
array (346) away from the eye’s (58) field of view, while also allowing views of the real world
(144) through the reflective surface (348).
Thus another configuration for creating the angular ity necessary for a
discretized aggregate wavefront display is ted. To optimize such a uration, the
sizes of the displays may be decreased to the maximum. Scanning fiber displays which may
be utilized as displays may have baseline diameters in the range of 1mm, but reduction in
ure and projection lens hardware may decrease the diameters of such displays to about
0.5mm or less, which is less disturbing for a user. Another downsizing geometric refinement
may be achieved by directly coupling a collimating lens (which may, for example, comprise a
gradient refractive index, or “GRIN”, lens, a conventional curved lens, or a diffractive lens) to
the tip of the scanning fiber itself in a case of a fiber scanning display array. For example,
referring to Figure 13D, a GRIN lens (354) is shown fused to the end of a single mode optical
fiber. An actuator (350; such as a lectric actuator) is coupled to the fiber (352) and may
be used to scan the fiber tip.
In another embodiment the end of the fiber may be shaped into a hemispherical
shape using a curved polishing treatment of an optical fiber to create a lensing effect. In
another embodiment a standard refractive lens may be coupled to the end of each optical fiber
using an ve. In another embodiment a lens may be built from a dab of transmissive
polymeric al or glass, such as epoxy. In another embodiment the end of an optical fiber
may be melted to create a curved surface for a lensing effect.
Figure 13C-2 shows an embodiment n display configurations (i.e., scanning
fiber ys with GRIN lenses; shown in close-up view of Figure 13C-1) such as that shown
in Figure 13D may be coupled together through a single transparent substrate (356) preferably
having a refractive index that y matches the cladding of the l fibers (352) so that the
fibers themselves are not very visible for g of the outside world across the depicted
assembly (if the index matching of the cladding is done precisely, then the larger
cladding/housing becomes transparent and only the tiny cores, which preferably are about 3
microns in diameter, will be obstructing the view. In one embodiment the matrix (358) of
ys may all be angled inward so they are directed toward the anatomic pupil of the user (in
another embodiment, they may stay parallel to each other, but such a configuration is less
efficient).
Referring to Figure 13E, another embodiment is depicted wherein rather than using
circular fibers to move cyclically, a thin series of planar waveguides (358) are configured to be
cantilevered relative to a larger substrate structure (356). In one variation, the substrate (356)
may be moved to produce cyclic motion (i.e., at the resonant frequency of the cantilevered
members 358) of the planar waveguides relative to the ate structure. In r
variation, the cantilevered waveguide portions (358) may be actuated with piezoelectric or other
actuators relative to the substrate. lmage illumination information may be injected, for example,
from the right side (360) of the substrate structure to be d into the cantilevered
waveguide portions (358). In one embodiment the substrate (356) may comprise a waveguide
configured (such as with an integrated DOE configuration as described above) to totally
internally t incoming light (360) along its length and then redirect it to the cantilevered
waveguide portions (358). As a person gazes toward the cantilevered waveguide portions
(358) and through to the real world (144) , the planar waveguides are configured to
minimize any dispersion and/or focus changes with their planar shape factors.
In the context of discussing discretized ate wavefront displays, there is value
placed in having some angular diversity created for every point in the exit pupil of the eye. In
other words, it is desirable to have multiple incoming beams to represent each pixel in a
displayed image. Referring to Figures 13F-1 and 13F-2, one way to gain further angular and
spatial diversity is to use a multicore fiber and place a lens at the exit point, such as a GRIN
lens, so that the exit beams are deflected through a single nodal point (366); that nodal point
may then be scanned back and forth in a scanned fiber type of arrangement (such as by a
piezoelectric actuator 368). If a retinal conjugate is placed at the plane defined at the end of
the GRIN lens, a display may be created that is functionally equivalent to the general case
tized aggregate ont configuration described above.
Referring to Figure 13G, a r effect may be achieved not by using a lens, but by
scanning the face of a multicore system at the correct conjugate of an optical system (372), the
goal being to create a higher angular and spatial diversity of beams. In other words, rather than
having a bunch of separately scanned fiber displays as in the bundled example of Figure 12A
described above, some of this requisite angular and spatial ity may be created h
the use of multiple cores to create a plane which may be relayed by a waveguide. Referring to
Figure 13H, a multicore fiber (362) may be scanned (such as by a piezoelectric actuator 368) to
create a set of beamlets with a multiplicity of angles of incidence and points of intersection
which may be relayed to the eye (58) by a waveguide (370). Thus in one ment a
collimated lightfield image may be injected into a waveguide, and without any additional
refocusing elements, that lightfield display may be translated directly to the human eye.
s 13l-13L depict certain commercially available multicore fiber (362)
configurations (from s such as Mitsubishi Cable ries, Ltd. of Japan), including one
variation (363) with a rectangular cross section, as well as variations with flat exit faces (372)
and angled exit faces (374).
Referring to Figure 13M, some additional angular diversity may be created by having
a waveguide (376) fed with a linear array of displays (378), such as scanning fiber displays.
Referring to Figures 14A-14F, another group of configurations for ng a fixed
viewpoint lightfield display is described. Referring back to Figure 11A, if a two-dimensional
plane was created that was intersecting all of the tiny beams coming in from the left, each
beamlet would have a certain point of ection with that plane. If another plane was created
at a different distance to the left, then all of the beamlets would ect that plane at a different
location. Then going back to Figure 14A, if s positions on each of two or more planes
can be allowed to selectively transmit or block the light radiation directed through it, such a
multi-planar configuration may be utilized to selectively create a lightfield by independently
modulating individual beamlets.
The basic ment of Figure 14A shows two spatial light modulators, such as
liquid l display panels (380, 382; in other embodiments they may be MEMS shutter
displays or DLP DMD arrays) which may be independently controlled to block or transmit
ent rays on a high-resolution basis. For example, referring to Figure 14A, if the second
panel (382) blocks or attenuates transmission of rays at point “a” (384), all of the depicted rays
will be blocked; but if only the first panel (380) blocks or attenuates transmission of rays at
point “b” (386), then only the lower incoming ray (388) will be blocked/attenuated, while the rest
will be transmitted toward the pupil (45). Each of the controllable panels or planes may be
deemed a “spatial light modulator” or “fatte”. The intensity of each transmitted beam passed
through a series of SLMs will be a function of the combination of the transparency of the
various pixels in the s SLM arrays. Thus without any sort of lens elements, a set of
beamlets with a multiplicity of angles and points of intersection (or a field”) may be created
using a plurality of stacked SLMs. Additional numbers of SLMs beyond two provides more
opportunities to control which beams are selectively attenuated.
As noted briefly above, in addition to using stacked liquid l ys as SLMs,
planes of DMD devices from DLP systems may be stacked to function as SLMs, and may be
preferred over liquid l systems as SLMs due to their ability to more efficiently pass light
(with a mirror element in a first state, reflectivity to the next element on the way to the eye may
be quite efficient; with a mirror element in a second state, the mirror angle may be moved by
an angle such as 12 degrees to direct the light away from the path to the eye). Referring to
Figure 14B, in one DMD embodiment, two DMDs (390, 390) may be utilized in series with a pair
of lenses (394, 396) in a periscope type of configuration to in a high amount of
transmission of light from the real world (144) to the eye (58) of the user. The embodiment of
Figure 14C provides six different DMD (402, 404, 406, 408, 410, 412) plane opportunities to
intercede from an SLM onality as beams are routed to the eye (58), along with two lenses
(398, 400) for beam control.
] Figure 14D illustrates a more cated periscope type arrangement with up to
four DMDs (422, 424, 426, 428) for SLM functionality and four lenses (414, 420, 416, 418); this
configuration is designed to ensure that the image does not become flipped upside down as it
travels through to the eye (58). Figure 14E illustrates in embodiment wherein light may be
reflected between two different DMD devices (430, 432) t any intervening lenses (the
lenses in the above designs are useful in such configurations for incorporating image
information from the real world), in a hall-of-mirrors type of arrangement wherein the display
may be viewed through the “hall of mirrors” and operates in a mode ntially similar to that
illustrated in Figure 14A. Figure 14F illustrates an embodiment wherein a the non-display
portions of two facing DMD chips (434, 436) may be covered with a reflective layer to
propagate light to and from active display regions (438, 440) of the DMD chips. In other
embodiments, in place of DMDs for SLM functionality, arrays of sliding MEMS shutters (such as
those available from vendors such as Pixtronics, a division of Qualcomm, Inc.) may be ed
to either pass or block light. In another ment, arrays of small louvers that move out of
place to t light-transmitting apertures may similarly be aggregated for SLM functionality.
A lightfield of many small beamlets (say, less than about 0.5mm in diameter) may be
ed into and propagated through a waveguide or other optical system. For example, a
conventional “birdbath” type of optical system may be suitable for transferring the light of a
lightfield input, or a freeform optics design, as described below, or any number of waveguide
configurations. Figures 15A—15C illustrate the use of a wedge type ide (442) along with
a plurality of light sources as another uration useful in creating a lightfield. Referring to
Figure 15A, light may be injected into the wedge-shaped waveguide (442) from two different
locations/displays (444, 446), and will emerge according to the total internal reflection
properties of the wedge-shaped waveguide at different angles (448) based upon the points of
injection into the waveguide.
Referring to Figure 15B, if one creates a linear array (450) of displays (such as
scanning fiber displays) projecting into the end of the waveguide as shown, then a large
angular diversity of beams (452) will be g the waveguide in one dimension, as shown in
Figure 15C. Indeed, if one contemplates adding yet another linear array of displays injecting
into the end of the waveguide but at a slightly ent angle, then an angular diversity of
beams may be created that exits similarly to the fanned out exit n shown in Figure 15C,
but at an orthogonal axis; together these may be utilized to create a two-dimensional fan of
rays exiting each location of the waveguide. Thus another configuration is presented for
creating angular diversity to form a ield display using one or more scanning fiber display
arrays (or alternatively using other displays which will meet the space requirements, such as
urized DLP projection configurations).
atively, as an input to the shaped waveguides shown herein, a stack of
SLM devices may be utilized, in which case rather than the direct view of SLM output as
described above, the lightfield output from the SLM configuration may be used as an input to a
configuration such as that shown in Figure 15C. One of the key concepts here is that while a
conventional waveguide is best suited to relay beams of collimated light successfully, with a
lightfield of small-diameter collimated beams, conventional ide technology may be
utilized to further manipulate the output of such a lightfield system as injected into the side of a
waveguide, such as a wedge-shaped waveguide, due to the beam size / collimation.
In another related embodiment, rather than projecting with multiple separate
displays, a multicore fiber may be used to generate a lightfield and inject it into the waveguide.
Further, a time-varying ield may be utilized as an input, such that rather than creating a
static distribution of beamlets coming out of a lightfield, one may have some dynamic elements
that are methodically changing the path of the set of beams. They may be done using
components such as waveguides with embedded DOEs (e.g., such as those described above
in reference to Figures SB-8N, or liquid crystal layers, as described in reference to Figure 78),
wherein two optical paths are created (one smaller total internal reflection path wherein a liquid
crystal layer is placed in a first voltage state to have a refractive index mismatch with the other
substrate material that causes total internal reflection down just the other substrate material’s
ide; one larger total internal reflection optical path wherein the liquid crystal layer is
placed in a second e state to have a matching refractive index with the other substrate
material, so that the light totally internally reflects through the composite waveguide which
includes both the liquid crystal portion and the other ate portion). Similarly a wedgeshaped
waveguide may be configured to have a bi-modal total internal reflection paradigm (for
example, in one variation, wedge-shaped elements may be configured such that when a liquid
crystal portion is activated, not only is the spacing changed, but also the angle at which the
beams are reflected).
One embodiment of a scanning light display may be characterized simply as a
ng fiber display with a lens at the end of the d fiber. Many lens varieties are
suitable, such as a GRIN lens, which may be used to collimate the light or to focus the light
down to a spot smaller than the fiber’s mode field diameter providing the advantage of
producing a numerical aperture (or “NA”) se and circumventing the optical invariant,
which is correlated inversely with spot size. r spot size generally tates a higher
resolution opportunity from a display ctive, which generally is preferred. In one
embodiment, a GRIN lens may be long enough relative to the fiber that it may comprise the
ing element (i.e., rather than the usual distal fiber tip vibration with a scanned fiber
display) — a configuration which may be deemed a “scanned GRIN lens display”.
In another embodiment, a diffractive lens may be utilized at the exit end of a
ng fiber display (i.e., patterned onto the fiber). In another embodiment, a curved mirror
may be positioned on the end of the fiber that operates in a reflecting configuration. Essentially
any of the urations known to collimate and focus a beam may be used at the end of a
scanning fiber to produce a suitable scanned light display.
Two significant utilities to having a lens coupled to or comprising the end of a
scanned fiber (i.e., as compared to configurations wherein an led lens may be utilized to
direct light after it exits a fiber) are a) the light exiting may be collimated to obviate the need to
use other external optics to do so; b) the NA, or the angle of the cone at which light sprays out
the end of the -mode fiber core, may be increased, thereby decreasing the associated
spot size for the fiber and increasing the ble resolution for the display.
As described above, a lens such as a GRIN lens may be fused to or otherwise
d to the end of an optical fiber or formed from a portion of the end of the fiber using
techniques such as polishing. In one embodiment, a typical optical fiber with an NA of about
0.13 or 0.14 may have a spot size (also known as the “mode field diameter” for the optical fiber
given the NA) of about 3 microns. This provides for relatively high resolution display
possibilities given the industry standard display resolution paradigms (for example, a typical
microdisplay logy such as LCD or organic light emitting diode, or “OLED” has a spot size
of about 5 microns). Thus the aforementioned ng light display may have 3/5 of the
smallest pixel pitch available with a conventional display; further, using a lens at the end of the
fiber, the aforementioned configuration may produce a spot size in the range of 1-2 microns.
In another embodiment, rather than using a scanned cylindrical fiber, a cantilevered
portion of a waveguide (such as a waveguide created using microfabrication processes such as
masking and etching, rather than drawn microfiber techniques) may be placed into scanning
atory , and may be fitted with lensing at the exit ends.
In another embodiment, an increased numerical aperture for a fiber to be scanned
may be created using a er (i.e., one configured to scatter light and create a larger NA)
covering the exit end of the fiber. In one variation, the diffuser may be created by etching the
end of the fiber to create small bits of terrain that scatter light; in another variation a bead or
sandblasting technique, or direct sanding/scuffing technique may be utilized to create ring
terrain. In another variation, an engineered diffuser, similar to a diffractive element, may be
created to maintain a clean spot size with desirable NA, which ties into the notion of using a
diffractive lens, as noted above.
Referring to Figure 16A, an array of l fibers (454) is shown coupled in to a
coupler (456) configured to hold them in parallel together so that their ends may be ground and
polished to have an output edge at a al angle (458; 42 s for most glass, for
example) to the longitudinal axes of the input fibers, such that the light exiting the angled faces
will exit as though it had been passing through a prism, and will bend and become nearly
parallel to the surfaces of the polished faces. The beams exiting the fibers (454) in the bundle
will become superimposed, but will be out of phase longitudinally due to the different path
lengths (referring to Figure 168, for example, the difference in path lengths from angled exit
face to ng lens for the different cores is visible).
What was an X axis type of separation in the bundle before exit from the angled
faces, will become a Z axis separation, a fact that is helpful in creating a multifocal light source
from such a configuration. In another embodiment, rather than using a bundled/coupled
plurality of single mode fibers, a multicore fiber, such as those ble from Mitsubishi Cable
Industries, Ltd. of Japan, may be angle polished.
In one embodiment, if a 45 degree angle is polished into a fiber and then covered
with a reflective element, such as a mirror coating, the exiting light may be reflected from the
polished surface and emerge from the side of the fiber (in one embodiment at a location
wherein a olished exit window has been created in the side of the fiber) such that as the
fiber is scanned in what would normally be an X-Y Cartesian coordinate system axis, that fiber
would now be functionally performing the equivalent of an X-Z scan, with the distance changing
during the course of the scan. Such a configuration may be beneficially ed to change the
focus of the display as well.
Multicore fibers may be configured to play a role in display resolution enhancement
(i.e., higher resolution). For example, in one embodiment, if te pixel data is sent down a
tight bundle of 19 cores in a multicore fiber, and that cluster is scanned around in a sparse
spiral pattern with the pitch of the spiral being approximately equal to the er of the
multicore, then sweeping around will effectively create a display resolution that is approximately
19x the resolution of a single core fiber being similarly scanned around. Indeed, it may be more
practical to have the fibers more sparsely positioned relative to each other, as in the
configuration of Figure 16C, which has 7 clusters (464; 7 is used for illustrative purposes
because it is an ent /hex pattern; other patterns or numbers may be utilized; for
example, a cluster of 19; the configuration is le up or down) of 3 fibers each housed
within a conduit (462).
] With a sparse configuration as shown in Figure 16C, ng of the multicore
scans each of the cores through its own local region, as d to a configuration wherein the
cores are all packed tightly together and scanned (wherein cores end up pping with
scanning; if the cores are too close to each other, the NA of the core is not large enough and
the very closely packed cores end up blurring together somewhat and not creating as
discriminable a spot for display). Thus, for resolution increases, it is preferable to have sparse
tiling rather than highly dense tiling, although both will work.
The notion that densely packed scanned cores can create blurring at the display
may be utilized as an advantage in one embodiment wherein a plurality (say a triad or cores to
carry red, green, and blue light) of cores may be intentionally packed together densely so that
each triad forms a triad of overlapped spots featuring red, green, and blue light. With such a
configuration, one is able to have an RGB display without having to combine red, green, and
blue into a single-mode core, which is an advantage, because conventional mechanisms for
combining a plurality (such as three) wavelets of light into a single core are subject to significant
losses in optical energy. Referring to Figure 16C, in one embodiment each tight cluster of 3
fiber cores contains one core that relays red light, one core that relays green light, and one core
that relays blue light, with the 3 fiber cores close enough together that their positional
differences are not able by the subsequent relay optics, g an effectively
superimposed RGB pixel; thus, the sparse tiling of 7 clusters produces resolution enhancement
while the tight packing of 3 cores within the clusters facilitates seamless color blending without
the need to utilize glossy RGB fiber ers (e.g., those using ngth division
multiplexing or evanescent coupling ques).
Referring to Figure 16D, in another more simple variation, one may have just one
cluster (464) housed in a conduit (468) for, say, red/green/blue (and in another embodiment,
another core may be added for infrared for uses such as eye tracking). In another embodiment,
onal cores may be placed in the tight cluster to carrying onal wavelengths of light to
comprise a multi-primary display for increased color gamut. Referring to Figure 16E, in another
embodiment, a sparse array of single cores (470); in one variation with red, green, and blue
combined down each of them) within a conduit (466) may be utilized; such a configuration is
workable albeit somewhat less efficient for tion increase, but not optimum for
red/green/blue combining.
Multicore fibers also may be utilized for creating lightfield displays. lndeed, rather
than keeping the cores separated enough from each other so that the cores do not scan on
each other’s local area at the display panel, as described above in the context of creating a
scanning light display, with a lightfield display, it is ble to scan around a densely packed
ity of fibers because each of the beams produced represents a specific part of the
lightfield. The light exiting from the bundled fiber tips can be relatively narrow if the fibers have
a small NA; lightfield configurations may take age of this and have an arrangement in
which at the anatomic pupil, a plurality of slightly different beams are being received from the
array. Thus there are optical configurations with scanning a multicore that are functionally
equivalent to an array of single scanning fiber modules, and thus a lightfield may be created by
scanning a multicore rather than scanning a group of single mode fibers.
In one embodiment, a core phased array approach may be used to create a
large exit pupil variable wavefront uration to tate three-dimensional perception. A
single laser configuration with phase modulators is described above. In a multicore
embodiment, phase delays may be d into different channels of a multicore fiber, such
that a single laser’s light is injected into all of the cores of the multicore configuration so that
there is mutual nce.
In one embodiment, a multi-core fiber may be combined with a lens, such as a GRIN
lens. Such lens may be, for example, a refractive lens, diffractive lens, or a polished edge
functioning as a lens. The lens may be a single optical surface, or may comprise multiple
optical surfaces stacked up. Indeed, in addition to having a single lens that extends the
diameter of the multicore, a smaller t array may be desirable at the exit point of light from
the cores of the multicore, for example. Figure 16F shows an embodimentwherein a multicore
fiber (470) is emitting multiple beams into a lens (472), such as a GRIN lens. The lens collects
the beams down to a focal point (474) in space in front of the lens. In many conventional
configurations, the beams would exit the multicore fiber as ing. The GRIN or other lens is
configured to function to direct them down to a single point and ate them, such that the
collimated result may be scanned around for a lightfield display, for instance.
Referring to Figure 16G, smaller lenses (478) may be placed in front of each of the
cores of a multicore (476) configuration, and these lenses may be utilized to collimate; then a
shared lens (480) may be configured to focus the ated beams down to a diffraction d
spot (482) that is aligned for all of the three spots. The net result of such a configuration: by
combining three collimated, narrow beams with narrow NA together as shown, one effectively
combines all three into a much larger angle of emission which translates to a smaller spot size
in, for example, a head mounted optical display system which may be next in the chain of light
delivery to the user.
Referring to Figure 16H, one embodiment es a multicore fiber (476) with a
lenslet (478) array feeding the light to a small prism array (484) that deflects the beams
generated by the dual cores to a common point. Alternatively one may have the small
lenslet array d relative to the cores such that the light is being deflected and focused down
to a single point. Such a configuration may be utilized to increase the numerical aperture.
Referring to Figure 16I, a two-step configuration is shown with a small lenslet (478)
array capturing light from the multicore fiber (476), followed sequentially by a shared lens (486)
to focus the beams to a single point (488). Such a configuration may be utilized to increase the
numerical aperture. As discussed above, a larger NA ponds to a smaller pixel size and
higher possible y resolution.
Referring to Figure 16J, a beveled fiber array which may be held together with a
r (456), such as those described above, may be scanned with a ting device (494;
such as a DMD module of a DLP system). With multiple single fibers (454) coupled into the
array, or a multicore instead, the superimposed light can be directed through one or more
focusing lenses (490, 492) to create a multifocal beam; with the superimposing and angulation
of the array, the different sources are ent distances from the focusing lens, which creates
different focus levels in the beams as they emerge from the lens (492) and are directed toward
the retina (54) of the eye (58) of the user. For example, the farthest optical route/beam may
be set up to be a collimated beam representative of optical infinity focal positions. Closer
routes/beams may be associated with diverging spherical wavefronts of closer focal locations.
The multifocal beam may be passed into a scanning mirror which may be configured
to create a raster scan (or, for example, a ous curve scan pattern or a spiral scan pattern)
of the multifocal beam which may be passed through a series of focusing lenses and then to
the cornea and crystalline lens of the eye. The various beams emerging from the lenses are
creating different pixels or voxels of g focal distances that are superimposed.
In one embodiment, one may write different data to each of the light modulation
ls at the front end, thereby creating an image that is projected to the eye with one or
more focus ts. By changing the focal distance of the crystalline lens (i.e., by
accommodating), the user can bring different incoming pixels into and out of focus, as shown in
Figures 16K and 16L wherein the crystalline lens is in different Z axis positions. In another
embodiment, the fiber array may be ed/moved around by a piezoelectric actuator. In
another embodiment, a relatively thin ribbon array may be resonated in evered form along
the axis perpendicular to the arrangement of the array fibers (i.e., in the thin direction of the
ribbon) when a piezoelectric actuator is activated. In one variation, a separate piezoelectric
actuator may be utilized to create a vibratory scan in the orthogonal long axis. In another
embodiment, a single mirror axis scan may be employed for a slow scan along the long axis
while the fiber ribbon is vibrated resonantly.
Referring to Figure 16M, an array (496) of ng fiber displays (498) may be
beneficially bundled/tiled for an effective resolution increase, the notion being that with such as
configuration, each scanning fiber of the bundle is configured to write to a different portion of
the image plane (500), as shown, for example, in Figure 16N, wherein each portion of the
image plane is addressed by the emissions from a least one bundle. In other ments,
optical configurations may be utilized that allow for slight magnification of the beams as they
exit the optical fiber so that there is some overlap in the nal, or other lattice pattern, that
hits the display plane, so there is a better fill factor while also maintaining an adequately small
spot size in the image plane and understanding that there is a subtle magnification in that
image plane.
Rather than having individual lenses at the end of each scanned fiber enclosure
housing, in one embodiment a monolithic t array may be utilized, so that the lenses can
be as closely packed as possible, which allows for even smaller spot sizes in the image plane
because one may use a lower amount of ication in the optical system. Thus arrays of
fiber scan displays may be used to increase the resolution of the y, or in other words, they
may be used to increase the field of view of the display, because each engine is being used to
scan a different portion of the field of view.
For a lightfield configuration, the emissions may be more desirably overlapped at the
image plane. In one embodiment, a lightfield display may be created using a plurality of small
diameterfibers scanned around in space. For example, instead of having all of the fibers
address a different part of an image plane as described above, have more pping, more
fibers angled inward, etc., or change the focal power of the lenses so that the small spot sizes
are not conjugate with a tiled image plane configuration. Such a configuration may be used to
create a lightfield display to scan lots of smaller diameter rays around that become intercepted
in the same al space.
Referring back to Figure 128, it was discussed that one way of ng a lightfield
display involves making the output of the elements on the left ated with narrow beams,
and then making the projecting array conjugate with the eye pupil on the right.
Referring to Figure 160, with a common substrate block (502), a single actuator
may be utilized to actuate a plurality of fibers (506) in unison together. A similar configuration is
discussed above in reference to Figures 13-C-1 and 13-C-2. It may be cally ult to
have all of the fibers retain the same resonant frequency, vibrate in a desirable phase
relationship to each other, or have the same ions of cantilevering from the substrate
block. To s this challenge, the tips of the fibers may be ically coupled with a
lattice or sheet (504), such as a graphene sheet that is very thin, rigid, and light in weight. With
such a coupling, the entire array may vibrate similarly and have the same phase relationship.
In another embodiment a matrix of carbon nanotubes may be utilized to couple the fibers, or a
piece of very thin planar glass (such as the kind used in creating liquid crystal display panels)
may be coupled to the fiber ends. Further, a laser or other precision cutting device may be
utilized to cut all associated fibers to the same cantilevered length.
Referring to Figure 17, in one embodiment it may be desirable to have a contact
lens ly interfaced with the cornea, and configured to tate the eye focusing on a
display that is quite close (such as the typical distance between a cornea and an eyeglasses
lens). Rather than placing an optical lens as a contact lens, in one variation the lens may
comprise a selective filter. Figure 17 depicts a plot (508) what may be deemed a “notch ”,
due to its design to block only certain wavelength bands, such as 450nm (peak blue), 530nm
(green), and 650nm, and generally pass or transmit other wavelengths. In one embodiment
several layers of tric coatings may be aggregated to provide the notch filtering
functionality.
Such a filtering configuration may be coupled with a scanning fiber display that is
producing a very narrow band illumination for red, green, and blue, and the contact lens with
the notch filtering will block out all of the light coming from the y (such as a minidisplay,
such as an OLED display, mounted in a position normally occupied by an sses lens)
except for the transmissive wavelengths. A narrow pinhole may be created in the middle of the
contact lens filtering layers/film such that the small aperture (i.e., less than about 1.5mm
diameter) does allow passage of the otherwise blocked wavelengths. Thus a pinhole lens
configuration is d that functions in a pinhole manner for red, green, and blue only to
intake images from the minidisplay, while light from the real world, which generally is
broadband illumination, will pass through the contact lens relatively unimpeded. Thus a large
depth of focus l display configuration may be assembled and operated. In another
embodiment, a collimated image exiting from a waveguide would be visible at the retina
because of the e large-depth-of-focus configuration.
It may be useful to create a display that can vary its depth of focus over time. For
example, in one embodiment, a display may be configured to have different display modes that
may be selected (preferably rapidly toggling between the two at the command of the operator)
by an or, such as a first mode combining a very large depth of focus with a small exit
pupil diameter (i.e., so that everything is in focus all of the time), and a second mode featuring a
larger exit pupil and a more narrow depth of focus. In ion, if a user is to play a three-
dimensional video game with s to be perceived at many depths of field, the operator may
select the first mode; alternatively, if a user is to type in a long essay (i.e., for a relatively long
period of time) using a two-dimensional word processing display configuration, it may be more
ble to switch to the second mode to have the convenience of a larger exit pupil, and a
sharper image.
In another embodiment, it may be desirable to have a multi-depth of focus y
configuration wherein some subimages are presented with a large depth of focus while other
subimages are presented with small depth of focus. For example, one configuration may have
red wavelength and blue wavelength channels ted with a very small exit pupil so that
they are always in focus. Then, a green channel only may be presented with a large exit pupil
configuration with multiple depth planes (i.e., because the human accommodation system tends
to preferentially target green ngths for optimizing focus level). Thus, in order to cut costs
associated with having too many elements to represent with full depth planes in red, green, and
blue, the green wavelength may be prioritized and represented with various different wavefront
levels. Red and blue may be relegated to being represented with a more Maxwellian approach
(and, as described above in reference to Maxwellian displays, software may be utilized to
induce Gaussian levels of blur). Such a display would simultaneously present multiple depths
of focus.
As described above, there are portions of the retina which have a higher density of
light sensors. The fovea portion, for example, generally is populated with approximately 120
cones per visual . Display systems have been created in the past that use eye or gaze
tracking as an input, and to save computation resources by only creating really high resolution
rendering for where the person is gazing at the time, while lower resolution rendering is
presented to the rest of the retina; the locations of the high versus low resolution portions may
be dynamically slaved to the tracked gaze location in such a configuration, which may be
termed a “foveated display”.
An improvement on such configurations may se a scanning fiber display with
pattern spacing that may be dynamically slaved to tracked eye gaze. For example, with a
typical scanning fiber display operating in a spiral n, as shown in Figure 18 (the st
n 510 of the image in Figure 18 illustrates a spiral motion n of a scanned multicore
fiber 514; the rightmost portion 512 of the image in Figure 18 illustrates a spiral motion pattern
of a scanned single fiber 516 for comparison), a constant pattern pitch provides for a uniform
display resolution.
In a ed display configuration, a non-uniform scanning pitch may be ed,
with r/tighter pitch (and therefore higher resolution) dynamically slaved to the detected
gaze location. For example, if the user’s gaze was detected as moving toward the edge of the
display screen, the spirals may be clustered more densely in such on, which would create
a toroid-type scanning pattern for the esolution portions, and the rest of the display being
in a lower-resolution mode. In a configuration wherein gaps may be created in the portions of
the y in a lower-resolution mode, blur could be intentionally dynamically d to
smooth out the transitions between scans, as well as between transitions from high-resolution
to lower-resolution scan pitch.
] The term lightfield may be used to describe a volumetric 3-D representation of light
traveling from an object to a viewer’s eye. However, an optical see-through display can only
reflect light to the eye, not the absence of light, and ambient light from the real world will add to
any light representing a virtual object. That is, if a virtual object presented to the eye contains a
black or very dark portion, the t light from the real world may pass through that dark
portion and obscure that it was intended to be dark.
It is nonetheless desirable to be able to present a dark virtual object over a bright
real ound, and for that dark virtual object to appear to occupy a volume at a desired
viewing distance; i.e., it is useful to create a “darkfield” representation of that dark virtual object,
in which the absence of light is perceived to be located at a particular point in space. With
regard to occlusion elements and the presentation of information to the eye of the user so that
he or she can perceive darkfield aspects of l objects, even in well lighted actual
environments, certain aspects of the aforementioned spatial light modulator, or “SLM”,
configurations are pertinent. As described above, with a light-sensing system such as the eye,
one way to get selective perception of dark field to ively ate light from such portions
of the display, because the subject y systems are about manipulation and presentation of
light; in other words, darkfield cannot be specifically projected — it’s the lack of illumination that
may be perceived as darkfield, and thus, configurations for selective attenuation of illumination
have been developed.
Referring back to the discussion of SLM configurations, one way to selectively
attenuate for a eld perception is to block all of the light coming from one angle, while
allowing light from other angles to be transmitted. This may be accomplished with a plurality of
SLM planes sing elements such as liquid crystal (which may not be the most optimal due
to its relatively low transparency when in the transmitting state), DMD elements of DLP systems
(which have relative high ission/reflection ratios when in such mode), and MEMS arrays
or shutters that are configured to controllably shutter or pass light radiation, as bed
above.
With regard to suitable liquid crystal display (“LCD”) configurations, a cholesteric
LCD array may be utilized for a controlled occlusion/blocking array. As opposed to the
conventional LCD paradigm wherein a polarization state is changed as a function of voltage,
with a teric LCD uration, a pigment is being bound to the liquid crystal molecule,
and then the molecule is physically tilted in response to an applied voltage. Such a
configuration may be designed to achieve greater transparency when in a transmissive mode
than conventional LCD, and a stack of polarizing films is not needed as it is with conventional
LCD.
In another embodiment, a plurality of layers of controllably interrupted patterns may
be utilized to controllably block selected presentation of light using moire effects. For example,
in one configuration, two arrays of ation patterns, each of which may comprise, for
example, fine-pitched sine waves printed or painted upon a transparent planar material such as
a glass substrate, may be presented to the eye of a user at a distance close enough that when
the viewer looks through either of the patterns alone, the view is essentially transparent, but if
the viewer looks through both ns lined up in sequence, the viewer will see a spatial beat
ncy moire attenuation pattern, even when the two attenuation patterns are placed in
sequence relatively close to the eye of the user.
The beat ncy is dependent upon the pitch of the patterns on the two
attenuation planes, so in one embodiment, an attenuation pattern for selectively blocking
certain light transmission for darkfield tion may be created using two sequential patterns,
each of which otherwise would be transparent to the user, but which together in series create a
l beat frequency moire attenuation pattern selected to attenuate in accordance with the
darkfield perception desired in the augmented reality .
In another ment a controlled occlusion paradigm for darkfield effect may be
created using a multi-view y style occluder. For e, one configuration may
comprise one pin-holed layer that fully occludes with the exception of small apertures or
pinholes, along with a selective attenuation layer in series, which may comprise an LCD, DLP
system, or other selective attenuation layer configuration, such as those described above. In
one scenario, with the pinhole array placed at a typical sses lens distance from the
cornea (about 30mm), and with a selective attenuation panel d opposite the pinhole array
from the eye, a perception of a sharp mechanical edge out in space may be created. In
essence, if the configuration will allow certain angles of light to pass, and others to be blocked
or occluded, than a perception of a very sharp pattern, such as a sharp edge projection, may be
created. In another related embodiment, the pinhole array layer may be replaced with a second
dynamic attenuation layer to e a somewhat similar configuration, but with more controls
than the static e array layer (the static pinhole layer could be simulated, but need not be).
In another related embodiment, the pinholes may be replaced with cylindrical
lenses. The same pattern of occlusion as in the pinhole array layer configuration may be
achieved, but with cylindrical lenses, the array is not restricted to the very tiny pinhole
geometries. To prevent the eye from being ted with distortions due to the lenses when
viewing through to the real world, a second lens array may be added on the side of the aperture
or lens array opposite of the side t the eye to compensate and provide the view-through
illumination with basically a zero power telescope configuration.
In another embodiment, rather than physically blocking light for occlusion and
creation of darkfield tion, the light may be bent or bounced, or a polarization of the light
may be changed if a liquid crystal layer is ed. For example, in one variation, each liquid
crystal layer may act as a polarization rotator such that if a ned polarizing material is
incorporated on one face of a panel, then the polarization of individual rays coming from the
real world may be selectively manipulated so they catch a portion of the patterned polarizer.
There are polarizers known in the art that have checkerboard patterns wherein half of the
“checker boxes” have vertical polarization and the other half have horizontal polarization. In
addition, if a al such as liquid crystal is used in which polarization may be selectively
manipulated, light may be selectively attenuated with this.
As described above, selective reflectors may provide greater transmission efficiency
than LCD. In one embodiment, if a lens system is placed such that it takes light coming in from
the real world and focuses a plane from the real world onto an image plane, and if a DMD (i.e.,
DLP technology) is placed at that image plane to t light when in an “on” state towards
another set of lenses that pass the light to the eye, and those lenses also have the DMD at their
focal length, the one may create an attenuation pattern that is in focus for the eye. In other
words, DMDs may be used in a selective reflector plane in a zero magnification telescope
configuration, such as is shown in Figure 19A, to controllably occlude and facilitate creating
darkfield perception.
As shown in Figure 19A, a lens (518) is taking light from the real world (144) and
focusing it down to an image plane (520); if a DMD (or other spatial attenuation device) (522) is
placed at the focal length of the lens (i.e., at the image plane 520), the lens (518) is going to
take whatever light is coming from optical infinity and focus that onto the image plane (520).
Then the spatial attenuator (522) may be utilized to selectively block out things that are to be
attenuated. Figure 19A shows the attenuator DMDs in the transmissive mode wherein they
pass the beams shown crossing the device. The image is then placed at the focal length of the
second lens (524). Preferably the two lenses (518, 524) have the same focal power so they
end up being a zero-power telescope, or a “relay”, that does not magnify views to the real world
(144). Such a configuration may be used to present unmagnified views of the world while also
allowing selective blocking/attenuation of certain pixels.
In another embodiment, as shown in Figures 198 and 19C, additional DMDs may be
added such that light reflects from each of four DMDs (526, 528, 530, 532) before g to
the eye. Figure 198 shows an ment with two lenses preferably with the same focal
power (focal length “F”) placed at a 2F relationship from one another (the focal length of the first
being conjugate to the focal length of the second) to have the zero-power telescope effect;
Figure 19C shows an embodiment without . The angles of orientation of the four
reflective panels (526, 528, 530, 532) in the depicted embodiments of Figures 198 and 19C are
shown to be around 45 degrees for simple ration purposes, but specific relative orientation
is ed (for example, a typical DMD reflect at about a 12 degree .
In another embodiment, the panels may also be ferroelectric, or may be any other
kind of reflective or selective attenuator panel or array. In one embodiment similar to those
depicted in Figures 198 and 19C, one of the three reflector arrays may be a simple mirror, such
that the other 3 are ive attenuators, thus still providing three independent planes to
controllably occlude portions of the incoming illumination in furtherance of darkfield perception.
By having multiple dynamic reflective ators in series, masks at different optical distances
relative to the real world may be created.
Alternatively, referring back to Figure 19C, one may create a uration wherein
one or more DMDs are placed in a reflective ope configuration without any lenses. Such
a uration may be driven in lightfield algorithms to selectively attenuate certain rays while
others are passed.
In another embodiment, a DMD or similar matrix of controllably movable devices
may be created upon a transparent substrate as opposed to a generally opaque substrate, for
use in a transmissive configuration such as virtual reality.
In another embodiment, two LCD panels may be ed as lightfield occluders. In
one variation, they may be thought of as attenuators due to their attenuating capability as
described above; alternatively they may be considered polarization rotators with a shared
polarizer stack. Suitable LCDs may comprise components such as blue phase liquid crystal,
cholesteric liquid crystal, lectric liquid l, and/or twisted nematic liquid crystal.
One embodiment may comprise an array of directionally-selective occlusion
elements, such as a MEMS device featuring a set of louvers that can change rotation such that
they pass the majority of light that is coming from a particular angle, but are presenting more of
a broad face to light that is coming from a different angle (somewhat akin to the manner in
which plantation shutters may be utilized with a typical human scale window). The
MEMS/louvers configuration may be placed upon an optically transparent substrate, with the
louvers substantially opaque. Ideally such a configuration would have a louver pitch fine
enough to selectably occlude light on a by-pixel basis. In another embodiment, two or
more layers or stacks of louvers may be combined to provide yet further controls. In another
ment, rather than selectively blocking light, the louvers may be polarizers configured to
change the polarization state of light on a controllably variable basis.
As described above, another embodiment for selective ion may se an
array of sliding panels in a MEMS device such that the sliding panels may be controllably
opened (i.e., by sliding in a planar fashion from a first position to a second position; or by
rotating from a first orientation to a second orientation; or, for example, combined onal
reorientation and displacement) to transmit light h a small frame or aperture, and
controllably closed to occlude the frame or re and prevent transmission. The array may
be configured to open or occlude the various frames or apertures such that they maximally
ate the rays that are to be attenuated, and only minimally attenuate the rays to be
transmitted.
In an embodiment wherein a fixed number of sliding panels can either occupy a first
position occluding a first aperture and opening a second aperture, or a second position
occluding the second aperture and opening the first aperture, there will always be the same
amount of light transmitted overall (because 50% of the apertures are occluded, and the other
50% are open, with such a configuration), but the local position s of the shutters or
doors may create targeted moire or other effects for darkfield perception with the dynamic
positioning of the various sliding panels. In one embodiment, the sliding panels may comprise
sliding polarizers, and if placed in a stacked configuration with other zing elements that
are either static or dynamic, may be utilized to selectively attenuate.
Referring to Figure 19D, another uration providing an opportunity for selective
reflection, such as via a DMD style reflector array (534), is shown, such that a stacked set of
two waveguides (536, 538) along with a pair of focus ts (540, 542) and a reflector (534;
such as a DMD) may be used to e a portion of incoming light with an entrance reflector
(544). The reflected light may be totally internally reflected down the length of the first
waveguide (536), into a focusing t (540) to bring the light into focus on a reflector (534)
such as a DMD array, after which the DMD may selectively attenuate and t a portion of
the light back through a focusing lens (542; the lens configured to facilitate injection of the light
back into the second waveguide) and into the second ide (538) for total internal
reflection down to an exit reflector (546) configured to exit the light out of the waveguide and
toward the eye (58).
Such a configuration may have a relatively thin shape factor, and is designed to
allow light from the real world (144) to be selectively attenuated. As waveguides work most
cleanly with collimated light, such a configuration may be well suited for virtual y
urations wherein focal lengths are in the range of optical infinity. For closer focal lengths,
a lightfield y may be used as a layer on top of the silhouette d by the
aforementioned selective attenuation / darkfield configuration to provide other cues to the eye
of the user that light is coming from another focal ce. An occlusion mask may be out of
focus, even nondesirably so, and then in one embodiment, a lightfield on top of the masking
layer may be used to hide the fact that the darkfield may be at the wrong focal distance.
Referring to Figure 19E, an embodiment is shown featuring two waveguides (552,
554) each having two angled reflectors (558, 544; 556, 546) for illustrative purposes shown at
imately 45 degrees; in actual configurations the angle may differ depending upon the
reflective surface, reflective/refractive properties of the waveguides, etc.) directing a portion of
light incoming from the real world down each side of a first waveguide (or down two separate
waveguides if the top layer is not monolithic) such that it hits a reflector (548, 550) at each end,
such as a DMD which may be used for selective attenuation, after which the reflected light may
be injected back into the second waveguide (or into two te ides if the bottom
layer is not monolithic) and back toward two angled reflectors (again, they need not be at 45
degrees as shown) for exit out toward the eye (58).
ng lenses may also be placed n the reflectors at each end and the
waveguides. In another embodiment the reflectors (548, 550) at each end may se
standard mirrors (such as alumized mirrors). Further, the reflectors may be wavelength
selective reflectors, such as dichroic mirrors or film interference filters. Further, the tors
may be diffractive elements configured to reflect incoming light.
Figure 19F illustrates a configuration wherein four reflective surfaces in a pyramid
type configuration are utilized to direct light through two waveguides (560, 562), in which
incoming light from the real world may be divided up and reflected to four difference axes. The
pyramid-shaped reflector (564) may have more than fourfacets, and may be resident within the
substrate prism, as with the reflectors of the configuration of Figure 19E. The configuration of
Figure 19F is an extension of that of Figure 19E.
Referring to Figure 19G, a single ide (566) may be utilized to capture light
from the world (144) with one or more reflective surfaces (574, 576, 578, 580, 582), relay it
(570) to a selective attenuator (568; such as a DMD array), and recouple it back into the same
waveguide so that it ates (572) and encounters one or more other reflective surfaces
(584, 586, 588, 590, 592) that cause it to at least lly exit (594) the waveguide on a path
toward the eye (58) of the user. Preferably the waveguide comprises selective reflectors such
that one group (574, 576, 578, 580, 582) may be switched on to capture incoming light and
direct it down to the selective attenuator, while separate another group (584, 586, 588, 590,
592) may be switched on to exit light returning from the selective ator out toward the eye
(58).
For city the selective attenuator is shown oriented ntially perpendicularly
to the ide; in other embodiments, various optics ents, such as refractive or
reflective optics, may be utilized to have the selective attenuator at a different and more
compact orientation relative to the waveguide.
Referring to Figure 19H, a variation on the configuration described in reference to
Figure 19D is illustrated. This configuration is somewhat analogous to that discussed above in
reference to Figure 5B, wherein a switchable array of reflectors may be embedded within each
of a pair of waveguides (602, 604). Referring to Figure 19H, a controller may be ured to
turn the reflectors (598, 600) on and off in sequence, such that multiple reflectors may be
operated on a frame sequential basis; then the DMD or other ive attenuator (594) may
also be sequentially driven in sync with the different mirrors being turned on and off.
Referring to Figure 19l, a pair of wedge-shaped waveguides similar to those
described above (for example, in reference to Figures 15A—15C) are shown in side or sectional
view to illustrate that the two long surfaces of each wedge-shaped waveguide (610, 612) are
not co-planar. A “turning film” (606, 608; such as that available from 3M ation under the
trade name, “TRAF”, which in essence comprises a microprism array), may be utilized on one
or more surfaces of the wedge-shaped waveguides to either turn incoming rays at an angle so
that they will be ed by total internal reflection, or to turn outgoing rays as they are exiting
the waveguide toward an eye or other target. Incoming rays are directed down the first wedge
and toward the selective attenuator (614) such as a DMD, LCD (such as a lectric LCD), or
an LCD stack to act as a mask).
After the selective attenuator (614), reflected light is coupled back into the second
wedge-shaped waveguide which then relays the light by total internal reflection along the
wedge. The properties of the wedge-shaped ide are intentionally such that each
bounce of light causes an angle change; the point at which the angle has changed enough to
be the critical angle to escape total internal reflection becomes the exit point from the wedge-
shaped waveguide. Typically the exit will be at an oblique angle, so another layer of turning
film may be used to “turn” the exiting light toward a targeted object such as the eye (58).
Referring to Figure 19J, l arcuate lenslet arrays (616, 620, 622) are
positioned relative to an eye and ured such that a spatial attenuator array (618) is
positioned at a focal/image plane so that it may be in focus with the eye (58). The first (616)
and second (620) arrays are configured such that in the aggregate, light passing from the real
world to the eye is essentially passed through a zero power telescope. The embodiment of
Figure 19J shows a third array (622) of lenslets which may be utilized for ed optical
compensation, but the general case does not require such a third layer. As discussed above,
having telescopic lenses that are the diameter of the viewing optic may create an undesirably
large form factor (somewhat akin to having a bunch of small sets of binoculars in front of the
eyes).
One way to ze the overall geometry is to reduce the diameter of the lenses by
splitting them out into smaller lenslets, as shown in Figure 19J (i.e., an array of lenses rather
than one single large lens). The lenslet arrays (616, 620, 622) are shown wrapped radially or
arcuately around the eye (58) to ensure that beams incoming to the pupil are d through
the appropriate lenslets (else the system may suffer from optical problems such as sion,
ng, and/or lack of focus). Thus all of the lenslets are oriented “toed in” and pointed at the
pupil of the eye (58), and the system facilitates avoidance of scenarios wherein rays are
propagated through unintended sets of lenses en route to the pupil.
Referring to Figures 19K-19N, various software approaches may be utilized to assist
in the tation of darkfield in a virtual or augmented reality displace scenario. Referring to
Figure 19K, a l nging scenario for augmented reality is depicted (632), with a
textured carpet (624) and non-uniform background architectural features (626), both of which
are lightly-colored. The black box (628) depicted indicates the region of the y in which
one or more augmented reality features are to be presented to the user for three-dimensional
perception, and in the black box a robot creature (630) is being presented that may, for
example, be part of an augmented reality game in which the user is d. In the depicted
example, the robot ter (630) is darkly-colored, which makes for a challenging
presentation in three-dimensional perception, particularly with the background selected for this
example scenario.
As discussed briefly above, one of the main challenges for a presenting darkfield
augmented reality object is that the system generally cannot add or paint in “darkness”;
generally the display is configured to add light. Thus, referring to Figure 19L, without any
specialized software treatments to enhance eld perception, tation of the robot
character in the augmented reality view results in a scene wherein portions of the robot
character that are to be essentially flat black in presentation are not visible, and portions of the
robot character that are to have some lighting (such as the lightly-pigmented cover of the
shoulder gun of the robot character) are only barely e (634) — they appear almost like a
light grayscale disruption to the otherwise normal background image.
Referring to Figure 19M, using a software-based global attenuation treatment (akin
to digitally g on a pair of sunglasses) provides enhanced visibility to the robot character
because the brightness of the nearly black robot character is effectively increased relative to
the rest of the space, which now s more dark (640). Also shown in Figure 19M is a
lly-added light halo (636) which may be added to enhance and distinguish the now-more-
visible robot ter shapes (638) from the background. With the halo treatment, even the
portions of the robot character that are to be presented as flat black become visible with the
contrast to the white halo, or “aura” presented around the robot character.
Preferably the halo may be presented to the user with a perceived focal distance
that is behind the focal distance of the robot character in three-dimensional space. In a
configuration wherein single panel occlusion techniques such as those described above is
being utilized to present darkfield, the light halo may be ted with an intensity gradient to
match the dark halo that may accompany the occlusion, minimizing the visibility of either
darkfield effect. Further, the halo may be presented with blurring to the background behind the
presented halo illumination for r distinguishing effect. A more subtle aura or halo effect
may be created by ng, at least in part, the color and/or brightness of a relatively light-
colored background.
Referring to Figure 19N, some or all of the black intonations of the robot character
may be changed to dark, cool blue colors to provide a r distinguishing effect relative to the
background, and relatively good visualization of the robot (642).
Wedge-shaped waveguides have been bed above, such as in reference to
Figures 15A—15D and Figure 19l. With a wedge-shaped waveguide, every time a ray s
off of one of the non-coplanar surfaces, it gets an angle change, which ultimately results in the
ray exiting total al reflection when its approach angle to one of the surfaces goes past the
critical angle. Turning films may be used to redirect exiting light so that exiting beams leave
with a trajectory that is more or less perpendicular to the exit surface, depending upon the
geometric and ergonomic issues at play.
With a series or array of displays injecting image information into a wedge-shaped
waveguide, as shown in Figure 15C, for example, the wedge-shaped waveguide may be
configured to create a itched array of angle-biased rays ng from the wedge.
Somewhat similarly, it has been discussed above that a lightfield display, or a variable
wavefront creating waveguide, both may produce a multiplicity of beamlets or beams to
represent a single pixel in space such that wherever the eye is positioned, the eye is hit by a
ity of different beamlets or beams that are unique to that particular eye position in front of
the y panel.
As was further discussed above in the context of ield displays, a plurality of
viewing zones may be d within a given pupil, and each may be used for a different focal
distance, with the aggregate producing a perception similar to that of a variable wavefront
creating waveguide, or similar to the actual optical physics of reality of the objects viewed were
real. Thus a wedge-shaped waveguide with multiple displays may be utilized to generate a
lightfield. In an embodiment similar to that of Figure 15C with a linear array of displays injecting
image information, a fan of exiting rays is created for each pixel. This concept may be
extended in an ment wherein le linear arrays are stacked to all inject image
information into the wedge-shaped waveguide (in one ion, one array may inject at one
angle relative to the wedge-shaped waveguide face, while the second array may inject at a
second angle ve to the wedge-shaped ide face), in which case exit beams fan out
at two different axes from the wedge.
Thus such a configuration may be utilized to produce pluralities of beams spraying
out at lots of different angles, and each beam may be driven separately due to the fact that
under such configuration, each beam is driven using a separate display. In another
embodiment, one or more arrays or displays may be configured to inject image information into
wedge-shaped waveguide through sides or faces of the wedge-shaped waveguide other than
that shown in Figure 15C, such as by using a diffractive optic to bend injected image
information into total an internal reflection configuration ve to the wedge-shaped
waveguide.
Various tors or reflecting surfaces may also be utilized in t with such a
wedge-shaped waveguide embodiment to outcouple and manage light from the wedge-shaped
waveguide. In one embodiment, an entrance aperture to a wedge-shaped waveguide, or
injection of image information through a different face other than shown in Figure 15C, may be
utilized to tate staggering (geometric and/or temporal) of different displays and arrays such
that a Z-axis delta may also be ped as a means for injecting three-dimensional
information into the shaped waveguide. For a greater than three-dimensions array
configuration, various displays may be configured to enter a shaped waveguide at
multiple edges in multiple stacks with staggering to get higher dimensional configurations.
Referring to Figure 20A, a uration similar to that depicted in Figure 8H is
shown wherein a waveguide (646) has a diffractive optical t (648; or “DOE”, as noted
above) sandwiched in the middle (alternatively, as described above, the diffractive optical
element may reside on the front or back face of the depicted waveguide). A ray may enter the
waveguide (646) from the tor or display (644). Once in the waveguide (646), each time
the ray intersects the DOE (648), part of it is exited out of the waveguide (646). As described
above, the DOE may be designed such that the exit illuminance across the length of the
waveguide (646) is somewhat uniform (for example, the first such DOE intersection may be
configured to exit about 10% of the light; then the second DOE intersection may be configured
to exit about 10% of the remaining light so that 81% is passed on, and so on; in r
embodied a DOE may be designed to have a variable diffraction efficiency, such as linearly-
decreasing ction efficiency, along its length to map out a more uniform exit illuminance
across the length of the waveguide).
To further bute ing light that reaches an end (and in one ment to
allow for selection of a relatively low diffraction efficiency DOE which would be favorable from a
o-the-world transparency perspective), a reflective element (650) at one or both ends may
be included. Further, referring to the embodiment of Figure 20B, additional distribution and
preservation may be achieved by including an elongate reflector (652) across the length of the
waveguide as shown (comprising, for example, a thin film dichroic coating that is wavelength-
selective); preferably such reflector would be blocking light that accidentally is reflected upward
(back toward the real world 144 for exit in a way that it would not be utilized by the viewer). In
some embodiments, such an elongate reflector may contribute to a “ghosting” effect perception
by the user.
In one embodiment, this ghosting effect may be eliminated by having a dual-
waveguide (646, 654) circulating reflection configuration, such as that shown in Figure 20C,
which is designed to keep the light moving around until it has been exited toward the eye (58) in
a preferably substantially equally distributed manner across the length of the waveguide
assembly. Referring to Figure 20C, light may be injected with a projector or display (644), and
as it travels across the DOE (656) of the first waveguide (654), it ejects a preferably
substantially m pattern of light out toward the eye (58); light that remains in the first
waveguide is reflected by a first reflector assembly (660) into the second waveguide (646). In
one embodiment, the second ide (646) may be configured to not have a DOE, such that
it merely transports or recycles the remaining light back to the first waveguide, using the second
reflector ly.
In another embodiment (as shown in Figure 20C) the second waveguide (646) may
also have a DOE (648) configured to uniformly eject fractions of travelling light to provide a
second plane of focus for three-dimensional perception. Unlike the urations of s
20A and 208, the configuration of Figure 20C is designed for light to travel the waveguide in
one direction, which avoids the aforementioned ghosting problem that is related to passing light
backwards through a waveguide with a DOE. Referring to Figure 20D, rather than having a
mirror or box style reflector ly (660) at the ends of a waveguide for recycling the light,
an array of smaller retroreflectors (662), or a eflective material, may be ed.
Referring to Figure 20E, an embodiment is shown that utilizes some of the light
recycling configurations of the embodiment of Figure 20C to “snake” the light down through a
waveguide (646) having a sandwiched DOE (648) after it has been injected with a display or
tor (644) so that it crosses the waveguide (646) many times back and forth before
reaching the , at which point it may be recycled back up to the top level for further
recycling. Such a configuration not only recycles the light and facilitates use of relatively low
diffraction efficiency DOE elements for exiting light toward the eye (58), but also distributes the
light, to provide for a large exit pupil configuration akin to that described in reference to Figure
Referring to Figure 20F, an illustrative configuration similar to that of Figure 5A is
shown, with incoming light injected along a conventional prism or beamsplitter substrate (104)
to a reflector (102) without total internal reflection (i.e., without the prism being considered a
waveguide) because the input projection (106), scanning or otherwise, is kept within the bounds
of the prism — which means that the geometry of such prism becomes a significant constraint.
In another embodiment, a waveguide may be utilized in place of the simple prism of Figure 20F,
which facilitates the use of total internal reflection to provide more geometric flexibility.
Other configurations describe above are configured to profit from the inclusion of
waveguides for similar manipulations and light. For example, referring back to Figure 7A, the
general concept illustrated therein is that a collimated image injected into a ide may be
refocused before er out toward an eye, in a uration also designed to facilitate
viewing light from the real world. In place of the refractive lens shown in Figure 7A, a diffractive
optical element may be used as a variable focus element.
Referring back to Figure 7B, r waveguide configuration is illustrated in the
context of having multiple layers stacked upon each other with controllable access toggling
between a smaller path (total internal reflection through a ide) and a larger path (total
internal reflection through a hybrid waveguide sing the original waveguide and a liquid
crystal isolated region with the liquid crystal switched to a mode wherein the refractive indices
are substantially matched between the main waveguide and the auxiliary waveguide), so that
the controller can tune on a frame-by-frame basis which path is being taken. High-speed
switching electro-active materials, such as lithium niobate, facilitate path changes with such a
configuration at gigahertz rates, which allows one to change the path of light on a pixel-by-pixel
basis.
Referring back to Figure 8A, a stack of waveguides paired with weak lenses is
illustrated to demonstrate a multifocal uration wherein the lens and waveguide elements
may be static. Each pair of waveguide and lens may be functionally replaced with waveguide
having an embedded DOE element (which may be static, in a closer analogy to the
configuration of Figure 8A, or dynamic), such as that described in reference to Figure 8|.
Referring to Figure 20G, if a transparent prism or block (104; Le, not a waveguide)
is utilized to hold a mirror or reflector (102) in a periscope type of configuration to receive light
from other components, such as a lens (662) and projector or y (644), the field of view is
limited by the size of that reflector (102; the bigger the reflector, the wider the field of view).
Thus to have a larger field of view with such uration, a thicker substrate may be needed
to hold a larger reflector; otherwise, the functionality of an ated plurality of reflectors
may be utilized to increase the functional field of view, as described in reference to Figures 80,
8P, and 8Q. Referring to Figure 20H, a stack (664) of planar waveguides (666), each fed with a
display or tor (644; or in another embodiment a lexing of a single display) and
having an exit reflector (668), may be utilized to aggregate toward the function of a larger single
reflector. The exit tors may be at the same angle in some cases, or not the same angle in
other cases, depending upon the positioning of the eye (58) relative to the assembly.
Figure 20l rates a d configuration, n the reflectors (680, 682, 684,
686, 688) in each of the planar ides (670, 672, 674, 676, 678) have been offset from
each other, and wherein each takes in light from a projector or display (644) which may be sent
through a lens (690) to ultimately contribute exiting light to the pupil (45) of the eye (58) by
virtue of the reflectors (680, 682, 684, 686, 688) in each of the planar waveguides (670, 672,
674, 676, 678). If one can create a total range of all of the angles that would be expected to be
seen in the scene (i.e., preferably without blind spots in the key field of view), then a useful field
of view has been achieved. As described above, the eye (58) functions based at least on what
angle light rays enter the eye, and this can be simulated. The rays need not pass through the
exact same point in space at the pupil — rather the light rays just need to get h the pupil
and be sensed by the retina. Figure 20K illustrates a variation wherein the shaded portion of
the optical assembly may be utilized as a compensating lens to functionally pass light from the
real world (144) through the assembly as though it has been passed through a zero power
ope.
Referring to Figure 20J, each of the aforementioned rays may also be a relative
wide beam that is being reflected through the pertinent waveguide (670, 672) by total internal
reflection. The reflector (680, 682) facet size will determine what the g beam width can
Referring to Figure 20L, a further discretization of the reflector is shown, wherein a
plurality of small straight angular reflectors may form a y parabolic reflecting surface
(694) in the ate h a waveguide or stack thereof (696). Light coming in from the
displays (644; or single MUXed display, for example), such as through a lens (690), is all
directed toward the same shared focal point at the pupil (45) of the eye (58).
Referring back to Figure 13M, a linear array of displays (378) injects light into a
shared waveguide (376). In another embodiment a single display may be multiplexed to a
series of entry lenses to provide similar functionality as the embodiment of Figure 13M, with the
entry lenses creating parallel paths of rays running through the waveguide.
In a conventional waveguide approach wherein total internal reflection is relied upon
for light propagation, the field of view is restricted because there is only a n angular range
of rays ating through the waveguide (others may escape out). In one embodiment, if a
red/green/blue (or “RGB”) laserline reflector is placed at one or both ends of the planar
surfaces, akin to a thin film interference filter that is highly reflective for only n
wavelengths and poorly reflective for other wavelengths, than one can functionally increase the
range of angles of light propagation. Windows (without the coating) may be provided for
allowing light to exit in predetermined locations. r, the coating may be selected to have a
directional selectivity hat like reflective elements that are only highly reflective for
certain angles of incidence). Such a coating may be most nt for the larger planes/sides
of a waveguide.
Referring back to Figure 13E, a variation on a scanning fiber display was discussed,
which may be deemed a scanning thin waveguide configuration, such that a plurality of very
thin planar waveguides (358) may be oscillated or ed such that if a variety of injected
beams is coming through with total internal reflection, the configuration functionally would
provide a linear array of beams escaping out of the edges of the vibrating elements (358). The
depicted configuration has approximately five externally-projecting planar waveguide portions
(358) in a host medium or substrate (356) that is transparent, but which preferably has a
different tive index so that the light will stay in total internal reflection within each of the
substrate-bound smaller waveguides that tely feed (in the depicted embodiment there is a
90 degree turn in each path at which point a planar, curved, or other reflector may be utilized to
bounce the light outward) the externally-projecting planar waveguide portions (358).
The externally-projecting planar waveguide portions (358) may be vibrated
individually, or as a group along with oscillatory motion of the substrate (356). Such scanning
motion may provide ntal scanning, and for vertical ng, the input (360) aspect of the
assembly (i.e., such as one or more scanning fiber displays scanning in the vertical axis) may
be utilized. Thus a variation of the scanning fiber display is presented.
] Referring back to Figure 13H, a ide (370) may be utilized to create a
lightfield. With ides working best with collimated beams that may be associated with
optical infinity from a perception perspective, all beams staying in focus may cause tion
discomfort (i.e., the eye will not make a discernible difference in dioptric blur as a function of
accommodation; in other words, the narrow diameter, such as 0.5mm or less, collimated
ts may open loop the eye’s accommodation/vergence , causing discomfort).
In one embodiment, a single beam may be fed in with a number of cone beamlets
coming out, but if the introduction vector of the entering beam is changed (i.e., laterally shift the
beam injection location for the projector/display relative to the waveguide), one may control
where the beam exits from the waveguide as it is directed toward the eye. Thus one may use a
waveguide to create a lightfield by creating a bunch of narrow er collimated beams, and
such a configuration is not reliant upon a true variation in a light wavefront to be associated with
the desired perception at the eye.
If a set of angularly and laterally diverse beamlets is injected into a ide (for
example, by using a multicore fiber and driving each core tely; another configuration may
utilize a plurality of fiber scanners coming from different angles; another configuration may
utilize a high-resolution panel display with a lenslet array on top of it), a number of exiting
beamlets can be created at ent exit angles and exit locations. Since the waveguide may
scramble the lightfield, the decoding is preferably predetermined.
Referring to s 20M and 20N, a waveguide (646) assembly (696) is shown that
comprises stacked waveguide components in the vertical or horizontal axis. Rather than
having one monolithic planar waveguide, the notion with these embodiments is to stack a
plurality of smaller ides (646) immediately adjacent each other such that light uced
into one waveguide, in addition to propagating down (i.e., propagating along a Z axis with total
internal tion in +X,-X) such waveguide by total internal reflection, also totally internally
reflects in the perpendicular axis (+y, -Y) as well, such that it is not spilling into other areas. In
other words, if total al reflection is from left to right and back during Z axis propagation,
the configuration will be set up to totally internally reflect any light that hits the top or bottom
sides as well; each layer may be driven separately without interference from other layers.
Each waveguide may have a DOE (648) embedded and configured to eject out light with a
predetermined distribution along the length of the waveguide, as bed above, with a
predetermined focal length configuration (shown in Figure 20M as ranging from 0.5 meters to
optical infinity).
In r variation, a very dense stack of waveguides with embedded DOEs may
be produced such that it spans the size of the anatomical pupil of the eye (i.e., such that
multiple layers 698 of the composite waveguide are ed to cross the exit pupil, as
illustrated in Figure 20N). With such a configuration, one may feed a collimated image for one
wavelength, and then the portion located the next millimeter down producing a diverging
wavefront that represents an object coming from a focal ce of, say, 15 meters away, and
so on, with the notion being that an exit pupil is coming from a number of different waveguides
as a result of the DOEs and total internal tion through the waveguides and across the
DOEs. Thus rather than creating one uniform exit pupil, such a configuration creates a plurality
of stripes that, in the aggregate, facilitate the perception of different focal depths with the
eye/brain.
Such a concept may be extended to configurations comprising a waveguide with a
switchable/controllable embedded DOE (i.e. that is switchable to different focal distances), such
as those described in relation to Figures 8B-8N, which allows more efficient light trapping in the
axis across each ide. Multiple ys may be coupled into each of the layers, and
each waveguide with DOE would emit rays along its own length. In another embodiment,
rather than relying on total internal reflection, a laserline reflector may be used to se
angular range. In between layers of the composite waveguide, a completely reflective
metallized coating may be utilized, such as aluminum, to ensure total reflection, or alternatively
dichroic style or narrow band reflectors may be utilized.
Referring to Figure 200, the whole composite waveguide assembly (696) maybe be
curved ely toward the eye (58) such that each of the individual waveguides is ed
toward the pupil. In other words, the configuration may be designed to more efficiently direct
the light toward the location where the pupil is likely to be present. Such a configuration also
may be utilized to increase the field of view.
As was discussed above in relation to Figures 8L, 8M, and 8N, a changeable
diffraction configuration allows for scanning in one axis, somewhat akin to a scanning light
display. Figure 21A illustrates a ide (698) having an embedded (i.e., sandwiched
within) DOE (700) with a linear g term that may be changed to alter the exit angle of
exiting light (702) from the waveguide, as shown. A high-frequency switching DOE material
such as lithium niobate may be utilized. In one embodiment, such a scanning configuration
may be used as the sole mechanism for scanning a beam in one axis; in another ment,
the scanning configuration may be combined with other scanning axes, and may be used to
create a larger field of view (i.e., if a normal field of view is 40 degrees, and by changing the
linear diffraction pitch one can steer over another 40 degrees, the effective usable field of view
for the system is 80 degrees).
Referring to Figure 21B, in a conventional uration, a ide (708) may be
placed perpendicular to a panel display (704), such as an LCD or OLED panel, such that
beams may be injected from the waveguide (708), through a lens (706), and into the panel
(704) in a scanning configuration to provide a viewable display for television or other purposes.
Thus the waveguide may be utilized in such configuration as a scanning image source, in
contrast to the configurations described in reference to Figure 21A, wherein a single beam of
light may be manipulated by a scanning fiber or other element to sweep through different
angular locations, and in addition, another direction may be scanned using the high-frequency
diffractive optical element.
In another embodiment, a uniaxial scanning fiber display (say scanning the fast line
scan, as the scanning fiber is relatively high frequency) may be used to inject the fast line scan
into the waveguide, and then the relatively slow DOE ing (i.e., in the range of 100 Hz)
may be used to scan lines in the other axis to form an image.
In another embodiment, a DOE with a g of fixed pitch may be combined with
an nt layer of o-active material having a c refractive index (such as liquid
crystal), so that light may be redirected into the grating at different angles. This is an
application of the basic multipath configuration described above in reference to Figure 7B, in
which an o-active layer comprising an electro-active material such as liquid crystal or
lithium niobate may change its refractive index such that it changes the angle at which a ray
emerges from the ide. A linear diffraction grating may be added to the configuration of
Figure 7B (in one embodiment, sandwiched within the glass or other material comprising the
larger lower waveguide) such that the diffraction g may remain at a fixed pitch, but the
light is biased before it hits the grating.
Figure 21C shows another embodiment featuring two wedge-like waveguide
elements (710, 712), wherein one or more of them may be electro-active so that the related
refractive index may be changed. The elements may be configured such that when the wedges
have matching refractive indices, the light totally internally reflects through the pair (which in the
aggregate performs akin to a planar waveguide with both wedges matching) while the wedge
interfaces have no effect. Then if one of the refractive indices is d to create a mismatch,
a beam deflection at the wedge interface (714) is caused, and there is total internal reflection
from that e back into the associated wedge. Then a llable DOE (716) with a linear
grating may be coupled along one of the long edges of the wedge to allow light to exit out and
reach the eye at a desirable exit angle.
In r embodiment, a DOE such as a Bragg grating, may be configured to
change pitch versus time, such as by a mechanical stretching of the g (for example, if the
grating resides on or comprises an elastic material), a moire beat pattern between two gratings
on two different planes (the gratings may be the same or different pitches), Z-axis motion (i.e.,
closer to the eye, or farther away from the eye) of the grating, which functionally is similar in
effect to hing of the grating, or electro-active gratings that may be switched on or off, such
as one created using a polymer dispersed liquid crystal approach wherein liquid l droplets
may be controllably activated to change the refractive index to become an active grating,
versus turning the voltage off and allowing a switch back to a refractive index that matches that
of the host medium.
In another embodiment, a time-varying grating may be utilized for field of view
expansion by ng a tiled display configuration. Further, a arying grating may be
utilized to address chromatic aberration (failure to focus all colors/wavelengths at the same
focal point). One ty of diffraction gratings is that they will deflect a beam as a function of
its angle of nce and wavelength (Le, a DOE will t different wavelengths by different
angles: somewhat akin to the manner in which a simple prism will divide out a beam into its
wavelength components).
One may use time-varying g control to compensate for chromatic aberration in
addition to field of view expansion. Thus, for example, in a ide with embedded DOE
type of configuration as described above, the DOE may be configured to drive the red
wavelength to a slightly different place than the green and blue to s unwanted chromatic
aberration. The DOE may be time-varied by having a stack of elements that switch on and off
(i.e. to get red, green, and blue to be diffracted outbound similarly).
In another embodiment, a time-varying grating may be utilized for exit pupil
expansion. For example, ing to Figure 21D, it is possible that a waveguide (718) with
embedded DOE (720) may be positioned relative to a target pupil such that none of the beams
exiting in a baseline mode actually enter the target pupil (45) — such that the pertinent pixel
would be missed by the user. A time-varying configuration may be utilized to fill in the gaps in
the outbound exit n by shifting the exit pattern laterally (shown in dashed/dotted lines) to
effectively scan each of the 5 exiting beams to better ensure that one of them hits the pupil of
the eye. In other words, the functional exit pupil of the display system is expanded.
In r embodiment, a time-varying grating may be utilized with a waveguide for
one, two, or three axis light scanning. In a manner akin to that described in reference to Figure
21A, one may use a term in a grating that is scanning a beam in the vertical axis, as well as a
grating that is ng in the horizontal axis. Further, if radial elements of a grating are
incorporated, as is sed above in relation to Figures SB-8N, one may have scanning of the
beam in the Z axis (i.e., toward/away from the eye), all of which may be time sequential
scanning.
hstanding the discussions herein regarding specialized treatments and uses
of DOEs lly in connection with waveguides, many of these uses of DOE are usable
r or not the DOE is embedded in a waveguide. For example, the output of a ide
may be separately manipulated using a DOE; or a beam may be manipulated by a DOE before
it is injected into a waveguide; further, one or more DOEs, such as a time-varying DOE, may
be utilized as an input for freeform optics configurations, as discussed below.
As discussed above in reference to Figures SB-8N, an element of a DOE may have
a circularly-symmetric term, which may be summed with a linear term to create a controlled exit
pattern (i.e., as described above, the same DOE that outcouples light may also focus it). In
another embodiment, the circular term of the DOE diffraction grating may be varied such that
the focus of the beams representing those pertinent pixels is modulated. In on, one
configuration may have a second/separate circular DOE, obviating the need to have a linear
term in the DOE.
Referring to Figure 21E, one may have a waveguide (722) outputting collimated
light with no DOE element embedded, and a second waveguide that has a circularly-symmetric
DOE that can be switched between multiple configurations — in one embodiment by having a
stack (724) of such DOE elements (Figure 21 F shows another configuration wherein a
functional stack 728 of DOE elements may comprise a stack of polymer dispersed liquid l
elements 726, as bed above, wherein without a voltage applied, a host medium refraction
index matches that of a sed molecules of liquid crystal; in another ment,
molecules of lithium niobate may be dispersed for faster se times; with voltage applied,
such as through arent indium tin oxide layers on either side of the host medium, the
dispersed molecules change index of refraction and functionally form a diffraction pattern within
the host medium) that can be switched on/off.
In r embodiment, a circular DOE may be layered in front of a waveguide for
focus modulation. Referring to Figure 21G, the waveguide (722) is outputting collimated light,
which will be ved as associated with a focal depth of optical infinity unless otherwise
modified. The collimated light from the waveguide may be input into a diffractive optical
element (730) which may be used for dynamic focus modulation (i.e., one may switch on and
off different circular DOE patterns to impart various different focuses to the exiting light). In a
related embodiment, a static DOE may be used to focus collimated light g from a
waveguide to a single depth of focus that may be useful for a particular user application.
In another embodiment, multiple stacked circular DOEs may be used for additive
power and many focus levels — from a vely small number of switchable DOE layers. In
other words, three different DOE layers may be switched on in various combinations relative to
each other; the optical powers of the DOEs that are switched on may be added. In one
embodiment wherein a range of up to 4 diopters is desired, for e, a first DOE may be
configured to provide half of the total diopter range desired (in this example, 2 diopters of
change in focus); a second DOE may be configured to induce a 1 diopter change in focus;
then a third DOE may be configured to induce a 1/2 diopter change in focus. These three
DOEs may be mixed and matched to provide 1/2, 1, 1.5, 2, 2.5, 3, and 3.5 rs of change in
focus. Thus a super large number of DOEs would not be required to get a relatively broad
range of control.
In one embodiment, a matrix of switchable DOE elements may be utilized for
ng, field of view expansion, and/or exit pupil expansion. Generally in the above
discussions of DOEs, it has been assume that a typical DOE is either all on or all off. In one
variation, a DOE (732) may be subdivided into a plurality of functional subsections (such as the
one labeled as t 734 in Figure 21H), each of which preferably is uniquely controllable to
be on or off (for example, referring to Figure 21 H, each subsection may be operated by its own
set of indium tin oxide, or other control lead material, voltage application leads 736 back to a
central controller). Given this level of control over a DOE paradigm, additional configurations
are facilitated.
Referring to Figure 21l, a waveguide (738) with ed DOE (740) is viewed
from the top down, with the user’s eye oned in front of the waveguide. A given pixel may
be represented as a beam coming into the waveguide and totally internally reflecting along until
it may be exited by a diffraction pattern to come out of the waveguide as a set of beams.
Depending upon the diffraction configuration, the beams may come out parallel/collimated (as
shown in Figure 21| for convenience), or in a diverging fan configuration if representing a focal
distance closer than optical infinity.
The depicted set of parallel exiting beams may represent, for example, the st
left pixel of what the user is seeing in the real world as viewed h the waveguide, and light
off to the rightmost extreme will be a different group of el exiting beams. , with
modular l of the DOE subsections as described above, one may spend more computing
resource or time creating and manipulating the small subset of beams that is likely to be
actively addressing the user’s pupil (i.e., because the other beams never reach the user’s eye
and are effectively wasted). Thus, referring to Figure 21J, a waveguide (738) configuration is
shown wherein only the two tions (740, 742) of the DOE (744) are deemed to be likely to
address the user’s pupil (45) are activated. Preferably one subsection may be configured to
direct light in one direction simultaneously as another subsection is directing light in a different
direction.
Figure 21 K shows an orthogonal view of two independently controlled subsections
(734, 746) of a DOE (732). Referring to the top view of Figure 21L, such independent control
may be used for scanning or focusing light. In the configuration depicted in Figure 21K, an
assembly (748) of three independently controlled veguide subsections (750, 752, 754)
may be used to scan, increase the field of view, and/or increase the exit pupil region. Such
functionality may arise from a single waveguide with such independently controllable DOE
subsections, or a vertical stack of these for additional complexity.
In one embodiment, if a circular DOE may be controllably stretched radially-
symmetrically, the ction pitch may be ted, and the DOE may be utilized as a
tunable lens with an analog type of control. In another embodiment, a single axis of stretch (for
example, to adjust an angle of a linear DOE term) may be utilized for DOE control. Further, in
another embodiment a membrane, akin to a drum head, may be vibrated, with oscillatory
motion in the Z-axis (i.e., toward/away from the eye) providing Z-axis control and focus change
over time.
] Referring to Figure 21 M, a stack of several DOEs (756) is shown receiving
collimated light from a waveguide (722) and refocusing it based upon the additive powers of the
activated DOEs. Linear and/or radial terms of DOEs may be modulated over time, such as on
a frame sequential basis, to produce a y of treatments (such as tiled display
configurations or ed field of view) for the light coming from the waveguide and exiting,
preferably toward the user’s eye. In configurations wherein the DOE or DOEs are embedded
within the waveguide, a low diffraction efficiency is desired to maximize transparency for light
passed from the real world; in configurations wherein the DOE or DOEs are not embedded, a
high diffraction efficiency may be desired, as described above. In one embodiment, both linear
and radial DOE terms may be combined outside of the ide, in which case high
diffraction efficiency would be d.
Referring to Figure 21 N, a segmented or parabolic reflector, such as those
discussed above in Figure SQ, is shown. Rather than executing a segmented reflector by
combining a ity of smaller reflectors, in one embodiment the same onality may result
from a single waveguide with a DOE having different phase profiles for each section of it, such
that it is controllable by subsection. In other words, while the entire segmented reflector
functionality may be turned on or off together, generally the DOE may be configured to direct
light toward the same region in space (i.e., the pupil of the user).
Referring to s 22A—22Z, optical urations known as “freeform optics” may
be utilized certain of the aforementioned nges. The term “freeform” generally is used in
reference to arbitrarily curved surfaces that may be utilized in situations wherein a spherical,
parabolic, or cylindrical lens does not meet a design complexity such as a geometric constraint.
For example, referring to Figure 22A, one of the common challenges with display (762)
configurations when a user is g through a mirror (and also sometimes a lens 760) is that
the field of view is limited by the area subtended by the final lens (760) of the system.
Referring to Figure 228, in more simple terms, if one has a display (762), which
may include some lens elements, there is a straightforward geometric relationship such that the
field of view cannot be larger than the angle subtended by the display (762). Referring to
Figure 22C, this challenge is exacerbated if the user is trying to have an ted reality
experience wherein light from the real world is also be to passed through the optical system,
because in such case, there often is a reflector (764) that leads to a lens (760); by interposing
a reflector, the overall path length to get to the lens from the eye is sed, which tightens
the angle and reduces the field of view.
Given this, if one wants to increase the field of view, he must increase the size of the
lens, but that might mean pushing a physical lens toward the forehead of the user from an
mic ctive. Further, the reflector may not catch all of the light from the larger lens.
Thus, there is a cal limitation d by human head geometry, and it generally is a
challenge to get more than a 40-degree field of view using conventional see-through displays
and lenses.
With freeform lenses, rather than having a standard planar tor as described
above, one has a combined reflector and lens with power (Le, a curved reflector 766), which
means that the curved lens geometry determines the field of view. Referring to Figure 22D,
without the circuitous path length of a conventional paradigm as described above in nce
to Figure 22C, it is possible for a freeform arrangement to realize a significantly larger field of
view for a given set of optical requirements.
Referring to Figure 22E, a typical freeform optic has three active surfaces. Referring
to Figure 22E, in one typical freeform optic (770) configuration, light may be directed toward the
freeform optic from an image plane, such as a flat panel display (768), into the first active
e (772), which typically is a primarily transmissive freeform surface that ts
transmitted light and imparts a focal change (such as an added stigmatism, because the final
bounce from the third surface will add a matching/opposite stigmatism and these are desirably
canceled). The incoming light may be directed from the first surface to a second surface (774),
wherein it may strike with an angle shallow enough to cause the light to be reflected under total
internal reflection toward the third surface (776).
The third surface may comprise a half-silvered, arbitrarily-curved surface configured
to bounce the light out through the second surface toward the eye, as shown in Figure 22E.
Thus in the depicted typical freeform configuration, the light enters h the first surface,
bounces from the second surface, bounces from the third surface, and is ed out of the
second e. Due to the optimization of the second surface to have the requisite reflective
properties on the first pass, as well as refractive properties on the second pass as the light is
exited toward the eye, a variety of curved surfaces with higher-order shapes than a simple
sphere or parabola are formed into the freeform optic.
Referring to Figure 22F, a compensating lens (780) may be added to the freeform
optic (770) such that the total thickness of the optic assembly is substantially m in
ess, and ably without magnification, to light ng from the real world (144) in an
augmented reality uration.
Referring to Figure 22G, a freeform optic (770) may be ed with a waveguide
(778) configured to facilitate total internal reflection of captured light within certain constraints.
For example, as shown in Figure 22G, light may be directed into the rm/waveguide
assembly from an image plane, such as a flat panel display, and totally internally reflected
within the waveguide until it hits the curved freeform surface and escapes toward the eye of the
user. Thus the light bounces several times in total internal reflection until it reaches the
freeform wedge portion.
One of the main objectives with such an assembly is to try to lengthen the optic
assembly while retaining as uniform a thickness as possible (to facilitate transport by total
internal reflection, and also viewing of the world through the assembly without further
compensation) for a larger field of view. Figure 22H depicts a configuration similar to that of
Figure 22G, with the exception that the configuration of Figure 22H also features a
compensating lens portion to further extend the thickness uniformity and assist with viewing the
world through the assembly without further compensation.
Referring to Figure 22l, in another embodiment, a freeform optic (782) is shown with
a small flat surface, or fourth face (784), at the lower left corner that is configured to facilitate
injection of image information at a different location than is lly used with freeform optics.
The input device (786) may se, for e, a ng fiber display, which may be
designed to have a very small output geometry. The fourth face may se s
geometries itself and have its own refractive power, such as by use planar or freeform surface
geometries.
Referring to Figure 22J, in practice, such a configuration may also feature a
reflective coating (788) along the first surface such that it directs light back to the second
surface, which then bounces the light to the third surface, which directs the light out across the
second surface and to the eye (58). The addition of the fourth small surface for injection of the
image information facilitates a more compact configuration. In an embodiment wherein a
classical freeform input configuration and a scanning fiber display (790) are utilized, some
lenses (792, 794) may be required in order to appropriately form an image plane (796) using
the output from the scanning fiber display; these hardware components add extra bulk that
may not be desired.
] Referring to Figure 22K, an embodiment is shown wherein light from a scanning
fiber display (790) is passed through an input optics assembly (792, 794) to an image plane
(796), and then directed across the first surface of the freeform optic (770) to a total al
reflection bounce off of the second surface, then another total al reflection bounce from
the third e results in the light exiting across the second surface and being directed toward
the eye (58).
An all-total-internal-reflection freeform waveguide may be created such that there
are no reflective coatings (i.e., such that total-internal-reflection is being relied upon for
propagation of light until a critical angle of incidence with a surface is met, at which point the
light exits in a manner akin to the wedge-shaped optics described above). In other words,
rather than having two planar surfaces, one may have a surface comprising one or more sub-
es from a set of conical curves, such as parabolas, spheres, ellipses, etc.).
Such a configuration still may produce a shallow-enough angles for total internal
reflection within the optic; thus an approach that is somewhat a hybrid between a conventional
freeform optic and a wedge-shaped waveguide is presented. One motivation to have such a
configuration is to get away from the use of tive coatings, which do help product reflection,
but also are known to prevent transmission of a relatively large portion (such as 50%) of the
light transmitting through from the real world (144). Further, such coatings also may block an
equivalent amount of the light coming into the freeform optic from the input device. Thus there
are reasons to develop designs that do not have reflective gs.
As described above, one of the surfaces of a conventional freeform optic may
comprise a half-silvered reflective surface. Generally such a reflective e will be of
al density”, meaning that it will generally reflect all wavelengths similarly. In another
embodiment, such as one n a scanning fiber display is utilized as an input, the
conventional reflector gm may be replaced with a narrow band reflector that is
wavelength ive, such as a thin film laserline reflector. Thus in one embodiment, a
configuration may reflect particular red/green/blue wavelength ranges and remain e to
other wavelengths, which generally will increase arency of the optic and therefore be
preferred for augmented reality configurations wherein transmission of image information from
the real world (144) across the optic also is valued.
Referring to Figure 22L, an ment is depicted wherein multiple rm optics
(770) may be stacked in the Z axis (i.e., along an axis substantially aligned with the optical axis
of the eye). In one variation, each of the three depicted freeform optics may have a
ngth-selective coating (for example, one highly selective for blue, the next for green, the
next for red) so that images may be injected into each to have blue reflected from one surface,
green from another, and red from a third surface. Such a configuration may be utilized, for
example, to address chromatic aberration issues, to create a lightfield, or to increase the
functional exit pupil size.
Referring to Figure 22M, an embodiment is shown wherein a single freeform optic
(798) has le reflective surfaces (800, 802, 804), each of which may be wavelength or
polarization selective so that their reflective ties may be individually controlled.
Referring to Figure 22N, in one embodiment, multiple microdisplays, such as
scanning light ys, (786) may be injected into a single rm optic to tile images (thereby
providing an increased field of view), increase the functional pupil size, or address challenges
such as tic aberration (i.e., by reflecting one wavelength per display). Each of the
depicted displays would inject light that would take a different path through the freeform optic
due to the different positioning of the displays relative to the freeform optic, which would e
a larger functional exit pupil .
In one embodiment, a packet or bundle of scanning fiber displays may be utilized as
an input to overcome one of the challenges in operatively coupling a scanning fiber display to a
rm optic. One such challenge with a scanning fiber display configuration is that the output
of an individual fiber is emitted with a certain numerical aperture, or “NA”, which is like the
projectional angle of light from the fiber; ultimately this angle determines the diameter of the
beam that passes through various optics, and ultimately determines the exit functional exit pupil
size; thus in order to maximize exit pupil size with a rm optic configuration, one may
either increase the NA of the fiber using optimized refractive relationships, such as between
core and cladding, or one may place a lens (Le, a refractive lens, such as a gradient refractive
index lens, or “GRIN” lens) at the end of the fiber or build one into the end of the fiber as
described above, or create an array of fibers that is feeding into the freeform optic, in which
case all of those NAs in the bundle remain small, and at the exit pupil an array of small exit
pupils is produced that in the ate forms the functional equivalent of a large exit pupil.
Alternatively, in another embodiment a more sparse array (i.e., not bundled y as
a packet) of scanning fiber displays or other displays may be utilized to functionally increase the
field of view of the virtual image through the freeform optic. Referring to Figure 220, in another
embodiment, a plurality of displays or displays (786) may be injected through the top of a
rm optic (770), as well as another plurality (786) through the lower ; the display
arrays may be two or three dimensional arrays. Referring to Figure 22P, in another related
embodiment, image information also may be injected in from the side (806) of the freeform optic
(770) as well.
In an ment wherein a plurality of smaller exit pupils is to be aggregated into a
functionally larger exit pupil, one may elect to have each of the scanning fibers monochromatic,
such that within a given bundle or plurality of projectors or displays, one may have a subgroup
of solely red fibers, a subgroup of solely blue fibers, and a subgroup of solely green fibers.
Such a configuration facilitates more efficiency in output coupling for ng light into the
l fibers; for instance, there would be no need in such an embodiment to superimpose
red, green, and blue into the same band.
ing to Figures 22Q-22V, various freeform optic tiling configurations are
depicted. Referring to Figure 22Q, an embodiment is depicted wherein two freeform optics are
tiled side-by-side and a microdisplay, such as a scanning light display, (786) on each side is
ured to inject image information from each side, such that one freeform optic wedge
represents each half of the field of view.
Referring to Figure 22R, a compensator lens (808) may be included to facilitate
views of the real world through the optics assembly. Figure 228 rates a uration
wherein freeform optics wedges are tiled side by side to increase the functional field of view
while keeping the thickness of such optical assembly relatively uniform.
Referring to Figure 22T, a star-shaped assembly comprises a plurality of freeform
optics wedges (also shown with a plurality of displays for inputting image ation) in a
uration that may provide a larger field of view expansion while also ining a
relatively thin overall optics assembly thickness.
With a tiled freeform optics assembly, the optics elements may be aggregated to
produce a larger field of view; the tiling configurations described above have addressed this
notion. For example, in a configuration wherein two freeform waveguides are aimed at the eye
such as that depicted in Figure 22R, there are several ways to increase the field of view. One
option is to “toe in” the freeform waveguides such that their outputs share, or are superimposed
in, the space of the pupil (for example, the user may see the left half of the visual field through
the left freeform ide, and the right half of the visual field through the right freeform
waveguide).
With such a configuration, the field of view has been increased with the tiled
freeform waveguides, but the exit pupil has not grown in size. Alternatively, the freeform
waveguides may be ed such that they do not toe in as much — so they create exit pupils
that are side-by—side at the eye’s anatomical pupil. In one example, the anatomical pupil may
be 8mm wide, and each of the side-by—side exit pupils may be 8mm, such that the functional
exit pupil is ed by about two times. Thus such a configuration provides an enlarged exit
pupil, but if the eye is moved around in the “eyebox” defined by that exit pupil, that eye may
lose parts of the visual field (i.e., lose either a portion of the left or right incoming light e
of the side-by—side nature of such configuration).
In one embodiment using such an approach for tiling freeform optics, especially in
the Z-axis relative to the eye of the user, red wavelengths may be driven through one freeform
optic, green through another, and blue through r, such red/green/blue chromatic
aberration may be addressed. Multiple freeform optics also may be provided to such a
configuration that are stacked up, each of which is configured to address a particular
wavelength.
Referring to Figure 22U, two oppositely-oriented freeform optics are shown stacked
in the Z-axis (i.e., they are upside down relative to each other). With such a configuration, a
compensating lens may not be required to tate accurate views of the world through the
assembly; in other words, rather than having a compensating lens such as in the embodiment
of Figure 22F or Figure 22R, an additional freeform optic may be ed, which may further
assist in routing light to the eye. Figure 22V shows another similar configuration wherein the
assembly of two rm optics is presented as a vertical stack.
To ensure that one surface is not interfering with another surface in the freeform
optics, one may use ngth or polarization selective reflector surfaces. For example,
referring to Figure 22V, red, green, and blue wavelengths in the form of 650nm, 530nm, and
450nm may be injected, as well as red, green, and blue wavelengths in the form of 620nm,
550nm, and 470nm; different selective reflectors may be utilized in each of the freeform optics
so that they do not interfere with each other. In a configuration wherein polarization filtering is
used for a similar purpose, the reflection/transmission selectivity for light that is polarized in a
particular axis may be varied (i.e., the images may be larized before they are sent to
each freeform ide, to work with reflector selectivity).
Referring to Figures 22W and 22X, configurations are illustrated wherein a plurality
of rm waveguides may be utilized together in series. Referring to Figure 22W, light may
enter from the real world and be ed tially through a first freeform optic (770),
through an optional lens (812) which may be configured to relay light to a tor (810) such
as a DMD from a DLP , which may be configured to reflect the light that has been
filtered on a pixel by pixel basis (i.e., an occlusion mask may be utilized to block out certain
elements of the real world, such as for darkfield perception, as described above; suitable
l light modulators may be used which comprise DMDs, LCDs, ferroelectric LCOSs,
MEMS shutter arrays, and the like, as described above) to another freeform optic (770) that is
relaying light to the eye (28) of the user. Such a uration may be more compact than one
using conventional lenses for spatial light modulation.
ing to Figure 22X, in a scenario wherein it is very important to keep overall
thickness minimized, a configuration may be utilized that has one surface that is highly-
reflective so that it may bounce light straight into another compactly positioned freeform optic.
In one embodiment a selective attenuator (814) may be interposed between the two freeform
optics (770).
Referring to Figure 22Y, an embodiment is depicted wherein a freeform optic (770)
may comprise one aspect of a contact lens system. A miniaturized rm optic is shown
engaged against the cornea of a user’s eye (58) with a miniaturized compensator lens portion
(780), akin to that bed in reference to Figure 22F. Signals may be injected into the
miniaturized freeform assembly using a tethered scanning fiber display which may, for e,
be coupled between the freeform optic and a tear duct area of the user, or between the
freeform optic and another head-mounted display configuration.
Various example embodiments of the invention are described . Reference is
made to these examples in a miting sense. They are provided to illustrate more broadly
applicable aspects of the invention. Various changes may be made to the ion described
and equivalents may be substituted without departing from the true spirit and scope of the
ion. In addition, many cations may be made to adapt a particular situation, material,
composition of matter, process, s act(s) or step(s) to the ive(s), spirit or scope of
the present invention. Further, as will be appreciated by those with skill in the art that each of
the individual variations described and illustrated herein has discrete components and features
which may be readily separated from or combined with the features of any of the other several
embodiments without departing from the scope or spirit of the present inventions. All such
modifications are intended to be within the scope of claims associated with this sure.
The invention includes methods that may be performed using the subject devices.
The methods may comprise the act of providing such a suitable device. Such provision may be
performed by the end user. In other words, the "providing" act merely requires the end user
obtain, access, approach, position, set-up, activate, power-up or otherwise act to e the
requisite device in the subject method. s recited herein may be carried out in any order
of the recited events which is logically possible, as well as in the recited order of events.
Example aspects of the invention, together with details regarding material selection
and manufacture have been set forth above. As for other details of the present invention, these
may be iated in connection with the above-referenced patents and publications as well
as generally known or appreciated by those with skill in the art. The same may hold true with
respect to method-based aspects of the invention in terms of additional acts as commonly or
logically employed.
In addition, though the invention has been described in reference to several
examples optionally incorporating various features, the invention is not to be limited to that
which is described or indicated as contemplated with respect to each variation of the invention.
Various changes may be made to the invention described and equivalents (whether recited
herein or not included for the sake of some brevity) may be substituted without departing from
the true spirit and scope of the invention. In addition, where a range of values is provided, it is
understood that every intervening value, n the upper and lower limit of that range and
any other stated or intervening value in that stated range, is encompassed within the ion.
Also, it is contemplated that any optional feature of the inventive variations
described may be set forth and claimed independently, or in combination with any one or more
of the features described . Reference to a singular item, es the possibility that there
are plural of the same items present. More specifically, as used herein and in claims associated
, the singularforms "a, an," "said," and "the" e plural nts unless the
specifically stated otherwise. In other words, use of the articles allow for "at least one" of the
subject item in the description above as well as claims ated with this disclosure. It is
further noted that such claims may be d to exclude any optional element. As such, this
statement is intended to serve as antecedent basis for use of such exclusive terminology as
y," "only" and the like in connection with the recitation of claim elements, or use of a
"negative" limitation.
Without the use of such exclusive terminology, the term ising" in claims
associated with this disclosure shall allow for the inclusion of any additional element--
irrespective of whether a given number of elements are ated in such claims, or the
addition of a feature could be regarded as transforming the nature of an element set forth in
such claims. Except as ically defined herein, all technical and scientific terms used herein
are to be given as broad a commonly understood meaning as le while
maintaining claim validity.
The breadth of the present invention is not to be limited to the es
provided and/or the subject specification, but rather only by the scope of claim
language associated with this disclosure.
The reference in this specification to any prior publication (or information
derived from it), or to any matter which is known, is not, and should not be taken as an
acknowledgment or admission or any form of suggestion that the prior publication (or
information derived from it) or known matter forms part of the common l
knowledge in the field of endeavour to which this specification relates.
Claims (20)
1. A system for displaying virtual t to a user, comprising: at least one light source to multiplex a plurality of light beams to display a respective plurality of light patterns associated with one or more frames of image data; first, second and third planar waveguides to receive the plurality of light beams and to direct the plurality of light beams toward an exit pupil, wherein the first, second and third planar waveguides are stacked along an l axis of the user; and at least one optical element to modify a focus of a light beam of the plurality of light beams directed by the first, second and third planar waveguides. n the first, second and third planar waveguides have first, second and third wavefront curvatures corresponding to respective first, second and third focal distances; wherein the first focal distance is a focal distance of a determined accommodation of the user’s eyes; wherein the second focal distance is based at least in part on the determined accommodation and is located at a first predetermined error margin r from the user’s eyes relative to the first focal ce; wherein the third focal ce is based at least in part on the determined accommodation and is located at a second predetermined error margin closer to the user’s eyes relative to the first focal distance; and, wherein the first, second and third planar waveguides are configured to allow a range of accommodation before a physical adjustment of the focus level is necessary due to a change in the accommodation of the user’s eyes.
2. The system of claim 1, wherein the at least one optical element modifies the plurality of light beams in a manner such that a ont curvature is created, and wherein the created wavefront curvature corresponds to a focal plane when the plurality of light beams are viewed by the user.
3. The system of claim 1 or 2, wherein the first, second and third planar waveguides are each switchable between on and off states.
4. The system of claim 1 or 2, wherein the first, second and third planar waveguides are static planar waveguides.
5. The system of any one of claims 1 to 4, wherein an optical element of the at least one optical elements is able between on and off states.
6. The system of any one of claims 1 to 4, wherein an optical element of the at least one optical elements is a static optical element.
7. The system of any one of claims 1 to 6, wherein an optical element of the at least one optical elements is a diffractive optical element (DOE) in one of the first, second or third planar waveguides, wherein the DOE is switchable between on and off states.
8. The system of any one of claims 1 to 7, wherein the at least one optical element comprises a weak lens.
9. The system of any one of claims 1 to 7, wherein the at least one optical element comprises a l zone plate.
10. The system of any one of claims 1 to 9, wherein first, second and third frames of the one or more frames of image data are red through the first, second and third planar ides to the user's eye aneously.
11. The system of any one of claims 1 to 9, wherein first, second and third frames of the one or more frames of image data are delivered through the first, second and third planar waveguides to the user's eye sequentially.
12. The system of any one of claims 1 to 11, wherein the first planar waveguide is disposed between the second and third planar waveguides, and wherein the first wavefront ure corresponds to the focal distance of the determined accommodation of the user’s eyes.
13. The system of claim 12, n the second planar waveguide is closer than the first planar waveguide to the user, and wherein the second wavefront ure corresponds to the first predetermined error margin from the user’s eyes relative to the first focal distance.
14. The system of claim 13, wherein the third planar waveguide is farther than the first planar waveguide from the user, wherein the third wavefront curvature corresponds to the second predetermined error margin closer to the user’s eyes ve to the first focal distance; wherein the third ont curvature is greater than the first wavefront curvature; and, wherein the first wavefront curvature is r than the second wavefront curvature.
15. The system of any one of claims 1 to 14, further comprising a compensating lens to compensate for a cumulative focus modification of the at least one optical elements on a light beam passing through the first, second and third planar waveguides along the optical axis of the user.
16. The system of claim 15, wherein the compensating lens is disposed on an opposite side of the plurality of planar waveguides from the user.
17. The system of any one of claims 1 to 16, wherein the at least one light source comprises a plurality of light sources corresponding to respective ones of the plurality of waveguides.
18. The system of any one of claims 1 to 16, wherein the at least one light source is asingle multiplexed light source.
19. The system of any one of claims 1 to 18, wherein the first predetermined error margin is an accommodation error margin, and wherein second ermined error margin is an error accommodation margin.
20. The system of claim 1, wherein the at least one light source comprises first, second and third light s respectively addressing the first, second and third planar waveguides along respective udinal axes of the first, second and third planar waveguides orthogonal to the optical axis of the user.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361909774P | 2013-11-27 | 2013-11-27 | |
US61/909,774 | 2013-11-27 | ||
NZ755273A NZ755273B2 (en) | 2013-11-27 | 2014-11-27 | Virtual and augmented reality systems and methods |
Publications (2)
Publication Number | Publication Date |
---|---|
NZ758961A NZ758961A (en) | 2021-02-26 |
NZ758961B2 true NZ758961B2 (en) | 2021-05-27 |
Family
ID=
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2022246404B2 (en) | Virtual and augmented reality systems and methods | |
NZ758961B2 (en) | Virtual and augmented reality systems and methods | |
NZ755266B2 (en) | Virtual and augmented reality systems and methods | |
NZ755273B2 (en) | Virtual and augmented reality systems and methods | |
NZ758961A (en) | Virtual and augmented reality systems and methods | |
NZ755269B2 (en) | Virtual and augmented reality systems and methods |