US20150241696A1 - Inducing phase delays to vary an aggregate wavefront for augmented or virtual reality - Google Patents
Inducing phase delays to vary an aggregate wavefront for augmented or virtual reality Download PDFInfo
- Publication number
- US20150241696A1 US20150241696A1 US14/706,596 US201514706596A US2015241696A1 US 20150241696 A1 US20150241696 A1 US 20150241696A1 US 201514706596 A US201514706596 A US 201514706596A US 2015241696 A1 US2015241696 A1 US 2015241696A1
- Authority
- US
- United States
- Prior art keywords
- light
- eye
- waveguide
- configuration
- display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000001934 delay Effects 0.000 title claims description 6
- 230000003190 augmentative effect Effects 0.000 title abstract description 35
- 230000001939 inductive effect Effects 0.000 title 1
- 239000000835 fiber Substances 0.000 claims description 157
- 230000003287 optical effect Effects 0.000 claims description 89
- 238000000034 method Methods 0.000 claims description 24
- 229920001410 Microfiber Polymers 0.000 claims description 2
- 239000003658 microfiber Substances 0.000 claims description 2
- 230000001066 destructive effect Effects 0.000 claims 1
- 239000000758 substrate Substances 0.000 abstract description 49
- 210000001508 eye Anatomy 0.000 description 232
- 210000000695 crystalline len Anatomy 0.000 description 212
- 241000219739 Lens Species 0.000 description 199
- 210000001747 pupil Anatomy 0.000 description 107
- 230000008447 perception Effects 0.000 description 46
- 230000004308 accommodation Effects 0.000 description 36
- 210000003128 head Anatomy 0.000 description 32
- 239000004973 liquid crystal related substance Substances 0.000 description 32
- 210000001525 retina Anatomy 0.000 description 32
- 230000033001 locomotion Effects 0.000 description 29
- 230000008859 change Effects 0.000 description 28
- 239000013307 optical fiber Substances 0.000 description 24
- 230000000694 effects Effects 0.000 description 22
- 238000003491 array Methods 0.000 description 20
- 210000004556 brain Anatomy 0.000 description 20
- 239000000463 material Substances 0.000 description 20
- 230000006870 function Effects 0.000 description 19
- 238000013459 approach Methods 0.000 description 18
- 238000005286 illumination Methods 0.000 description 18
- 238000000576 coating method Methods 0.000 description 17
- 239000011295 pitch Substances 0.000 description 17
- 239000011521 glass Substances 0.000 description 16
- 230000010287 polarization Effects 0.000 description 14
- 230000005540 biological transmission Effects 0.000 description 13
- GQYHUHYESMUTHG-UHFFFAOYSA-N lithium niobate Chemical compound [Li+].[O-][Nb](=O)=O GQYHUHYESMUTHG-UHFFFAOYSA-N 0.000 description 13
- 238000012545 processing Methods 0.000 description 13
- 230000003068 static effect Effects 0.000 description 13
- 239000011248 coating agent Substances 0.000 description 12
- 238000010168 coupling process Methods 0.000 description 10
- 230000000007 visual effect Effects 0.000 description 10
- 210000004087 cornea Anatomy 0.000 description 9
- 238000005859 coupling reaction Methods 0.000 description 9
- 125000001475 halogen functional group Chemical group 0.000 description 9
- 238000012800 visualization Methods 0.000 description 9
- 230000008901 benefit Effects 0.000 description 8
- 230000008878 coupling Effects 0.000 description 8
- 238000005516 engineering process Methods 0.000 description 8
- 238000002347 injection Methods 0.000 description 8
- 239000007924 injection Substances 0.000 description 8
- 239000011263 electroactive material Substances 0.000 description 7
- 230000004044 response Effects 0.000 description 7
- 238000011282 treatment Methods 0.000 description 7
- 230000004075 alteration Effects 0.000 description 6
- 230000002238 attenuated effect Effects 0.000 description 6
- 230000000903 blocking effect Effects 0.000 description 6
- 238000013461 design Methods 0.000 description 6
- 238000009826 distribution Methods 0.000 description 6
- 210000000887 face Anatomy 0.000 description 6
- 239000010408 film Substances 0.000 description 6
- 230000000873 masking effect Effects 0.000 description 6
- 108091008695 photoreceptors Proteins 0.000 description 6
- 241000282412 Homo Species 0.000 description 5
- 230000001886 ciliary effect Effects 0.000 description 5
- 238000005253 cladding Methods 0.000 description 5
- 239000002131 composite material Substances 0.000 description 5
- 238000001914 filtration Methods 0.000 description 5
- 239000004983 Polymer Dispersed Liquid Crystal Substances 0.000 description 4
- 230000001427 coherent effect Effects 0.000 description 4
- AMGQUBHHOARCQH-UHFFFAOYSA-N indium;oxotin Chemical compound [In].[Sn]=O AMGQUBHHOARCQH-UHFFFAOYSA-N 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 239000012528 membrane Substances 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 230000001902 propagating effect Effects 0.000 description 4
- 230000005855 radiation Effects 0.000 description 4
- 230000002123 temporal effect Effects 0.000 description 4
- 230000007704 transition Effects 0.000 description 4
- 206010025421 Macule Diseases 0.000 description 3
- 230000035559 beat frequency Effects 0.000 description 3
- 239000003086 colorant Substances 0.000 description 3
- 230000000875 corresponding effect Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 238000002156 mixing Methods 0.000 description 3
- 230000003534 oscillatory effect Effects 0.000 description 3
- 238000004064 recycling Methods 0.000 description 3
- 238000009877 rendering Methods 0.000 description 3
- 230000000638 stimulation Effects 0.000 description 3
- 239000010409 thin film Substances 0.000 description 3
- KRQUFUKTQHISJB-YYADALCUSA-N 2-[(E)-N-[2-(4-chlorophenoxy)propoxy]-C-propylcarbonimidoyl]-3-hydroxy-5-(thian-3-yl)cyclohex-2-en-1-one Chemical compound CCC\C(=N/OCC(C)OC1=CC=C(Cl)C=C1)C1=C(O)CC(CC1=O)C1CCCSC1 KRQUFUKTQHISJB-YYADALCUSA-N 0.000 description 2
- 201000004569 Blindness Diseases 0.000 description 2
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 2
- 239000000654 additive Substances 0.000 description 2
- 230000000996 additive effect Effects 0.000 description 2
- 230000004931 aggregating effect Effects 0.000 description 2
- 230000002776 aggregation Effects 0.000 description 2
- 238000004220 aggregation Methods 0.000 description 2
- 230000003416 augmentation Effects 0.000 description 2
- 230000003098 cholesteric effect Effects 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 238000005520 cutting process Methods 0.000 description 2
- 230000003247 decreasing effect Effects 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 239000006185 dispersion Substances 0.000 description 2
- 210000000613 ear canal Anatomy 0.000 description 2
- 238000005530 etching Methods 0.000 description 2
- 239000012530 fluid Substances 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 210000003205 muscle Anatomy 0.000 description 2
- 230000002688 persistence Effects 0.000 description 2
- 239000004033 plastic Substances 0.000 description 2
- 238000005498 polishing Methods 0.000 description 2
- 229920000642 polymer Polymers 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 230000011514 reflex Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 238000010408 sweeping Methods 0.000 description 2
- 230000032258 transport Effects 0.000 description 2
- 230000004462 vestibulo-ocular reflex Effects 0.000 description 2
- 239000004986 Cholesteric liquid crystals (ChLC) Substances 0.000 description 1
- 239000004593 Epoxy Substances 0.000 description 1
- 206010019233 Headaches Diseases 0.000 description 1
- 230000005374 Kerr effect Effects 0.000 description 1
- RAQQRQCODVNJCK-JLHYYAGUSA-N N-[(4-amino-2-methylpyrimidin-5-yl)methyl]-N-[(E)-5-hydroxy-3-(2-hydroxyethyldisulfanyl)pent-2-en-2-yl]formamide Chemical compound C\C(N(Cc1cnc(C)nc1N)C=O)=C(\CCO)SSCCO RAQQRQCODVNJCK-JLHYYAGUSA-N 0.000 description 1
- 239000004988 Nematic liquid crystal Substances 0.000 description 1
- 206010033799 Paralysis Diseases 0.000 description 1
- 241000270295 Serpentes Species 0.000 description 1
- 239000000853 adhesive Substances 0.000 description 1
- 230000001070 adhesive effect Effects 0.000 description 1
- 229910052782 aluminium Inorganic materials 0.000 description 1
- XAGFODPZIPBFFR-UHFFFAOYSA-N aluminium Chemical compound [Al] XAGFODPZIPBFFR-UHFFFAOYSA-N 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 238000000429 assembly Methods 0.000 description 1
- 208000003464 asthenopia Diseases 0.000 description 1
- 239000011324 bead Substances 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000002041 carbon nanotube Substances 0.000 description 1
- 229910021393 carbon nanotube Inorganic materials 0.000 description 1
- 210000003161 choroid Anatomy 0.000 description 1
- 239000002642 cobra venom Substances 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000001447 compensatory effect Effects 0.000 description 1
- 238000002591 computed tomography Methods 0.000 description 1
- 230000008602 contraction Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 239000013013 elastic material Substances 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000004424 eye movement Effects 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 239000005262 ferroelectric liquid crystals (FLCs) Substances 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 229910021389 graphene Inorganic materials 0.000 description 1
- 230000004886 head movement Effects 0.000 description 1
- 231100000869 headache Toxicity 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000000670 limiting effect Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000004459 microsaccades Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000008450 motivation Effects 0.000 description 1
- 230000007433 nerve pathway Effects 0.000 description 1
- 230000006855 networking Effects 0.000 description 1
- 230000007935 neutral effect Effects 0.000 description 1
- 238000010428 oil painting Methods 0.000 description 1
- 210000001328 optic nerve Anatomy 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 238000012856 packing Methods 0.000 description 1
- 239000003973 paint Substances 0.000 description 1
- 239000002245 particle Substances 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 230000002085 persistent effect Effects 0.000 description 1
- 239000000049 pigment Substances 0.000 description 1
- 238000004321 preservation Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000002310 reflectometry Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 230000002207 retinal effect Effects 0.000 description 1
- 230000004256 retinal image Effects 0.000 description 1
- 238000005096 rolling process Methods 0.000 description 1
- 238000005488 sandblasting Methods 0.000 description 1
- 210000003786 sclera Anatomy 0.000 description 1
- 238000012163 sequencing technique Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 239000007921 spray Substances 0.000 description 1
- 238000005507 spraying Methods 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 239000002435 venom Substances 0.000 description 1
- 210000001048 venom Anatomy 0.000 description 1
- 231100000611 venom Toxicity 0.000 description 1
- 230000004470 vergence movement Effects 0.000 description 1
- 230000016776 visual perception Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/42—Diffraction optics, i.e. systems including a diffractive element being designed for providing a diffractive effect
- G02B27/4205—Diffraction optics, i.e. systems including a diffractive element being designed for providing a diffractive effect having a diffractive optical element [DOE] contributing to image formation, e.g. whereby modulation transfer function MTF or optical aberrations are relevant
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B17/00—Systems with reflecting surfaces, with or without refracting elements
- G02B17/08—Catadioptric systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B26/00—Optical devices or arrangements for the control of light using movable or deformable optical elements
- G02B26/08—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light
- G02B26/0808—Optical devices or arrangements for the control of light using movable or deformable optical elements for controlling the direction of light by means of one or more diffracting elements
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0087—Phased arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0093—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0176—Head mounted characterised by mechanical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B3/00—Simple or compound lenses
- G02B3/0006—Arrays
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B3/00—Simple or compound lenses
- G02B3/0006—Arrays
- G02B3/0037—Arrays characterized by the distribution or form of lenses
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/22—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the stereoscopic type
- G02B30/24—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes of the stereoscopic type involving temporal multiplexing, e.g. using sequentially activated left and right shutters
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B30/00—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images
- G02B30/20—Optical systems or apparatus for producing three-dimensional [3D] effects, e.g. stereoscopic images by providing first and second parallax images to an observer's left and right eyes
- G02B30/34—Stereoscopes providing a stereoscopic pair of separated images corresponding to parallactically displaced views of the same object, e.g. 3D slide viewers
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/005—Diaphragms
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/18—Diffraction gratings
- G02B5/1814—Diffraction gratings structurally combined with one or more further optical elements, e.g. lenses, mirrors, prisms or other diffraction gratings
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/18—Diffraction gratings
- G02B5/1828—Diffraction gratings having means for producing variable diffraction
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/20—Filters
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/02—Optical fibres with cladding with or without a coating
- G02B6/02042—Multicore optical fibres
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/04—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres
- G02B6/06—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres the relative position of the fibres being the same at both ends, e.g. for transporting images
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/24—Coupling light guides
- G02B6/26—Optical coupling means
- G02B6/32—Optical coupling means having lens focusing means positioned between opposed fibre ends
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/24—Coupling light guides
- G02B6/26—Optical coupling means
- G02B6/34—Optical coupling means utilising prism or grating
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/24—Coupling light guides
- G02B6/36—Mechanical coupling means
- G02B6/40—Mechanical coupling means having fibre bundle mating means
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C7/00—Optical parts
- G02C7/02—Lenses; Lens systems ; Methods of designing lenses
- G02C7/04—Contact lenses for the eyes
- G02C7/049—Contact lenses having special fitting or structural features achieved by special materials or material structures
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C7/00—Optical parts
- G02C7/10—Filters, e.g. for facilitating adaptation of the eyes to the dark; Sunglasses
- G02C7/104—Filters, e.g. for facilitating adaptation of the eyes to the dark; Sunglasses having spectral characteristics for purposes other than sun-protection
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/0102—Constructional details, not otherwise provided for in this subclass
- G02F1/0105—Illuminating devices
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/13—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on liquid crystals, e.g. single liquid crystal display cells
- G02F1/133—Constructional arrangements; Operation of liquid crystal cells; Circuit arrangements
- G02F1/1333—Constructional arrangements; Manufacturing methods
- G02F1/1334—Constructional arrangements; Manufacturing methods based on polymer dispersed liquid crystals, e.g. microencapsulated liquid crystals
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/01—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour
- G02F1/17—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the intensity, phase, polarisation or colour based on variable-absorption elements not provided for in groups G02F1/015 - G02F1/169
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
- G02F1/292—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection by controlled diffraction or phased-array beam steering
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
- G02F1/294—Variable focal length devices
-
- G—PHYSICS
- G02—OPTICS
- G02F—OPTICAL DEVICES OR ARRANGEMENTS FOR THE CONTROL OF LIGHT BY MODIFICATION OF THE OPTICAL PROPERTIES OF THE MEDIA OF THE ELEMENTS INVOLVED THEREIN; NON-LINEAR OPTICS; FREQUENCY-CHANGING OF LIGHT; OPTICAL LOGIC ELEMENTS; OPTICAL ANALOGUE/DIGITAL CONVERTERS
- G02F1/00—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics
- G02F1/29—Devices or arrangements for the control of the intensity, colour, phase, polarisation or direction of light arriving from an independent light source, e.g. switching, gating or modulating; Non-linear optics for the control of the position or the direction of light beams, i.e. deflection
- G02F1/31—Digital deflection, i.e. optical switching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/14—Digital output to display device ; Cooperation and interconnection of the display device with other functional units
- G06F3/1423—Digital output to display device ; Cooperation and interconnection of the display device with other functional units controlling a plurality of local displays, e.g. CRT and flat panel display
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/506—Illumination models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/10—Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/003—Navigation within 3D models or images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/73—Deblurring; Sharpening
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/02—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/10—Intensity circuits
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/18—Timing circuits for raster scan displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/111—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation
- H04N13/117—Transformation of image signals corresponding to virtual viewpoints, e.g. spatial image interpolation the virtual viewpoint locations being selected by the viewers or determined by viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/286—Image signal generators having separate monoscopic and stereoscopic modes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/327—Calibration thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/341—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] using temporal multiplexing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/344—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/361—Reproducing mixed stereoscopic images; Reproducing mixed monoscopic and stereoscopic images, e.g. a stereoscopic image overlay window on a monoscopic image background
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/363—Image reproducers using image projection screens
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/366—Image reproducers using viewer tracking
- H04N13/383—Image reproducers using viewer tracking for tracking with gaze detection, i.e. detecting the lines of sight of the viewer's eyes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/14—Picture signal circuitry for video frequency region
- H04N5/21—Circuitry for suppressing or minimising disturbance, e.g. moiré or halo
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/74—Projection arrangements for image reproduction, e.g. using eidophor
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B2006/0098—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings for scanning
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0123—Head-up displays characterised by optical features comprising devices increasing the field of view
- G02B2027/0125—Field-of-view increase by wavefront division
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0132—Head-up displays characterised by optical features comprising binocular systems
- G02B2027/0134—Head-up displays characterised by optical features comprising binocular systems of stereoscopic type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0138—Head-up displays characterised by optical features comprising image capture systems, e.g. camera
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/014—Head-up displays characterised by optical features comprising information/image processing systems
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0149—Head-up displays characterised by mechanical features
- G02B2027/0161—Head-up displays characterised by mechanical features characterised by the relative positioning of the constitutive elements
- G02B2027/0163—Electric or electronic control thereof
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0179—Display position adjusting means not related to the information to be displayed
- G02B2027/0187—Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C2202/00—Generic optical aspects applicable to one or more of the subgroups of G02C7/00
- G02C2202/16—Laminated or compound lenses
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
- G06T2207/10021—Stereoscopic video; Stereoscopic image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2213/00—Indexing scheme for animation
- G06T2213/08—Animation software package
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N2013/0074—Stereoscopic image analysis
- H04N2013/0085—Motion estimation from stereoscopic image signals
Definitions
- the present disclosure relates to virtual reality and augmented reality imaging and visualization systems.
- a virtual reality, or “VR”, scenario typically involves presentation of digital or virtual image information without transparency to other actual real-world visual input;
- an augmented reality, or “AR”, scenario typically involves presentation of digital or virtual image information as an augmentation to visualization of the actual world around the user. For example, referring to FIG.
- an augmented reality scene ( 4 ) is depicted wherein a user of an AR technology sees a real-world park-like setting ( 6 ) featuring people, trees, buildings in the background, and a concrete platform ( 1120 ).
- the user of the AR technology also perceives that he “sees” a robot statue ( 1110 ) standing upon the real-world platform ( 1120 ), and a cartoon-like avatar character ( 2 ) flying by which seems to be a personification of a bumble bee, even though these elements ( 2 , 1110 ) do not exist in the real world.
- the human visual perception system is very complex, and producing a VR or AR technology that facilitates a comfortable, natural-feeling, rich presentation of virtual image elements amongst other virtual or real-world imagery elements is challenging.
- stereoscopic wearable glasses ( 8 ) type configurations have been developed which generally feature two displays ( 10 , 12 ) that are configured to display images with slightly different element presentation such that a three-dimensional perspective is perceived by the human visual system.
- Such configurations have been found to be uncomfortable for many users due to a mismatch between vergence and accommodation which must be overcome to perceive the images in three dimensions; indeed, some users are not able to tolerate stereoscopic configurations.
- FIG. 2B shows another stereoscopic wearable glasses ( 14 ) type configuration featuring two forward-oriented cameras ( 16 , 18 ) configured to capture images for an augmented reality presentation to the user through stereoscopic displays.
- the position of the cameras ( 16 , 18 ) and displays generally blocks the natural field of view of the user when the glasses ( 14 ) are mounted on the user's head.
- an augmented reality configuration ( 20 ) which features a visualization module ( 26 ) coupled to a glasses frame ( 24 ) which also holds conventional glasses lenses ( 22 ).
- the user is able to see an at least partially unobstructed view of the real world with such a system, and has a small display ( 28 ) with which digital imagery may be presented in an AR configuration to one eye—for a monocular AR presentation.
- FIG. 2D features a configuration wherein a visualization module ( 32 ) may be coupled to a hat or helmet ( 30 ) and configured to present monocular augmented digital imagery to a user through a small display ( 34 ).
- FIG. 2E illustrates another similar configuration wherein a frame ( 36 ) couple-able to a user's head in a manner similar to an eyeglasses coupling so that a visualization module ( 38 ) may be utilized to capture images and also present monocular augmented digital imagery to a user through a small display ( 40 ).
- a visualization module 38
- Such a configuration is available, for example, from Google, Inc., of Mountain View, Calif. under the trade name GoogleGlass®.
- None of these configurations is optimally suited for presenting a rich, binocular, three-dimensional augmented reality experience in a manner that will be comfortable and maximally useful to the user, in part because prior systems fail to address some of the fundamental aspects of the human perception system, including the photoreceptors of the retina and their interoperation with the brain to produce the perception of visualization to the user.
- FIG. 3 a simplified cross-sectional view of a human eye is depicted featuring a cornea ( 42 ), iris ( 44 ), lens—or “crystalline lens” ( 46 ), sclera ( 48 ), choroid layer ( 50 ), macula ( 52 ), retina ( 54 ), and optic nerve pathway ( 56 ) to the brain.
- the macula is the center of the retina, which is utilized to see moderate detail; at the center of the macula is a portion of the retina that is referred to as the “fovea”, which is utilized for seeing the finest details, and which contains more photoreceptors (approximately 120 cones per visual degree) than any other portion of the retina.
- the human visual system is not a passive sensor type of system; it is configured to actively scan the environment.
- the photoreceptors of the eye fire in response to changes in stimulation, rather than constantly responding to a constant state of stimulation.
- motion is required to present photoreceptor information to the brain (as is motion of the linear scanner array across a piece of paper in a flatbed scanner, or motion of a finger across a word of Braille imprinted into a paper).
- the fovea of the retina contains the greatest density of photoreceptors, and while humans typically have the perception that they have high-resolution visualization capabilities throughout their field of view, they generally actually have only a small high-resolution center that they are mechanically sweeping around a lot, along with a persistent memory of the high-resolution information recently captured with the fovea.
- the focal distance control mechanism of the eye (ciliary muscles operatively coupled to the crystalline lens in a manner wherein ciliary relaxation causes taut ciliary connective fibers to flatten out the lens for more distant focal lengths; ciliary contraction causes loose ciliary connective fibers, which allow the lens to assume a more rounded geometry for more close-in focal lengths) dithers back and forth by approximately 1 ⁇ 4 to 1 ⁇ 2 diopter to cyclically induce a small amount of what is called “dioptric blur” on both the close side and far side of the targeted focal length; this is utilized by the accommodation control circuits of the brain as cyclical negative feedback that helps to constantly correct course and keep the retinal image of a fixated object approximately in focus.
- the visualization center of the brain also gains valuable perception information from the motion of both eyes and components thereof relative to each other.
- Vergence movements i.e., rolling movements of the pupils toward or away from each other to converge the lines of sight of the eyes to fixate upon an object
- vergence movements of the two eyes relative to each other are closely associated with focusing (or “accommodation”) of the lenses of the eyes.
- accommodation movements i.e., rolling movements of the pupils toward or away from each other to converge the lines of sight of the eyes to fixate upon an object
- accommodation or “accommodation”
- Movement of the head which houses the eyes, also has a key impact upon visualization of objects.
- Humans move their heads to visualize the world around them; they often are in a fairly constant state of repositioning and reorienting the head relative to an object of interest. Further, most people prefer to move their heads when their eye gaze needs to move more than about 20 degrees off center to focus on a particular object (i.e., people do not typically like to look at things “from the corner of the eye”). Humans also typically scan or move their heads in relation to sounds—to improve audio signal capture and utilize the geometry of the ears relative to the head.
- the human visual system gains powerful depth cues from what is called “head motion parallax”, which is related to the relative motion of objects at different distances as a function of head motion and eye vergence distance (i.e., if a person moves his head from side to side and maintains fixation on an object, items farther out from that object will move in the same direction as the head; items in front of that object will move opposite the head motion; these are very salient cues for where things are spatially in the environment relative to the person—perhaps as powerful as stereopsis). Head motion also is utilized to look around objects, of course.
- head and eye motion are coordinated with something called the “vestibulo-ocular reflex”, which stabilizes image information relative to the retina during head rotations, thus keeping the object image information approximately centered on the retina.
- the vestibulo-ocular reflex In response to a head rotation, the eyes are reflexively and proportionately rotated in the opposite direction to maintain stable fixation on an object.
- many humans can read a book while shaking their head back and forth (interestingly, if the book is panned back and forth at the same speed with the head approximately stationary, the same generally is not true—the person is not likely to be able to read the moving book; the vestibulo-ocular reflex is one of head and eye motion coordination, generally not developed for hand motion).
- This paradigm may be important for augmented reality systems, because head motions of the user may be associated relatively directly with eye motions, and the system preferably will be ready to work with this relationship.
- the 2-D oil painting object may be head-centric, in which case the object moves around along with the user's head (e.g., as in a GoogleGlass approach); or the object may be world-centric, in which case it may be presented as though it is part of the real world coordinate system, so that the user may move his head or eyes without moving the position of the object relative to the real world.
- the object should be presented as world centric (i.e., the virtual object stays in position in the real world so that the user may move his body, head, eyes around it without changing its position relative to the real world objects surrounding it, such as a real world wall); body, or torso, centric, in which case a virtual element may be fixed relative to the user's torso, so that the user can move his head or eyes without moving the object, but that is slaved to torso movements; head centric, in which case the displayed object (and/or display itself) may be moved along with head movements, as described above in reference to GoogleGlass; or eye centric, as in a “foveated display” configuration, as is described below, wherein content is slewed around as a function of what the eye position is.
- world centric i.e., the virtual object stays in position in the real world so that the user may move his body, head, eyes around it without changing its position relative to the real world objects surrounding it, such as a real world wall
- Embodiments of the present invention are directed to devices, systems and methods for facilitating virtual reality and/or augmented reality interaction for one or more users.
- a system for displaying virtual content is disclosed.
- the system comprises an image-generating source to provide one or more frames of image data in a time-sequential manner, a light modulator configured to transmit light associated with the one or more frames of image data, a substrate to direct image information to a user's eye, wherein the substrate houses a plurality of reflectors, a first reflector of the plurality of reflectors to reflect light associated with a first frame of image data at a first angle to the user's eye, and a second reflector of the plurality of reflectors to reflect light associated with a second frame of image data at a second angle to the user's eye.
- a system for displaying virtual content comprises an image-generating source to provide one or more frames of image data in a time-sequential manner, a display assembly to project light rays associated with the one or more frames of image data, the display assembly comprises a first display element corresponding to a first frame-rate and a first bit depth, and a second display element corresponding to a second frame-rate and a second bit depth, and a variable focus element (VFE) configurable to vary a focus of the projected light and transmit the light to the user's eye.
- VFE variable focus element
- a system for displaying virtual content comprises an array of optical fibers to transmit light beams associated with an image to be presented to a user, and a lens coupled to the array of the optical fibers to deflect a plurality of light beams output by the array of optical fibers through a single nodal point, wherein the lens is physically attached to the optical fibers such that a movement of the optical fiber causes the lens to move, and wherein the single nodal point is scanned.
- a virtual reality display system comprises a plurality of optical fibers to generate light beams associated with one or more images to be presented to a user, and a plurality of phase modulators coupled to the plurality of optical fibers to modulate the light beams, wherein the plurality of phase modulators modulate the light in a manner that affects a wavefront generated as a result of the plurality of light beams.
- a system for displaying virtual content to a user comprises a light projection system to project light associated with one or more frames of image data to a user's eyes, the light project system configured to project light corresponding to a plurality of pixels associated with the image data and a processor to modulate a size of the plurality of pixels displayed to the user.
- a system of displaying virtual content to a user comprises an image-generating source to provide one or more frames of image data, a multicore assembly comprising a plurality of multicore fibers to project light associated with the one or more frames of image data, a multicore fiber of the plurality of multicore fibers emitting light in a wavefront, such that the multicore assembly produces an aggregate wavefront of the projected light, and a phase modulator to induce phase delays between the multicore fibers in a manner such that the aggregate wavefront emitted by the multicore assembly is varied, thereby varying a focal distance at which the user perceives the one or more frames of image data.
- a system for displaying virtual content to a user comprises an array of microprojectors to project light beams associated with one or more frames of image data to be presented to the user, wherein the microprojector is configurable to be movable relative to one or more microprojectors of the array of the microprojectors, a frame to house the array of microprojectors, a processor operatively coupled to the one or more microprojectors of the array of microprojectors to control one or more light beams transmitted from the one or more projectors in a manner such that the one or more light beams are modulated as a function of a position of the one or more microprojectors relative to the array of microprojectors, thereby enabling delivery of a lightfield image to the user.
- FIG. 1 illustrates a user's view of augmented reality (AR) through a wearable AR user device, in one illustrated embodiment.
- AR augmented reality
- FIGS. 2A-2E illustrates various embodiments of wearable AR devices.
- FIG. 3 illustrates a cross-sectional view of the human eye, in one illustrated embodiment.
- FIGS. 4A-4D illustrate one or more embodiments of various internal processing components of the wearable AR device.
- FIGS. 5A-5H illustrate embodiments of transmitting focused light to a user through a transmissive beamsplitter substrate.
- FIGS. 6A and 6B illustrate embodiments of coupling a lens element with the transmissive beamsplitter substrate of FIGS. 5A-5H .
- FIGS. 7A and 7B illustrate embodiments of using one or more waveguides to transmit light to a user.
- FIGS. 8A-8Q illustrate embodiments of a diffractive optical element (DOE).
- DOE diffractive optical element
- FIGS. 9A and 9B illustrate a wavefront produced from a light projector, according to one illustrated embodiment.
- FIG. 10 illustrates an embodiment of a stacked configuration of multiple transmissive beamsplitter substrate coupled with optical elements, according to one illustrated embodiment.
- FIGS. 11A-11C illustrate a set of beamlets projected into a user's pupil, according to the illustrated embodiments.
- FIGS. 12A and 12B illustrate configurations of an array of microprojectors, according to the illustrated embodiments.
- FIGS. 13A-13M illustrate embodiments of coupling microprojectors with optical elements, according to the illustrated embodiments.
- FIGS. 14A-14F illustrate embodiments of spatial light modulators coupled with optical elements, according to the illustrated embodiments.
- FIGS. 15A-15C illustrate the use of a wedge type waveguides along with a plurality of light sources, according to the illustrated embodiments.
- FIGS. 16A-160 illustrate embodiments of coupling optical elements to optical fibers, according to the illustrated embodiments.
- FIG. 17 illustrates a notch filter, according to one illustrated embodiment.
- FIG. 18 illustrates a spiral pattern of a fiber scanning display, according to one illustrated embodiment.
- FIGS. 19A-19N illustrate occlusion effects in presenting a darkfield to a user, according to the illustrated embodiments.
- FIGS. 20A-200 illustrate embodiments of various waveguide assemblies, according to the illustrated embodiments.
- FIGS. 21A-21N illustrate various configurations of DOEs coupled to other optical elements, according to the illustrated embodiments.
- FIGS. 22A-22Y illustrate various configurations of freeform optics, according to the illustrated embodiments.
- FIGS. 4A-4D some general componentry options are illustrated.
- various systems, subsystems, and components are presented for addressing the objectives of providing a high-quality, comfortably-perceived display system for human VR and/or AR.
- an AR system user ( 60 ) is depicted wearing a frame ( 64 ) structure coupled to a display system ( 62 ) positioned in front of the eyes of the user.
- a speaker ( 66 ) is coupled to the frame ( 64 ) in the depicted configuration and positioned adjacent the ear canal of the user (in one embodiment, another speaker, not shown, is positioned adjacent the other ear canal of the user to provide for stereo/shapeable sound control).
- the display ( 62 ) is operatively coupled ( 68 ), such as by a wired lead or wireless connectivity, to a local processing and data module ( 70 ) which may be mounted in a variety of configurations, such as fixedly attached to the frame ( 64 ), fixedly attached to a helmet or hat ( 80 ) as shown in the embodiment of FIG. 4B , embedded in headphones, removably attached to the torso ( 82 ) of the user ( 60 ) in a backpack-style configuration as shown in the embodiment of FIG. 4C , or removably attached to the hip ( 84 ) of the user ( 60 ) in a belt-coupling style configuration as shown in the embodiment of FIG. 4D .
- the local processing and data module ( 70 ) may comprise a power-efficient processor or controller, as well as digital memory, such as flash memory, both of which may be utilized to assist in the processing, caching, and storage of data a) captured from sensors which may be operatively coupled to the frame ( 64 ), such as image capture devices (such as cameras), microphones, inertial measurement units, accelerometers, compasses, GPS units, radio devices, and/or gyros; and/or b) acquired and/or processed using the remote processing module ( 72 ) and/or remote data repository ( 74 ), possibly for passage to the display ( 62 ) after such processing or retrieval.
- image capture devices such as cameras
- microphones such as inertial measurement units
- accelerometers compasses
- GPS units GPS units
- radio devices radio devices
- the local processing and data module ( 70 ) may be operatively coupled ( 76 , 78 ), such as via a wired or wireless communication links, to the remote processing module ( 72 ) and remote data repository ( 74 ) such that these remote modules ( 72 , 74 ) are operatively coupled to each other and available as resources to the local processing and data module ( 70 ).
- the remote processing module ( 72 ) may comprise one or more relatively powerful processors or controllers configured to analyze and process data and/or image information.
- the remote data repository ( 74 ) may comprise a relatively large-scale digital data storage facility, which may be available through the internet or other networking configuration in a “cloud” resource configuration. In one embodiment, all data is stored and all computation is performed in the local processing and data module, allowing fully autonomous use from any remote modules.
- FIGS. 5A through 22Y various display configurations are presented that are designed to present the human eyes with photon-based radiation patterns that can be comfortably perceived as augmentations to physical reality, with high-levels of image quality and three-dimensional perception, as well as being capable of presenting two-dimensional content.
- a transmissive beamsplitter substrate ( 104 ) with a 45-degree reflecting surface ( 102 ) directs incoming radiation ( 106 ), which may be output from a lens (not shown), through the pupil ( 45 ) of the eye ( 58 ) and to the retina ( 54 ).
- incoming radiation 106
- the field of view for such a system is limited by the geometry of the beamsplitter ( 104 ).
- a larger field of view can be created by aggregating the outputs/reflections of various different reflective and/or diffractive surfaces and using, e.g., a frame-sequential configuration wherein eye ( 58 ) is presented with a sequence of frames at high frequency that provides the perception of a single coherent scene.
- the reflectors may separate content by other means, such as polarization selectivity or wavelength selectivity.
- the reflectors can relay the three-dimensional wavefronts associated with true-three-dimensional viewing of actual physical objects.
- a substrate ( 108 ) comprising a plurality of reflectors at a plurality of angles ( 110 ) is shown, with each reflector actively reflecting in the depicted configuration for illustrative purposes.
- the reflectors may be switchable elements to facilitate temporal selectivity.
- the reflective surfaces would intentionally be sequentially activated with frame-sequential input information ( 106 ), in which each reflective surface presents a narrow field of view sub-image which is tiled with other narrow field of view sub-images presented by the other reflective surfaces to form a composite wide field of view image.
- frame-sequential input information 106
- surface ( 110 ), about in the middle of substrate ( 108 ), is switched “on” to a reflecting state, such that it reflects incoming image information ( 106 ) to present a relatively narrow field of view sub-image in the middle of a larger field of view, while the other potential reflective surfaces are in a transmissive state.
- FIG. 5C incoming image information ( 106 ) coming from the right of the narrow field of view sub-image (as shown by the angle of incoming beams 106 relative to the substrate 108 input interface 112 , and the resultant angle at which they exit the substrate 108 ) is reflected toward the eye ( 58 ) from reflective surface ( 110 ).
- FIG. 5D illustrates the same reflector ( 110 ) active, with image information coming from the middle of the narrow field of view sub-image, as shown by the angle of the input information ( 106 ) at the input interface ( 112 ) and its angle as it exits substrate ( 108 ).
- FIG. 5E illustrates the same reflector ( 110 ) active, with image information coming from the left of the field of view, as shown by the angle of the input information ( 106 ) at the input interface ( 112 ) and the resultant exit angle at the surface of the substrate ( 108 ).
- FIG. 5F illustrates a configuration wherein the bottom reflector ( 110 ) is active, with image information ( 106 ) coming in from the far right of the overall field of view.
- FIGS. 5C , 5 D, and 5 E can illustrate one frame representing the center of a frame-sequential tiled image
- FIG. 5F can illustrate a second frame representing the far right of that tiled image.
- the light carrying the image information ( 106 ) may strike the reflective surface ( 110 ) directly after entering substrate ( 108 ) at input interface ( 112 ), without first reflecting from the surfaces of substrate ( 108 ).
- the light carrying the image information ( 106 ) may reflect from one or more surfaces of substrate ( 108 ) after entering at input interface ( 112 ) and before striking the reflective surface ( 110 ); for instance, substrate ( 108 ) may act as a planar waveguide, propagating the light carrying image information ( 106 ) by total internal reflection.
- Light may also reflect from one or more surfaces of the substrate ( 108 ) from a partially reflective coating, a wavelength-selective coating, an angle-selective coating, and/or a polarization-selective coating.
- the angled reflectors may be constructed using an electro-active material, such that upon application of a voltage and/or current to a particular reflector, the refractive index of the material comprising such reflector changes from an index substantially matched to the rest of the substrate ( 108 ), in which case the reflector is in a transmissive configuration, to a reflective configuration wherein the refractive index of the reflector mismatches the refractive index of the substrate ( 108 ) such that a reflection effect is created.
- Example electro-active material includes lithium niobate and electro-active polymers.
- Suitable substantially transparent electrodes for controlling a plurality of such reflectors may comprise materials such as indium tin oxide, which is utilized in liquid crystal displays.
- the electro-active reflectors ( 110 ) may comprise liquid crystal, embedded in a substrate ( 108 ) host medium such as glass or plastic.
- liquid crystal may be selected that changes refractive index as a function of an applied electric signal, so that more analog changes may be accomplished as opposed to binary (from one transmissive state to one reflective state).
- an input display that can refresh at the rate of about 360 Hz, with an electro-active reflector array that can keep up with such frequency.
- lithium niobate may be utilized as an electro-active reflective material as opposed to liquid crystal; lithium niobate is utilized in the photonics industry for high-speed switches and fiber optic networks and has the capability to switch refractive index in response to an applied voltage at a very high frequency; this high frequency may be used to steer line-sequential or pixel-sequential sub-image information, especially if the input display is a scanned light display, such as a fiber-scanned display or scanning mirror-based display.
- a variable switchable angled mirror configuration may comprise one or more high-speed mechanically repositionable reflective surfaces, such as a MEMS (micro-electro-mechanical system) device.
- a MEMS device may include what is known as a “digital mirror device”, or “DMD”, (often part of a “digital light processing”, or “DLP” system, such as those available from Texas Instruments, Inc.).
- DMD digital mirror device
- DLP digital light processing
- a plurality of air-gapped (or in vacuum) reflective surfaces could be mechanically moved in and out of place at high frequency.
- a single reflective surface may be moved up and down and re-pitched at very high frequency.
- the switchable variable angle reflector configurations described herein are capable of passing not only collimated or flat wavefront information to the retina ( 54 ) of the eye ( 58 ), but also curved wavefront ( 122 ) image information, as shown in the illustration of FIG. 5G .
- the ability to pass curved wavefront information facilitates the ability of configurations such as those shown in FIGS. 5B-5H to provide the retina ( 54 ) with input perceived as focused at various distances from the eye ( 58 ), not just optical infinity (which would be the interpretation of collimated light absent other cues).
- an array of static partially reflective surfaces ( 116 ) may be embedded in a substrate ( 114 ) with a high-frequency gating layer ( 118 ) controlling outputs to the eye ( 58 ) by only allowing transmission through an aperture ( 120 ) which is controllably movable. In other words, everything may be selectively blocked except for transmissions through the aperture ( 120 ).
- the gating layer ( 118 ) may comprise a liquid crystal array, a lithium niobate array, an array of MEMS shutter elements, an array of DLP DMD elements, or an array of other MEMS devices configured to pass or transmit with relatively high-frequency switching and high transmissibility upon being switched to transmission mode.
- FIGS. 6A-6B other embodiments are depicted wherein arrayed optical elements may be combined with exit pupil expansion configurations to assist with the comfort of the virtual or augmented reality experience of the user.
- exit pupil for the optics configuration, the user's eye positioning relative to the display (which, as in FIGS. 4A-4D , may be mounted on the user's head in an eyeglasses sort of configuration) is not as likely to disrupt his experience—because due to the larger exit pupil of the system, there is a larger acceptable area wherein the user's anatomical pupil may be located to still receive the information from the display system as desired.
- the system is less likely to be sensitive to slight misalignments of the display relative to the user's anatomical pupil, and greater comfort for the user may be achieved through less geometric constraint on his or her relationship with the display/glasses.
- the display ( 140 ) on the left feeds a set of parallel rays into the substrate ( 124 ).
- the display may be a scanned fiber display scanning a narrow beam of light back and forth at an angle as shown to project an image through the lens or other optical element ( 142 ), which may be utilized to collect the angularly-scanned light and convert it to a parallel bundle of rays.
- the rays may be reflected from a series of reflective surfaces ( 126 , 128 , 130 , 132 , 134 , 136 ) which may be configured to partially reflect and partially transmit incoming light so that the light may be shared across the group of reflective surfaces ( 126 , 128 , 130 , 132 , 134 , 136 ) approximately equally.
- the exiting light rays may be steered through a nodal point and scanned out toward the eye ( 58 ) to provide an array of exit pupils, or the functional equivalent of one large exit pupil that is usable by the user as he or she gazes toward the display system.
- a similar set of lenses ( 139 ) may be presented on the opposite side of the waveguide ( 124 ) to compensate for the lower set of lenses; thus creating a the equivalent of a zero-magnification telescope.
- the reflective surfaces ( 126 , 128 , 130 , 132 , 134 , 136 ) each may be aligned at approximately 45 degrees as shown, or may be configured to have different alignments, akin to the configurations of FIGS. 5B-5H , for example).
- the reflective surfaces may comprise wavelength-selective reflectors, band pass reflectors, half silvered mirrors, or other reflective configurations.
- the lenses ( 138 , 139 ) shown are refractive lenses, but diffractive lens elements may also be utilized.
- FIG. 6B a somewhat similar configuration is depicted wherein a plurality of curved reflective surfaces ( 148 , 150 , 152 , 154 , 156 , 158 ) may be utilized to effectively combine the lens (element 138 of FIG. 6A ) and reflector (elements 126 , 128 , 130 , 132 , 134 , 136 of FIG. 6A ) functionality of the embodiment of FIG. 6A , thereby obviating the need for the two groups of lenses (element 138 of FIG. 6A ).
- the curved reflective surfaces may be various curved configurations selected to both reflect and impart angular change, such as parabolic or elliptical curved surfaces.
- a parabolic shape a parallel set of incoming rays will be collected into a single output point; with an elliptical configuration, a set of rays diverging from a single point of origin are collected to a single output point.
- an elliptical configuration a set of rays diverging from a single point of origin are collected to a single output point.
- the curved reflective surfaces ( 148 , 150 , 152 , 154 , 156 , 158 ) preferably are configured to partially reflect and partially transmit so that the incoming light is shared across the length of the waveguide ( 146 ).
- the curved reflective surfaces ( 148 , 150 , 152 , 154 , 156 , 158 ) may comprise wavelength-selective notch reflectors, half silvered mirrors, or other reflective configurations.
- the curved reflective surfaces ( 148 , 150 , 152 , 154 , 156 , 158 ) may be replaced with diffractive reflectors configured to reflect and also deflect.
- perceptions of Z-axis difference may be facilitated by using a waveguide in conjunction with a variable focus optical element configuration.
- image information from a display ( 160 ) may be collimated and injected into a waveguide ( 164 ) and distributed in a large exit pupil manner using, e.g., configurations such as those described in reference to FIGS.
- variable focus optical element capability may be utilized to change the focus of the wavefront of light emerging from the waveguide and provide the eye with the perception that the light coming from the waveguide ( 164 ) is from a particular focal distance.
- variable focus optical element capability since the incoming light has been collimated to avoid challenges in total internal reflection waveguide configurations, it will exit in collimated fashion, requiring a viewer's eye to accommodate to the far point to bring it into focus on the retina, and naturally be interpreted as being from optical infinity—unless some other intervention causes the light to be refocused and perceived as from a different viewing distance; one suitable such intervention is a variable focus lens.
- collimated image information is injected into a piece of glass ( 162 ) or other material at an angle such that it totally internally reflects and is passed into the adjacent waveguide ( 164 ).
- the waveguide ( 164 ) may be configured akin to the waveguides of FIGS. 6A or 6 B ( 124 , 146 , respectively) so that the collimated light from the display is distributed to exit somewhat uniformly across the distribution of reflectors or diffractive features along the length of the waveguide.
- variable focus lens element ( 166 ) Upon exit toward the eye ( 58 ), in the depicted configuration the exiting light is passed through a variable focus lens element ( 166 ) wherein, depending upon the controlled focus of the variable focus lens element ( 166 ), the light exiting the variable focus lens element ( 166 ) and entering the eye ( 58 ) will have various levels of focus (a collimated flat wavefront to represent optical infinity, more and more beam divergence/wavefront curvature to represent closer viewing distance relative to the eye 58 ).
- variable focus lens element ( 166 ) between the eye ( 58 ) and the waveguide ( 164 ) another similar variable focus lens element ( 167 ) is placed on the opposite side of the waveguide ( 164 ) to cancel out the optical effects of the lenses ( 166 ) for light coming from the world ( 144 ) for augmented reality (i.e., as described above, one lens compensates for the other, producing the functional equivalent of a zero-magnification telescope).
- the variable focus lens element ( 166 ) may be a refractive element, such as a liquid crystal lens, an electro-active lens, a conventional refractive lens with moving elements, a mechanical-deformation-based lens (such as a fluid-filled membrane lens, or a lens akin to the human crystalline lens wherein a flexible element is flexed and relaxed by actuators), an electrowetting lens, or a plurality of fluids with different refractive indices.
- a refractive element such as a liquid crystal lens, an electro-active lens, a conventional refractive lens with moving elements, a mechanical-deformation-based lens (such as a fluid-filled membrane lens, or a lens akin to the human crystalline lens wherein a flexible element is flexed and relaxed by actuators), an electrowetting lens, or a plurality of fluids with different refractive indices.
- variable focus lens element ( 166 ) may also comprise a switchable diffractive optical element (such as one featuring a polymer dispersed liquid crystal approach wherein a host medium, such as a polymeric material, has microdroplets of liquid crystal dispersed within the material; when a voltage is applied, the molecules reorient so that their refractive indices no longer match that of the host medium, thereby creating a high-frequency switchable diffraction pattern).
- a switchable diffractive optical element such as one featuring a polymer dispersed liquid crystal approach wherein a host medium, such as a polymeric material, has microdroplets of liquid crystal dispersed within the material; when a voltage is applied, the molecules reorient so that their refractive indices no longer match that of the host medium, thereby creating a high-frequency switchable diffraction pattern).
- One embodiment includes a host medium in which microdroplets of a Kerr effect-based electro-active material, such as lithium niobate, is dispersed within the host medium, enabling refocusing of image information on a pixel-by-pixel or line-by-line basis, when coupled with a scanning light display, such as a fiber-scanned display or scanning-mirror-based display.
- a scanning light display such as a fiber-scanned display or scanning-mirror-based display.
- the pattern spacing may be modulated to not only change the focal power of the variable focus lens element ( 166 ), but also to change the focal power of the overall optical system—for a zoom lens type of functionality.
- the lenses ( 166 ) could be telecentric, in that focus of the display imagery can be altered while keeping magnification constant—in the same way that a photography zoom lens may be configured to decouple focus from zoom position.
- the lenses ( 166 ) may be non-telecentric, so that focus changes will also slave zoom changes. With such a configuration, such magnification changes may be compensated for in software with dynamic scaling of the output from the graphics system in sync with focus changes).
- a stack of sequential two-dimensional images may be fed to the display sequentially to produce three-dimensional perception over time; in a manner akin to the manner in which a computed tomography system uses stacked image slices to represent a three-dimensional structure.
- a series of two-dimensional image slices may be presented to the eye, each at a different focal distance to the eye, and the eye/brain would integrate such a stack into a perception of a coherent three-dimensional volume.
- line-by-line, or even pixel-by-pixel sequencing may be conducted to produce the perception of three-dimensional viewing. For example, with a scanned light display (such as a scanning fiber display or scanning mirror display), then the display is presenting the waveguide ( 164 ) with one line or one pixel at a time in a sequential fashion.
- variable focus lens element ( 166 ) is able to keep up with the high-frequency of pixel-by-pixel or line-by-line presentation, then each line or pixel may be presented and dynamically focused through the variable focus lens element ( 166 ) to be perceived at a different focal distance from the eye ( 58 ).
- Pixel-by-pixel focus modulation generally requires an extremely fast/high-frequency variable focus lens element ( 166 ).
- a 1080 P resolution display with an overall frame rate of 60 frames per second typically presents around 125 million pixels per second.
- Such a configuration also may be constructed using a solid state switchable lens, such as one using an electro-active material, e.g., lithium niobate or an electro-active polymer.
- a frame sequential multi-focal display driving approach may be used in conjunction with a number of the display system and optics embodiments described in this disclosure.
- an electro-active layer ( 172 ) (such as one comprising liquid crystal or lithium niobate) surrounded by functional electrodes ( 170 , 174 ) which may be made of indium tin oxide
- a waveguide ( 168 ) with a conventional transmissive substrate ( 176 , such as one made from glass or plastic with known total internal reflection characteristics and an index of refraction that matches the on or off state of the electro-active layer 172 ) may be controlled such that the paths of entering beams may be dynamically altered to essentially create a time-varying light field.
- a stacked waveguide assembly may be utilized to provide three-dimensional perception to the eye/brain by having a plurality of waveguides ( 182 , 184 , 186 , 188 , 190 ) and a plurality of weak lenses ( 198 , 196 , 194 , 192 ) configured together to send image information to the eye with various levels of wavefront curvature for each waveguide level indicative of focal distance to be perceived for that waveguide level.
- a plurality of displays ( 200 , 202 , 204 , 206 , 208 ), or in another embodiment a single multiplexed display, may be utilized to inject collimated image information into the waveguides ( 182 , 184 , 186 , 188 , 190 ), each of which may be configured, as described above, to distribute incoming light substantially equally across the length of each waveguide, for exit down toward the eye.
- the waveguide ( 182 ) nearest the eye is configured to deliver collimated light, as injected into such waveguide ( 182 ), to the eye, which may be representative of the optical infinity focal plane.
- the next waveguide up ( 184 ) is configured to send out collimated light which passes through the first weak lens ( 192 ; e.g., a weak negative lens) before it can reach the eye ( 58 ); such first weak lens ( 192 ) may be configured to create a slight convex wavefront curvature so that the eye/brain interprets light coming from that next waveguide up ( 184 ) as coming from a first focal plane closer inward toward the person from optical infinity.
- the third up waveguide ( 186 ) passes its output light through both the first ( 192 ) and second ( 194 ) lenses before reaching the eye ( 58 ); the combined optical power of the first ( 192 ) and second ( 194 ) lenses may be configured to create another incremental amount of wavefront divergence so that the eye/brain interprets light coming from that third waveguide up ( 186 ) as coming from a second focal plane even closer inward toward the person from optical infinity than was light from the next waveguide up ( 184 ).
- the other waveguide layers ( 188 , 190 ) and weak lenses ( 196 , 198 ) are similarly configured, with the highest waveguide ( 190 ) in the stack sending its output through all of the weak lenses between it and the eye for an aggregate focal power representative of the closest focal plane to the person.
- a compensating lens layer ( 180 ) is disposed at the top of the stack to compensate for the aggregate power of the lens stack ( 198 , 196 , 194 , 192 ) below.
- Such a configuration provides as many perceived focal planes as there are available waveguide/lens pairings, again with a relatively large exit pupil configuration as described above.
- Both the reflective aspects of the waveguides and the focusing aspects of the lenses may be static (i.e., not dynamic or electro-active). In an alternative embodiment they may be dynamic using electro-active features as described above, enabling a small number of waveguides to be multiplexed in a time sequential fashion to produce a larger number of effective focal planes.
- FIGS. 8B-8N various aspects of diffraction configurations for focusing and/or redirecting collimated beams are depicted. Other aspects of diffraction systems for such purposes are disclosed in U.S. Patent Application Ser. No. 61/845,907 (U.S. patent application Ser. No. 14/331,218), which is incorporated by reference herein in its entirety.
- a linear diffraction pattern such as a Bragg grating
- FIG. 8C illustrates the deflection effect of passing a collimated beam through a linear diffraction pattern ( 210 );
- FIG. 8D illustrates the focusing effect of passing a collimated beam through a radially symmetric diffraction pattern ( 212 ).
- a combination diffraction pattern that has both linear and radial elements ( 214 ) produces both deflection and focusing of a collimated input beam.
- These deflection and focusing effects can be produced in a reflective as well as transmissive mode.
- These principles may be applied with waveguide configurations to allow for additional optical system control, as shown in FIGS. 8G-8N , for example. As shown in FIGS.
- a diffraction pattern ( 220 ), or “diffractive optical element” (or “DOE”) has been embedded within a planar waveguide ( 216 ) such that as a collimated beam is totally internally reflected along the planar waveguide ( 216 ), it intersects the diffraction pattern ( 220 ) at a multiplicity of locations.
- DOE diffractive optical element
- the DOE ( 220 ) has a relatively low diffraction efficiency so that only a portion of the light of the beam is deflected away toward the eye ( 58 ) with each intersection of the DOE ( 220 ) while the rest continues to move through the planar waveguide ( 216 ) via total internal reflection; the light carrying the image information is thus divided into a number of related light beams that exit the waveguide at a multiplicity of locations and the result is a fairly uniform pattern of exit emission toward the eye ( 58 ) for this particular collimated beam bouncing around within the planar waveguide ( 216 ), as shown in FIG. 8H .
- the exit beams toward the eye ( 58 ) are shown in FIG.
- the exit beam pattern is more divergent, which would require the eye to accommodation to a closer distance to bring it into focus on the retina and would be interpreted by the brain as light from a viewing distance closer to the eye than optical infinity.
- a DOE ( 221 ) embedded in this other waveguide ( 218 ), such as a linear diffraction pattern, may function to spread the light across the entire larger planar waveguide ( 216 ), which functions to provide the eye ( 58 ) with a very large incoming field of incoming light that exits from the larger planar waveguide ( 216 ), i.e., a large eye box, in accordance with the particular DOE configurations at work.
- the DOEs ( 220 , 221 ) are depicted bisecting the associated waveguides ( 216 , 218 ) but this need not be the case; they could be placed closer to, or upon, either side of either of the waveguides ( 216 , 218 ) to have the same functionality.
- FIG. 8K with the injection of a single collimated beam, an entire field of cloned collimated beams may be directed toward the eye ( 58 ).
- a combined linear diffraction pattern/radially symmetric diffraction pattern scenario such as that depicted in FIGS.
- a beam distribution waveguide optic for functionality such as exit pupil functional expansion; with a configuration such as that of FIG. 8K , the exit pupil can be as large as the optical element itself, which can be a very significant advantage for user comfort and ergonomics
- Z-axis focusing capability is presented, in which both the divergence angle of the cloned beams and the wavefront curvature of each beam represent light coming from a point closer than optical infinity.
- one or more DOEs are switchable between “on” states in which they actively diffract, and “off” states in which they do not significantly diffract.
- a switchable DOE may comprise a layer of polymer dispersed liquid crystal, in which microdroplets comprise a diffraction pattern in a host medium, and the refractive index of the microdroplets can be switched to substantially match the refractive index of the host material (in which case the pattern does not appreciably diffract incident light) or the microdroplet can be switched to an index that does not match that of the host medium (in which case the pattern actively diffracts incident light).
- the diffraction terms such as the linear diffraction pitch term as in FIGS.
- a beam scanning or tiling functionality may be achieved.
- it is desirable to have a relatively low diffraction grating efficiency in each of the DOEs ( 220 , 221 ) because it facilitates distribution of the light, and also because light coming through the waveguides that is desirably transmitted (for example, light coming from the world 144 toward the eye 58 in an augmented reality configuration) is less affected when the diffraction efficiency of the DOE that it crosses ( 220 ) is lower—so a better view of the real world through such a configuration is achieved.
- Configurations such as those illustrated in FIG. 8K preferably are driven with injection of image information in a time sequential approach, with frame sequential driving being the most straightforward to implement.
- an image of the sky at optical infinity may be injected at time 1 and the diffraction grating retaining collimation of light may be utilized; then an image of a closer tree branch may be injected at time 2 while a DOE controllably imparts a focal change, say one diopter or 1 meter away, to provide the eye/brain with the perception that the branch light information is coming from the closer focal range.
- This kind of paradigm can be repeated in rapid time sequential fashion such that the eye/brain perceives the input to be all part of the same image.
- This kind of configuration generally assumes that the DOE is switched at a relatively low speed (i.e., in sync with the frame-rate of the display that is injecting the images—in the range of tens to hundreds of cycles/second).
- the opposite extreme may be a configuration wherein DOE elements can shift focus at tens to hundreds of MHz or greater, which facilitates switching of the focus state of the DOE elements on a pixel-by-pixel basis as the pixels are scanned into the eye ( 58 ) using a scanned light display type of approach.
- This is desirable because it means that the overall display frame-rate can be kept quite low; just low enough to make sure that “flicker” is not a problem (in the range of about 60-120 frames/sec).
- the DOEs can be switched at KHz rates, then on a line-by-line basis the focus on each scan line may be adjusted, which may afford the user with a visible benefit in terms of temporal artifacts during an eye motion relative to the display, for example.
- the different focal planes in a scene may, in this manner, be interleaved, to minimize visible artifacts in response to a head motion (as is discussed in greater detail later in this disclosure).
- a line-by-line focus modulator may be operatively coupled to a line scan display, such as a grating light valve display, in which a linear array of pixels is swept to form an image; and may be operatively coupled to scanned light displays, such as fiber-scanned displays and mirror-scanned light displays.
- a line scan display such as a grating light valve display, in which a linear array of pixels is swept to form an image
- scanned light displays such as fiber-scanned displays and mirror-scanned light displays.
- a stacked configuration may use dynamic DOEs (rather than the static waveguides and lenses of the embodiment of FIG. 8A ) to provide multi-planar focusing simultaneously.
- dynamic DOEs rather than the static waveguides and lenses of the embodiment of FIG. 8A
- a primary focus plane (based upon measured eye accommodation, for example) could be presented to the user, and a +margin and ⁇ margin (i.e., one focal plane closer, one farther out) could be utilized to provide a large focal range in which the user can accommodate before the planes need be updated.
- This increased focal range can provide a temporal advantage if the user switches to a closer or farther focus (i.e., as determined by accommodation measurement); then the new plane of focus could be made to be the middle depth of focus, with the + and ⁇ margins again ready for a fast switchover to either one while the system catches up.
- a stack ( 222 ) of planar waveguides ( 244 , 246 , 248 , 250 , 252 ) is shown, each having a reflector ( 254 , 256 , 258 , 260 , 262 ) at the end and being configured such that collimated image information injected in one end by a display ( 224 , 226 , 228 , 230 , 232 ) bounces by total internal reflection down to the reflector, at which point some or all of the light is reflected out toward an eye or other target.
- Each of the reflectors may have slightly different angles so that they all reflect exiting light toward a common destination such as a pupil. Such a configuration is somewhat similar to that of FIG.
- each different angled reflector in the embodiment of FIG. 8O has its own waveguide for less interference when projected light is travelling to the targeted reflector.
- Lenses ( 234 , 236 , 238 , 240 , 242 ) may be interposed between the displays and waveguides for beam steering and/or focusing.
- FIG. 8P illustrates a geometrically staggered version wherein reflectors ( 276 , 278 , 280 , 282 , 284 ) are positioned at staggered lengths in the waveguides ( 266 , 268 , 270 , 272 , 274 ) so that exiting beams may be relatively easily aligned with objects such as an anatomical pupil.
- the geometries of the reflectors ( 276 , 278 , 280 , 282 , 284 ) and waveguides ( 266 , 268 , 270 , 272 , 274 ) may be set up to fill the eye pupil (typically about 8 mm across or less) with exiting light.
- the viewer may make eye movements while retaining the ability to see the displayed imagery. Referring back to the discussion related to FIGS. 5A and 5B about field of view expansion and reflector size, an expanded field of view is presented by the configuration of FIG. 8P as well, and it does not involve the complexity of the switchable reflective elements of the embodiment of FIG. 5B .
- FIG. 8Q illustrates a version wherein many reflectors ( 298 ) form a relatively continuous curved reflection surface in the aggregate or discrete flat facets that are oriented to align with an overall curve.
- the curve could a parabolic or elliptical curve and is shown cutting across a plurality of waveguides ( 288 , 290 , 292 , 294 , 296 ) to minimize any crosstalk issues, although it also could be utilized with a monolithic waveguide configuration.
- a high-frame-rate and lower persistence display may be combined with a lower-frame-rate and higher persistence display and a variable focus element to comprise a relatively high-frequency frame sequential volumetric display.
- the high-frame-rate display has a lower bit depth and the lower-frame-rate display has a higher bit depth, and are combined to comprise an effective high-frame-rate and high bit depth display, that is well suited to presenting image slices in a frame sequential fashion.
- a three-dimensional volume that is desirably represented is functionally divided into a series of two-dimensional slices. Each of those two-dimensional slices is projected to the eye frame sequentially, and in sync with this presentation, the focus of a variable focus element is changed.
- two display elements may be integrated: a full-color, high-resolution liquid crystal display (“LCD”; a backlighted ferroelectric panel display also may be utilized in another embodiment; in a further embodiment a scanning fiber display may be utilized) operating at 60 frames per second, and aspects of a higher-frequency DLP system.
- LCD liquid crystal display
- a backlighted ferroelectric panel display also may be utilized in another embodiment; in a further embodiment a scanning fiber display may be utilized
- the conventional lighting configuration may be removed to accommodate using the DLP projector to project a mask pattern on the back of the LCD (in one embodiment, the mask pattern may be binary in that the DLP either projects illumination, or not-illumination; in another embodiment described below, the DLP may be utilized to project a grayscale mask image).
- DLP projection systems can operate at very high frame rates; in one embodiment for 6 depth planes at 60 frames per second, a DLP projection system can be operated against the back of the LCD display at 360 frames/second. Then the DLP projector is utilized to selectively illuminate portions of the LCD panel in sync with a high-frequency variable focus element (such as a deformable membrane mirror) that is disposed between the viewing side of the LCD panel and the eye of the user, the variable focus element being used to change the global display focus on a frame by frame basis at 360 frames/second.
- a high-frequency variable focus element such as a deformable membrane mirror
- variable focus element is positioned to be optically conjugate to the exit pupil, to enable adjustments of focus without simultaneously affecting image magnification or “zoom.”
- variable focus element is not conjugate to the exit pupil, such that image magnification changes accompany focus adjustments, and software is used to compensate for these optical magnification changes and any distortions by pre-scaling or warping the images to be presented.
- the system may be configured to present on an LCD a full-color, all in-focus image of the tree branch in front the sky.
- the DLP projector in a binary masking configuration (i.e., illumination or absence of illumination) may be used to only illuminate the portion of the LCD that represents the cloudy sky while functionally black-masking (i.e., failing to illuminate) the portion of the LCD that represents the tree branch and other elements that are not to be perceived at the same focal distance as the sky, and the variable focus element (such as a deformable membrane mirror) may be utilized to position the focal plane at optical infinity so that the eye sees a sub-image at subframe 1 as being clouds that are infinitely far away.
- a binary masking configuration i.e., illumination or absence of illumination
- functionally black-masking i.e., failing to illuminate
- the variable focus element such as a deformable membrane mirror
- the variable focus element may be switched to focusing on a point about 1 meter away from the user's eyes (or whatever distance is required; here 1 meter for the branch location is used for illustrative purposes), the pattern of illumination from the DLP can be switched so that the system only illuminates the portion of the LCD that represents the tree branch while functionally black-masking (i.e., failing to illuminate) the portion of the LCD that represents the sky and other elements that are not to be perceived at the same focal distance as the tree branch.
- the eye gets a quick flash of cloud at optical infinity followed by a quick flash of tree at 1 meter, and the sequence is integrated by the eye/brain to form a three-dimensional perception.
- the branch may be positioned diagonally relative to the viewer, such that it extends through a range of viewing distances, e.g., it may join with the trunk at around 2 meters viewing distance while the tips of the branch are at the closer position of 1 meter.
- the display system can divide the 3-D volume of the tree branch into multiple slices, rather than a single slice at 1 meter.
- one focus slice may be used to represent the sky (using the DLP to mask all areas of the tree during presentation of this slice), while the tree branch is divided across 5 focus slices (using the DLP to mask the sky and all portions of the tree except one, for each part of the tree branch to be presented).
- the depth slices are positioned with a spacing equal to or smaller than the depth of focus of the eye, such that the viewer will be unlikely to notice the transition between slices, and instead perceive a smooth and continuous flow of the branch through the focus range.
- the DLP in a binary (illumination or darkfield only) mode, it may be utilized to project a grayscale (for example, 256 shades of grayscale) mask onto the back of the LCD panel to enhance three-dimensional perception.
- the grayscale shades may be utilized to impart to the eye/brain a perception that something resides in between adjacent depth or focal planes.
- that portion branch on the LCD may be lit up with full intensity white from the DLP system with the variable focus element at focalplane 1 .
- grayscale masking can be utilized.
- the DLP can project an illumination mask to that portion during both subframe 1 and subframe 2 , but at half-illumination (such as at level 128 out of 256 grayscale) for each subframe. This provides the perception of a blending of depth of focus layers, with the perceived focal distance being proportional to the illuminance ratio between subframe 1 and subframe 2 .
- an about 25% intensity grayscale mask can be used to illuminate that portion of the LCD at subframe 1 and an about 75% grayscale mask can be used to illuminate the same portion of the LCD at subframe 2 .
- bit depths of both the low-frame-rate display and the high-frame-rate display can be combined for image modulation, to create a high dynamic range display.
- the high dynamic range driving may be conducted in tandem with the focus plane addressing function described above, to comprise a high dynamic range multi-focal 3-D display.
- only a certain portion of the display (i.e., LCD) output may be mask-illuminated by the DMD and variably focused en route to the user's eye.
- the middle portion of the display may be mask illuminated, with the periphery of the display not providing varying accommodation cues to the user (i.e. the periphery could be uniformly illuminated by the DLP DMD, while a central portion is actively masked and variably focused en route to the eye).
- a refresh rate of about 360 Hz allows for 6 depth planes at about 60 frames/second each.
- even higher refresh rates may be achieved by increasing the operating frequency of the DLP.
- a standard DLP configuration uses a MEMS device and an array of micro-mirrors that toggle between a mode of reflecting light toward the display or user to a mode of reflecting light away from the display or user, such as into a light trap—thus they are inherently binary.
- DLPs typically create grayscale images using a pulse width modulation schema wherein the mirror is left in the “on” state for a variable amount of time for a variable duty cycle in order to create a brighter pixel, or pixel of interim brightness.
- pulse width modulation schema wherein the mirror is left in the “on” state for a variable amount of time for a variable duty cycle in order to create a brighter pixel, or pixel of interim brightness.
- a typical pulse width modulation scheme for a Texas Instruments DLP system has an 8-bit command signal (first bit is the first long pulse of the mirror; second bit is a pulse that is half as long as the first; third bit is half as long again; and so on)—so that the configuration can create 2 to the 8th power different illumination levels.
- the backlighting from the DLP may have its intensity varied in sync with the different pulses of the DMD to equalize the brightness of the subimages that are created, which is a practical workaround to get existing DMD drive electronics to produce significantly higher frame rates.
- direct control changes to the DMD drive electronics and software may be utilized to have the mirrors always have an equal on-time instead of the variable on-time configuration that is conventional, which would facilitate higher frame rates.
- the DMD drive electronics may be configured to present low bit depth images at a frame rate above that of high bit depth images but lower than the binary frame rate, enabling some grayscale blending between focus planes, while moderately increasing the number of focus planes.
- the virtual monster when limited to a finite number of depth planes, such as 6 in the example above, it is desirable to functionally move these 6 depth planes around to be maximally useful in the scene that is being presented to the user. For example, if a user is standing in a room and a virtual monster is to be placed into his augmented reality view, the virtual monster being about 2 feet deep in the Z axis straight away from the user's eyes, then it makes sense to cluster all 6 depth planes around the center of the monster's current location (and dynamically move them with him as he moves relative to the user)—so that more rich accommodation cues may be provided for the user, with all six depth planes in the direct region of the monster (for example, 3 in front of the center of the monster, 3 in back of the center of the monster). Such allocation of depth planes is content dependent.
- the same monster is to be presented in the same room, but also to be presented to the user is a virtual window frame element, and then a virtual view to optical infinity out of the virtual window frame, it will be useful to spend at least one depth plane on optical infinity, one on the depth of the wall that is to house the virtual window frame, and then perhaps the remaining four depth planes on the monster in the room. If the content causes the virtual window to disappear, then the two depth planes may be dynamically reallocated to the region around the monster, and so on—content-based dynamic allocation of focal plane resources to provide the most rich experience to the user given the computing and presentation resources.
- phase delays in a multicore fiber or an array of single-core fibers may be utilized to create variable focus light wavefronts.
- a multicore fiber ( 300 ) may comprise the aggregation of multiple individual fibers ( 302 );
- FIG. 9B shows a close-up view of a multicore assembly, which emits light from each core in the form of a spherical wavefront ( 304 ) from each.
- these small spherical wavefronts ultimately constructively and destructively interfere with each other, and if they were emitted from the multicore fiber in phase, they will develop an approximately planar wavefront ( 306 ) in the aggregate, as shown.
- phase delays are induced between the cores (using a conventional phase modulator such as one using lithium niobate, for example, to slow the path of some cores relative to others)
- a curved or spherical wavefront may be created in the aggregate, to represent at the eyes/brain an object coming from a point closer than optical infinity, which presents another option that may be used in place of the variable focus elements described above.
- phased multicore configuration, or phased array may be utilized to create multiple optical focus levels from a light source.
- a known Fourier transform aspect of multi-mode optical fiber or light guiding rods or pipes may be utilized for control of the wavefronts that are output from such fiber.
- Optical fibers typically are available in two categories: single mode and multi-mode.
- Multi-mode optical fiber typically has larger core diameters and allows light to propagate along multiple angular paths, rather than just the one of single mode optical fiber. It is known that if an image is injected into one end of a multi-mode fiber, that angular differences that are encoded into that image will be retained to some degree as it propagates through the multi-mode fiber, and for some configurations the output from the fiber will be significantly similar to a Fourier transform of the image that was input.
- the inverse Fourier transform of a wavefront may be input so that, after passing through the fiber that optically imparts a Fourier transform, the output is the desired shaped, or focused, wavefront.
- a wavefront such as a diverging spherical wavefront to represent a focal plane nearer to the user than optical infinity
- the output is the desired shaped, or focused, wavefront.
- Such output end may be scanned about to be used as a scanned fiber display, or may be used as a light source for a scanning mirror to form an image, for instance.
- Such a configuration may be utilized as yet another focus modulation subsystem.
- Other kinds of light patterns and wavefronts may be injected into a multi-mode fiber, such that on the output end, a certain spatial pattern is emitted.
- the Fourier transform of a hologram may be injected into the input end of a multi-mode fiber to output a wavefront that may be used for three-dimensional focus modulation and/or resolution enhancement.
- Certain single fiber core, multi-core fibers, or concentric core+cladding configurations also may be utilized in the aforementioned inverse Fourier transform configurations.
- a system may be configured to monitor the user's accommodation and rather than presenting a set of multiple different light wavefronts, present a single wavefront at a time that corresponds to the accommodation state of the eye.
- Accommodation may be measured directly (such as by infrared autorefractor or eccentric photorefraction) or indirectly (such as by measuring the convergence level of the two eyes of the user; as described above, vergence and accommodation are strongly linked neurologically, so an estimate of accommodation can be made based upon vergence geometry).
- the wavefront presentations at the eye may be configured for a 1 meter focal distance using any of the above variable focus configurations. If an accommodation change to focus at 2 meters is detected, the wavefront presentation at the eye may be reconfigured for a 2 meter focal distance, and so on.
- variable focus element may be placed in the optical path between an outputting combiner (e.g., a waveguide or beamsplitter) and the eye of the user, so that the focus may be changed along with (i.e., preferably at the same rate as) accommodation changes of the eye.
- Software effects may be utilized to produce variable amounts blur (e.g., Gaussian) to objects which should not be in focus to simulate the dioptric blur expected at the retina if an object were at that viewing distance and enhance the three-dimensional perception by the eyes/brain.
- a simple embodiment is a single plane whose focus level is slaved to the viewer's accommodation level, however the performance demands on the accommodation tracking system can be relaxed if even a low number of multiple planes are used.
- a stack ( 328 ) of about 3 waveguides ( 318 , 320 , 322 ) may be utilized to create three focal planes worth of wavefronts simultaneously.
- the weak lenses ( 324 , 326 ) may have static focal distances, and a variable focal lens ( 316 ) may be slaved to the accommodation tracking of the eyes such that one of the three waveguides (say the middle waveguide 320 ) outputs what is deemed to be the in-focus wavefront, while the other two waveguides ( 322 , 318 ) output a +margin wavefront and a ⁇ margin wavefront (i.e., a little farther than detected focal distance, a little closer than detected focal distance) which may improve the three-dimensional perception and also provide enough difference for the brain/eye accommodation control system to sense some blur as negative feedback, which enhances the perception of reality, and allows a range of accommodation before an physical adjustment of the focus levels is necessary.
- a variable focal lens ( 316 ) may be slaved to the accommodation tracking of the eyes such that one of the three waveguides (say the middle waveguide 320 ) outputs what is deemed to be the in-focus wavefront, while the other two waveguides (
- variable focus compensating lens ( 314 ) is also shown to ensure that light coming in from the real world ( 144 ) in an augmented reality configuration is not refocused or magnified by the assembly of the stack ( 328 ) and output lens ( 316 ).
- the variable focus in the lenses ( 316 , 314 ) may be achieved, as discussed above, with refractive, diffractive, or reflective techniques.
- each of the waveguides in a stack may contain their own capability for changing focus (such as by having an included electronically switchable DOE) so that the variable focus element need not be centralized as in the stack ( 328 ) of the configuration of FIG. 10 .
- variable focus elements may be interleaved between the waveguides of a stack (i.e., rather than fixed focus weak lenses as in the embodiment of FIG. 10 ) to obviate the need for a combination of fixed focus weak lenses plus whole-stack-refocusing variable focus element.
- stacking configurations may be used in accommodation tracked variations as described herein, and also in a frame-sequential multi-focal display approach.
- a pinhole lens configuration wherein the beam is always interpreted as in-focus by the eyes/brain—e.g., a scanned light display using a 0.5 mm diameter beam to scan images to the eye.
- accommodation tracking input may be utilized to induce blur using software to image information that is to be perceived as at a focal plane behind or in front of the focal plane determined from the accommodation tracking.
- simulated dioptric blur may be induced with software, and may be slaved to the accommodation tracking status.
- a scanning fiber display is well suited to such configuration because it may be configured to only output small-diameter beams in a Maxwellian form.
- an array of small exit pupils may be created to increase the functional eye box of the system (and also to reduce the impact of a light-blocking particle which may reside in the vitreous or cornea of the eye), such as by one or more scanning fiber displays, or by a DOE configuration such as that described in reference to FIG. 8K , with a pitch in the array of presented exit pupils that ensure that only one will hit the anatomical pupil of the user at any given time (for example, if the average anatomical pupil diameter is 4 mm, one configuration may comprise 1 ⁇ 2 mm exit pupils spaced at intervals of approximate 4 mm apart).
- Such exit pupils may also be switchable in response to eye position, such that only the eye always receives one, and only one, active small exit pupil at a time; allowing a denser array of exit pupils.
- Such user will have a large depth of focus to which software-based blur techniques may be added to enhance perceived depth perception.
- an object at optical infinity creates a substantially planar wavefront
- an object closer such as 1 m away from the eye
- the eye's optical system needs to have enough optical power to bend the incoming rays of light so that they end up focused on the retina (convex wavefront gets turned into concave, and then down to a focal point on the retina).
- light directed to the eye has been treated as being part of one continuous wavefront, some subset of which would hit the pupil of the particular eye.
- light directed to the eye may be effectively discretized or broken down into a plurality of beamlets or individual rays, each of which has a diameter less than about 0.5 mm and a unique propagation pathway as part of a greater aggregated wavefront that may be functionally created with the an aggregation of the beamlets or rays.
- a curved wavefront may be approximated by aggregating a plurality of discrete neighboring collimated beams, each of which is approaching the eye from an appropriate angle to represent a point of origin that matches the center of the radius of curvature of the desired aggregate wavefront.
- each individual beamlet is always in relative focus on the retina, independent of the accommodation state of the eye—however the trajectory of each beamlet will be affected by the accommodation state. For instance, if the beamlets approach the eye in parallel, representing a discretized collimated aggregate wavefront, then an eye that is correctly accommodated to infinity will deflect the beamlets to all converge upon the same shared spot on the retina, and will appear in focus. If the eye accommodates to, say, 1 m, the beams will be converged to a spot in front of the retina, cross paths, and fall on multiple neighboring or partially overlapping spots on the retina—appearing blurred.
- the beamlets approach the eye in a diverging configuration, with a shared point of origin 1 meter from the viewer, then an accommodation of 1 m will steer the beams to a single spot on the retina, and will appear in focus; if the viewer accommodates to infinity, the beamlets will converge to a spot behind the retina, and produce multiple neighboring or partially overlapping spots on the retina, producing a blurred image.
- the accommodation of the eye determines the degree of overlap of the spots on the retina, and a given pixel is “in focus” when all of the spots are directed to the same spot on the retina and “defocused” when the spots are offset from one another.
- a set of multiple narrow beams may be used to emulate what is going on with a larger diameter variable focus beam, and if the beamlet diameters are kept to a maximum of about 0.5 mm, then they maintain a relatively static focus level, and to produce the perception of out-of-focus when desired, the beamlet angular trajectories may be selected to create an effect much like a larger out-of-focus beam (such a defocussing treatment may not be the same as a Gaussian blur treatment as for the larger beam, but will create a multimodal point spread function that may be interpreted in a similar fashion to a Gaussian blur).
- the beamlets are not mechanically deflected to form this aggregate focus effect, but rather the eye receives a superset of many beamlets that includes both a multiplicity of incident angles and a multiplicity of locations at which the beamlets intersect the pupil; to represent a given pixel from a particular viewing distance, a subset of beamlets from the superset that comprise the appropriate angles of incidence and points of intersection with the pupil (as if they were being emitted from the same shared point of origin in space) are turned on with matching color and intensity, to represent that aggregate wavefront, while beamlets in the superset that are inconsistent with the shared point of origin are not turned on with that color and intensity (but some of them may be turned on with some other color and intensity level to represent, e.g., a different pixel).
- each of a multiplicity of incoming beamlets ( 332 ) is passing through a small exit pupil ( 330 ) relative to the eye ( 58 ) in a discretized wavefront display configuration.
- a subset ( 334 ) of the group of beamlets ( 332 ) may be driven with matching color and intensity levels to be perceived as though they are part of the same larger-sized ray (the bolded subgroup 334 may be deemed an “aggregated beam”).
- the subset of beamlets are parallel to one another, representing a collimated aggregate beam from optical infinity (such as light coming from a distant mountain).
- the eye is accommodated to infinity, so the subset of beamlets are deflected by the eye's cornea and lens to all fall substantially upon the same location of the retina and are perceived to comprise a single in focus pixel.
- FIG. 11C shows another subset of beamlets representing an aggregated collimated beam ( 336 ) coming in from the right side of the field of view of the user's eye ( 58 ) if the eye ( 58 ) is viewed in a coronal-style planar view from above. Again, the eye is shown accommodated to infinity, so the beamlets fall on the same spot of the retina, and the pixel is perceived to be in focus. If, in contrast, a different subset of beamlets were chosen that were reaching the eye as a diverging fan of rays, those beamlets would not fall on the same location of the retina (and be perceived as in focus) until the eye were to shift accommodation to a near point that matches the geometrical point of origin of that fan of rays.
- patterns of points of intersection of beamlets with the anatomical pupil of the eye i.e., the pattern of exit pupils
- they may be organized in configurations such as a cross-sectionally efficient hex-lattice (for example, as shown in FIG. 12A ) or a square lattice or other two-dimensional array.
- a three-dimensional array of exit pupils could be created, as well as time-varying arrays of exit pupils.
- Discretized aggregate wavefronts may be created using several configurations, such as an array of microdisplays or microprojectors placed optically conjugate with the exit pupil of viewing optics, microdisplay or microprojector arrays coupled to a direct field of view substrate (such as an eyeglasses lens) such that they project light to the eye directly, without additional intermediate viewing optics, successive spatial light modulation array techniques, or waveguide techniques such as those described in relation to FIG. 8K .
- a direct field of view substrate such as an eyeglasses lens
- a lightfield may be created by bundling a group of small projectors or display units (such as scanned fiber displays).
- FIG. 12A depicts a hexagonal lattice projection bundle ( 338 ) which may, for example, create a 7 mm-diameter hex array with each fiber display outputting a sub-image ( 340 ). If such an array has an optical system, such as a lens, placed in front of it such that the array is placed optically conjugate with the eye's entrance pupil, this will create an image of the array at the eye's pupil, as shown in FIG. 12B , which essentially provides the same optical arrangement as the embodiment of FIG. 11A .
- Each of the small exit pupils of the configuration is created by a dedicated small display in the bundle ( 338 ), such as a scanning fiber display. Optically, it's as though the entire hex array ( 338 ) is positioned right into the anatomical pupil ( 45 ).
- Such embodiments are means for driving different subimages to different small exit pupils within the larger anatomical entrance pupil ( 45 ) of the eye, comprising a superset of beamlets with a multiplicity of incident angles and points of intersection with the eye pupil.
- Each of the separate projectors or displays may be driven with a slightly different image, such that subimages may be created that pull out different sets of rays to be driven at different light intensities and colors.
- a strict image conjugate may be created, as in the embodiment of FIG. 12B , wherein there is direct 1-to-1 mapping of the array ( 338 ) with the pupil ( 45 ).
- the spacing may be changed between displays in the array and the optical system (lens 342 , in FIG. 12B ) so that instead of getting a conjugate mapping of the array to the eye pupil, the eye pupil may be catching the rays from the array at some other distance.
- FIG. 13A another lightfield creating embodiment is depicted wherein an array of microdisplays or microprojectors ( 346 ) may be coupled to a frame ( 344 ; such as an eyeglasses frame) to be positioned in front of the eye ( 58 ).
- the depicted configuration is a nonconjugate arrangement wherein there are no large-scale optical elements interposed between the displays (for example, scanning fiber displays) of the array ( 346 ) and the eye ( 58 ).
- the displays for example, scanning fiber displays
- Each display may be configured to create a set of rays representing different elements of the beamlet superset.
- FIG. 13B illustrates a nonconjugate configuration similar to that of FIG. 13A , with the exception that the embodiment of FIG. 13B features a reflecting surface ( 348 ) to facilitate moving the display array ( 346 ) away from the eye's ( 58 ) field of view, while also allowing views of the real world ( 144 ) through the reflective surface ( 348 ).
- Scanning fiber displays which may be utilized as displays may have baseline diameters in the range of 1 mm, but reduction in enclosure and projection lens hardware may decrease the diameters of such displays to about 0.5 mm or less, which is less disturbing for a user.
- Another downsizing geometric refinement may be achieved by directly coupling a collimating lens (which may, for example, comprise a gradient refractive index, or “GRIN”, lens, a conventional curved lens, or a diffractive lens) to the tip of the scanning fiber itself in a case of a fiber scanning display array.
- GRIN gradient refractive index
- a GRIN lens ( 354 ) is shown fused to the end of a single mode optical fiber.
- An actuator ( 350 ; such as a piezoelectric actuator) is coupled to the fiber ( 352 ) and may be used to scan the fiber tip.
- the end of the fiber may be shaped into a hemispherical shape using a curved polishing treatment of an optical fiber to create a lensing effect.
- a standard refractive lens may be coupled to the end of each optical fiber using an adhesive.
- a lens may be built from a dab of transmissive polymeric material or glass, such as epoxy.
- the end of an optical fiber may be melted to create a curved surface for a lensing effect.
- FIG. 13C-2 shows an embodiment wherein display configurations (i.e., scanning fiber displays with GRIN lenses; shown in close-up view of FIG. 13C-1 ) such as that shown in FIG. 13D may be coupled together through a single transparent substrate ( 356 ) preferably having a refractive index that closely matches the cladding of the optical fibers ( 352 ) so that the fibers themselves are not very visible for viewing of the outside world across the depicted assembly (if the index matching of the cladding is done precisely, then the larger cladding/housing becomes transparent and only the tiny cores, which preferably are about 3 microns in diameter, will be obstructing the view.
- the matrix ( 358 ) of displays may all be angled inward so they are directed toward the anatomic pupil of the user (in another embodiment, they may stay parallel to each other, but such a configuration is less efficient).
- a thin series of planar waveguides ( 358 ) are configured to be cantilevered relative to a larger substrate structure ( 356 ).
- the substrate ( 356 ) may be moved to produce cyclic motion (i.e., at the resonant frequency of the cantilevered members 358 ) of the planar waveguides relative to the substrate structure.
- the cantilevered waveguide portions ( 358 ) may be actuated with piezoelectric or other actuators relative to the substrate.
- Image illumination information may be injected, for example, from the right side ( 360 ) of the substrate structure to be coupled into the cantilevered waveguide portions ( 358 ).
- the substrate ( 356 ) may comprise a waveguide configured (such as with an integrated DOE configuration as described above) to totally internally reflect incoming light ( 360 ) along its length and then redirect it to the cantilevered waveguide portions ( 358 ).
- the planar waveguides are configured to minimize any dispersion and/or focus changes with their planar shape factors.
- one way to gain further angular and spatial diversity is to use a multicore fiber and place a lens at the exit point, such as a GRIN lens, so that the exit beams are deflected through a single nodal point ( 366 ); that nodal point may then be scanned back and forth in a scanned fiber type of arrangement (such as by a piezoelectric actuator 368 ). If a retinal conjugate is placed at the plane defined at the end of the GRIN lens, a display may be created that is functionally equivalent to the general case discretized aggregate wavefront configuration described above.
- a similar effect may be achieved not by using a lens, but by scanning the face of a multicore system at the correct conjugate of an optical system ( 372 ), the goal being to create a higher angular and spatial diversity of beams.
- some of this requisite angular and spatial diversity may be created through the use of multiple cores to create a plane which may be relayed by a waveguide.
- a multicore fiber ( 362 ) may be scanned (such as by a piezoelectric actuator 368 ) to create a set of beamlets with a multiplicity of angles of incidence and points of intersection which may be relayed to the eye ( 58 ) by a waveguide ( 370 ).
- a collimated lightfield image may be injected into a waveguide, and without any additional refocusing elements, that lightfield display may be translated directly to the human eye.
- FIGS. 131-13L depict certain commercially available multicore fiber ( 362 ) configurations (from vendors such as Mitsubishi Cable Industries, Ltd. of Japan), including one variation ( 363 ) with a rectangular cross section, as well as variations with flat exit faces ( 372 ) and angled exit faces ( 374 ).
- some additional angular diversity may be created by having a waveguide ( 376 ) fed with a linear array of displays ( 378 ), such as scanning fiber displays.
- FIGS. 14A-14F another group of configurations for creating a fixed viewpoint lightfield display is described.
- FIG. 11A if a two-dimensional plane was created that was intersecting all of the tiny beams coming in from the left, each beamlet would have a certain point of intersection with that plane. If another plane was created at a different distance to the left, then all of the beamlets would intersect that plane at a different location. Then going back to FIG. 14A , if various positions on each of two or more planes can be allowed to selectively transmit or block the light radiation directed through it, such a multi-planar configuration may be utilized to selectively create a lightfield by independently modulating individual beamlets.
- FIG. 14A shows two spatial light modulators, such as liquid crystal display panels ( 380 , 382 ; in other embodiments they may be MEMS shutter displays or DLP DMD arrays) which may be independently controlled to block or transmit different rays on a high-resolution basis.
- two spatial light modulators such as liquid crystal display panels ( 380 , 382 ; in other embodiments they may be MEMS shutter displays or DLP DMD arrays) which may be independently controlled to block or transmit different rays on a high-resolution basis.
- a set of beamlets with a multiplicity of angles and points of intersection may be created using a plurality of stacked SLMs. Additional numbers of SLMs beyond two provides more opportunities to control which beams are selectively attenuated.
- planes of DMD devices from DLP systems may be stacked to function as SLMs, and may be preferred over liquid crystal systems as SLMs due to their ability to more efficiently pass light (with a mirror element in a first state, reflectivity to the next element on the way to the eye may be quite efficient; with a mirror element in a second state, the mirror angle may be moved by an angle such as 12 degrees to direct the light away from the path to the eye).
- two DMDs may be utilized in series with a pair of lenses ( 394 , 396 ) in a periscope type of configuration to maintain a high amount of transmission of light from the real world ( 144 ) to the eye ( 58 ) of the user.
- the embodiment of FIG. 14C provides six different DMD ( 402 , 404 , 406 , 408 , 410 , 412 ) plane opportunities to intercede from an SLM functionality as beams are routed to the eye ( 58 ), along with two lenses ( 398 , 400 ) for beam control.
- FIG. 14D illustrates a more complicated periscope type arrangement with up to four DMDs ( 422 , 424 , 426 , 428 ) for SLM functionality and four lenses ( 414 , 420 , 416 , 418 ); this configuration is designed to ensure that the image does not become flipped upside down as it travels through to the eye ( 58 ).
- FIG. 14D illustrates a more complicated periscope type arrangement with up to four DMDs ( 422 , 424 , 426 , 428 ) for SLM functionality and four lenses ( 414 , 420 , 416 , 418 ); this configuration is designed to ensure that the image does not become flipped upside down as it travels through to the eye ( 58 ).
- FIG. 14D illustrates a more complicated periscope type arrangement with up to four DMDs ( 422 , 424 , 426 , 428 ) for SLM functionality and four lenses ( 414 , 420 , 416 , 418 ); this configuration is
- FIG. 14E illustrates in embodiment wherein light may be reflected between two different DMD devices ( 430 , 432 ) without any intervening lenses (the lenses in the above designs are useful in such configurations for incorporating image information from the real world), in a hall-of-mirrors type of arrangement wherein the display may be viewed through the “hall of mirrors” and operates in a mode substantially similar to that illustrated in FIG. 14A .
- FIG. 14F illustrates an embodiment wherein a the non-display portions of two facing DMD chips ( 434 , 436 ) may be covered with a reflective layer to propagate light to and from active display regions ( 438 , 440 ) of the DMD chips.
- arrays of sliding MEMS shutters may be utilized to either pass or block light.
- arrays of small louvers that move out of place to present light-transmitting apertures may similarly be aggregated for SLM functionality.
- a lightfield of many small beamlets may be injected into and propagated through a waveguide or other optical system.
- a conventional “birdbath” type of optical system may be suitable for transferring the light of a lightfield input, or a freeform optics design, as described below, or any number of waveguide configurations.
- FIGS. 15A-15C illustrate the use of a wedge type waveguide ( 442 ) along with a plurality of light sources as another configuration useful in creating a lightfield. Referring to FIG.
- light may be injected into the wedge-shaped waveguide ( 442 ) from two different locations/displays ( 444 , 446 ), and will emerge according to the total internal reflection properties of the wedge-shaped waveguide at different angles ( 448 ) based upon the points of injection into the waveguide.
- FIG. 15B if one creates a linear array ( 450 ) of displays (such as scanning fiber displays) projecting into the end of the waveguide as shown, then a large angular diversity of beams ( 452 ) will be exiting the waveguide in one dimension, as shown in FIG. 15C .
- displays such as scanning fiber displays
- FIG. 15C if one contemplates adding yet another linear array of displays injecting into the end of the waveguide but at a slightly different angle, then an angular diversity of beams may be created that exits similarly to the fanned out exit pattern shown in FIG. 15C , but at an orthogonal axis; together these may be utilized to create a two-dimensional fan of rays exiting each location of the waveguide.
- another configuration is presented for creating angular diversity to form a lightfield display using one or more scanning fiber display arrays (or alternatively using other displays which will meet the space requirements, such as miniaturized DLP projection configurations).
- a stack of SLM devices may be utilized, in which case rather than the direct view of SLM output as described above, the lightfield output from the SLM configuration may be used as an input to a configuration such as that shown in FIG. 15C .
- a conventional waveguide is best suited to relay beams of collimated light successfully, with a lightfield of small-diameter collimated beams
- conventional waveguide technology may be utilized to further manipulate the output of such a lightfield system as injected into the side of a waveguide, such as a wedge-shaped waveguide, due to the beam size/collimation.
- a multicore fiber may be used to generate a lightfield and inject it into the waveguide.
- a time-varying lightfield may be utilized as an input, such that rather than creating a static distribution of beamlets coming out of a lightfield, one may have some dynamic elements that are methodically changing the path of the set of beams. They may be done using components such as waveguides with embedded DOEs (e.g., such as those described above in reference to FIGS. 8B-8N , or liquid crystal layers, as described in reference to FIG.
- a wedge-shaped waveguide may be configured to have a bi-modal total internal reflection paradigm (for example, in one variation, wedge-shaped elements may be configured such that when a liquid crystal portion is activated, not only is the spacing changed, but also the angle at which the beams are reflected).
- a scanning light display may be characterized simply as a scanning fiber display with a lens at the end of the scanned fiber.
- Many lens varieties are suitable, such as a GRIN lens, which may be used to collimate the light or to focus the light down to a spot smaller than the fiber's mode field diameter providing the advantage of producing a numerical aperture (or “NA”) increase and circumventing the optical invariant, which is correlated inversely with spot size. Smaller spot size generally facilitates a higher resolution opportunity from a display perspective, which generally is preferred.
- a GRIN lens may be long enough relative to the fiber that it may comprise the vibrating element (i.e., rather than the usual distal fiber tip vibration with a scanned fiber display)—a configuration which may be deemed a “scanned GRIN lens display”.
- a diffractive lens may be utilized at the exit end of a scanning fiber display (i.e., patterned onto the fiber).
- a curved mirror may be positioned on the end of the fiber that operates in a reflecting configuration. Essentially any of the configurations known to collimate and focus a beam may be used at the end of a scanning fiber to produce a suitable scanned light display.
- Two significant utilities to having a lens coupled to or comprising the end of a scanned fiber are a) the light exiting may be collimated to obviate the need to use other external optics to do so; b) the NA, or the angle of the cone at which light sprays out the end of the single-mode fiber core, may be increased, thereby decreasing the associated spot size for the fiber and increasing the available resolution for the display.
- a lens such as a GRIN lens may be fused to or otherwise coupled to the end of an optical fiber or formed from a portion of the end of the fiber using techniques such as polishing.
- a typical optical fiber with an NA of about 0.13 or 0.14 may have a spot size (also known as the “mode field diameter” for the optical fiber given the NA) of about 3 microns.
- This provides for relatively high resolution display possibilities given the industry standard display resolution paradigms (for example, a typical microdisplay technology such as LCD or organic light emitting diode, or “OLED” has a spot size of about 5 microns).
- the aforementioned scanning light display may have 3 ⁇ 5 of the smallest pixel pitch available with a conventional display; further, using a lens at the end of the fiber, the aforementioned configuration may produce a spot size in the range of 1-2 microns.
- a cantilevered portion of a waveguide (such as a waveguide created using microfabrication processes such as masking and etching, rather than drawn microfiber techniques) may be placed into scanning oscillatory motion, and may be fitted with lensing at the exit ends.
- an increased numerical aperture for a fiber to be scanned may be created using a diffuser (i.e., one configured to scatter light and create a larger NA) covering the exit end of the fiber.
- the diffuser may be created by etching the end of the fiber to create small bits of terrain that scatter light; in another variation a bead or sandblasting technique, or direct sanding/scuffing technique may be utilized to create scattering terrain.
- an engineered diffuser similar to a diffractive element, may be created to maintain a clean spot size with desirable NA, which ties into the notion of using a diffractive lens, as noted above.
- an array of optical fibers ( 454 ) is shown coupled in to a coupler ( 456 ) configured to hold them in parallel together so that their ends may be ground and polished to have an output edge at a critical angle ( 458 ; 42 degrees for most glass, for example) to the longitudinal axes of the input fibers, such that the light exiting the angled faces will exit as though it had been passing through a prism, and will bend and become nearly parallel to the surfaces of the polished faces.
- the beams exiting the fibers ( 454 ) in the bundle will become superimposed, but will be out of phase longitudinally due to the different path lengths (referring to FIG. 16B , for example, the difference in path lengths from angled exit face to focusing lens for the different cores is visible).
- a multicore fiber such as those available from Mitsubishi Cable Industries, Ltd. of Japan, may be angle polished.
- the exiting light may be reflected from the polished surface and emerge from the side of the fiber (in one embodiment at a location wherein a flat-polished exit window has been created in the side of the fiber) such that as the fiber is scanned in what would normally be an X-Y Cartesian coordinate system axis, that fiber would now be functionally performing the equivalent of an X-Z scan, with the distance changing during the course of the scan.
- a reflective element such as a mirror coating
- Multicore fibers may be configured to play a role in display resolution enhancement (i.e., higher resolution). For example, in one embodiment, if separate pixel data is sent down a tight bundle of 19 cores in a multicore fiber, and that cluster is scanned around in a sparse spiral pattern with the pitch of the spiral being approximately equal to the diameter of the multicore, then sweeping around will effectively create a display resolution that is approximately 19 ⁇ the resolution of a single core fiber being similarly scanned around. Indeed, it may be more practical to have the fibers more sparsely positioned relative to each other, as in the configuration of FIG.
- display resolution enhancement i.e., higher resolution
- 16C which has 7 clusters ( 464 ; 7 is used for illustrative purposes because it is an efficient tiling/hex pattern; other patterns or numbers may be utilized; for example, a cluster of 19; the configuration is scalable up or down) of 3 fibers each housed within a conduit ( 462 ).
- densely packed scanned cores can create blurring at the display may be utilized as an advantage in one embodiment wherein a plurality (say a triad or cores to carry red, green, and blue light) of cores may be intentionally packed together densely so that each triad forms a triad of overlapped spots featuring red, green, and blue light.
- a plurality say a triad or cores to carry red, green, and blue light
- cores may be intentionally packed together densely so that each triad forms a triad of overlapped spots featuring red, green, and blue light.
- each tight cluster of 3 fiber cores contains one core that relays red light, one core that relays green light, and one core that relays blue light, with the 3 fiber cores close enough together that their positional differences are not resolvable by the subsequent relay optics, forming an effectively superimposed RGB pixel; thus, the sparse tiling of 7 clusters produces resolution enhancement while the tight packing of 3 cores within the clusters facilitates seamless color blending without the need to utilize glossy RGB fiber combiners (e.g., those using wavelength division multiplexing or evanescent coupling techniques).
- one may have just one cluster ( 464 ) housed in a conduit ( 468 ) for, say, red/green/blue (and in another embodiment, another core may be added for infrared for uses such as eye tracking). In another embodiment, additional cores may be placed in the tight cluster to carrying additional wavelengths of light to comprise a multi-primary display for increased color gamut.
- a sparse array of single cores ( 470 ); in one variation with red, green, and blue combined down each of them) within a conduit ( 466 ) may be utilized; such a configuration is workable albeit somewhat less efficient for resolution increase, but not optimum for red/green/blue combining.
- Multicore fibers also may be utilized for creating lightfield displays. Indeed, rather than keeping the cores separated enough from each other so that the cores do not scan on each other's local area at the display panel, as described above in the context of creating a scanning light display, with a lightfield display, it is desirable to scan around a densely packed plurality of fibers because each of the beams produced represents a specific part of the lightfield.
- the light exiting from the bundled fiber tips can be relatively narrow if the fibers have a small NA; lightfield configurations may take advantage of this and have an arrangement in which at the anatomic pupil, a plurality of slightly different beams are being received from the array.
- there are optical configurations with scanning a multicore that are functionally equivalent to an array of single scanning fiber modules, and thus a lightfield may be created by scanning a multicore rather than scanning a group of single mode fibers.
- a multi-core phased array approach may be used to create a large exit pupil variable wavefront configuration to facilitate three-dimensional perception.
- a single laser configuration with phase modulators is described above.
- phase delays may be induced into different channels of a multicore fiber, such that a single laser's light is injected into all of the cores of the multicore configuration so that there is mutual coherence.
- a multi-core fiber may be combined with a lens, such as a GRIN lens.
- a lens such as a GRIN lens.
- Such lens may be, for example, a refractive lens, diffractive lens, or a polished edge functioning as a lens.
- the lens may be a single optical surface, or may comprise multiple optical surfaces stacked up. Indeed, in addition to having a single lens that extends the diameter of the multicore, a smaller lenslet array may be desirable at the exit point of light from the cores of the multicore, for example.
- FIG. 16F shows an embodiment wherein a multicore fiber ( 470 ) is emitting multiple beams into a lens ( 472 ), such as a GRIN lens.
- the lens collects the beams down to a focal point ( 474 ) in space in front of the lens.
- the beams would exit the multicore fiber as diverging.
- the GRIN or other lens is configured to function to direct them down to a single point and collimate them, such that the collimated result may be scanned around for a lightfield display, for instance.
- smaller lenses ( 478 ) may be placed in front of each of the cores of a multicore ( 476 ) configuration, and these lenses may be utilized to collimate; then a shared lens ( 480 ) may be configured to focus the collimated beams down to a diffraction limited spot ( 482 ) that is aligned for all of the three spots.
- a shared lens ( 480 ) may be configured to focus the collimated beams down to a diffraction limited spot ( 482 ) that is aligned for all of the three spots.
- one embodiment features a multicore fiber ( 476 ) with a lenslet ( 478 ) array feeding the light to a small prism array ( 484 ) that deflects the beams generated by the individual cores to a common point.
- a small prism array 484
- the small lenslet array shifted relative to the cores such that the light is being deflected and focused down to a single point.
- Such a configuration may be utilized to increase the numerical aperture.
- a two-step configuration is shown with a small lenslet ( 478 ) array capturing light from the multicore fiber ( 476 ), followed sequentially by a shared lens ( 486 ) to focus the beams to a single point ( 488 ).
- a small lenslet ( 478 ) array capturing light from the multicore fiber ( 476 ), followed sequentially by a shared lens ( 486 ) to focus the beams to a single point ( 488 ).
- Such a configuration may be utilized to increase the numerical aperture. As discussed above, a larger NA corresponds to a smaller pixel size and higher possible display resolution.
- a beveled fiber array which may be held together with a coupler ( 456 ), such as those described above, may be scanned with a reflecting device ( 494 ; such as a DMD module of a DLP system).
- a reflecting device 494 ; such as a DMD module of a DLP system.
- the superimposed light can be directed through one or more focusing lenses ( 490 , 492 ) to create a multifocal beam; with the superimposing and angulation of the array, the different sources are different distances from the focusing lens, which creates different focus levels in the beams as they emerge from the lens ( 492 ) and are directed toward the retina ( 54 ) of the eye ( 58 ) of the user.
- the farthest optical route/beam may be set up to be a collimated beam representative of optical infinity focal positions. Closer routes/beams may be associated with diverging spherical wavefronts of closer focal locations.
- the multifocal beam may be passed into a scanning mirror which may be configured to create a raster scan (or, for example, a Lissajous curve scan pattern or a spiral scan pattern) of the multifocal beam which may be passed through a series of focusing lenses and then to the cornea and crystalline lens of the eye.
- a scanning mirror which may be configured to create a raster scan (or, for example, a Lissajous curve scan pattern or a spiral scan pattern) of the multifocal beam which may be passed through a series of focusing lenses and then to the cornea and crystalline lens of the eye.
- the various beams emerging from the lenses are creating different pixels or voxels of varying focal distances that are superimposed.
- one may write different data to each of the light modulation channels at the front end, thereby creating an image that is projected to the eye with one or more focus elements.
- the fiber array may be actuated/moved around by a piezoelectric actuator.
- a relatively thin ribbon array may be resonated in cantilevered form along the axis perpendicular to the arrangement of the array fibers (i.e., in the thin direction of the ribbon) when a piezoelectric actuator is activated.
- a separate piezoelectric actuator may be utilized to create a vibratory scan in the orthogonal long axis.
- a single mirror axis scan may be employed for a slow scan along the long axis while the fiber ribbon is vibrated resonantly.
- an array ( 496 ) of scanning fiber displays ( 498 ) may be beneficially bundled/tiled for an effective resolution increase, the notion being that with such as configuration, each scanning fiber of the bundle is configured to write to a different portion of the image plane ( 500 ), as shown, for example, in FIG. 16N , wherein each portion of the image plane is addressed by the emissions from a least one bundle.
- optical configurations may be utilized that allow for slight magnification of the beams as they exit the optical fiber so that there is some overlap in the hexagonal, or other lattice pattern, that hits the display plane, so there is a better fill factor while also maintaining an adequately small spot size in the image plane and understanding that there is a subtle magnification in that image plane.
- a monolithic lenslet array may be utilized, so that the lenses can be as closely packed as possible, which allows for even smaller spot sizes in the image plane because one may use a lower amount of magnification in the optical system.
- arrays of fiber scan displays may be used to increase the resolution of the display, or in other words, they may be used to increase the field of view of the display, because each engine is being used to scan a different portion of the field of view.
- a lightfield display may be created using a plurality of small diameter fibers scanned around in space. For example, instead of having all of the fibers address a different part of an image plane as described above, have more overlapping, more fibers angled inward, etc., or change the focal power of the lenses so that the small spot sizes are not conjugate with a tiled image plane configuration. Such a configuration may be used to create a lightfield display to scan lots of smaller diameter rays around that become intercepted in the same physical space.
- one way of creating a lightfield display involves making the output of the elements on the left collimated with narrow beams, and then making the projecting array conjugate with the eye pupil on the right.
- a single actuator may be utilized to actuate a plurality of fibers ( 506 ) in unison together.
- a similar configuration is discussed above in reference to FIGS. 13 -C- 1 and 13 -C- 2 . It may be practically difficult to have all of the fibers retain the same resonant frequency, vibrate in a desirable phase relationship to each other, or have the same dimensions of cantilevering from the substrate block.
- the tips of the fibers may be mechanically coupled with a lattice or sheet ( 504 ), such as a graphene sheet that is very thin, rigid, and light in weight.
- the entire array may vibrate similarly and have the same phase relationship.
- a matrix of carbon nanotubes may be utilized to couple the fibers, or a piece of very thin planar glass (such as the kind used in creating liquid crystal display panels) may be coupled to the fiber ends.
- a laser or other precision cutting device may be utilized to cut all associated fibers to the same cantilevered length.
- FIG. 17 it may be desirable to have a contact lens directly interfaced with the cornea, and configured to facilitate the eye focusing on a display that is quite close (such as the typical distance between a cornea and an eyeglasses lens).
- the lens may comprise a selective filter.
- FIG. 17 depicts a plot ( 508 ) what may be deemed a “notch filter”, due to its design to block only certain wavelength bands, such as 450 nm (peak blue), 530 nm (green), and 650 nm, and generally pass or transmit other wavelengths.
- several layers of dielectric coatings may be aggregated to provide the notch filtering functionality.
- Such a filtering configuration may be coupled with a scanning fiber display that is producing a very narrow band illumination for red, green, and blue, and the contact lens with the notch filtering will block out all of the light coming from the display (such as a minidisplay, such as an OLED display, mounted in a position normally occupied by an eyeglasses lens) except for the transmissive wavelengths.
- a scanning fiber display that is producing a very narrow band illumination for red, green, and blue
- the contact lens with the notch filtering will block out all of the light coming from the display (such as a minidisplay, such as an OLED display, mounted in a position normally occupied by an eyeglasses lens) except for the transmissive wavelengths.
- a narrow pinhole may be created in the middle of the contact lens filtering layers/film such that the small aperture (i.e., less than about 1.5 mm diameter) does allow passage of the otherwise blocked wavelengths.
- a pinhole lens configuration that functions in a pinhole manner for red, green, and blue only to intake images from the minidisplay, while light from the real world, which generally is broadband illumination, will pass through the contact lens relatively unimpeded.
- a large depth of focus virtual display configuration may be assembled and operated.
- a collimated image exiting from a waveguide would be visible at the retina because of the pinhole large-depth-of-focus configuration.
- a display may be configured to have different display modes that may be selected (preferably rapidly toggling between the two at the command of the operator) by an operator, such as a first mode combining a very large depth of focus with a small exit pupil diameter (i.e., so that everything is in focus all of the time), and a second mode featuring a larger exit pupil and a more narrow depth of focus.
- the operator may select the first mode; alternatively, if a user is to type in a long essay (i.e., for a relatively long period of time) using a two-dimensional word processing display configuration, it may be more desirable to switch to the second mode to have the convenience of a larger exit pupil, and a sharper image.
- one configuration may have red wavelength and blue wavelength channels presented with a very small exit pupil so that they are always in focus.
- a green channel only may be presented with a large exit pupil configuration with multiple depth planes (i.e., because the human accommodation system tends to preferentially target green wavelengths for optimizing focus level).
- the green wavelength may be prioritized and represented with various different wavefront levels. Red and blue may be relegated to being represented with a more Maxwellian approach (and, as described above in reference to Maxwellian displays, software may be utilized to induce Gaussian levels of blur). Such a display would simultaneously present multiple depths of focus.
- the fovea portion generally is populated with approximately 120 cones per visual degree.
- Display systems have been created in the past that use eye or gaze tracking as an input, and to save computation resources by only creating really high resolution rendering for where the person is gazing at the time, while lower resolution rendering is presented to the rest of the retina; the locations of the high versus low resolution portions may be dynamically slaved to the tracked gaze location in such a configuration, which may be termed a “foveated display”.
- An improvement on such configurations may comprise a scanning fiber display with pattern spacing that may be dynamically slaved to tracked eye gaze.
- a scanning fiber display with pattern spacing that may be dynamically slaved to tracked eye gaze.
- a typical scanning fiber display operating in a spiral pattern as shown in FIG. 18 (the leftmost portion 510 of the image in FIG. 18 illustrates a spiral motion pattern of a scanned multicore fiber 514 ; the rightmost portion 512 of the image in FIG. 18 illustrates a spiral motion pattern of a scanned single fiber 516 for comparison), a constant pattern pitch provides for a uniform display resolution.
- a non-uniform scanning pitch may be utilized, with smaller/tighter pitch (and therefore higher resolution) dynamically slaved to the detected gaze location. For example, if the user's gaze was detected as moving toward the edge of the display screen, the spirals may be clustered more densely in such location, which would create a toroid-type scanning pattern for the high-resolution portions, and the rest of the display being in a lower-resolution mode. In a configuration wherein gaps may be created in the portions of the display in a lower-resolution mode, blur could be intentionally dynamically created to smooth out the transitions between scans, as well as between transitions from high-resolution to lower-resolution scan pitch.
- the term lightfield may be used to describe a volumetric 3-D representation of light traveling from an object to a viewer's eye.
- an optical see-through display can only reflect light to the eye, not the absence of light, and ambient light from the real world will add to any light representing a virtual object. That is, if a virtual object presented to the eye contains a black or very dark portion, the ambient light from the real world may pass through that dark portion and obscure that it was intended to be dark.
- one way to selectively attenuate for a darkfield perception is to block all of the light coming from one angle, while allowing light from other angles to be transmitted.
- This may be accomplished with a plurality of SLM planes comprising elements such as liquid crystal (which may not be the most optimal due to its relatively low transparency when in the transmitting state), DMD elements of DLP systems (which have relative high transmission/reflection ratios when in such mode), and MEMS arrays or shutters that are configured to controllably shutter or pass light radiation, as described above.
- a cholesteric LCD array may be utilized for a controlled occlusion/blocking array.
- a cholesteric LCD configuration a pigment is being bound to the liquid crystal molecule, and then the molecule is physically tilted in response to an applied voltage.
- Such a configuration may be designed to achieve greater transparency when in a transmissive mode than conventional LCD, and a stack of polarizing films is not needed as it is with conventional LCD.
- a plurality of layers of controllably interrupted patterns may be utilized to controllably block selected presentation of light using moire effects.
- two arrays of attenuation patterns each of which may comprise, for example, fine-pitched sine waves printed or painted upon a transparent planar material such as a glass substrate, may be presented to the eye of a user at a distance close enough that when the viewer looks through either of the patterns alone, the view is essentially transparent, but if the viewer looks through both patterns lined up in sequence, the viewer will see a spatial beat frequency moire attenuation pattern, even when the two attenuation patterns are placed in sequence relatively close to the eye of the user.
- the beat frequency is dependent upon the pitch of the patterns on the two attenuation planes, so in one embodiment, an attenuation pattern for selectively blocking certain light transmission for darkfield perception may be created using two sequential patterns, each of which otherwise would be transparent to the user, but which together in series create a spatial beat frequency moire attenuation pattern selected to attenuate in accordance with the darkfield perception desired in the augmented reality system.
- a controlled occlusion paradigm for darkfield effect may be created using a multi-view display style occluder.
- one configuration may comprise one pin-holed layer that fully occludes with the exception of small apertures or pinholes, along with a selective attenuation layer in series, which may comprise an LCD, DLP system, or other selective attenuation layer configuration, such as those described above.
- a selective attenuation layer in series which may comprise an LCD, DLP system, or other selective attenuation layer configuration, such as those described above.
- a perception of a sharp mechanical edge out in space may be created.
- the pinhole array layer may be replaced with a second dynamic attenuation layer to provide a somewhat similar configuration, but with more controls than the static pinhole array layer (the static pinhole layer could be simulated, but need not be).
- the pinholes may be replaced with cylindrical lenses.
- the same pattern of occlusion as in the pinhole array layer configuration may be achieved, but with cylindrical lenses, the array is not restricted to the very tiny pinhole geometries.
- a second lens array may be added on the side of the aperture or lens array opposite of the side nearest the eye to compensate and provide the view-through illumination with basically a zero power telescope configuration.
- each liquid crystal layer may act as a polarization rotator such that if a patterned polarizing material is incorporated on one face of a panel, then the polarization of individual rays coming from the real world may be selectively manipulated so they catch a portion of the patterned polarizer.
- polarizers known in the art that have checkerboard patterns wherein half of the “checker boxes” have vertical polarization and the other half have horizontal polarization.
- polarizers known in the art that have checkerboard patterns wherein half of the “checker boxes” have vertical polarization and the other half have horizontal polarization.
- a material such as liquid crystal is used in which polarization may be selectively manipulated, light may be selectively attenuated with this.
- selective reflectors may provide greater transmission efficiency than LCD.
- a lens system if a lens system is placed such that it takes light coming in from the real world and focuses a plane from the real world onto an image plane, and if a DMD (i.e., DLP technology) is placed at that image plane to reflect light when in an “on” state towards another set of lenses that pass the light to the eye, and those lenses also have the DMD at their focal length, the one may create an attenuation pattern that is in focus for the eye.
- DMDs may be used in a selective reflector plane in a zero magnification telescope configuration, such as is shown in FIG. 19A , to controllably occlude and facilitate creating darkfield perception.
- a lens ( 518 ) is taking light from the real world ( 144 ) and focusing it down to an image plane ( 520 ); if a DMD (or other spatial attenuation device) ( 522 ) is placed at the focal length of the lens (i.e., at the image plane 520 ), the lens ( 518 ) is going to take whatever light is coming from optical infinity and focus that onto the image plane ( 520 ). Then the spatial attenuator ( 522 ) may be utilized to selectively block out things that are to be attenuated.
- FIG. 19A shows the attenuator DMDs in the transmissive mode wherein they pass the beams shown crossing the device.
- the image is then placed at the focal length of the second lens ( 524 ).
- the two lenses ( 518 , 524 ) have the same focal power so they end up being a zero-power telescope, or a “relay”, that does not magnify views to the real world ( 144 ).
- Such a configuration may be used to present unmagnified views of the world while also allowing selective blocking/attenuation of certain pixels.
- FIGS. 19B and 19C additional DMDs may be added such that light reflects from each of four DMDs ( 526 , 528 , 530 , 532 ) before passing to the eye.
- FIG. 19B shows an embodiment with two lenses preferably with the same focal power (focal length “F”) placed at a 2 F relationship from one another (the focal length of the first being conjugate to the focal length of the second) to have the zero-power telescope effect;
- FIG. 19C shows an embodiment without lenses.
- the angles of orientation of the four reflective panels ( 526 , 528 , 530 , 532 ) in the depicted embodiments of FIGS. 19B and 19C are shown to be around 45 degrees for simple illustration purposes, but specific relative orientation is required (for example, a typical DMD reflect at about a 12 degree angle).
- the panels may also be ferroelectric, or may be any other kind of reflective or selective attenuator panel or array.
- one of the three reflector arrays may be a simple mirror, such that the other 3 are selective attenuators, thus still providing three independent planes to controllably occlude portions of the incoming illumination in furtherance of darkfield perception.
- Such a configuration may be driven in lightfield algorithms to selectively attenuate certain rays while others are passed.
- a DMD or similar matrix of controllably movable devices may be created upon a transparent substrate as opposed to a generally opaque substrate, for use in a transmissive configuration such as virtual reality.
- two LCD panels may be utilized as lightfield occluders.
- they may be thought of as attenuators due to their attenuating capability as described above; alternatively they may be considered polarization rotators with a shared polarizer stack.
- Suitable LCDs may comprise components such as blue phase liquid crystal, cholesteric liquid crystal, ferroelectric liquid crystal, and/or twisted nematic liquid crystal.
- One embodiment may comprise an array of directionally-selective occlusion elements, such as a MEMS device featuring a set of louvers that can change rotation such that they pass the majority of light that is coming from a particular angle, but are presenting more of a broad face to light that is coming from a different angle (somewhat akin to the manner in which plantation shutters may be utilized with a typical human scale window).
- the MEMS/louvers configuration may be placed upon an optically transparent substrate, with the louvers substantially opaque. Ideally such a configuration would have a louver pitch fine enough to selectably occlude light on a pixel-by-pixel basis.
- two or more layers or stacks of louvers may be combined to provide yet further controls.
- the louvers may be polarizers configured to change the polarization state of light on a controllably variable basis.
- another embodiment for selective occlusion may comprise an array of sliding panels in a MEMS device such that the sliding panels may be controllably opened (i.e., by sliding in a planar fashion from a first position to a second position; or by rotating from a first orientation to a second orientation; or, for example, combined rotational reorientation and displacement) to transmit light through a small frame or aperture, and controllably closed to occlude the frame or aperture and prevent transmission.
- the array may be configured to open or occlude the various frames or apertures such that they maximally attenuate the rays that are to be attenuated, and only minimally attenuate the rays to be transmitted.
- the sliding panels may comprise sliding polarizers, and if placed in a stacked configuration with other polarizing elements that are either static or dynamic, may be utilized to selectively attenuate.
- FIG. 19D another configuration providing an opportunity for selective reflection, such as via a DMD style reflector array ( 534 ), is shown, such that a stacked set of two waveguides ( 536 , 538 ) along with a pair of focus elements ( 540 , 542 ) and a reflector ( 534 ; such as a DMD) may be used to capture a portion of incoming light with an entrance reflector ( 544 ).
- a DMD style reflector array 534
- the reflected light may be totally internally reflected down the length of the first waveguide ( 536 ), into a focusing element ( 540 ) to bring the light into focus on a reflector ( 534 ) such as a DMD array, after which the DMD may selectively attenuate and reflect a portion of the light back through a focusing lens ( 542 ; the lens configured to facilitate injection of the light back into the second waveguide) and into the second waveguide ( 538 ) for total internal reflection down to an exit reflector ( 546 ) configured to exit the light out of the waveguide and toward the eye ( 58 ).
- a reflector 534
- the DMD may selectively attenuate and reflect a portion of the light back through a focusing lens ( 542 ; the lens configured to facilitate injection of the light back into the second waveguide) and into the second waveguide ( 538 ) for total internal reflection down to an exit reflector ( 546 ) configured to exit the light out of the waveguide and toward the eye ( 58 ).
- Such a configuration may have a relatively thin shape factor, and is designed to allow light from the real world ( 144 ) to be selectively attenuated.
- waveguides work most cleanly with collimated light
- such a configuration may be well suited for virtual reality configurations wherein focal lengths are in the range of optical infinity.
- a lightfield display may be used as a layer on top of the silhouette created by the aforementioned selective attenuation/darkfield configuration to provide other cues to the eye of the user that light is coming from another focal distance.
- An occlusion mask may be out of focus, even nondesirably so, and then in one embodiment, a lightfield on top of the masking layer may be used to hide the fact that the darkfield may be at the wrong focal distance.
- FIG. 19E an embodiment is shown featuring two waveguides ( 552 , 554 ) each having two angled reflectors ( 558 , 544 ; 556 , 546 ) for illustrative purposes shown at approximately 45 degrees; in actual configurations the angle may differ depending upon the reflective surface, reflective/refractive properties of the waveguides, etc.) directing a portion of light incoming from the real world down each side of a first waveguide (or down two separate waveguides if the top layer is not monolithic) such that it hits a reflector ( 548 , 550 ) at each end, such as a DMD which may be used for selective attenuation, after which the reflected light may be injected back into the second waveguide (or into two separate waveguides if the bottom layer is not monolithic) and back toward two angled reflectors (again, they need not be at 45 degrees as shown) for exit out toward the eye ( 58 ).
- a reflector 548 , 550
- Focusing lenses may also be placed between the reflectors at each end and the waveguides.
- the reflectors ( 548 , 550 ) at each end may comprise standard mirrors (such as alumized mirrors).
- the reflectors may be wavelength selective reflectors, such as dichroic mirrors or film interference filters.
- the reflectors may be diffractive elements configured to reflect incoming light.
- FIG. 19F illustrates a configuration wherein four reflective surfaces in a pyramid type configuration are utilized to direct light through two waveguides ( 560 , 562 ), in which incoming light from the real world may be divided up and reflected to four difference axes.
- the pyramid-shaped reflector ( 564 ) may have more than four facets, and may be resident within the substrate prism, as with the reflectors of the configuration of FIG. 19E .
- the configuration of FIG. 19F is an extension of that of FIG. 19E .
- a single waveguide ( 566 ) may be utilized to capture light from the world ( 144 ) with one or more reflective surfaces ( 574 , 576 , 578 , 580 , 582 ), relay it ( 570 ) to a selective attenuator ( 568 ; such as a DMD array), and recouple it back into the same waveguide so that it propagates ( 572 ) and encounters one or more other reflective surfaces ( 584 , 586 , 588 , 590 , 592 ) that cause it to at least partially exit ( 594 ) the waveguide on a path toward the eye ( 58 ) of the user.
- a selective attenuator 568 ; such as a DMD array
- the waveguide comprises selective reflectors such that one group ( 574 , 576 , 578 , 580 , 582 ) may be switched on to capture incoming light and direct it down to the selective attenuator, while separate another group ( 584 , 586 , 588 , 590 , 592 ) may be switched on to exit light returning from the selective attenuator out toward the eye ( 58 ).
- one group 574 , 576 , 578 , 580 , 582
- another group 584 , 586 , 588 , 590 , 592
- the selective attenuator is shown oriented substantially perpendicularly to the waveguide; in other embodiments, various optics components, such as refractive or reflective optics, may be utilized to have the selective attenuator at a different and more compact orientation relative to the waveguide.
- FIG. 19H a variation on the configuration described in reference to FIG. 19D is illustrated.
- This configuration is somewhat analogous to that discussed above in reference to FIG. 5B , wherein a switchable array of reflectors may be embedded within each of a pair of waveguides ( 602 , 604 ).
- a controller may be configured to turn the reflectors ( 598 , 600 ) on and off in sequence, such that multiple reflectors may be operated on a frame sequential basis; then the DMD or other selective attenuator ( 594 ) may also be sequentially driven in sync with the different mirrors being turned on and off.
- FIG. 19I a pair of wedge-shaped waveguides similar to those described above (for example, in reference to FIGS. 15A-15C ) are shown in side or sectional view to illustrate that the two long surfaces of each wedge-shaped waveguide ( 610 , 612 ) are not co-planar.
- a “turning film” ( 606 , 608 ; such as that available from 3M corporation under the trade name, “TRAF”, which in essence comprises a microprism array), may be utilized on one or more surfaces of the wedge-shaped waveguides to either turn incoming rays at an angle so that they will be captured by total internal reflection, or to turn outgoing rays as they are exiting the waveguide toward an eye or other target.
- Incoming rays are directed down the first wedge and toward the selective attenuator ( 614 ) such as a DMD, LCD (such as a ferroelectric LCD), or an LCD stack to act as a mask).
- reflected light is coupled back into the second wedge-shaped waveguide which then relays the light by total internal reflection along the wedge.
- the properties of the wedge-shaped waveguide are intentionally such that each bounce of light causes an angle change; the point at which the angle has changed enough to be the critical angle to escape total internal reflection becomes the exit point from the wedge-shaped waveguide.
- the exit will be at an oblique angle, so another layer of turning film may be used to “turn” the exiting light toward a targeted object such as the eye ( 58 ).
- FIG. 19J several arcuate lenslet arrays ( 616 , 620 , 622 ) are positioned relative to an eye and configured such that a spatial attenuator array ( 618 ) is positioned at a focal/image plane so that it may be in focus with the eye ( 58 ).
- the first ( 616 ) and second ( 620 ) arrays are configured such that in the aggregate, light passing from the real world to the eye is essentially passed through a zero power telescope.
- the embodiment of FIG. 19J shows a third array ( 622 ) of lenslets which may be utilized for improved optical compensation, but the general case does not require such a third layer.
- having telescopic lenses that are the diameter of the viewing optic may create an undesirably large form factor (somewhat akin to having a bunch of small sets of binoculars in front of the eyes).
- One way to optimize the overall geometry is to reduce the diameter of the lenses by splitting them out into smaller lenslets, as shown in FIG. 19J (i.e., an array of lenses rather than one single large lens).
- the lenslet arrays ( 616 , 620 , 622 ) are shown wrapped radially or arcuately around the eye ( 58 ) to ensure that beams incoming to the pupil are aligned through the appropriate lenslets (else the system may suffer from optical problems such as dispersion, aliasing, and/or lack of focus).
- all of the lenslets are oriented “toed in” and pointed at the pupil of the eye ( 58 ), and the system facilitates avoidance of scenarios wherein rays are propagated through unintended sets of lenses en route to the pupil.
- FIGS. 19K-19N various software approaches may be utilized to assist in the presentation of darkfield in a virtual or augmented reality displace scenario.
- a typical challenging scenario for augmented reality is depicted ( 632 ), with a textured carpet ( 624 ) and non-uniform background architectural features ( 626 ), both of which are lightly-colored.
- the black box ( 628 ) depicted indicates the region of the display in which one or more augmented reality features are to be presented to the user for three-dimensional perception, and in the black box a robot creature ( 630 ) is being presented that may, for example, be part of an augmented reality game in which the user is engaged.
- the robot character ( 630 ) is darkly-colored, which makes for a challenging presentation in three-dimensional perception, particularly with the background selected for this example scenario.
- one of the main challenges for a presenting darkfield augmented reality object is that the system generally cannot add or paint in “darkness”; generally the display is configured to add light.
- presentation of the robot character in the augmented reality view results in a scene wherein portions of the robot character that are to be essentially flat black in presentation are not visible, and portions of the robot character that are to have some lighting (such as the lightly-pigmented cover of the shoulder gun of the robot character) are only barely visible ( 634 )—they appear almost like a light grayscale disruption to the otherwise normal background image.
- using a software-based global attenuation treatment provides enhanced visibility to the robot character because the brightness of the nearly black robot character is effective increased relative to the rest of the space, which now appears more dark ( 640 ).
- a digitally-added light halo 636
- the halo treatment even the portions of the robot character that are to be presented as flat black become visible with the contrast to the white halo, or “aura” presented around the robot character.
- the halo may be presented to the user with a perceived focal distance that is behind the focal distance of the robot character in three-dimensional space.
- the light halo may be presented with an intensity gradient to match the dark halo that may accompany the occlusion, minimizing the visibility of either darkfield effect.
- the halo may be presented with blurring to the background behind the presented halo illumination for further distinguishing effect. A more subtle aura or halo effect may be created by matching, at least in part, the color and/or brightness of a relatively light-colored background.
- some or all of the black intonations of the robot character may be changed to dark, cool blue colors to provide a further distinguishing effect relative to the background, and relatively good visualization of the robot ( 642 ).
- Wedge-shaped waveguides have been described above, such as in reference to FIGS. 15A-15D and FIG. 19I .
- a wedge-shaped waveguide every time a ray bounces off of one of the non-coplanar surfaces, it gets an angle change, which ultimately results in the ray exiting total internal reflection when its approach angle to one of the surfaces goes past the critical angle.
- Turning films may be used to redirect exiting light so that exiting beams leave with a trajectory that is more or less perpendicular to the exit surface, depending upon the geometric and ergonomic issues at play.
- the wedge-shaped waveguide may be configured to create a fine-pitched array of angle-biased rays emerging from the wedge.
- a lightfield display, or a variable wavefront creating waveguide both may produce a multiplicity of beamlets or beams to represent a single pixel in space such that wherever the eye is positioned, the eye is hit by a plurality of different beamlets or beams that are unique to that particular eye position in front of the display panel.
- a plurality of viewing zones may be created within a given pupil, and each may be used for a different focal distance, with the aggregate producing a perception similar to that of a variable wavefront creating waveguide, or similar to the actual optical physics of reality of the objects viewed were real.
- a wedge-shaped waveguide with multiple displays may be utilized to generate a lightfield.
- a fan of exiting rays is created for each pixel.
- This concept may be extended in an embodiment wherein multiple linear arrays are stacked to all inject image information into the wedge-shaped waveguide (in one variation, one array may inject at one angle relative to the wedge-shaped waveguide face, while the second array may inject at a second angle relative to the wedge-shaped waveguide face), in which case exit beams fan out at two different axes from the wedge.
- one or more arrays or displays may be configured to inject image information into wedge-shaped waveguide through sides or faces of the wedge-shaped waveguide other than that shown in FIG. 15C , such as by using a diffractive optic to bend injected image information into total an internal reflection configuration relative to the wedge-shaped waveguide.
- Various reflectors or reflecting surfaces may also be utilized in concert with such a wedge-shaped waveguide embodiment to outcouple and manage light from the wedge-shaped waveguide.
- an entrance aperture to a wedge-shaped waveguide, or injection of image information through a different face other than shown in FIG. 15C may be utilized to facilitate staggering (geometric and/or temporal) of different displays and arrays such that a Z-axis delta may also be developed as a means for injecting three-dimensional information into the wedge-shaped waveguide.
- various displays may be configured to enter a wedge-shaped waveguide at multiple edges in multiple stacks with staggering to get higher dimensional configurations.
- a configuration similar to that depicted in FIG. 8H is shown wherein a waveguide ( 646 ) has a diffractive optical element ( 648 ; or “DOE”, as noted above) sandwiched in the middle (alternatively, as described above, the diffractive optical element may reside on the front or back face of the depicted waveguide).
- a ray may enter the waveguide ( 646 ) from the projector or display ( 644 ). Once in the waveguide ( 646 ), each time the ray intersects the DOE ( 648 ), part of it is exited out of the waveguide ( 646 ).
- the DOE may be designed such that the exit illuminance across the length of the waveguide ( 646 ) is somewhat uniform (for example, the first such DOE intersection may be configured to exit about 10% of the light; then the second DOE intersection may be configured to exit about 10% of the remaining light so that 81% is passed on, and so on; in another embodied a DOE may be designed to have a variable diffraction efficiency, such as linearly-decreasing diffraction efficiency, along its length to map out a more uniform exit illuminance across the length of the waveguide).
- a variable diffraction efficiency such as linearly-decreasing diffraction efficiency
- a reflective element ( 650 ) at one or both ends may be included.
- additional distribution and preservation may be achieved by including an elongate reflector ( 652 ) across the length of the waveguide as shown (comprising, for example, a thin film dichroic coating that is wavelength-selective); preferably such reflector would be blocking light that accidentally is reflected upward (back toward the real world 144 for exit in a way that it would not be utilized by the viewer).
- an elongate reflector may contribute to a “ghosting” effect perception by the user.
- this ghosting effect may be eliminated by having a dual-waveguide ( 646 , 654 ) circulating reflection configuration, such as that shown in FIG. 20C , which is designed to keep the light moving around until it has been exited toward the eye ( 58 ) in a preferably substantially equally distributed manner across the length of the waveguide assembly.
- light may be injected with a projector or display ( 644 ), and as it travels across the DOE ( 656 ) of the first waveguide ( 654 ), it ejects a preferably substantially uniform pattern of light out toward the eye ( 58 ); light that remains in the first waveguide is reflected by a first reflector assembly ( 660 ) into the second waveguide ( 646 ).
- the second waveguide ( 646 ) may be configured to not have a DOE, such that it merely transports or recycles the remaining light back to the first waveguide, using the second reflector assembly.
- the second waveguide ( 646 ) may also have a DOE ( 648 ) configured to uniformly eject fractions of travelling light to provide a second plane of focus for three-dimensional perception.
- the configuration of FIG. 20C is designed for light to travel the waveguide in one direction, which avoids the aforementioned ghosting problem that is related to passing light backwards through a waveguide with a DOE.
- an array of smaller retroreflectors ( 662 ), or a retroreflective material may be utilized.
- FIG. 20E an embodiment is shown that utilizes some of the light recycling configurations of the embodiment of FIG. 20C to “snake” the light down through a waveguide ( 646 ) having a sandwiched DOE ( 648 ) after it has been injected with a display or projector ( 644 ) so that it crosses the waveguide ( 646 ) many times back and forth before reaching the bottom, at which point it may be recycled back up to the top level for further recycling.
- Such a configuration not only recycles the light and facilitates use of relatively low diffraction efficiency DOE elements for exiting light toward the eye ( 58 ), but also distributes the light, to provide for a large exit pupil configuration akin to that described in reference to FIG. 8K .
- FIG. 20F an illustrative configuration similar to that of FIG. 5A is shown, with incoming light injected along a conventional prism or beamsplitter substrate ( 104 ) to a reflector ( 102 ) without total internal reflection (i.e., without the prism being considered a waveguide) because the input projection ( 106 ), scanning or otherwise, is kept within the bounds of the prism—which means that the geometry of such prism becomes a significant constraint.
- a waveguide may be utilized in place of the simple prism of FIG. 20F , which facilitates the use of total internal reflection to provide more geometric flexibility.
- a collimated image injected into a waveguide may be refocused before transfer out toward an eye, in a configuration also designed to facilitate viewing light from the real world.
- a diffractive optical element may be used as a variable focus element.
- FIG. 7B another waveguide configuration is illustrated in the context of having multiple layers stacked upon each other with controllable access toggling between a smaller path (total internal reflection through a waveguide) and a larger path (total internal reflection through a hybrid waveguide comprising the original waveguide and a liquid crystal isolated region with the liquid crystal switched to a mode wherein the refractive indices are substantially matched between the main waveguide and the auxiliary waveguide), so that the controller can tune on a frame-by-frame basis which path is being taken.
- High-speed switching electro-active materials such as lithium niobate, facilitate path changes with such a configuration at gigahertz rates, which allows one to change the path of light on a pixel-by-pixel basis.
- a stack of waveguides paired with weak lenses is illustrated to demonstrate a multifocal configuration wherein the lens and waveguide elements may be static.
- Each pair of waveguide and lens may be functionally replaced with waveguide having an embedded DOE element (which may be static, in a closer analogy to the configuration of FIG. 8A , or dynamic), such as that described in reference to FIG. 8I .
- a transparent prism or block ( 104 ; i.e., not a waveguide) is utilized to hold a mirror or reflector ( 102 ) in a periscope type of configuration to receive light from other components, such as a lens ( 662 ) and projector or display ( 644 ), the field of view is limited by the size of that reflector ( 102 ; the bigger the reflector, the wider the field of view).
- a thicker substrate may be needed to hold a larger reflector; otherwise, the functionality of an aggregated plurality of reflectors may be utilized to increase the functional field of view, as described in reference to FIGS. 8O , 8 P, and 8 Q.
- a stack ( 664 ) of planar waveguides ( 666 ), each fed with a display or projector ( 644 ; or in another embodiment a multiplexing of a single display) and having an exit reflector ( 668 ), may be utilized to aggregate toward the function of a larger single reflector.
- the exit reflectors may be at the same angle in some cases, or not the same angle in other cases, depending upon the positioning of the eye ( 58 ) relative to the assembly.
- FIG. 20I illustrates a related configuration, wherein the reflectors ( 680 , 682 , 684 , 686 , 688 ) in each of the planar waveguides ( 670 , 672 , 674 , 676 , 678 ) have been offset from each other, and wherein each takes in light from a projector or display ( 644 ) which may be sent through a lens ( 690 ) to ultimately contribute exiting light to the pupil ( 45 ) of the eye ( 58 ) by virtue of the reflectors ( 680 , 682 , 684 , 686 , 688 ) in each of the planar waveguides ( 670 , 672 , 674 , 676 , 678 ).
- FIG. 20K illustrates a variation wherein the shaded portion of the optical assembly may be utilized as a compensating lens to functionally pass light from the real world ( 144 ) through the assembly as though it has been passed through a zero power telescope.
- each of the aforementioned rays may also be a relative wide beam that is being reflected through the pertinent waveguide ( 670 , 672 ) by total internal reflection.
- the reflector ( 680 , 682 ) facet size will determine what the exiting beam width can be.
- a further discretization of the reflector is shown, wherein a plurality of small straight angular reflectors may form a roughly parabolic reflecting surface ( 694 ) in the aggregate through a waveguide or stack thereof ( 696 ).
- a linear array of displays injects light into a shared waveguide ( 376 ).
- a single display may be multiplexed to a series of entry lenses to provide similar functionality as the embodiment of FIG. 13M , with the entry lenses creating parallel paths of rays running through the waveguide.
- a red/green/blue (or “RGB”) laserline reflector is placed at one or both ends of the planar surfaces, akin to a thin film interference filter that is highly reflective for only certain wavelengths and poorly reflective for other wavelengths, than one can functionally increase the range of angles of light propagation.
- Windows without the coating
- the coating may be selected to have a directional selectivity (somewhat like reflective elements that are only highly reflective for certain angles of incidence). Such a coating may be most relevant for the larger planes/sides of a waveguide.
- a variation on a scanning fiber display was discussed, which may be deemed a scanning thin waveguide configuration, such that a plurality of very thin planar waveguides ( 358 ) may be oscillated or vibrated such that if a variety of injected beams is coming through with total internal reflection, the configuration functionally would provide a linear array of beams escaping out of the edges of the vibrating elements ( 358 ).
- the depicted configuration has approximately five externally-projecting planar waveguide portions ( 358 ) in a host medium or substrate ( 356 ) that is transparent, but which preferably has a different refractive index so that the light will stay in total internal reflection within each of the substrate-bound smaller waveguides that ultimately feed (in the depicted embodiment there is a 90 degree turn in each path at which point a planar, curved, or other reflector may be utilized to bounce the light outward) the externally-projecting planar waveguide portions ( 358 ).
- the externally-projecting planar waveguide portions ( 358 ) may be vibrated individually, or as a group along with oscillatory motion of the substrate ( 356 ). Such scanning motion may provide horizontal scanning, and for vertical scanning, the input ( 360 ) aspect of the assembly (i.e., such as one or more scanning fiber displays scanning in the vertical axis) may be utilized. Thus a variation of the scanning fiber display is presented.
- a waveguide ( 370 ) may be utilized to create a lightfield.
- all beams staying in focus may cause perception discomfort (i.e., the eye will not make a discernible difference in dioptric blur as a function of accommodation; in other words, the narrow diameter, such as 0.5 mm or less, collimated beamlets may open loop the eye's accommodation/vergence system, causing discomfort).
- a single beam may be fed in with a number of cone beamlets coming out, but if the introduction vector of the entering beam is changed (i.e., laterally shift the beam injection location for the projector/display relative to the waveguide), one may control where the beam exits from the waveguide as it is directed toward the eye.
- a waveguide to create a lightfield by creating a bunch of narrow diameter collimated beams, and such a configuration is not reliant upon a true variation in a light wavefront to be associated with the desired perception at the eye.
- a set of angularly and laterally diverse beamlets is injected into a waveguide (for example, by using a multicore fiber and driving each core separately; another configuration may utilize a plurality of fiber scanners coming from different angles; another configuration may utilize a high-resolution panel display with a lenslet array on top of it), a number of exiting beamlets can be created at different exit angles and exit locations. Since the waveguide may scramble the lightfield, the decoding is preferably predetermined.
- a waveguide ( 646 ) assembly ( 696 ) that comprises stacked waveguide components in the vertical or horizontal axis.
- the notion with these embodiments is to stack a plurality of smaller waveguides ( 646 ) immediately adjacent each other such that light introduced into one waveguide, in addition to propagating down (i.e., propagating along a Z axis with total internal reflection in +X, ⁇ X) such waveguide by total internal reflection, also totally internally reflects in the perpendicular axis (+y, ⁇ Y) as well, such that it is not spilling into other areas.
- each waveguide may have a DOE ( 648 ) embedded and configured to eject out light with a predetermined distribution along the length of the waveguide, as described above, with a predetermined focal length configuration (shown in FIG. 20M as ranging from 0.5 meters to optical infinity).
- a very dense stack of waveguides with embedded DOEs may be produced such that it spans the size of the anatomical pupil of the eye (i.e., such that multiple layers 698 of the composite waveguide are required to cross the exit pupil, as illustrated in FIG. 20N ).
- one may feed a collimated image for one wavelength, and then the portion located the next millimeter down producing a diverging wavefront that represents an object coming from a focal distance of, say, 15 meters away, and so on, with the notion being that an exit pupil is coming from a number of different waveguides as a result of the DOEs and total internal reflection through the waveguides and across the DOEs.
- such a configuration creates a plurality of stripes that, in the aggregate, facilitate the perception of different focal depths with the eye/brain.
- Such a concept may be extended to configurations comprising a waveguide with a switchable/controllable embedded DOE (i.e. that is switchable to different focal distances), such as those described in relation to FIGS. 8B-8N , which allows more efficient light trapping in the axis across each waveguide.
- Multiple displays may be coupled into each of the layers, and each waveguide with DOE would emit rays along its own length.
- a laserline reflector may be used to increase angular range.
- a completely reflective metallized coating may be utilized, such as aluminum, to ensure total reflection, or alternatively dichroic style or narrow band reflectors may be utilized.
- the whole composite waveguide assembly ( 696 ) may be curved concavely toward the eye ( 58 ) such that each of the individual waveguides is directed toward the pupil.
- the configuration may be designed to more efficiently direct the light toward the location where the pupil is likely to be present.
- Such a configuration also may be utilized to increase the field of view.
- FIG. 21A illustrates a waveguide ( 698 ) having an embedded (i.e., sandwiched within) DOE ( 700 ) with a linear grating term that may be changed to alter the exit angle of exiting light ( 702 ) from the waveguide, as shown.
- a high-frequency switching DOE material such as lithium niobate may be utilized.
- such a scanning configuration may be used as the sole mechanism for scanning a beam in one axis; in another embodiment, the scanning configuration may be combined with other scanning axes, and may be used to create a larger field of view (i.e., if a normal field of view is 40 degrees, and by changing the linear diffraction pitch one can steer over another 40 degrees, the effective usable field of view for the system is 80 degrees).
- a waveguide ( 708 ) may be placed perpendicular to a panel display ( 704 ), such as an LCD or OLED panel, such that beams may be injected from the waveguide ( 708 ), through a lens ( 706 ), and into the panel ( 704 ) in a scanning configuration to provide a viewable display for television or other purposes.
- the waveguide may be utilized in such configuration as a scanning image source, in contrast to the configurations described in reference to FIG. 21A , wherein a single beam of light may be manipulated by a scanning fiber or other element to sweep through different angular locations, and in addition, another direction may be scanned using the high-frequency diffractive optical element.
- a uniaxial scanning fiber display (say scanning the fast line scan, as the scanning fiber is relatively high frequency) may be used to inject the fast line scan into the waveguide, and then the relatively slow DOE switching (i.e., in the range of 100 Hz) may be used to scan lines in the other axis to form an image.
- the relatively slow DOE switching i.e., in the range of 100 Hz
- a DOE with a grating of fixed pitch may be combined with an adjacent layer of electro-active material having a dynamic refractive index (such as liquid crystal), so that light may be redirected into the grating at different angles.
- a dynamic refractive index such as liquid crystal
- an electro-active layer comprising an electro-active material such as liquid crystal or lithium niobate may change its refractive index such that it changes the angle at which a ray emerges from the waveguide.
- a linear diffraction grating may be added to the configuration of FIG. 7B (in one embodiment, sandwiched within the glass or other material comprising the larger lower waveguide) such that the diffraction grating may remain at a fixed pitch, but the light is biased before it hits the grating.
- FIG. 21C shows another embodiment featuring two wedge-like waveguide elements ( 710 , 712 ), wherein one or more of them may be electro-active so that the related refractive index may be changed.
- the elements may be configured such that when the wedges have matching refractive indices, the light totally internally reflects through the pair (which in the aggregate performs akin to a planar waveguide with both wedges matching) while the wedge interfaces have no effect. Then if one of the refractive indices is changed to create a mismatch, a beam deflection at the wedge interface ( 714 ) is caused, and there is total internal reflection from that surface back into the associated wedge. Then a controllable DOE ( 716 ) with a linear grating may be coupled along one of the long edges of the wedge to allow light to exit out and reach the eye at a desirable exit angle.
- a DOE such as a Bragg grating
- a time-varying grating may be utilized for field of view expansion by creating a tiled display configuration. Further, a time-varying grating may be utilized to address chromatic aberration (failure to focus all colors/wavelengths at the same focal point).
- chromatic aberration frailure to focus all colors/wavelengths at the same focal point.
- One property of diffraction gratings is that they will deflect a beam as a function of its angle of incidence and wavelength (i.e., a DOE will deflect different wavelengths by different angles: somewhat akin to the manner in which a simple prism will divide out a beam into its wavelength components).
- the DOE may be configured to drive the red wavelength to a slightly different place than the green and blue to address unwanted chromatic aberration.
- the DOE may be time-varied by having a stack of elements that switch on and off (i.e. to get red, green, and blue to be diffracted outbound similarly).
- a time-varying grating may be utilized for exit pupil expansion.
- a waveguide ( 718 ) with embedded DOE ( 720 ) may be positioned relative to a target pupil such that none of the beams exiting in a baseline mode actually enter the target pupil ( 45 )—such that the pertinent pixel would be missed by the user.
- a time-varying configuration may be utilized to fill in the gaps in the outbound exit pattern by shifting the exit pattern laterally (shown in dashed/dotted lines) to effectively scan each of the 5 exiting beams to better ensure that one of them hits the pupil of the eye. In other words, the functional exit pupil of the display system is expanded.
- a time-varying grating may be utilized with a waveguide for one, two, or three axis light scanning.
- a term in a grating that is scanning a beam in the vertical axis as well as a grating that is scanning in the horizontal axis.
- radial elements of a grating are incorporated, as is discussed above in relation to FIGS. 8B-8N , one may have scanning of the beam in the Z axis (i.e., toward/away from the eye), all of which may be time sequential scanning.
- DOEs are usable whether or not the DOE is embedded in a waveguide.
- the output of a waveguide may be separately manipulated using a DOE; or a beam may be manipulated by a DOE before it is injected into a waveguide; further, one or more DOEs, such as a time-varying DOE, may be utilized as an input for freeform optics configurations, as discussed below.
- an element of a DOE may have a circularly-symmetric term, which may be summed with a linear term to create a controlled exit pattern (i.e., as described above, the same DOE that outcouples light may also focus it).
- the circular term of the DOE diffraction grating may be varied such that the focus of the beams representing those pertinent pixels is modulated.
- one configuration may have a second/separate circular DOE, obviating the need to have a linear term in the DOE.
- a functional stack 728 of DOE elements may comprise a stack of polymer dispersed liquid crystal elements 726 , as described above, wherein without a voltage applied, a host medium refraction index matches that of a dispersed molecules of liquid crystal; in another embodiment, molecules of lithium niobate may be dispersed for faster response times; with voltage applied, such as through transparent indium tin oxide layers on either side of the host medium, the dispersed molecules change index of refraction and functionally form a diffraction pattern within the host medium) that can be switched on/off.
- a circular DOE may be layered in front of a waveguide for focus modulation.
- the waveguide ( 722 ) is outputting collimated light, which will be perceived as associated with a focal depth of optical infinity unless otherwise modified.
- the collimated light from the waveguide may be input into a diffractive optical element ( 730 ) which may be used for dynamic focus modulation (i.e., one may switch on and off different circular DOE patterns to impart various different focuses to the exiting light).
- a static DOE may be used to focus collimated light exiting from a waveguide to a single depth of focus that may be useful for a particular user application.
- multiple stacked circular DOEs may be used for additive power and many focus levels—from a relatively small number of switchable DOE layers.
- three different DOE layers may be switched on in various combinations relative to each other; the optical powers of the DOEs that are switched on may be added.
- a first DOE may be configured to provide half of the total diopter range desired (in this example, 2 diopters of change in focus);
- a second DOE may be configured to induce a 1 diopter change in focus; then a third DOE may be configured to induce a 1 ⁇ 2 diopter change in focus.
- These three DOEs may be mixed and matched to provide 1 ⁇ 2, 1, 1.5, 2, 2.5, 3, and 3.5 diopters of change in focus. Thus a super large number of DOEs would not be required to get a relatively broad range of control.
- a matrix of switchable DOE elements may be utilized for scanning, field of view expansion, and/or exit pupil expansion.
- a typical DOE is either all on or all off.
- a DOE ( 732 ) may be subdivided into a plurality of functional subsections (such as the one labeled as element 734 in FIG. 21H ), each of which preferably is uniquely controllable to be on or off (for example, referring to FIG. 21H , each subsection may be operated by its own set of indium tin oxide, or other control lead material, voltage application leads 736 back to a central controller). Given this level of control over a DOE paradigm, additional configurations are facilitated.
- a waveguide ( 738 ) with embedded DOE ( 740 ) is viewed from the top down, with the user's eye positioned in front of the waveguide.
- a given pixel may be represented as a beam coming into the waveguide and totally internally reflecting along until it may be exited by a diffraction pattern to come out of the waveguide as a set of beams.
- the beams may come out parallel/collimated (as shown in FIG. 21I for convenience), or in a diverging fan configuration if representing a focal distance closer than optical infinity.
- the depicted set of parallel exiting beams may represent, for example, the farthest left pixel of what the user is seeing in the real world as viewed through the waveguide, and light off to the rightmost extreme will be a different group of parallel exiting beams.
- the DOE subsections as described above, one may spend more computing resource or time creating and manipulating the small subset of beams that is likely to be actively addressing the user's pupil (i.e., because the other beams never reach the user's eye and are effectively wasted).
- a waveguide ( 738 ) configuration is shown wherein only the two subsections ( 740 , 742 ) of the DOE ( 744 ) are deemed to be likely to address the user's pupil ( 45 ) are activated.
- one subsection may be configured to direct light in one direction simultaneously as another subsection is directing light in a different direction.
- FIG. 21K shows an orthogonal view of two independently controlled subsections ( 734 , 746 ) of a DOE ( 732 ).
- such independent control may be used for scanning or focusing light.
- an assembly ( 748 ) of three independently controlled DOE/waveguide subsections ( 750 , 752 , 754 ) may be used to scan, increase the field of view, and/or increase the exit pupil region.
- Such functionality may arise from a single waveguide with such independently controllable DOE subsections, or a vertical stack of these for additional complexity.
- a circular DOE may be controllably stretched radially-symmetrically, the diffraction pitch may be modulated, and the DOE may be utilized as a tunable lens with an analog type of control.
- a single axis of stretch (for example, to adjust an angle of a linear DOE term) may be utilized for DOE control.
- a membrane, akin to a drum head may be vibrated, with oscillatory motion in the Z-axis (i.e., toward/away from the eye) providing Z-axis control and focus change over time.
- a stack of several DOEs ( 756 ) is shown receiving collimated light from a waveguide ( 722 ) and refocusing it based upon the additive powers of the activated DOEs.
- Linear and/or radial terms of DOEs may be modulated over time, such as on a frame sequential basis, to produce a variety of treatments (such as tiled display configurations or expanded field of view) for the light coming from the waveguide and exiting, preferably toward the user's eye.
- a low diffraction efficiency is desired to maximize transparency for light passed from the real world; in configurations wherein the DOE or DOEs are not embedded, a high diffraction efficiency may be desired, as described above.
- both linear and radial DOE terms may be combined outside of the waveguide, in which case high diffraction efficiency would be desired.
- a segmented or parabolic reflector such as those discussed above in FIG. 8Q .
- the same functionality may result from a single waveguide with a DOE having different phase profiles for each section of it, such that it is controllable by subsection.
- the DOE may be configured to direct light toward the same region in space (i.e., the pupil of the user).
- optical configurations known as “freeform optics” may be utilized certain of the aforementioned challenges.
- the term “freeform” generally is used in reference to arbitrarily curved surfaces that may be utilized in situations wherein a spherical, parabolic, or cylindrical lens does not meet a design complexity such as a geometric constraint.
- a spherical, parabolic, or cylindrical lens does not meet a design complexity such as a geometric constraint.
- one of the common challenges with display ( 762 ) configurations when a user is looking through a mirror (and also sometimes a lens 760 ) is that the field of view is limited by the area subtended by the final lens ( 760 ) of the system.
- a typical freeform optic has three active surfaces.
- light may be directed toward the freeform optic from an image plane, such as a flat panel display ( 768 ), into the first active surface ( 772 ), which typically is a primarily transmissive freeform surface that refracts transmitted light and imparts a focal change (such as an added stigmatism, because the final bounce from the third surface will add a matching/opposite stigmatism and these are desirably canceled).
- the incoming light may be directed from the first surface to a second surface ( 774 ), wherein it may strike with an angle shallow enough to cause the light to be reflected under total internal reflection toward the third surface ( 776 ).
- the third surface may comprise a half-silvered, arbitrarily-curved surface configured to bounce the light out through the second surface toward the eye, as shown in FIG. 22E .
- the light enters through the first surface, bounces from the second surface, bounces from the third surface, and is directed out of the second surface. Due to the optimization of the second surface to have the requisite reflective properties on the first pass, as well as refractive properties on the second pass as the light is exited toward the eye, a variety of curved surfaces with higher-order shapes than a simple sphere or parabola are formed into the freeform optic.
- a compensating lens ( 780 ) may be added to the freeform optic ( 770 ) such that the total thickness of the optic assembly is substantially uniform in thickness, and preferably without magnification, to light incoming from the real world ( 144 ) in an augmented reality configuration.
- a freeform optic ( 770 ) may be combined with a waveguide ( 778 ) configured to facilitate total internal reflection of captured light within certain constraints.
- a waveguide ( 778 ) configured to facilitate total internal reflection of captured light within certain constraints.
- light may be directed into the freeform/waveguide assembly from an image plane, such as a flat panel display, and totally internally reflected within the waveguide until it hits the curved freeform surface and escapes toward the eye of the user. Thus the light bounces several times in total internal reflection until it reaches the freeform wedge portion.
- FIG. 22H depicts a configuration similar to that of FIG. 22G , with the exception that the configuration of FIG. 22H also features a compensating lens portion to further extend the thickness uniformity and assist with viewing the world through the assembly without further compensation.
- a freeform optic ( 782 ) is shown with a small flat surface, or fourth face ( 784 ), at the lower left corner that is configured to facilitate injection of image information at a different location than is typically used with freeform optics.
- the input device ( 786 ) may comprise, for example, a scanning fiber display, which may be designed to have a very small output geometry.
- the fourth face may comprise various geometries itself and have its own refractive power, such as by use planar or freeform surface geometries.
- such a configuration may also feature a reflective coating ( 788 ) along the first surface such that it directs light back to the second surface, which then bounces the light to the third surface, which directs the light out across the second surface and to the eye ( 58 ).
- a reflective coating 788
- the addition of the fourth small surface for injection of the image information facilitates a more compact configuration.
- some lenses ( 792 , 794 ) may be required in order to appropriately form an image plane ( 796 ) using the output from the scanning fiber display; these hardware components add extra bulk that may not be desired.
- FIG. 22K an embodiment is shown wherein light from a scanning fiber display ( 790 ) is passed through an input optics assembly ( 792 , 794 ) to an image plane ( 796 ), and then directed across the first surface of the freeform optic ( 770 ) to a total internal reflection bounce off of the second surface, then another total internal reflection bounce from the third surface results in the light exiting across the second surface and being directed toward the eye ( 58 ).
- An all-total-internal-reflection freeform waveguide may be created such that there are no reflective coatings (i.e., such that total-internal-reflection is being relied upon for propagation of light until a critical angle of incidence with a surface is met, at which point the light exits in a manner akin to the wedge-shaped optics described above).
- a surface comprising one or more sub-surfaces from a set of conical curves, such as parabolas, spheres, ellipses, etc.).
- Such a configuration still may produce a shallow-enough angles for total internal reflection within the optic; thus an approach that is somewhat a hybrid between a conventional freeform optic and a wedge-shaped waveguide is presented.
- One motivation to have such a configuration is to get away from the use of reflective coatings, which do help product reflection, but also are known to prevent transmission of a relatively large portion (such as 50%) of the light transmitting through from the real world ( 144 ). Further, such coatings also may block an equivalent amount of the light coming into the freeform optic from the input device. Thus there are reasons to develop designs that do not have reflective coatings.
- one of the surfaces of a conventional freeform optic may comprise a half-silvered reflective surface.
- a reflective surface will be of “neutral density”, meaning that it will generally reflect all wavelengths similarly.
- the conventional reflector paradigm may be replaced with a narrow band reflector that is wavelength sensitive, such as a thin film laserline reflector.
- a configuration may reflect particular red/green/blue wavelength ranges and remain passive to other wavelengths, which generally will increase transparency of the optic and therefore be preferred for augmented reality configurations wherein transmission of image information from the real world ( 144 ) across the optic also is valued.
- each of the three depicted freeform optics may have a wavelength-selective coating (for example, one highly selective for blue, the next for green, the next for red) so that images may be injected into each to have blue reflected from one surface, green from another, and red from a third surface.
- a wavelength-selective coating for example, one highly selective for blue, the next for green, the next for red
- Such a configuration may be utilized, for example, to address chromatic aberration issues, to create a lightfield, or to increase the functional exit pupil size.
- a single freeform optic ( 798 ) has multiple reflective surfaces ( 800 , 802 , 804 ), each of which may be wavelength or polarization selective so that their reflective properties may be individually controlled.
- multiple microdisplays such as scanning light displays, ( 786 ) may be injected into a single freeform optic to tile images (thereby providing an increased field of view), increase the functional pupil size, or address challenges such as chromatic aberration (i.e., by reflecting one wavelength per display).
- Each of the depicted displays would inject light that would take a different path through the freeform optic due to the different positioning of the displays relative to the freeform optic, which would provide a larger functional exit pupil output.
- a packet or bundle of scanning fiber displays may be utilized as an input to overcome one of the challenges in operatively coupling a scanning fiber display to a freeform optic.
- One such challenge with a scanning fiber display configuration is that the output of an individual fiber is emitted with a certain numerical aperture, or “NA”, which is like the projectional angle of light from the fiber; ultimately this angle determines the diameter of the beam that passes through various optics, and ultimately determines the exit functional exit pupil size; thus in order to maximize exit pupil size with a freeform optic configuration, one may either increase the NA of the fiber using optimized refractive relationships, such as between core and cladding, or one may place a lens (i.e., a refractive lens, such as a gradient refractive index lens, or “GRIN” lens) at the end of the fiber or build one into the end of the fiber as described above, or create an array of fibers that is feeding into the freeform optic, in which case all of those NAs in the bundle remain small, and at the exit pupil an array of small exit
- a more sparse array (i.e., not bundled tightly as a packet) of scanning fiber displays or other displays may be utilized to functionally increase the field of view of the virtual image through the freeform optic.
- a plurality of displays or displays ( 786 ) may be injected through the top of a freeform optic ( 770 ), as well as another plurality ( 786 ) through the lower corner; the display arrays may be two or three dimensional arrays.
- image information also may be injected in from the side ( 806 ) of the freeform optic ( 770 ) as well.
- each of the scanning fibers monochromatic, such that within a given bundle or plurality of projectors or displays, one may have a subgroup of solely red fibers, a subgroup of solely blue fibers, and a subgroup of solely green fibers.
- Such a configuration facilitates more efficiency in output coupling for bringing light into the optical fibers; for instance, there would be no need in such an embodiment to superimpose red, green, and blue into the same band.
- FIGS. 22Q-22V various freeform optic tiling configurations are depicted.
- FIG. 22Q an embodiment is depicted wherein two freeform optics are tiled side-by-side and a microdisplay, such as a scanning light display, ( 786 ) on each side is configured to inject image information from each side, such that one freeform optic wedge represents each half of the field of view.
- a microdisplay such as a scanning light display
- a compensator lens ( 808 ) may be included to facilitate views of the real world through the optics assembly.
- FIG. 22S illustrates a configuration wherein freeform optics wedges are tiled side by side to increase the functional field of view while keeping the thickness of such optical assembly relatively uniform.
- a star-shaped assembly comprises a plurality of freeform optics wedges (also shown with a plurality of displays for inputting image information) in a configuration that may provide a larger field of view expansion while also maintaining a relatively thin overall optics assembly thickness.
- the optics elements may be aggregated to produce a larger field of view; the tiling configurations described above have addressed this notion.
- the freeform waveguides such that their outputs share, or are superimposed in, the space of the pupil (for example, the user may see the left half of the visual field through the left freeform waveguide, and the right half of the visual field through the right freeform waveguide).
- the freeform waveguides may be oriented such that they do not toe in as much—so they create exit pupils that are side-by-side at the eye's anatomical pupil.
- the anatomical pupil may be 8 mm wide, and each of the side-by-side exit pupils may be 8 mm, such that the functional exit pupil is expanded by about two times.
- Such a configuration provides an enlarged exit pupil, but if the eye is moved around in the “eyebox” defined by that exit pupil, that eye may lose parts of the visual field (i.e., lose either a portion of the left or right incoming light because of the side-by-side nature of such configuration).
- red wavelengths may be driven through one freeform optic, green through another, and blue through another, such red/green/blue chromatic aberration may be addressed.
- Multiple freeform optics also may be provided to such a configuration that are stacked up, each of which is configured to address a particular wavelength.
- FIG. 22U two oppositely-oriented freeform optics are shown stacked in the Z-axis (i.e., they are upside down relative to each other).
- a compensating lens may not be required to facilitate accurate views of the world through the assembly; in other words, rather than having a compensating lens such as in the embodiment of FIG. 22F or FIG. 22R , an additional freeform optic may be utilized, which may further assist in routing light to the eye.
- FIG. 22V shows another similar configuration wherein the assembly of two freeform optics is presented as a vertical stack.
- wavelength or polarization selective reflector surfaces For example, referring to FIG. 22V , red, green, and blue wavelengths in the form of 650 nm, 530 nm, and 450 nm may be injected, as well as red, green, and blue wavelengths in the form of 620 nm, 550 nm, and 470 nm; different selective reflectors may be utilized in each of the freeform optics so that they do not interfere with each other.
- the reflection/transmission selectivity for light that is polarized in a particular axis may be varied (i.e., the images may be pre-polarized before they are sent to each freeform waveguide, to work with reflector selectivity).
- a plurality of freeform waveguides may be utilized together in series.
- light may enter from the real world and be directed sequentially through a first freeform optic ( 770 ), through an optional lens ( 812 ) which may be configured to relay light to a reflector ( 810 ) such as a DMD from a DLP system, which may be configured to reflect the light that has been filtered on a pixel by pixel basis (i.e., an occlusion mask may be utilized to block out certain elements of the real world, such as for darkfield perception, as described above; suitable spatial light modulators may be used which comprise DMDs, LCDs, ferroelectric LCOSs, MEMS shutter arrays, and the like, as described above) to another freeform optic ( 770 ) that is relaying light to the eye ( 28 ) of the user.
- a configuration may be more compact than one using conventional lenses for spatial light modulation.
- a configuration may be utilized that has one surface that is highly-reflective so that it may bounce light straight into another compactly positioned freeform optic.
- a selective attenuator ( 814 ) may be interposed between the two freeform optics ( 770 ).
- a freeform optic may comprise one aspect of a contact lens system.
- a miniaturized freeform optic is shown engaged against the cornea of a user's eye ( 58 ) with a miniaturized compensator lens portion ( 780 ), akin to that described in reference to FIG. 22F .
- Signals may be injected into the miniaturized freeform assembly using a tethered scanning fiber display which may, for example, be coupled between the freeform optic and a tear duct area of the user, or between the freeform optic and another head-mounted display configuration.
- the invention includes methods that may be performed using the subject devices.
- the methods may comprise the act of providing such a suitable device. Such provision may be performed by the end user.
- the “providing” act merely requires the end user obtain, access, approach, position, set-up, activate, power-up or otherwise act to provide the requisite device in the subject method.
- Methods recited herein may be carried out in any order of the recited events which is logically possible, as well as in the recited order of events.
- any optional feature of the inventive variations described may be set forth and claimed independently, or in combination with any one or more of the features described herein.
- Reference to a singular item includes the possibility that there are plural of the same items present. More specifically, as used herein and in claims associated hereto, the singular forms “a,” “an,” “said,” and “the” include plural referents unless the specifically stated otherwise.
- use of the articles allow for “at least one” of the subject item in the description above as well as claims associated with this disclosure. It is further noted that such claims may be drafted to exclude any optional element. As such, this statement is intended to serve as antecedent basis for use of such exclusive terminology as “solely,” “only” and the like in connection with the recitation of claim elements, or use of a “negative” limitation.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Optics & Photonics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Nonlinear Science (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Computer Graphics (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Ophthalmology & Optometry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Geometry (AREA)
- General Health & Medical Sciences (AREA)
- Chemical & Material Sciences (AREA)
- Architecture (AREA)
- Dispersion Chemistry (AREA)
- Mathematical Physics (AREA)
- Crystallography & Structural Chemistry (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Mechanical Optical Scanning Systems (AREA)
- Lenses (AREA)
- Eyeglasses (AREA)
- Liquid Crystal (AREA)
- User Interface Of Digital Computer (AREA)
- Mechanical Light Control Or Optical Switches (AREA)
- Optical Elements Other Than Lenses (AREA)
- Light Guides In General And Applications Therefor (AREA)
- Devices For Indicating Variable Information By Combining Individual Elements (AREA)
- Processing Or Creating Images (AREA)
- Stereoscopic And Panoramic Photography (AREA)
- Optical Modulation, Optical Deflection, Nonlinear Optics, Optical Demodulation, Optical Logic Elements (AREA)
- Control Of Indicators Other Than Cathode Ray Tubes (AREA)
Abstract
Description
- This application is a continuation of pending U.S. patent application Ser. No. 14/555,585, filed Nov. 27, 2014, entitled “VIRTUAL AND AUGMENTED REALITY SYSTEMS AND METHODS, which claims priority from U.S. Provisional Application Ser. No. 61/909,774, filed Nov. 27, 2013, entitled “VIRTUAL AND AUGMENTED REALITY SYSTEMS AND METHODS”. The contents of the aforementioned applications are hereby expressly incorporated by reference into the present application their entireties.
- The present disclosure relates to virtual reality and augmented reality imaging and visualization systems.
- Modern computing and display technologies have facilitated the development of systems for so called “virtual reality” or “augmented reality” experiences, wherein digitally reproduced images or portions thereof are presented to a user in a manner wherein they seem to be, or may be perceived as, real. A virtual reality, or “VR”, scenario typically involves presentation of digital or virtual image information without transparency to other actual real-world visual input; an augmented reality, or “AR”, scenario typically involves presentation of digital or virtual image information as an augmentation to visualization of the actual world around the user. For example, referring to
FIG. 1 , an augmented reality scene (4) is depicted wherein a user of an AR technology sees a real-world park-like setting (6) featuring people, trees, buildings in the background, and a concrete platform (1120). In addition to these items, the user of the AR technology also perceives that he “sees” a robot statue (1110) standing upon the real-world platform (1120), and a cartoon-like avatar character (2) flying by which seems to be a personification of a bumble bee, even though these elements (2, 1110) do not exist in the real world. As it turns out, the human visual perception system is very complex, and producing a VR or AR technology that facilitates a comfortable, natural-feeling, rich presentation of virtual image elements amongst other virtual or real-world imagery elements is challenging. - Referring to
FIG. 2A , stereoscopic wearable glasses (8) type configurations have been developed which generally feature two displays (10, 12) that are configured to display images with slightly different element presentation such that a three-dimensional perspective is perceived by the human visual system. Such configurations have been found to be uncomfortable for many users due to a mismatch between vergence and accommodation which must be overcome to perceive the images in three dimensions; indeed, some users are not able to tolerate stereoscopic configurations.FIG. 2B shows another stereoscopic wearable glasses (14) type configuration featuring two forward-oriented cameras (16, 18) configured to capture images for an augmented reality presentation to the user through stereoscopic displays. The position of the cameras (16, 18) and displays generally blocks the natural field of view of the user when the glasses (14) are mounted on the user's head. - Referring to
FIG. 2C , an augmented reality configuration (20) is shown which features a visualization module (26) coupled to a glasses frame (24) which also holds conventional glasses lenses (22). The user is able to see an at least partially unobstructed view of the real world with such a system, and has a small display (28) with which digital imagery may be presented in an AR configuration to one eye—for a monocular AR presentation.FIG. 2D features a configuration wherein a visualization module (32) may be coupled to a hat or helmet (30) and configured to present monocular augmented digital imagery to a user through a small display (34).FIG. 2E illustrates another similar configuration wherein a frame (36) couple-able to a user's head in a manner similar to an eyeglasses coupling so that a visualization module (38) may be utilized to capture images and also present monocular augmented digital imagery to a user through a small display (40). Such a configuration is available, for example, from Google, Inc., of Mountain View, Calif. under the trade name GoogleGlass®. None of these configurations is optimally suited for presenting a rich, binocular, three-dimensional augmented reality experience in a manner that will be comfortable and maximally useful to the user, in part because prior systems fail to address some of the fundamental aspects of the human perception system, including the photoreceptors of the retina and their interoperation with the brain to produce the perception of visualization to the user. - Referring to
FIG. 3 , a simplified cross-sectional view of a human eye is depicted featuring a cornea (42), iris (44), lens—or “crystalline lens” (46), sclera (48), choroid layer (50), macula (52), retina (54), and optic nerve pathway (56) to the brain. The macula is the center of the retina, which is utilized to see moderate detail; at the center of the macula is a portion of the retina that is referred to as the “fovea”, which is utilized for seeing the finest details, and which contains more photoreceptors (approximately 120 cones per visual degree) than any other portion of the retina. The human visual system is not a passive sensor type of system; it is configured to actively scan the environment. In a manner somewhat akin to use of a flatbed scanner to capture an image, or use of a finger to read Braille from a paper, the photoreceptors of the eye fire in response to changes in stimulation, rather than constantly responding to a constant state of stimulation. Thus motion is required to present photoreceptor information to the brain (as is motion of the linear scanner array across a piece of paper in a flatbed scanner, or motion of a finger across a word of Braille imprinted into a paper). Indeed, experiments with substances such as cobra venom, which has been utilized to paralyze the muscles of the eye, have shown that a human subject will experience blindness if positioned with his eyes open, viewing a static scene with venom-induced paralysis of the eyes. In other words, without changes in stimulation, the photoreceptors do not provide input to the brain and blindness is experienced. It is believed that this is at least one reason that the eyes of normal humans have been observed to move back and forth, or dither, in side-to-side motion in what are called “microsaccades”. - As noted above, the fovea of the retina contains the greatest density of photoreceptors, and while humans typically have the perception that they have high-resolution visualization capabilities throughout their field of view, they generally actually have only a small high-resolution center that they are mechanically sweeping around a lot, along with a persistent memory of the high-resolution information recently captured with the fovea. In a somewhat similar manner, the focal distance control mechanism of the eye (ciliary muscles operatively coupled to the crystalline lens in a manner wherein ciliary relaxation causes taut ciliary connective fibers to flatten out the lens for more distant focal lengths; ciliary contraction causes loose ciliary connective fibers, which allow the lens to assume a more rounded geometry for more close-in focal lengths) dithers back and forth by approximately ¼ to ½ diopter to cyclically induce a small amount of what is called “dioptric blur” on both the close side and far side of the targeted focal length; this is utilized by the accommodation control circuits of the brain as cyclical negative feedback that helps to constantly correct course and keep the retinal image of a fixated object approximately in focus.
- The visualization center of the brain also gains valuable perception information from the motion of both eyes and components thereof relative to each other. Vergence movements (i.e., rolling movements of the pupils toward or away from each other to converge the lines of sight of the eyes to fixate upon an object) of the two eyes relative to each other are closely associated with focusing (or “accommodation”) of the lenses of the eyes. Under normal conditions, changing the focus of the lenses of the eyes, or accommodating the eyes, to focus upon an object at a different distance will automatically cause a matching change in vergence to the same distance, under a relationship known as the “accommodation-vergence reflex.” Likewise, a change in vergence will trigger a matching change in accommodation, under normal conditions. Working against this reflex, as do most conventional stereoscopic AR or VR configurations, is known to produce eye fatigue, headaches, or other forms of discomfort in users.
- Movement of the head, which houses the eyes, also has a key impact upon visualization of objects. Humans move their heads to visualize the world around them; they often are in a fairly constant state of repositioning and reorienting the head relative to an object of interest. Further, most people prefer to move their heads when their eye gaze needs to move more than about 20 degrees off center to focus on a particular object (i.e., people do not typically like to look at things “from the corner of the eye”). Humans also typically scan or move their heads in relation to sounds—to improve audio signal capture and utilize the geometry of the ears relative to the head. The human visual system gains powerful depth cues from what is called “head motion parallax”, which is related to the relative motion of objects at different distances as a function of head motion and eye vergence distance (i.e., if a person moves his head from side to side and maintains fixation on an object, items farther out from that object will move in the same direction as the head; items in front of that object will move opposite the head motion; these are very salient cues for where things are spatially in the environment relative to the person—perhaps as powerful as stereopsis). Head motion also is utilized to look around objects, of course.
- Further, head and eye motion are coordinated with something called the “vestibulo-ocular reflex”, which stabilizes image information relative to the retina during head rotations, thus keeping the object image information approximately centered on the retina. In response to a head rotation, the eyes are reflexively and proportionately rotated in the opposite direction to maintain stable fixation on an object. As a result of this compensatory relationship, many humans can read a book while shaking their head back and forth (interestingly, if the book is panned back and forth at the same speed with the head approximately stationary, the same generally is not true—the person is not likely to be able to read the moving book; the vestibulo-ocular reflex is one of head and eye motion coordination, generally not developed for hand motion). This paradigm may be important for augmented reality systems, because head motions of the user may be associated relatively directly with eye motions, and the system preferably will be ready to work with this relationship.
- Indeed, given these various relationships, when placing digital content (e.g., 3-D content such as a virtual chandelier object presented to augment a real-world view of a room; or 2-D content such as a planar/flat virtual oil painting object presented to augment a real-world view of a room), design choices may be made to control behavior of the objects. For example, the 2-D oil painting object may be head-centric, in which case the object moves around along with the user's head (e.g., as in a GoogleGlass approach); or the object may be world-centric, in which case it may be presented as though it is part of the real world coordinate system, so that the user may move his head or eyes without moving the position of the object relative to the real world.
- Thus when placing virtual content into the augmented reality world presented with an augmented reality system, whether the object should be presented as world centric (i.e., the virtual object stays in position in the real world so that the user may move his body, head, eyes around it without changing its position relative to the real world objects surrounding it, such as a real world wall); body, or torso, centric, in which case a virtual element may be fixed relative to the user's torso, so that the user can move his head or eyes without moving the object, but that is slaved to torso movements; head centric, in which case the displayed object (and/or display itself) may be moved along with head movements, as described above in reference to GoogleGlass; or eye centric, as in a “foveated display” configuration, as is described below, wherein content is slewed around as a function of what the eye position is.
- With world-centric configurations, it may be desirable to have inputs such as accurate head pose measurement, accurate representation and/or measurement of real world objects and geometries around the user, low-latency dynamic rendering in the augmented reality display as a function of head pose, and a generally low-latency display.
- The systems and techniques described herein are configured to work with the visual configuration of the typical human to address these challenges.
- Embodiments of the present invention are directed to devices, systems and methods for facilitating virtual reality and/or augmented reality interaction for one or more users. In one aspect, a system for displaying virtual content is disclosed.
- In one or more embodiment, the system comprises an image-generating source to provide one or more frames of image data in a time-sequential manner, a light modulator configured to transmit light associated with the one or more frames of image data, a substrate to direct image information to a user's eye, wherein the substrate houses a plurality of reflectors, a first reflector of the plurality of reflectors to reflect light associated with a first frame of image data at a first angle to the user's eye, and a second reflector of the plurality of reflectors to reflect light associated with a second frame of image data at a second angle to the user's eye.
- In another embodiment, a system for displaying virtual content comprises an image-generating source to provide one or more frames of image data in a time-sequential manner, a display assembly to project light rays associated with the one or more frames of image data, the display assembly comprises a first display element corresponding to a first frame-rate and a first bit depth, and a second display element corresponding to a second frame-rate and a second bit depth, and a variable focus element (VFE) configurable to vary a focus of the projected light and transmit the light to the user's eye.
- In yet another embodiment, a system for displaying virtual content comprises an array of optical fibers to transmit light beams associated with an image to be presented to a user, and a lens coupled to the array of the optical fibers to deflect a plurality of light beams output by the array of optical fibers through a single nodal point, wherein the lens is physically attached to the optical fibers such that a movement of the optical fiber causes the lens to move, and wherein the single nodal point is scanned.
- In another embodiment, a virtual reality display system comprises a plurality of optical fibers to generate light beams associated with one or more images to be presented to a user, and a plurality of phase modulators coupled to the plurality of optical fibers to modulate the light beams, wherein the plurality of phase modulators modulate the light in a manner that affects a wavefront generated as a result of the plurality of light beams.
- In one embodiment, a system for displaying virtual content to a user comprises a light projection system to project light associated with one or more frames of image data to a user's eyes, the light project system configured to project light corresponding to a plurality of pixels associated with the image data and a processor to modulate a size of the plurality of pixels displayed to the user.
- In one embodiment, a system of displaying virtual content to a user, comprises an image-generating source to provide one or more frames of image data, a multicore assembly comprising a plurality of multicore fibers to project light associated with the one or more frames of image data, a multicore fiber of the plurality of multicore fibers emitting light in a wavefront, such that the multicore assembly produces an aggregate wavefront of the projected light, and a phase modulator to induce phase delays between the multicore fibers in a manner such that the aggregate wavefront emitted by the multicore assembly is varied, thereby varying a focal distance at which the user perceives the one or more frames of image data.
- In another embodiment, a system for displaying virtual content to a user comprises an array of microprojectors to project light beams associated with one or more frames of image data to be presented to the user, wherein the microprojector is configurable to be movable relative to one or more microprojectors of the array of the microprojectors, a frame to house the array of microprojectors, a processor operatively coupled to the one or more microprojectors of the array of microprojectors to control one or more light beams transmitted from the one or more projectors in a manner such that the one or more light beams are modulated as a function of a position of the one or more microprojectors relative to the array of microprojectors, thereby enabling delivery of a lightfield image to the user.
- Additional and other objects, features, and advantages of the invention are described in the detail description, figures and claims.
-
FIG. 1 illustrates a user's view of augmented reality (AR) through a wearable AR user device, in one illustrated embodiment. -
FIGS. 2A-2E illustrates various embodiments of wearable AR devices. -
FIG. 3 illustrates a cross-sectional view of the human eye, in one illustrated embodiment. -
FIGS. 4A-4D illustrate one or more embodiments of various internal processing components of the wearable AR device. -
FIGS. 5A-5H illustrate embodiments of transmitting focused light to a user through a transmissive beamsplitter substrate. -
FIGS. 6A and 6B illustrate embodiments of coupling a lens element with the transmissive beamsplitter substrate ofFIGS. 5A-5H . -
FIGS. 7A and 7B illustrate embodiments of using one or more waveguides to transmit light to a user. -
FIGS. 8A-8Q illustrate embodiments of a diffractive optical element (DOE). -
FIGS. 9A and 9B illustrate a wavefront produced from a light projector, according to one illustrated embodiment. -
FIG. 10 illustrates an embodiment of a stacked configuration of multiple transmissive beamsplitter substrate coupled with optical elements, according to one illustrated embodiment. -
FIGS. 11A-11C illustrate a set of beamlets projected into a user's pupil, according to the illustrated embodiments. -
FIGS. 12A and 12B illustrate configurations of an array of microprojectors, according to the illustrated embodiments. -
FIGS. 13A-13M illustrate embodiments of coupling microprojectors with optical elements, according to the illustrated embodiments. -
FIGS. 14A-14F illustrate embodiments of spatial light modulators coupled with optical elements, according to the illustrated embodiments. -
FIGS. 15A-15C illustrate the use of a wedge type waveguides along with a plurality of light sources, according to the illustrated embodiments. -
FIGS. 16A-160 illustrate embodiments of coupling optical elements to optical fibers, according to the illustrated embodiments. -
FIG. 17 illustrates a notch filter, according to one illustrated embodiment. -
FIG. 18 illustrates a spiral pattern of a fiber scanning display, according to one illustrated embodiment. -
FIGS. 19A-19N illustrate occlusion effects in presenting a darkfield to a user, according to the illustrated embodiments. -
FIGS. 20A-200 illustrate embodiments of various waveguide assemblies, according to the illustrated embodiments. -
FIGS. 21A-21N illustrate various configurations of DOEs coupled to other optical elements, according to the illustrated embodiments. -
FIGS. 22A-22Y illustrate various configurations of freeform optics, according to the illustrated embodiments. - Referring to
FIGS. 4A-4D , some general componentry options are illustrated. In the portions of the detailed description which follow the discussion ofFIGS. 4A-4D , various systems, subsystems, and components are presented for addressing the objectives of providing a high-quality, comfortably-perceived display system for human VR and/or AR. - As shown in
FIG. 4A , an AR system user (60) is depicted wearing a frame (64) structure coupled to a display system (62) positioned in front of the eyes of the user. A speaker (66) is coupled to the frame (64) in the depicted configuration and positioned adjacent the ear canal of the user (in one embodiment, another speaker, not shown, is positioned adjacent the other ear canal of the user to provide for stereo/shapeable sound control). The display (62) is operatively coupled (68), such as by a wired lead or wireless connectivity, to a local processing and data module (70) which may be mounted in a variety of configurations, such as fixedly attached to the frame (64), fixedly attached to a helmet or hat (80) as shown in the embodiment ofFIG. 4B , embedded in headphones, removably attached to the torso (82) of the user (60) in a backpack-style configuration as shown in the embodiment ofFIG. 4C , or removably attached to the hip (84) of the user (60) in a belt-coupling style configuration as shown in the embodiment ofFIG. 4D . - The local processing and data module (70) may comprise a power-efficient processor or controller, as well as digital memory, such as flash memory, both of which may be utilized to assist in the processing, caching, and storage of data a) captured from sensors which may be operatively coupled to the frame (64), such as image capture devices (such as cameras), microphones, inertial measurement units, accelerometers, compasses, GPS units, radio devices, and/or gyros; and/or b) acquired and/or processed using the remote processing module (72) and/or remote data repository (74), possibly for passage to the display (62) after such processing or retrieval. The local processing and data module (70) may be operatively coupled (76, 78), such as via a wired or wireless communication links, to the remote processing module (72) and remote data repository (74) such that these remote modules (72, 74) are operatively coupled to each other and available as resources to the local processing and data module (70).
- In one embodiment, the remote processing module (72) may comprise one or more relatively powerful processors or controllers configured to analyze and process data and/or image information. In one embodiment, the remote data repository (74) may comprise a relatively large-scale digital data storage facility, which may be available through the internet or other networking configuration in a “cloud” resource configuration. In one embodiment, all data is stored and all computation is performed in the local processing and data module, allowing fully autonomous use from any remote modules.
- Referring to
FIGS. 5A through 22Y , various display configurations are presented that are designed to present the human eyes with photon-based radiation patterns that can be comfortably perceived as augmentations to physical reality, with high-levels of image quality and three-dimensional perception, as well as being capable of presenting two-dimensional content. - Referring to
FIG. 5A , in a simplified example, a transmissive beamsplitter substrate (104) with a 45-degree reflecting surface (102) directs incoming radiation (106), which may be output from a lens (not shown), through the pupil (45) of the eye (58) and to the retina (54). The field of view for such a system is limited by the geometry of the beamsplitter (104). To accommodate the desire to have comfortable viewing with minimal hardware, in one embodiment, a larger field of view can be created by aggregating the outputs/reflections of various different reflective and/or diffractive surfaces and using, e.g., a frame-sequential configuration wherein eye (58) is presented with a sequence of frames at high frequency that provides the perception of a single coherent scene. As an alternative to, or in addition to, presenting different image data via different reflectors in a time-sequential fashion, the reflectors may separate content by other means, such as polarization selectivity or wavelength selectivity. In addition to being capable of relaying two-dimensional images, the reflectors can relay the three-dimensional wavefronts associated with true-three-dimensional viewing of actual physical objects. - Referring to
FIG. 5B , a substrate (108) comprising a plurality of reflectors at a plurality of angles (110) is shown, with each reflector actively reflecting in the depicted configuration for illustrative purposes. The reflectors may be switchable elements to facilitate temporal selectivity. In one embodiment, the reflective surfaces would intentionally be sequentially activated with frame-sequential input information (106), in which each reflective surface presents a narrow field of view sub-image which is tiled with other narrow field of view sub-images presented by the other reflective surfaces to form a composite wide field of view image. For example, referring toFIGS. 5C , 5D, and 5E, surface (110), about in the middle of substrate (108), is switched “on” to a reflecting state, such that it reflects incoming image information (106) to present a relatively narrow field of view sub-image in the middle of a larger field of view, while the other potential reflective surfaces are in a transmissive state. - Referring to
FIG. 5C , incoming image information (106) coming from the right of the narrow field of view sub-image (as shown by the angle ofincoming beams 106 relative to thesubstrate 108input interface 112, and the resultant angle at which they exit the substrate 108) is reflected toward the eye (58) from reflective surface (110).FIG. 5D illustrates the same reflector (110) active, with image information coming from the middle of the narrow field of view sub-image, as shown by the angle of the input information (106) at the input interface (112) and its angle as it exits substrate (108).FIG. 5E illustrates the same reflector (110) active, with image information coming from the left of the field of view, as shown by the angle of the input information (106) at the input interface (112) and the resultant exit angle at the surface of the substrate (108).FIG. 5F illustrates a configuration wherein the bottom reflector (110) is active, with image information (106) coming in from the far right of the overall field of view. For example,FIGS. 5C , 5D, and 5E can illustrate one frame representing the center of a frame-sequential tiled image, andFIG. 5F can illustrate a second frame representing the far right of that tiled image. - In one embodiment, the light carrying the image information (106) may strike the reflective surface (110) directly after entering substrate (108) at input interface (112), without first reflecting from the surfaces of substrate (108). In one embodiment, the light carrying the image information (106) may reflect from one or more surfaces of substrate (108) after entering at input interface (112) and before striking the reflective surface (110); for instance, substrate (108) may act as a planar waveguide, propagating the light carrying image information (106) by total internal reflection. Light may also reflect from one or more surfaces of the substrate (108) from a partially reflective coating, a wavelength-selective coating, an angle-selective coating, and/or a polarization-selective coating.
- In one embodiment, the angled reflectors may be constructed using an electro-active material, such that upon application of a voltage and/or current to a particular reflector, the refractive index of the material comprising such reflector changes from an index substantially matched to the rest of the substrate (108), in which case the reflector is in a transmissive configuration, to a reflective configuration wherein the refractive index of the reflector mismatches the refractive index of the substrate (108) such that a reflection effect is created. Example electro-active material includes lithium niobate and electro-active polymers. Suitable substantially transparent electrodes for controlling a plurality of such reflectors may comprise materials such as indium tin oxide, which is utilized in liquid crystal displays.
- In one embodiment, the electro-active reflectors (110) may comprise liquid crystal, embedded in a substrate (108) host medium such as glass or plastic. In some variations, liquid crystal may be selected that changes refractive index as a function of an applied electric signal, so that more analog changes may be accomplished as opposed to binary (from one transmissive state to one reflective state). In an embodiment wherein 6 sub-images are to be presented to the eye frame-sequential to form a large tiled image with an overall refresh rate of 60 frames per second, it is desirable to have an input display that can refresh at the rate of about 360 Hz, with an electro-active reflector array that can keep up with such frequency. In one embodiment, lithium niobate may be utilized as an electro-active reflective material as opposed to liquid crystal; lithium niobate is utilized in the photonics industry for high-speed switches and fiber optic networks and has the capability to switch refractive index in response to an applied voltage at a very high frequency; this high frequency may be used to steer line-sequential or pixel-sequential sub-image information, especially if the input display is a scanned light display, such as a fiber-scanned display or scanning mirror-based display.
- In another embodiment, a variable switchable angled mirror configuration may comprise one or more high-speed mechanically repositionable reflective surfaces, such as a MEMS (micro-electro-mechanical system) device. A MEMS device may include what is known as a “digital mirror device”, or “DMD”, (often part of a “digital light processing”, or “DLP” system, such as those available from Texas Instruments, Inc.). In another electromechanical embodiment, a plurality of air-gapped (or in vacuum) reflective surfaces could be mechanically moved in and out of place at high frequency. In another electromechanical embodiment, a single reflective surface may be moved up and down and re-pitched at very high frequency.
- Referring to
FIG. 5G , it is notable that the switchable variable angle reflector configurations described herein are capable of passing not only collimated or flat wavefront information to the retina (54) of the eye (58), but also curved wavefront (122) image information, as shown in the illustration ofFIG. 5G . This generally is not the case with other waveguide-based configurations, wherein total internal reflection of curved wavefront information causes undesirable complications, and therefore the inputs generally must be collimated. The ability to pass curved wavefront information facilitates the ability of configurations such as those shown inFIGS. 5B-5H to provide the retina (54) with input perceived as focused at various distances from the eye (58), not just optical infinity (which would be the interpretation of collimated light absent other cues). - Referring to
FIG. 5H , in another embodiment, an array of static partially reflective surfaces (116) (i.e., always in a reflective mode; in another embodiment, they may be electro-active, as above) may be embedded in a substrate (114) with a high-frequency gating layer (118) controlling outputs to the eye (58) by only allowing transmission through an aperture (120) which is controllably movable. In other words, everything may be selectively blocked except for transmissions through the aperture (120). The gating layer (118) may comprise a liquid crystal array, a lithium niobate array, an array of MEMS shutter elements, an array of DLP DMD elements, or an array of other MEMS devices configured to pass or transmit with relatively high-frequency switching and high transmissibility upon being switched to transmission mode. - Referring to
FIGS. 6A-6B , other embodiments are depicted wherein arrayed optical elements may be combined with exit pupil expansion configurations to assist with the comfort of the virtual or augmented reality experience of the user. With a larger “exit pupil” for the optics configuration, the user's eye positioning relative to the display (which, as inFIGS. 4A-4D , may be mounted on the user's head in an eyeglasses sort of configuration) is not as likely to disrupt his experience—because due to the larger exit pupil of the system, there is a larger acceptable area wherein the user's anatomical pupil may be located to still receive the information from the display system as desired. In other words, with a larger exit pupil, the system is less likely to be sensitive to slight misalignments of the display relative to the user's anatomical pupil, and greater comfort for the user may be achieved through less geometric constraint on his or her relationship with the display/glasses. - As shown in
FIG. 6A , the display (140) on the left feeds a set of parallel rays into the substrate (124). In one embodiment, the display may be a scanned fiber display scanning a narrow beam of light back and forth at an angle as shown to project an image through the lens or other optical element (142), which may be utilized to collect the angularly-scanned light and convert it to a parallel bundle of rays. The rays may be reflected from a series of reflective surfaces (126, 128, 130, 132, 134, 136) which may be configured to partially reflect and partially transmit incoming light so that the light may be shared across the group of reflective surfaces (126, 128, 130, 132, 134, 136) approximately equally. With a small lens (138) placed at each exit point from the waveguide (124), the exiting light rays may be steered through a nodal point and scanned out toward the eye (58) to provide an array of exit pupils, or the functional equivalent of one large exit pupil that is usable by the user as he or she gazes toward the display system. - For virtual reality configurations wherein it is desirable to also be able to see through the waveguide to the real world (144), a similar set of lenses (139) may be presented on the opposite side of the waveguide (124) to compensate for the lower set of lenses; thus creating a the equivalent of a zero-magnification telescope. The reflective surfaces (126, 128, 130, 132, 134, 136) each may be aligned at approximately 45 degrees as shown, or may be configured to have different alignments, akin to the configurations of
FIGS. 5B-5H , for example). The reflective surfaces (126, 128, 130, 132, 134, 136) may comprise wavelength-selective reflectors, band pass reflectors, half silvered mirrors, or other reflective configurations. The lenses (138, 139) shown are refractive lenses, but diffractive lens elements may also be utilized. - Referring to
FIG. 6B , a somewhat similar configuration is depicted wherein a plurality of curved reflective surfaces (148, 150, 152, 154, 156, 158) may be utilized to effectively combine the lens (element 138 ofFIG. 6A ) and reflector (elements FIG. 6A ) functionality of the embodiment ofFIG. 6A , thereby obviating the need for the two groups of lenses (element 138 ofFIG. 6A ). The curved reflective surfaces (148, 150, 152, 154, 156, 158) may be various curved configurations selected to both reflect and impart angular change, such as parabolic or elliptical curved surfaces. With a parabolic shape, a parallel set of incoming rays will be collected into a single output point; with an elliptical configuration, a set of rays diverging from a single point of origin are collected to a single output point. As with the configuration ofFIG. 6A , the curved reflective surfaces (148, 150, 152, 154, 156, 158) preferably are configured to partially reflect and partially transmit so that the incoming light is shared across the length of the waveguide (146). The curved reflective surfaces (148, 150, 152, 154, 156, 158) may comprise wavelength-selective notch reflectors, half silvered mirrors, or other reflective configurations. In another embodiment, the curved reflective surfaces (148, 150, 152, 154, 156, 158) may be replaced with diffractive reflectors configured to reflect and also deflect. - Referring to
FIG. 7A , perceptions of Z-axis difference (i.e., distance straight out from the eye along the optical axis) may be facilitated by using a waveguide in conjunction with a variable focus optical element configuration. As shown inFIG. 7A , image information from a display (160) may be collimated and injected into a waveguide (164) and distributed in a large exit pupil manner using, e.g., configurations such as those described in reference toFIGS. 6A and 6B , or other substrate-guided optics methods known to those skilled in the art—and then variable focus optical element capability may be utilized to change the focus of the wavefront of light emerging from the waveguide and provide the eye with the perception that the light coming from the waveguide (164) is from a particular focal distance. In other words, since the incoming light has been collimated to avoid challenges in total internal reflection waveguide configurations, it will exit in collimated fashion, requiring a viewer's eye to accommodate to the far point to bring it into focus on the retina, and naturally be interpreted as being from optical infinity—unless some other intervention causes the light to be refocused and perceived as from a different viewing distance; one suitable such intervention is a variable focus lens. - In the embodiment of
FIG. 7A , collimated image information is injected into a piece of glass (162) or other material at an angle such that it totally internally reflects and is passed into the adjacent waveguide (164). The waveguide (164) may be configured akin to the waveguides ofFIGS. 6A or 6B (124, 146, respectively) so that the collimated light from the display is distributed to exit somewhat uniformly across the distribution of reflectors or diffractive features along the length of the waveguide. Upon exit toward the eye (58), in the depicted configuration the exiting light is passed through a variable focus lens element (166) wherein, depending upon the controlled focus of the variable focus lens element (166), the light exiting the variable focus lens element (166) and entering the eye (58) will have various levels of focus (a collimated flat wavefront to represent optical infinity, more and more beam divergence/wavefront curvature to represent closer viewing distance relative to the eye 58). - To compensate for the variable focus lens element (166) between the eye (58) and the waveguide (164), another similar variable focus lens element (167) is placed on the opposite side of the waveguide (164) to cancel out the optical effects of the lenses (166) for light coming from the world (144) for augmented reality (i.e., as described above, one lens compensates for the other, producing the functional equivalent of a zero-magnification telescope).
- The variable focus lens element (166) may be a refractive element, such as a liquid crystal lens, an electro-active lens, a conventional refractive lens with moving elements, a mechanical-deformation-based lens (such as a fluid-filled membrane lens, or a lens akin to the human crystalline lens wherein a flexible element is flexed and relaxed by actuators), an electrowetting lens, or a plurality of fluids with different refractive indices. The variable focus lens element (166) may also comprise a switchable diffractive optical element (such as one featuring a polymer dispersed liquid crystal approach wherein a host medium, such as a polymeric material, has microdroplets of liquid crystal dispersed within the material; when a voltage is applied, the molecules reorient so that their refractive indices no longer match that of the host medium, thereby creating a high-frequency switchable diffraction pattern).
- One embodiment includes a host medium in which microdroplets of a Kerr effect-based electro-active material, such as lithium niobate, is dispersed within the host medium, enabling refocusing of image information on a pixel-by-pixel or line-by-line basis, when coupled with a scanning light display, such as a fiber-scanned display or scanning-mirror-based display. In a variable focus lens element (166) configuration wherein liquid crystal, lithium niobate, or other technology is utilized to present a pattern, the pattern spacing may be modulated to not only change the focal power of the variable focus lens element (166), but also to change the focal power of the overall optical system—for a zoom lens type of functionality.
- In one embodiment, the lenses (166) could be telecentric, in that focus of the display imagery can be altered while keeping magnification constant—in the same way that a photography zoom lens may be configured to decouple focus from zoom position. In another embodiment, the lenses (166) may be non-telecentric, so that focus changes will also slave zoom changes. With such a configuration, such magnification changes may be compensated for in software with dynamic scaling of the output from the graphics system in sync with focus changes).
- Referring back to the projector or other video display unit (160) and the issue of how to feed images into the optical display system, in a “frame sequential” configuration, a stack of sequential two-dimensional images may be fed to the display sequentially to produce three-dimensional perception over time; in a manner akin to the manner in which a computed tomography system uses stacked image slices to represent a three-dimensional structure. A series of two-dimensional image slices may be presented to the eye, each at a different focal distance to the eye, and the eye/brain would integrate such a stack into a perception of a coherent three-dimensional volume. Depending upon the display type, line-by-line, or even pixel-by-pixel sequencing may be conducted to produce the perception of three-dimensional viewing. For example, with a scanned light display (such as a scanning fiber display or scanning mirror display), then the display is presenting the waveguide (164) with one line or one pixel at a time in a sequential fashion.
- If the variable focus lens element (166) is able to keep up with the high-frequency of pixel-by-pixel or line-by-line presentation, then each line or pixel may be presented and dynamically focused through the variable focus lens element (166) to be perceived at a different focal distance from the eye (58). Pixel-by-pixel focus modulation generally requires an extremely fast/high-frequency variable focus lens element (166). For example, a 1080P resolution display with an overall frame rate of 60 frames per second typically presents around 125 million pixels per second. Such a configuration also may be constructed using a solid state switchable lens, such as one using an electro-active material, e.g., lithium niobate or an electro-active polymer. In addition to its compatibility with the system illustrated in
FIG. 7A , a frame sequential multi-focal display driving approach may be used in conjunction with a number of the display system and optics embodiments described in this disclosure. - Referring to
FIG. 7B , with an electro-active layer (172) (such as one comprising liquid crystal or lithium niobate) surrounded by functional electrodes (170, 174) which may be made of indium tin oxide, a waveguide (168) with a conventional transmissive substrate (176, such as one made from glass or plastic with known total internal reflection characteristics and an index of refraction that matches the on or off state of the electro-active layer 172) may be controlled such that the paths of entering beams may be dynamically altered to essentially create a time-varying light field. - Referring to
FIG. 8A , a stacked waveguide assembly (178) may be utilized to provide three-dimensional perception to the eye/brain by having a plurality of waveguides (182, 184, 186, 188, 190) and a plurality of weak lenses (198, 196, 194, 192) configured together to send image information to the eye with various levels of wavefront curvature for each waveguide level indicative of focal distance to be perceived for that waveguide level. A plurality of displays (200, 202, 204, 206, 208), or in another embodiment a single multiplexed display, may be utilized to inject collimated image information into the waveguides (182, 184, 186, 188, 190), each of which may be configured, as described above, to distribute incoming light substantially equally across the length of each waveguide, for exit down toward the eye. - The waveguide (182) nearest the eye is configured to deliver collimated light, as injected into such waveguide (182), to the eye, which may be representative of the optical infinity focal plane. The next waveguide up (184) is configured to send out collimated light which passes through the first weak lens (192; e.g., a weak negative lens) before it can reach the eye (58); such first weak lens (192) may be configured to create a slight convex wavefront curvature so that the eye/brain interprets light coming from that next waveguide up (184) as coming from a first focal plane closer inward toward the person from optical infinity. Similarly, the third up waveguide (186) passes its output light through both the first (192) and second (194) lenses before reaching the eye (58); the combined optical power of the first (192) and second (194) lenses may be configured to create another incremental amount of wavefront divergence so that the eye/brain interprets light coming from that third waveguide up (186) as coming from a second focal plane even closer inward toward the person from optical infinity than was light from the next waveguide up (184).
- The other waveguide layers (188, 190) and weak lenses (196, 198) are similarly configured, with the highest waveguide (190) in the stack sending its output through all of the weak lenses between it and the eye for an aggregate focal power representative of the closest focal plane to the person. To compensate for the stack of lenses (198, 196, 194, 192) when viewing/interpreting light coming from the world (144) on the other side of the stacked waveguide assembly (178), a compensating lens layer (180) is disposed at the top of the stack to compensate for the aggregate power of the lens stack (198, 196, 194, 192) below. Such a configuration provides as many perceived focal planes as there are available waveguide/lens pairings, again with a relatively large exit pupil configuration as described above. Both the reflective aspects of the waveguides and the focusing aspects of the lenses may be static (i.e., not dynamic or electro-active). In an alternative embodiment they may be dynamic using electro-active features as described above, enabling a small number of waveguides to be multiplexed in a time sequential fashion to produce a larger number of effective focal planes.
- Referring to
FIGS. 8B-8N , various aspects of diffraction configurations for focusing and/or redirecting collimated beams are depicted. Other aspects of diffraction systems for such purposes are disclosed in U.S. Patent Application Ser. No. 61/845,907 (U.S. patent application Ser. No. 14/331,218), which is incorporated by reference herein in its entirety. Referring toFIG. 8B , passing a collimated beam through a linear diffraction pattern (210), such as a Bragg grating, will deflect, or “steer”, the beam. Passing a collimated beam through a radially symmetric diffraction pattern (212), or “Fresnel zone plate”, will change the focal point of the beam.FIG. 8C illustrates the deflection effect of passing a collimated beam through a linear diffraction pattern (210);FIG. 8D illustrates the focusing effect of passing a collimated beam through a radially symmetric diffraction pattern (212). - Referring to
FIGS. 8E and 8F , a combination diffraction pattern that has both linear and radial elements (214) produces both deflection and focusing of a collimated input beam. These deflection and focusing effects can be produced in a reflective as well as transmissive mode. These principles may be applied with waveguide configurations to allow for additional optical system control, as shown inFIGS. 8G-8N , for example. As shown inFIGS. 8G-8N , a diffraction pattern (220), or “diffractive optical element” (or “DOE”) has been embedded within a planar waveguide (216) such that as a collimated beam is totally internally reflected along the planar waveguide (216), it intersects the diffraction pattern (220) at a multiplicity of locations. - Preferably, the DOE (220) has a relatively low diffraction efficiency so that only a portion of the light of the beam is deflected away toward the eye (58) with each intersection of the DOE (220) while the rest continues to move through the planar waveguide (216) via total internal reflection; the light carrying the image information is thus divided into a number of related light beams that exit the waveguide at a multiplicity of locations and the result is a fairly uniform pattern of exit emission toward the eye (58) for this particular collimated beam bouncing around within the planar waveguide (216), as shown in
FIG. 8H . The exit beams toward the eye (58) are shown inFIG. 8H as substantially parallel, because, in this case, the DOE (220) has only a linear diffraction pattern. As shown in the comparison betweenFIGS. 8L , 8M, and 8N, changes to this linear diffraction pattern pitch may be utilized to controllably deflect the exiting parallel beams, thereby producing a scanning or tiling functionality. - Referring back to
FIG. 8I , with changes in the radially symmetric diffraction pattern component of the embedded DOE (220), the exit beam pattern is more divergent, which would require the eye to accommodation to a closer distance to bring it into focus on the retina and would be interpreted by the brain as light from a viewing distance closer to the eye than optical infinity. Referring toFIG. 8J , with the addition of another waveguide (218) into which the beam may be injected (by a projector or display, for example), a DOE (221) embedded in this other waveguide (218), such as a linear diffraction pattern, may function to spread the light across the entire larger planar waveguide (216), which functions to provide the eye (58) with a very large incoming field of incoming light that exits from the larger planar waveguide (216), i.e., a large eye box, in accordance with the particular DOE configurations at work. - The DOEs (220, 221) are depicted bisecting the associated waveguides (216, 218) but this need not be the case; they could be placed closer to, or upon, either side of either of the waveguides (216, 218) to have the same functionality. Thus, as shown in
FIG. 8K , with the injection of a single collimated beam, an entire field of cloned collimated beams may be directed toward the eye (58). In addition, with a combined linear diffraction pattern/radially symmetric diffraction pattern scenario such as that depicted inFIGS. 8F (214) and 8I (220), a beam distribution waveguide optic (for functionality such as exit pupil functional expansion; with a configuration such as that ofFIG. 8K , the exit pupil can be as large as the optical element itself, which can be a very significant advantage for user comfort and ergonomics) with Z-axis focusing capability is presented, in which both the divergence angle of the cloned beams and the wavefront curvature of each beam represent light coming from a point closer than optical infinity. - In one embodiment, one or more DOEs are switchable between “on” states in which they actively diffract, and “off” states in which they do not significantly diffract. For instance, a switchable DOE may comprise a layer of polymer dispersed liquid crystal, in which microdroplets comprise a diffraction pattern in a host medium, and the refractive index of the microdroplets can be switched to substantially match the refractive index of the host material (in which case the pattern does not appreciably diffract incident light) or the microdroplet can be switched to an index that does not match that of the host medium (in which case the pattern actively diffracts incident light). Further, with dynamic changes to the diffraction terms, such as the linear diffraction pitch term as in
FIGS. 8L-8N , a beam scanning or tiling functionality may be achieved. As noted above, it is desirable to have a relatively low diffraction grating efficiency in each of the DOEs (220, 221) because it facilitates distribution of the light, and also because light coming through the waveguides that is desirably transmitted (for example, light coming from theworld 144 toward theeye 58 in an augmented reality configuration) is less affected when the diffraction efficiency of the DOE that it crosses (220) is lower—so a better view of the real world through such a configuration is achieved. - Configurations such as those illustrated in
FIG. 8K preferably are driven with injection of image information in a time sequential approach, with frame sequential driving being the most straightforward to implement. For example, an image of the sky at optical infinity may be injected at time1 and the diffraction grating retaining collimation of light may be utilized; then an image of a closer tree branch may be injected at time2 while a DOE controllably imparts a focal change, say one diopter or 1 meter away, to provide the eye/brain with the perception that the branch light information is coming from the closer focal range. This kind of paradigm can be repeated in rapid time sequential fashion such that the eye/brain perceives the input to be all part of the same image. This is just a two focal plane example; preferably the system will be configured to have more focal planes to provide a smoother transition between objects and their focal distances. This kind of configuration generally assumes that the DOE is switched at a relatively low speed (i.e., in sync with the frame-rate of the display that is injecting the images—in the range of tens to hundreds of cycles/second). - The opposite extreme may be a configuration wherein DOE elements can shift focus at tens to hundreds of MHz or greater, which facilitates switching of the focus state of the DOE elements on a pixel-by-pixel basis as the pixels are scanned into the eye (58) using a scanned light display type of approach. This is desirable because it means that the overall display frame-rate can be kept quite low; just low enough to make sure that “flicker” is not a problem (in the range of about 60-120 frames/sec).
- In between these ranges, if the DOEs can be switched at KHz rates, then on a line-by-line basis the focus on each scan line may be adjusted, which may afford the user with a visible benefit in terms of temporal artifacts during an eye motion relative to the display, for example. For instance, the different focal planes in a scene may, in this manner, be interleaved, to minimize visible artifacts in response to a head motion (as is discussed in greater detail later in this disclosure). A line-by-line focus modulator may be operatively coupled to a line scan display, such as a grating light valve display, in which a linear array of pixels is swept to form an image; and may be operatively coupled to scanned light displays, such as fiber-scanned displays and mirror-scanned light displays.
- A stacked configuration, similar to those of
FIG. 8A , may use dynamic DOEs (rather than the static waveguides and lenses of the embodiment ofFIG. 8A ) to provide multi-planar focusing simultaneously. For example, with three simultaneous focal planes, a primary focus plane (based upon measured eye accommodation, for example) could be presented to the user, and a +margin and −margin (i.e., one focal plane closer, one farther out) could be utilized to provide a large focal range in which the user can accommodate before the planes need be updated. This increased focal range can provide a temporal advantage if the user switches to a closer or farther focus (i.e., as determined by accommodation measurement); then the new plane of focus could be made to be the middle depth of focus, with the + and − margins again ready for a fast switchover to either one while the system catches up. - Referring to
FIG. 8O , a stack (222) of planar waveguides (244, 246, 248, 250, 252) is shown, each having a reflector (254, 256, 258, 260, 262) at the end and being configured such that collimated image information injected in one end by a display (224, 226, 228, 230, 232) bounces by total internal reflection down to the reflector, at which point some or all of the light is reflected out toward an eye or other target. Each of the reflectors may have slightly different angles so that they all reflect exiting light toward a common destination such as a pupil. Such a configuration is somewhat similar to that ofFIG. 5B , with the exception that each different angled reflector in the embodiment ofFIG. 8O has its own waveguide for less interference when projected light is travelling to the targeted reflector. Lenses (234, 236, 238, 240, 242) may be interposed between the displays and waveguides for beam steering and/or focusing. -
FIG. 8P illustrates a geometrically staggered version wherein reflectors (276, 278, 280, 282, 284) are positioned at staggered lengths in the waveguides (266, 268, 270, 272, 274) so that exiting beams may be relatively easily aligned with objects such as an anatomical pupil. With knowledge of how far the stack (264) is going to be from the eye (such as 28 mm between the cornea of the eye and an eyeglasses lens, a typical comfortable geometry), the geometries of the reflectors (276, 278, 280, 282, 284) and waveguides (266, 268, 270, 272, 274) may be set up to fill the eye pupil (typically about 8 mm across or less) with exiting light. By directing light to an eye box larger than the diameter of the eye pupil, the viewer may make eye movements while retaining the ability to see the displayed imagery. Referring back to the discussion related toFIGS. 5A and 5B about field of view expansion and reflector size, an expanded field of view is presented by the configuration ofFIG. 8P as well, and it does not involve the complexity of the switchable reflective elements of the embodiment ofFIG. 5B . -
FIG. 8Q illustrates a version wherein many reflectors (298) form a relatively continuous curved reflection surface in the aggregate or discrete flat facets that are oriented to align with an overall curve. The curve could a parabolic or elliptical curve and is shown cutting across a plurality of waveguides (288, 290, 292, 294, 296) to minimize any crosstalk issues, although it also could be utilized with a monolithic waveguide configuration. - In one implementation, a high-frame-rate and lower persistence display may be combined with a lower-frame-rate and higher persistence display and a variable focus element to comprise a relatively high-frequency frame sequential volumetric display. In one embodiment, the high-frame-rate display has a lower bit depth and the lower-frame-rate display has a higher bit depth, and are combined to comprise an effective high-frame-rate and high bit depth display, that is well suited to presenting image slices in a frame sequential fashion. With such an approach, a three-dimensional volume that is desirably represented is functionally divided into a series of two-dimensional slices. Each of those two-dimensional slices is projected to the eye frame sequentially, and in sync with this presentation, the focus of a variable focus element is changed.
- In one embodiment, to get enough frame rate to support such a configuration, two display elements may be integrated: a full-color, high-resolution liquid crystal display (“LCD”; a backlighted ferroelectric panel display also may be utilized in another embodiment; in a further embodiment a scanning fiber display may be utilized) operating at 60 frames per second, and aspects of a higher-frequency DLP system. Instead of illuminating the back of the LCD panel in a conventional manner (i.e., with a full size fluorescent lamp or LED array), the conventional lighting configuration may be removed to accommodate using the DLP projector to project a mask pattern on the back of the LCD (in one embodiment, the mask pattern may be binary in that the DLP either projects illumination, or not-illumination; in another embodiment described below, the DLP may be utilized to project a grayscale mask image).
- DLP projection systems can operate at very high frame rates; in one embodiment for 6 depth planes at 60 frames per second, a DLP projection system can be operated against the back of the LCD display at 360 frames/second. Then the DLP projector is utilized to selectively illuminate portions of the LCD panel in sync with a high-frequency variable focus element (such as a deformable membrane mirror) that is disposed between the viewing side of the LCD panel and the eye of the user, the variable focus element being used to change the global display focus on a frame by frame basis at 360 frames/second. In one embodiment, the variable focus element is positioned to be optically conjugate to the exit pupil, to enable adjustments of focus without simultaneously affecting image magnification or “zoom.” In another embodiment, the variable focus element is not conjugate to the exit pupil, such that image magnification changes accompany focus adjustments, and software is used to compensate for these optical magnification changes and any distortions by pre-scaling or warping the images to be presented.
- Operationally, it's useful to consider an example again wherein a three-dimensional scene is to be presented to a user wherein the sky in the background is to be at a viewing distance of optical infinity, and wherein a branch coupled to a tree located at a certain location closer to the user than optical infinity extends from the tree trunk in a direction toward the user, so that the tip of the branch is closer to the user than is the proximal portion of the branch that joins the tree trunk.
- In one embodiment, for a given global frame, the system may be configured to present on an LCD a full-color, all in-focus image of the tree branch in front the sky. Then at subframe1, within the global frame, the DLP projector in a binary masking configuration (i.e., illumination or absence of illumination) may be used to only illuminate the portion of the LCD that represents the cloudy sky while functionally black-masking (i.e., failing to illuminate) the portion of the LCD that represents the tree branch and other elements that are not to be perceived at the same focal distance as the sky, and the variable focus element (such as a deformable membrane mirror) may be utilized to position the focal plane at optical infinity so that the eye sees a sub-image at subframe1 as being clouds that are infinitely far away.
- Then at subframe2, the variable focus element may be switched to focusing on a point about 1 meter away from the user's eyes (or whatever distance is required; here 1 meter for the branch location is used for illustrative purposes), the pattern of illumination from the DLP can be switched so that the system only illuminates the portion of the LCD that represents the tree branch while functionally black-masking (i.e., failing to illuminate) the portion of the LCD that represents the sky and other elements that are not to be perceived at the same focal distance as the tree branch. Thus the eye gets a quick flash of cloud at optical infinity followed by a quick flash of tree at 1 meter, and the sequence is integrated by the eye/brain to form a three-dimensional perception. The branch may be positioned diagonally relative to the viewer, such that it extends through a range of viewing distances, e.g., it may join with the trunk at around 2 meters viewing distance while the tips of the branch are at the closer position of 1 meter.
- In this case, the display system can divide the 3-D volume of the tree branch into multiple slices, rather than a single slice at 1 meter. For instance, one focus slice may be used to represent the sky (using the DLP to mask all areas of the tree during presentation of this slice), while the tree branch is divided across 5 focus slices (using the DLP to mask the sky and all portions of the tree except one, for each part of the tree branch to be presented). Preferably, the depth slices are positioned with a spacing equal to or smaller than the depth of focus of the eye, such that the viewer will be unlikely to notice the transition between slices, and instead perceive a smooth and continuous flow of the branch through the focus range.
- In another embodiment, rather than utilizing the DLP in a binary (illumination or darkfield only) mode, it may be utilized to project a grayscale (for example, 256 shades of grayscale) mask onto the back of the LCD panel to enhance three-dimensional perception. The grayscale shades may be utilized to impart to the eye/brain a perception that something resides in between adjacent depth or focal planes. Back to the branch and clouds scenario, if the leading edge of the branch closest to the user is to be in focalplane1, then at subframe1, that portion branch on the LCD may be lit up with full intensity white from the DLP system with the variable focus element at focalplane1.
- Then at subframe2, with the variable focus element at focalplane2 right behind the part that was lit up, there would be no illumination. These are similar steps to the binary DLP masking configuration above. However, if there is a portion of the branch that is to be perceived at a position between focalplane1 and focalplane1, e.g., halfway, grayscale masking can be utilized. The DLP can project an illumination mask to that portion during both subframe1 and subframe2, but at half-illumination (such as at
level 128 out of 256 grayscale) for each subframe. This provides the perception of a blending of depth of focus layers, with the perceived focal distance being proportional to the illuminance ratio between subframe1 and subframe2. For instance, for a portion of the tree branch that should lie ¾ths of the way between focalplane1 and focalplane2, an about 25% intensity grayscale mask can be used to illuminate that portion of the LCD at subframe1 and an about 75% grayscale mask can be used to illuminate the same portion of the LCD at subframe2. - In one embodiment, the bit depths of both the low-frame-rate display and the high-frame-rate display can be combined for image modulation, to create a high dynamic range display. The high dynamic range driving may be conducted in tandem with the focus plane addressing function described above, to comprise a high dynamic range multi-focal 3-D display.
- In another embodiment that may be more efficient on computation resources, only a certain portion of the display (i.e., LCD) output may be mask-illuminated by the DMD and variably focused en route to the user's eye. For example, the middle portion of the display may be mask illuminated, with the periphery of the display not providing varying accommodation cues to the user (i.e. the periphery could be uniformly illuminated by the DLP DMD, while a central portion is actively masked and variably focused en route to the eye).
- In the above described embodiment, a refresh rate of about 360 Hz allows for 6 depth planes at about 60 frames/second each. In another embodiment, even higher refresh rates may be achieved by increasing the operating frequency of the DLP. A standard DLP configuration uses a MEMS device and an array of micro-mirrors that toggle between a mode of reflecting light toward the display or user to a mode of reflecting light away from the display or user, such as into a light trap—thus they are inherently binary. DLPs typically create grayscale images using a pulse width modulation schema wherein the mirror is left in the “on” state for a variable amount of time for a variable duty cycle in order to create a brighter pixel, or pixel of interim brightness. Thus, to create grayscale images at moderate frame rate, they are running at a much higher binary rate.
- In the above described configurations, such setup works well for creating grayscale masking. However, if the DLP drive scheme is adapted so that it is flashing subimages in a binary pattern, then the frame rate may be increased significantly—by thousands of frames per second, which allows for hundreds to thousands of depth planes being refreshed at 60 frames/second, which may be utilized to obviate the between-depth-plane grayscale interpolating as described above. A typical pulse width modulation scheme for a Texas Instruments DLP system has an 8-bit command signal (first bit is the first long pulse of the mirror; second bit is a pulse that is half as long as the first; third bit is half as long again; and so on)—so that the configuration can create 2 to the 8th power different illumination levels. In one embodiment, the backlighting from the DLP may have its intensity varied in sync with the different pulses of the DMD to equalize the brightness of the subimages that are created, which is a practical workaround to get existing DMD drive electronics to produce significantly higher frame rates.
- In another embodiment, direct control changes to the DMD drive electronics and software may be utilized to have the mirrors always have an equal on-time instead of the variable on-time configuration that is conventional, which would facilitate higher frame rates. In another embodiment, the DMD drive electronics may be configured to present low bit depth images at a frame rate above that of high bit depth images but lower than the binary frame rate, enabling some grayscale blending between focus planes, while moderately increasing the number of focus planes.
- In another embodiment, when limited to a finite number of depth planes, such as 6 in the example above, it is desirable to functionally move these 6 depth planes around to be maximally useful in the scene that is being presented to the user. For example, if a user is standing in a room and a virtual monster is to be placed into his augmented reality view, the virtual monster being about 2 feet deep in the Z axis straight away from the user's eyes, then it makes sense to cluster all 6 depth planes around the center of the monster's current location (and dynamically move them with him as he moves relative to the user)—so that more rich accommodation cues may be provided for the user, with all six depth planes in the direct region of the monster (for example, 3 in front of the center of the monster, 3 in back of the center of the monster). Such allocation of depth planes is content dependent.
- For example, in the scene above the same monster is to be presented in the same room, but also to be presented to the user is a virtual window frame element, and then a virtual view to optical infinity out of the virtual window frame, it will be useful to spend at least one depth plane on optical infinity, one on the depth of the wall that is to house the virtual window frame, and then perhaps the remaining four depth planes on the monster in the room. If the content causes the virtual window to disappear, then the two depth planes may be dynamically reallocated to the region around the monster, and so on—content-based dynamic allocation of focal plane resources to provide the most rich experience to the user given the computing and presentation resources.
- In another embodiment, phase delays in a multicore fiber or an array of single-core fibers may be utilized to create variable focus light wavefronts. Referring to
FIG. 9A , a multicore fiber (300) may comprise the aggregation of multiple individual fibers (302);FIG. 9B shows a close-up view of a multicore assembly, which emits light from each core in the form of a spherical wavefront (304) from each. If the cores are transmitting coherent light, e.g., from a shared laser light source, these small spherical wavefronts ultimately constructively and destructively interfere with each other, and if they were emitted from the multicore fiber in phase, they will develop an approximately planar wavefront (306) in the aggregate, as shown. However, if phase delays are induced between the cores (using a conventional phase modulator such as one using lithium niobate, for example, to slow the path of some cores relative to others), then a curved or spherical wavefront may be created in the aggregate, to represent at the eyes/brain an object coming from a point closer than optical infinity, which presents another option that may be used in place of the variable focus elements described above. In other words, such a phased multicore configuration, or phased array, may be utilized to create multiple optical focus levels from a light source. - In another embodiment related to the use of optical fibers, a known Fourier transform aspect of multi-mode optical fiber or light guiding rods or pipes may be utilized for control of the wavefronts that are output from such fiber. Optical fibers typically are available in two categories: single mode and multi-mode. Multi-mode optical fiber typically has larger core diameters and allows light to propagate along multiple angular paths, rather than just the one of single mode optical fiber. It is known that if an image is injected into one end of a multi-mode fiber, that angular differences that are encoded into that image will be retained to some degree as it propagates through the multi-mode fiber, and for some configurations the output from the fiber will be significantly similar to a Fourier transform of the image that was input.
- Thus in one embodiment, the inverse Fourier transform of a wavefront (such as a diverging spherical wavefront to represent a focal plane nearer to the user than optical infinity) may be input so that, after passing through the fiber that optically imparts a Fourier transform, the output is the desired shaped, or focused, wavefront. Such output end may be scanned about to be used as a scanned fiber display, or may be used as a light source for a scanning mirror to form an image, for instance. Thus such a configuration may be utilized as yet another focus modulation subsystem. Other kinds of light patterns and wavefronts may be injected into a multi-mode fiber, such that on the output end, a certain spatial pattern is emitted. This may be utilized to have the equivalent of a wavelet pattern (in optics, an optical system may be analyzed in terms of what are called the Zernicke coefficients; images may be similarly characterized and decomposed into smaller principal components, or a weighted combination of comparatively simpler image components). Thus if light is scanned into the eye using the principal components on the input side, a higher resolution image may be recovered at the output end of the multi-mode fiber.
- In another embodiment, the Fourier transform of a hologram may be injected into the input end of a multi-mode fiber to output a wavefront that may be used for three-dimensional focus modulation and/or resolution enhancement. Certain single fiber core, multi-core fibers, or concentric core+cladding configurations also may be utilized in the aforementioned inverse Fourier transform configurations.
- In another embodiment, rather than physically manipulating the wavefronts approaching the eye of the user at a high frame rate without regard to the user's particular state of accommodation or eye gaze, a system may be configured to monitor the user's accommodation and rather than presenting a set of multiple different light wavefronts, present a single wavefront at a time that corresponds to the accommodation state of the eye. Accommodation may be measured directly (such as by infrared autorefractor or eccentric photorefraction) or indirectly (such as by measuring the convergence level of the two eyes of the user; as described above, vergence and accommodation are strongly linked neurologically, so an estimate of accommodation can be made based upon vergence geometry). Thus with a determined accommodation of, say, 1 meter from the user, then the wavefront presentations at the eye may be configured for a 1 meter focal distance using any of the above variable focus configurations. If an accommodation change to focus at 2 meters is detected, the wavefront presentation at the eye may be reconfigured for a 2 meter focal distance, and so on.
- Thus in one embodiment incorporating accommodation tracking, a variable focus element may be placed in the optical path between an outputting combiner (e.g., a waveguide or beamsplitter) and the eye of the user, so that the focus may be changed along with (i.e., preferably at the same rate as) accommodation changes of the eye. Software effects may be utilized to produce variable amounts blur (e.g., Gaussian) to objects which should not be in focus to simulate the dioptric blur expected at the retina if an object were at that viewing distance and enhance the three-dimensional perception by the eyes/brain.
- A simple embodiment is a single plane whose focus level is slaved to the viewer's accommodation level, however the performance demands on the accommodation tracking system can be relaxed if even a low number of multiple planes are used. Referring to
FIG. 10 , in another embodiment, a stack (328) of about 3 waveguides (318, 320, 322) may be utilized to create three focal planes worth of wavefronts simultaneously. In one embodiment, the weak lenses (324, 326) may have static focal distances, and a variable focal lens (316) may be slaved to the accommodation tracking of the eyes such that one of the three waveguides (say the middle waveguide 320) outputs what is deemed to be the in-focus wavefront, while the other two waveguides (322, 318) output a +margin wavefront and a −margin wavefront (i.e., a little farther than detected focal distance, a little closer than detected focal distance) which may improve the three-dimensional perception and also provide enough difference for the brain/eye accommodation control system to sense some blur as negative feedback, which enhances the perception of reality, and allows a range of accommodation before an physical adjustment of the focus levels is necessary. - A variable focus compensating lens (314) is also shown to ensure that light coming in from the real world (144) in an augmented reality configuration is not refocused or magnified by the assembly of the stack (328) and output lens (316). The variable focus in the lenses (316, 314) may be achieved, as discussed above, with refractive, diffractive, or reflective techniques.
- In another embodiment, each of the waveguides in a stack may contain their own capability for changing focus (such as by having an included electronically switchable DOE) so that the variable focus element need not be centralized as in the stack (328) of the configuration of
FIG. 10 . - In another embodiment, variable focus elements may be interleaved between the waveguides of a stack (i.e., rather than fixed focus weak lenses as in the embodiment of
FIG. 10 ) to obviate the need for a combination of fixed focus weak lenses plus whole-stack-refocusing variable focus element. - Such stacking configurations may be used in accommodation tracked variations as described herein, and also in a frame-sequential multi-focal display approach.
- In a configuration wherein light enters the pupil with a small exit pupil, such as ½ mm diameter or less, one has the equivalent of a pinhole lens configuration wherein the beam is always interpreted as in-focus by the eyes/brain—e.g., a scanned light display using a 0.5 mm diameter beam to scan images to the eye. Such a configuration is known as a Maxwellian view configuration, and in one embodiment, accommodation tracking input may be utilized to induce blur using software to image information that is to be perceived as at a focal plane behind or in front of the focal plane determined from the accommodation tracking. In other words, if one starts with a display presenting a Maxwellian view, then everything theoretically can be in focus, and to provide a rich and natural three-dimensional perception, simulated dioptric blur may be induced with software, and may be slaved to the accommodation tracking status.
- In one embodiment a scanning fiber display is well suited to such configuration because it may be configured to only output small-diameter beams in a Maxwellian form. In another embodiment, an array of small exit pupils may be created to increase the functional eye box of the system (and also to reduce the impact of a light-blocking particle which may reside in the vitreous or cornea of the eye), such as by one or more scanning fiber displays, or by a DOE configuration such as that described in reference to
FIG. 8K , with a pitch in the array of presented exit pupils that ensure that only one will hit the anatomical pupil of the user at any given time (for example, if the average anatomical pupil diameter is 4 mm, one configuration may comprise ½ mm exit pupils spaced at intervals of approximate 4 mm apart). Such exit pupils may also be switchable in response to eye position, such that only the eye always receives one, and only one, active small exit pupil at a time; allowing a denser array of exit pupils. Such user will have a large depth of focus to which software-based blur techniques may be added to enhance perceived depth perception. - As discussed above, an object at optical infinity creates a substantially planar wavefront; an object closer, such as 1 m away from the eye, creates a curved wavefront (with about 1 m convex radius of curvature). The eye's optical system needs to have enough optical power to bend the incoming rays of light so that they end up focused on the retina (convex wavefront gets turned into concave, and then down to a focal point on the retina). These are basic functions of the eye.
- In many of the embodiments described above, light directed to the eye has been treated as being part of one continuous wavefront, some subset of which would hit the pupil of the particular eye. In another approach, light directed to the eye may be effectively discretized or broken down into a plurality of beamlets or individual rays, each of which has a diameter less than about 0.5 mm and a unique propagation pathway as part of a greater aggregated wavefront that may be functionally created with the an aggregation of the beamlets or rays. For example, a curved wavefront may be approximated by aggregating a plurality of discrete neighboring collimated beams, each of which is approaching the eye from an appropriate angle to represent a point of origin that matches the center of the radius of curvature of the desired aggregate wavefront.
- When the beamlets have a diameter of about 0.5 mm or less, it is as though it is coming through a pinhole lens configuration, which means that each individual beamlet is always in relative focus on the retina, independent of the accommodation state of the eye—however the trajectory of each beamlet will be affected by the accommodation state. For instance, if the beamlets approach the eye in parallel, representing a discretized collimated aggregate wavefront, then an eye that is correctly accommodated to infinity will deflect the beamlets to all converge upon the same shared spot on the retina, and will appear in focus. If the eye accommodates to, say, 1 m, the beams will be converged to a spot in front of the retina, cross paths, and fall on multiple neighboring or partially overlapping spots on the retina—appearing blurred.
- If the beamlets approach the eye in a diverging configuration, with a shared point of
origin 1 meter from the viewer, then an accommodation of 1 m will steer the beams to a single spot on the retina, and will appear in focus; if the viewer accommodates to infinity, the beamlets will converge to a spot behind the retina, and produce multiple neighboring or partially overlapping spots on the retina, producing a blurred image. Stated more generally, the accommodation of the eye determines the degree of overlap of the spots on the retina, and a given pixel is “in focus” when all of the spots are directed to the same spot on the retina and “defocused” when the spots are offset from one another. This notion that all of the 0.5 mm diameter or less beamlets are always in focus, and that they may be aggregated to be perceived by the eyes/brain as though they are substantially the same as coherent wavefronts, may be utilized in producing configurations for comfortable three-dimensional virtual or augmented reality perception. - In other words, a set of multiple narrow beams may be used to emulate what is going on with a larger diameter variable focus beam, and if the beamlet diameters are kept to a maximum of about 0.5 mm, then they maintain a relatively static focus level, and to produce the perception of out-of-focus when desired, the beamlet angular trajectories may be selected to create an effect much like a larger out-of-focus beam (such a defocussing treatment may not be the same as a Gaussian blur treatment as for the larger beam, but will create a multimodal point spread function that may be interpreted in a similar fashion to a Gaussian blur).
- In a preferred embodiment, the beamlets are not mechanically deflected to form this aggregate focus effect, but rather the eye receives a superset of many beamlets that includes both a multiplicity of incident angles and a multiplicity of locations at which the beamlets intersect the pupil; to represent a given pixel from a particular viewing distance, a subset of beamlets from the superset that comprise the appropriate angles of incidence and points of intersection with the pupil (as if they were being emitted from the same shared point of origin in space) are turned on with matching color and intensity, to represent that aggregate wavefront, while beamlets in the superset that are inconsistent with the shared point of origin are not turned on with that color and intensity (but some of them may be turned on with some other color and intensity level to represent, e.g., a different pixel).
- Referring to
FIG. 11A , each of a multiplicity of incoming beamlets (332) is passing through a small exit pupil (330) relative to the eye (58) in a discretized wavefront display configuration. Referring toFIG. 11B , a subset (334) of the group of beamlets (332) may be driven with matching color and intensity levels to be perceived as though they are part of the same larger-sized ray (thebolded subgroup 334 may be deemed an “aggregated beam”). In this case, the subset of beamlets are parallel to one another, representing a collimated aggregate beam from optical infinity (such as light coming from a distant mountain). The eye is accommodated to infinity, so the subset of beamlets are deflected by the eye's cornea and lens to all fall substantially upon the same location of the retina and are perceived to comprise a single in focus pixel. -
FIG. 11C shows another subset of beamlets representing an aggregated collimated beam (336) coming in from the right side of the field of view of the user's eye (58) if the eye (58) is viewed in a coronal-style planar view from above. Again, the eye is shown accommodated to infinity, so the beamlets fall on the same spot of the retina, and the pixel is perceived to be in focus. If, in contrast, a different subset of beamlets were chosen that were reaching the eye as a diverging fan of rays, those beamlets would not fall on the same location of the retina (and be perceived as in focus) until the eye were to shift accommodation to a near point that matches the geometrical point of origin of that fan of rays. - As regards patterns of points of intersection of beamlets with the anatomical pupil of the eye (i.e., the pattern of exit pupils), they may be organized in configurations such as a cross-sectionally efficient hex-lattice (for example, as shown in
FIG. 12A ) or a square lattice or other two-dimensional array. Further, a three-dimensional array of exit pupils could be created, as well as time-varying arrays of exit pupils. - Discretized aggregate wavefronts may be created using several configurations, such as an array of microdisplays or microprojectors placed optically conjugate with the exit pupil of viewing optics, microdisplay or microprojector arrays coupled to a direct field of view substrate (such as an eyeglasses lens) such that they project light to the eye directly, without additional intermediate viewing optics, successive spatial light modulation array techniques, or waveguide techniques such as those described in relation to
FIG. 8K . - Referring to
FIG. 12A , in one embodiment, a lightfield may be created by bundling a group of small projectors or display units (such as scanned fiber displays).FIG. 12A depicts a hexagonal lattice projection bundle (338) which may, for example, create a 7 mm-diameter hex array with each fiber display outputting a sub-image (340). If such an array has an optical system, such as a lens, placed in front of it such that the array is placed optically conjugate with the eye's entrance pupil, this will create an image of the array at the eye's pupil, as shown inFIG. 12B , which essentially provides the same optical arrangement as the embodiment ofFIG. 11A . - Each of the small exit pupils of the configuration is created by a dedicated small display in the bundle (338), such as a scanning fiber display. Optically, it's as though the entire hex array (338) is positioned right into the anatomical pupil (45). Such embodiments are means for driving different subimages to different small exit pupils within the larger anatomical entrance pupil (45) of the eye, comprising a superset of beamlets with a multiplicity of incident angles and points of intersection with the eye pupil. Each of the separate projectors or displays may be driven with a slightly different image, such that subimages may be created that pull out different sets of rays to be driven at different light intensities and colors.
- In one variation, a strict image conjugate may be created, as in the embodiment of
FIG. 12B , wherein there is direct 1-to-1 mapping of the array (338) with the pupil (45). In another variation, the spacing may be changed between displays in the array and the optical system (lens 342, inFIG. 12B ) so that instead of getting a conjugate mapping of the array to the eye pupil, the eye pupil may be catching the rays from the array at some other distance. With such a configuration, one would still get an angular diversity of beams through which one could create a discretized aggregate wavefront representation, but the mathematics regarding how to drive which ray and at which power and intensity may become more complex (although, on the other hand, such a configuration may be considered simpler from a viewing optics perspective). The mathematics involved with light field image capture may be leveraged for these calculations. - Referring to
FIG. 13A , another lightfield creating embodiment is depicted wherein an array of microdisplays or microprojectors (346) may be coupled to a frame (344; such as an eyeglasses frame) to be positioned in front of the eye (58). The depicted configuration is a nonconjugate arrangement wherein there are no large-scale optical elements interposed between the displays (for example, scanning fiber displays) of the array (346) and the eye (58). One can imagine a pair of glasses, and coupled to those glasses are a plurality of displays, such as scanning fiber engines, positioned orthogonal to the eyeglasses surface, and all angled inward so they are pointing at the pupil of the user. Each display may be configured to create a set of rays representing different elements of the beamlet superset. - With such a configuration, at the anatomical pupil (45) the user is going to receive a similar result as received in the embodiments discussed in reference to
FIG. 11A , in which every point at the user's pupil is receiving rays with a multiplicity of angles of incidence and points of intersection that are being contributed from the different displays.FIG. 13B illustrates a nonconjugate configuration similar to that ofFIG. 13A , with the exception that the embodiment ofFIG. 13B features a reflecting surface (348) to facilitate moving the display array (346) away from the eye's (58) field of view, while also allowing views of the real world (144) through the reflective surface (348). - Thus another configuration for creating the angular diversity necessary for a discretized aggregate wavefront display is presented. To optimize such a configuration, the sizes of the displays may be decreased to the maximum. Scanning fiber displays which may be utilized as displays may have baseline diameters in the range of 1 mm, but reduction in enclosure and projection lens hardware may decrease the diameters of such displays to about 0.5 mm or less, which is less disturbing for a user. Another downsizing geometric refinement may be achieved by directly coupling a collimating lens (which may, for example, comprise a gradient refractive index, or “GRIN”, lens, a conventional curved lens, or a diffractive lens) to the tip of the scanning fiber itself in a case of a fiber scanning display array. For example, referring to
FIG. 13D , a GRIN lens (354) is shown fused to the end of a single mode optical fiber. An actuator (350; such as a piezoelectric actuator) is coupled to the fiber (352) and may be used to scan the fiber tip. - In another embodiment the end of the fiber may be shaped into a hemispherical shape using a curved polishing treatment of an optical fiber to create a lensing effect. In another embodiment a standard refractive lens may be coupled to the end of each optical fiber using an adhesive. In another embodiment a lens may be built from a dab of transmissive polymeric material or glass, such as epoxy. In another embodiment the end of an optical fiber may be melted to create a curved surface for a lensing effect.
-
FIG. 13C-2 shows an embodiment wherein display configurations (i.e., scanning fiber displays with GRIN lenses; shown in close-up view ofFIG. 13C-1 ) such as that shown inFIG. 13D may be coupled together through a single transparent substrate (356) preferably having a refractive index that closely matches the cladding of the optical fibers (352) so that the fibers themselves are not very visible for viewing of the outside world across the depicted assembly (if the index matching of the cladding is done precisely, then the larger cladding/housing becomes transparent and only the tiny cores, which preferably are about 3 microns in diameter, will be obstructing the view. In one embodiment the matrix (358) of displays may all be angled inward so they are directed toward the anatomic pupil of the user (in another embodiment, they may stay parallel to each other, but such a configuration is less efficient). - Referring to
FIG. 13E , another embodiment is depicted wherein rather than using circular fibers to move cyclically, a thin series of planar waveguides (358) are configured to be cantilevered relative to a larger substrate structure (356). In one variation, the substrate (356) may be moved to produce cyclic motion (i.e., at the resonant frequency of the cantilevered members 358) of the planar waveguides relative to the substrate structure. In another variation, the cantilevered waveguide portions (358) may be actuated with piezoelectric or other actuators relative to the substrate. Image illumination information may be injected, for example, from the right side (360) of the substrate structure to be coupled into the cantilevered waveguide portions (358). In one embodiment the substrate (356) may comprise a waveguide configured (such as with an integrated DOE configuration as described above) to totally internally reflect incoming light (360) along its length and then redirect it to the cantilevered waveguide portions (358). As a person gazes toward the cantilevered waveguide portions (358) and through to the real world (144) behind, the planar waveguides are configured to minimize any dispersion and/or focus changes with their planar shape factors. - In the context of discussing discretized aggregate wavefront displays, there is value placed in having some angular diversity created for every point in the exit pupil of the eye. In other words, it is desirable to have multiple incoming beams to represent each pixel in a displayed image. Referring to
FIGS. 13F-1 and 13F-2, one way to gain further angular and spatial diversity is to use a multicore fiber and place a lens at the exit point, such as a GRIN lens, so that the exit beams are deflected through a single nodal point (366); that nodal point may then be scanned back and forth in a scanned fiber type of arrangement (such as by a piezoelectric actuator 368). If a retinal conjugate is placed at the plane defined at the end of the GRIN lens, a display may be created that is functionally equivalent to the general case discretized aggregate wavefront configuration described above. - Referring to
FIG. 13G , a similar effect may be achieved not by using a lens, but by scanning the face of a multicore system at the correct conjugate of an optical system (372), the goal being to create a higher angular and spatial diversity of beams. In other words, rather than having a bunch of separately scanned fiber displays as in the bundled example ofFIG. 12A described above, some of this requisite angular and spatial diversity may be created through the use of multiple cores to create a plane which may be relayed by a waveguide. Referring toFIG. 13H , a multicore fiber (362) may be scanned (such as by a piezoelectric actuator 368) to create a set of beamlets with a multiplicity of angles of incidence and points of intersection which may be relayed to the eye (58) by a waveguide (370). Thus in one embodiment a collimated lightfield image may be injected into a waveguide, and without any additional refocusing elements, that lightfield display may be translated directly to the human eye. -
FIGS. 131-13L depict certain commercially available multicore fiber (362) configurations (from vendors such as Mitsubishi Cable Industries, Ltd. of Japan), including one variation (363) with a rectangular cross section, as well as variations with flat exit faces (372) and angled exit faces (374). - Referring to
FIG. 13M , some additional angular diversity may be created by having a waveguide (376) fed with a linear array of displays (378), such as scanning fiber displays. - Referring to
FIGS. 14A-14F , another group of configurations for creating a fixed viewpoint lightfield display is described. Referring back toFIG. 11A , if a two-dimensional plane was created that was intersecting all of the tiny beams coming in from the left, each beamlet would have a certain point of intersection with that plane. If another plane was created at a different distance to the left, then all of the beamlets would intersect that plane at a different location. Then going back toFIG. 14A , if various positions on each of two or more planes can be allowed to selectively transmit or block the light radiation directed through it, such a multi-planar configuration may be utilized to selectively create a lightfield by independently modulating individual beamlets. - The basic embodiment of
FIG. 14A shows two spatial light modulators, such as liquid crystal display panels (380, 382; in other embodiments they may be MEMS shutter displays or DLP DMD arrays) which may be independently controlled to block or transmit different rays on a high-resolution basis. For example, referring toFIG. 14A , if the second panel (382) blocks or attenuates transmission of rays at point “a” (384), all of the depicted rays will be blocked; but if only the first panel (380) blocks or attenuates transmission of rays at point “b” (386), then only the lower incoming ray (388) will be blocked/attenuated, while the rest will be transmitted toward the pupil (45). Each of the controllable panels or planes may be deemed a “spatial light modulator” or “fatte”. The intensity of each transmitted beam passed through a series of SLMs will be a function of the combination of the transparency of the various pixels in the various SLM arrays. Thus without any sort of lens elements, a set of beamlets with a multiplicity of angles and points of intersection (or a “lightfield”) may be created using a plurality of stacked SLMs. Additional numbers of SLMs beyond two provides more opportunities to control which beams are selectively attenuated. - As noted briefly above, in addition to using stacked liquid crystal displays as SLMs, planes of DMD devices from DLP systems may be stacked to function as SLMs, and may be preferred over liquid crystal systems as SLMs due to their ability to more efficiently pass light (with a mirror element in a first state, reflectivity to the next element on the way to the eye may be quite efficient; with a mirror element in a second state, the mirror angle may be moved by an angle such as 12 degrees to direct the light away from the path to the eye). Referring to
FIG. 14B , in one DMD embodiment, two DMDs (390, 390) may be utilized in series with a pair of lenses (394, 396) in a periscope type of configuration to maintain a high amount of transmission of light from the real world (144) to the eye (58) of the user. The embodiment ofFIG. 14C provides six different DMD (402, 404, 406, 408, 410, 412) plane opportunities to intercede from an SLM functionality as beams are routed to the eye (58), along with two lenses (398, 400) for beam control. -
FIG. 14D illustrates a more complicated periscope type arrangement with up to four DMDs (422, 424, 426, 428) for SLM functionality and four lenses (414, 420, 416, 418); this configuration is designed to ensure that the image does not become flipped upside down as it travels through to the eye (58).FIG. 14E illustrates in embodiment wherein light may be reflected between two different DMD devices (430, 432) without any intervening lenses (the lenses in the above designs are useful in such configurations for incorporating image information from the real world), in a hall-of-mirrors type of arrangement wherein the display may be viewed through the “hall of mirrors” and operates in a mode substantially similar to that illustrated inFIG. 14A .FIG. 14F illustrates an embodiment wherein a the non-display portions of two facing DMD chips (434, 436) may be covered with a reflective layer to propagate light to and from active display regions (438, 440) of the DMD chips. In other embodiments, in place of DMDs for SLM functionality, arrays of sliding MEMS shutters (such as those available from vendors such as Pixtronics, a division of Qualcomm, Inc.) may be utilized to either pass or block light. In another embodiment, arrays of small louvers that move out of place to present light-transmitting apertures may similarly be aggregated for SLM functionality. - A lightfield of many small beamlets (say, less than about 0.5 mm in diameter) may be injected into and propagated through a waveguide or other optical system. For example, a conventional “birdbath” type of optical system may be suitable for transferring the light of a lightfield input, or a freeform optics design, as described below, or any number of waveguide configurations.
FIGS. 15A-15C illustrate the use of a wedge type waveguide (442) along with a plurality of light sources as another configuration useful in creating a lightfield. Referring toFIG. 15A , light may be injected into the wedge-shaped waveguide (442) from two different locations/displays (444, 446), and will emerge according to the total internal reflection properties of the wedge-shaped waveguide at different angles (448) based upon the points of injection into the waveguide. - Referring to
FIG. 15B , if one creates a linear array (450) of displays (such as scanning fiber displays) projecting into the end of the waveguide as shown, then a large angular diversity of beams (452) will be exiting the waveguide in one dimension, as shown inFIG. 15C . Indeed, if one contemplates adding yet another linear array of displays injecting into the end of the waveguide but at a slightly different angle, then an angular diversity of beams may be created that exits similarly to the fanned out exit pattern shown inFIG. 15C , but at an orthogonal axis; together these may be utilized to create a two-dimensional fan of rays exiting each location of the waveguide. Thus another configuration is presented for creating angular diversity to form a lightfield display using one or more scanning fiber display arrays (or alternatively using other displays which will meet the space requirements, such as miniaturized DLP projection configurations). - Alternatively, as an input to the wedge-shaped waveguides shown herein, a stack of SLM devices may be utilized, in which case rather than the direct view of SLM output as described above, the lightfield output from the SLM configuration may be used as an input to a configuration such as that shown in
FIG. 15C . One of the key concepts here is that while a conventional waveguide is best suited to relay beams of collimated light successfully, with a lightfield of small-diameter collimated beams, conventional waveguide technology may be utilized to further manipulate the output of such a lightfield system as injected into the side of a waveguide, such as a wedge-shaped waveguide, due to the beam size/collimation. - In another related embodiment, rather than projecting with multiple separate displays, a multicore fiber may be used to generate a lightfield and inject it into the waveguide. Further, a time-varying lightfield may be utilized as an input, such that rather than creating a static distribution of beamlets coming out of a lightfield, one may have some dynamic elements that are methodically changing the path of the set of beams. They may be done using components such as waveguides with embedded DOEs (e.g., such as those described above in reference to
FIGS. 8B-8N , or liquid crystal layers, as described in reference toFIG. 7B ), wherein two optical paths are created (one smaller total internal reflection path wherein a liquid crystal layer is placed in a first voltage state to have a refractive index mismatch with the other substrate material that causes total internal reflection down just the other substrate material's waveguide; one larger total internal reflection optical path wherein the liquid crystal layer is placed in a second voltage state to have a matching refractive index with the other substrate material, so that the light totally internally reflects through the composite waveguide which includes both the liquid crystal portion and the other substrate portion). Similarly a wedge-shaped waveguide may be configured to have a bi-modal total internal reflection paradigm (for example, in one variation, wedge-shaped elements may be configured such that when a liquid crystal portion is activated, not only is the spacing changed, but also the angle at which the beams are reflected). - One embodiment of a scanning light display may be characterized simply as a scanning fiber display with a lens at the end of the scanned fiber. Many lens varieties are suitable, such as a GRIN lens, which may be used to collimate the light or to focus the light down to a spot smaller than the fiber's mode field diameter providing the advantage of producing a numerical aperture (or “NA”) increase and circumventing the optical invariant, which is correlated inversely with spot size. Smaller spot size generally facilitates a higher resolution opportunity from a display perspective, which generally is preferred. In one embodiment, a GRIN lens may be long enough relative to the fiber that it may comprise the vibrating element (i.e., rather than the usual distal fiber tip vibration with a scanned fiber display)—a configuration which may be deemed a “scanned GRIN lens display”.
- In another embodiment, a diffractive lens may be utilized at the exit end of a scanning fiber display (i.e., patterned onto the fiber). In another embodiment, a curved mirror may be positioned on the end of the fiber that operates in a reflecting configuration. Essentially any of the configurations known to collimate and focus a beam may be used at the end of a scanning fiber to produce a suitable scanned light display.
- Two significant utilities to having a lens coupled to or comprising the end of a scanned fiber (i.e., as compared to configurations wherein an uncoupled lens may be utilized to direct light after it exits a fiber) are a) the light exiting may be collimated to obviate the need to use other external optics to do so; b) the NA, or the angle of the cone at which light sprays out the end of the single-mode fiber core, may be increased, thereby decreasing the associated spot size for the fiber and increasing the available resolution for the display.
- As described above, a lens such as a GRIN lens may be fused to or otherwise coupled to the end of an optical fiber or formed from a portion of the end of the fiber using techniques such as polishing. In one embodiment, a typical optical fiber with an NA of about 0.13 or 0.14 may have a spot size (also known as the “mode field diameter” for the optical fiber given the NA) of about 3 microns. This provides for relatively high resolution display possibilities given the industry standard display resolution paradigms (for example, a typical microdisplay technology such as LCD or organic light emitting diode, or “OLED” has a spot size of about 5 microns). Thus the aforementioned scanning light display may have ⅗ of the smallest pixel pitch available with a conventional display; further, using a lens at the end of the fiber, the aforementioned configuration may produce a spot size in the range of 1-2 microns.
- In another embodiment, rather than using a scanned cylindrical fiber, a cantilevered portion of a waveguide (such as a waveguide created using microfabrication processes such as masking and etching, rather than drawn microfiber techniques) may be placed into scanning oscillatory motion, and may be fitted with lensing at the exit ends.
- In another embodiment, an increased numerical aperture for a fiber to be scanned may be created using a diffuser (i.e., one configured to scatter light and create a larger NA) covering the exit end of the fiber. In one variation, the diffuser may be created by etching the end of the fiber to create small bits of terrain that scatter light; in another variation a bead or sandblasting technique, or direct sanding/scuffing technique may be utilized to create scattering terrain. In another variation, an engineered diffuser, similar to a diffractive element, may be created to maintain a clean spot size with desirable NA, which ties into the notion of using a diffractive lens, as noted above.
- Referring to
FIG. 16A , an array of optical fibers (454) is shown coupled in to a coupler (456) configured to hold them in parallel together so that their ends may be ground and polished to have an output edge at a critical angle (458; 42 degrees for most glass, for example) to the longitudinal axes of the input fibers, such that the light exiting the angled faces will exit as though it had been passing through a prism, and will bend and become nearly parallel to the surfaces of the polished faces. The beams exiting the fibers (454) in the bundle will become superimposed, but will be out of phase longitudinally due to the different path lengths (referring toFIG. 16B , for example, the difference in path lengths from angled exit face to focusing lens for the different cores is visible). - What was an X axis type of separation in the bundle before exit from the angled faces, will become a Z axis separation, a fact that is helpful in creating a multifocal light source from such a configuration. In another embodiment, rather than using a bundled/coupled plurality of single mode fibers, a multicore fiber, such as those available from Mitsubishi Cable Industries, Ltd. of Japan, may be angle polished.
- In one embodiment, if a 45 degree angle is polished into a fiber and then covered with a reflective element, such as a mirror coating, the exiting light may be reflected from the polished surface and emerge from the side of the fiber (in one embodiment at a location wherein a flat-polished exit window has been created in the side of the fiber) such that as the fiber is scanned in what would normally be an X-Y Cartesian coordinate system axis, that fiber would now be functionally performing the equivalent of an X-Z scan, with the distance changing during the course of the scan. Such a configuration may be beneficially utilized to change the focus of the display as well.
- Multicore fibers may be configured to play a role in display resolution enhancement (i.e., higher resolution). For example, in one embodiment, if separate pixel data is sent down a tight bundle of 19 cores in a multicore fiber, and that cluster is scanned around in a sparse spiral pattern with the pitch of the spiral being approximately equal to the diameter of the multicore, then sweeping around will effectively create a display resolution that is approximately 19× the resolution of a single core fiber being similarly scanned around. Indeed, it may be more practical to have the fibers more sparsely positioned relative to each other, as in the configuration of
FIG. 16C , which has 7 clusters (464; 7 is used for illustrative purposes because it is an efficient tiling/hex pattern; other patterns or numbers may be utilized; for example, a cluster of 19; the configuration is scalable up or down) of 3 fibers each housed within a conduit (462). - With a sparse configuration as shown in
FIG. 16C , scanning of the multicore scans each of the cores through its own local region, as opposed to a configuration wherein the cores are all packed tightly together and scanned (wherein cores end up overlapping with scanning; if the cores are too close to each other, the NA of the core is not large enough and the very closely packed cores end up blurring together somewhat and not creating as discriminable a spot for display). Thus, for resolution increases, it is preferable to have sparse tiling rather than highly dense tiling, although both will work. - The notion that densely packed scanned cores can create blurring at the display may be utilized as an advantage in one embodiment wherein a plurality (say a triad or cores to carry red, green, and blue light) of cores may be intentionally packed together densely so that each triad forms a triad of overlapped spots featuring red, green, and blue light. With such a configuration, one is able to have an RGB display without having to combine red, green, and blue into a single-mode core, which is an advantage, because conventional mechanisms for combining a plurality (such as three) wavelets of light into a single core are subject to significant losses in optical energy. Referring to
FIG. 16C , in one embodiment each tight cluster of 3 fiber cores contains one core that relays red light, one core that relays green light, and one core that relays blue light, with the 3 fiber cores close enough together that their positional differences are not resolvable by the subsequent relay optics, forming an effectively superimposed RGB pixel; thus, the sparse tiling of 7 clusters produces resolution enhancement while the tight packing of 3 cores within the clusters facilitates seamless color blending without the need to utilize glossy RGB fiber combiners (e.g., those using wavelength division multiplexing or evanescent coupling techniques). - Referring to
FIG. 16D , in another more simple variation, one may have just one cluster (464) housed in a conduit (468) for, say, red/green/blue (and in another embodiment, another core may be added for infrared for uses such as eye tracking). In another embodiment, additional cores may be placed in the tight cluster to carrying additional wavelengths of light to comprise a multi-primary display for increased color gamut. Referring toFIG. 16E , in another embodiment, a sparse array of single cores (470); in one variation with red, green, and blue combined down each of them) within a conduit (466) may be utilized; such a configuration is workable albeit somewhat less efficient for resolution increase, but not optimum for red/green/blue combining. - Multicore fibers also may be utilized for creating lightfield displays. Indeed, rather than keeping the cores separated enough from each other so that the cores do not scan on each other's local area at the display panel, as described above in the context of creating a scanning light display, with a lightfield display, it is desirable to scan around a densely packed plurality of fibers because each of the beams produced represents a specific part of the lightfield. The light exiting from the bundled fiber tips can be relatively narrow if the fibers have a small NA; lightfield configurations may take advantage of this and have an arrangement in which at the anatomic pupil, a plurality of slightly different beams are being received from the array. Thus there are optical configurations with scanning a multicore that are functionally equivalent to an array of single scanning fiber modules, and thus a lightfield may be created by scanning a multicore rather than scanning a group of single mode fibers.
- In one embodiment, a multi-core phased array approach may be used to create a large exit pupil variable wavefront configuration to facilitate three-dimensional perception. A single laser configuration with phase modulators is described above. In a multicore embodiment, phase delays may be induced into different channels of a multicore fiber, such that a single laser's light is injected into all of the cores of the multicore configuration so that there is mutual coherence.
- In one embodiment, a multi-core fiber may be combined with a lens, such as a GRIN lens. Such lens may be, for example, a refractive lens, diffractive lens, or a polished edge functioning as a lens. The lens may be a single optical surface, or may comprise multiple optical surfaces stacked up. Indeed, in addition to having a single lens that extends the diameter of the multicore, a smaller lenslet array may be desirable at the exit point of light from the cores of the multicore, for example.
FIG. 16F shows an embodiment wherein a multicore fiber (470) is emitting multiple beams into a lens (472), such as a GRIN lens. The lens collects the beams down to a focal point (474) in space in front of the lens. In many conventional configurations, the beams would exit the multicore fiber as diverging. The GRIN or other lens is configured to function to direct them down to a single point and collimate them, such that the collimated result may be scanned around for a lightfield display, for instance. - Referring to
FIG. 16G , smaller lenses (478) may be placed in front of each of the cores of a multicore (476) configuration, and these lenses may be utilized to collimate; then a shared lens (480) may be configured to focus the collimated beams down to a diffraction limited spot (482) that is aligned for all of the three spots. The net result of such a configuration: by combining three collimated, narrow beams with narrow NA together as shown, one effectively combines all three into a much larger angle of emission which translates to a smaller spot size in, for example, a head mounted optical display system which may be next in the chain of light delivery to the user. - Referring to
FIG. 16H , one embodiment features a multicore fiber (476) with a lenslet (478) array feeding the light to a small prism array (484) that deflects the beams generated by the individual cores to a common point. Alternatively one may have the small lenslet array shifted relative to the cores such that the light is being deflected and focused down to a single point. Such a configuration may be utilized to increase the numerical aperture. - Referring to
FIG. 16I , a two-step configuration is shown with a small lenslet (478) array capturing light from the multicore fiber (476), followed sequentially by a shared lens (486) to focus the beams to a single point (488). Such a configuration may be utilized to increase the numerical aperture. As discussed above, a larger NA corresponds to a smaller pixel size and higher possible display resolution. - Referring to
FIG. 16J , a beveled fiber array which may be held together with a coupler (456), such as those described above, may be scanned with a reflecting device (494; such as a DMD module of a DLP system). With multiple single fibers (454) coupled into the array, or a multicore instead, the superimposed light can be directed through one or more focusing lenses (490, 492) to create a multifocal beam; with the superimposing and angulation of the array, the different sources are different distances from the focusing lens, which creates different focus levels in the beams as they emerge from the lens (492) and are directed toward the retina (54) of the eye (58) of the user. For example, the farthest optical route/beam may be set up to be a collimated beam representative of optical infinity focal positions. Closer routes/beams may be associated with diverging spherical wavefronts of closer focal locations. - The multifocal beam may be passed into a scanning mirror which may be configured to create a raster scan (or, for example, a Lissajous curve scan pattern or a spiral scan pattern) of the multifocal beam which may be passed through a series of focusing lenses and then to the cornea and crystalline lens of the eye. The various beams emerging from the lenses are creating different pixels or voxels of varying focal distances that are superimposed.
- In one embodiment, one may write different data to each of the light modulation channels at the front end, thereby creating an image that is projected to the eye with one or more focus elements. By changing the focal distance of the crystalline lens (i.e., by accommodating), the user can bring different incoming pixels into and out of focus, as shown in
FIGS. 16K and 16L wherein the crystalline lens is in different Z axis positions. In another embodiment, the fiber array may be actuated/moved around by a piezoelectric actuator. In another embodiment, a relatively thin ribbon array may be resonated in cantilevered form along the axis perpendicular to the arrangement of the array fibers (i.e., in the thin direction of the ribbon) when a piezoelectric actuator is activated. In one variation, a separate piezoelectric actuator may be utilized to create a vibratory scan in the orthogonal long axis. In another embodiment, a single mirror axis scan may be employed for a slow scan along the long axis while the fiber ribbon is vibrated resonantly. - Referring to
FIG. 16M , an array (496) of scanning fiber displays (498) may be beneficially bundled/tiled for an effective resolution increase, the notion being that with such as configuration, each scanning fiber of the bundle is configured to write to a different portion of the image plane (500), as shown, for example, inFIG. 16N , wherein each portion of the image plane is addressed by the emissions from a least one bundle. In other embodiments, optical configurations may be utilized that allow for slight magnification of the beams as they exit the optical fiber so that there is some overlap in the hexagonal, or other lattice pattern, that hits the display plane, so there is a better fill factor while also maintaining an adequately small spot size in the image plane and understanding that there is a subtle magnification in that image plane. - Rather than having individual lenses at the end of each scanned fiber enclosure housing, in one embodiment a monolithic lenslet array may be utilized, so that the lenses can be as closely packed as possible, which allows for even smaller spot sizes in the image plane because one may use a lower amount of magnification in the optical system. Thus arrays of fiber scan displays may be used to increase the resolution of the display, or in other words, they may be used to increase the field of view of the display, because each engine is being used to scan a different portion of the field of view.
- For a lightfield configuration, the emissions may be more desirably overlapped at the image plane. In one embodiment, a lightfield display may be created using a plurality of small diameter fibers scanned around in space. For example, instead of having all of the fibers address a different part of an image plane as described above, have more overlapping, more fibers angled inward, etc., or change the focal power of the lenses so that the small spot sizes are not conjugate with a tiled image plane configuration. Such a configuration may be used to create a lightfield display to scan lots of smaller diameter rays around that become intercepted in the same physical space.
- Referring back to
FIG. 12B , it was discussed that one way of creating a lightfield display involves making the output of the elements on the left collimated with narrow beams, and then making the projecting array conjugate with the eye pupil on the right. - Referring to
FIG. 16O , with a common substrate block (502), a single actuator may be utilized to actuate a plurality of fibers (506) in unison together. A similar configuration is discussed above in reference to FIGS. 13-C-1 and 13-C-2. It may be practically difficult to have all of the fibers retain the same resonant frequency, vibrate in a desirable phase relationship to each other, or have the same dimensions of cantilevering from the substrate block. To address this challenge, the tips of the fibers may be mechanically coupled with a lattice or sheet (504), such as a graphene sheet that is very thin, rigid, and light in weight. With such a coupling, the entire array may vibrate similarly and have the same phase relationship. In another embodiment a matrix of carbon nanotubes may be utilized to couple the fibers, or a piece of very thin planar glass (such as the kind used in creating liquid crystal display panels) may be coupled to the fiber ends. Further, a laser or other precision cutting device may be utilized to cut all associated fibers to the same cantilevered length. - Referring to
FIG. 17 , in one embodiment it may be desirable to have a contact lens directly interfaced with the cornea, and configured to facilitate the eye focusing on a display that is quite close (such as the typical distance between a cornea and an eyeglasses lens). Rather than placing an optical lens as a contact lens, in one variation the lens may comprise a selective filter.FIG. 17 depicts a plot (508) what may be deemed a “notch filter”, due to its design to block only certain wavelength bands, such as 450 nm (peak blue), 530 nm (green), and 650 nm, and generally pass or transmit other wavelengths. In one embodiment several layers of dielectric coatings may be aggregated to provide the notch filtering functionality. - Such a filtering configuration may be coupled with a scanning fiber display that is producing a very narrow band illumination for red, green, and blue, and the contact lens with the notch filtering will block out all of the light coming from the display (such as a minidisplay, such as an OLED display, mounted in a position normally occupied by an eyeglasses lens) except for the transmissive wavelengths. A narrow pinhole may be created in the middle of the contact lens filtering layers/film such that the small aperture (i.e., less than about 1.5 mm diameter) does allow passage of the otherwise blocked wavelengths. Thus a pinhole lens configuration is created that functions in a pinhole manner for red, green, and blue only to intake images from the minidisplay, while light from the real world, which generally is broadband illumination, will pass through the contact lens relatively unimpeded. Thus a large depth of focus virtual display configuration may be assembled and operated. In another embodiment, a collimated image exiting from a waveguide would be visible at the retina because of the pinhole large-depth-of-focus configuration.
- It may be useful to create a display that can vary its depth of focus over time. For example, in one embodiment, a display may be configured to have different display modes that may be selected (preferably rapidly toggling between the two at the command of the operator) by an operator, such as a first mode combining a very large depth of focus with a small exit pupil diameter (i.e., so that everything is in focus all of the time), and a second mode featuring a larger exit pupil and a more narrow depth of focus. In operation, if a user is to play a three-dimensional video game with objects to be perceived at many depths of field, the operator may select the first mode; alternatively, if a user is to type in a long essay (i.e., for a relatively long period of time) using a two-dimensional word processing display configuration, it may be more desirable to switch to the second mode to have the convenience of a larger exit pupil, and a sharper image.
- In another embodiment, it may be desirable to have a multi-depth of focus display configuration wherein some subimages are presented with a large depth of focus while other subimages are presented with small depth of focus. For example, one configuration may have red wavelength and blue wavelength channels presented with a very small exit pupil so that they are always in focus. Then, a green channel only may be presented with a large exit pupil configuration with multiple depth planes (i.e., because the human accommodation system tends to preferentially target green wavelengths for optimizing focus level). Thus, in order to cut costs associated with having too many elements to represent with full depth planes in red, green, and blue, the green wavelength may be prioritized and represented with various different wavefront levels. Red and blue may be relegated to being represented with a more Maxwellian approach (and, as described above in reference to Maxwellian displays, software may be utilized to induce Gaussian levels of blur). Such a display would simultaneously present multiple depths of focus.
- As described above, there are portions of the retina which have a higher density of light sensors. The fovea portion, for example, generally is populated with approximately 120 cones per visual degree. Display systems have been created in the past that use eye or gaze tracking as an input, and to save computation resources by only creating really high resolution rendering for where the person is gazing at the time, while lower resolution rendering is presented to the rest of the retina; the locations of the high versus low resolution portions may be dynamically slaved to the tracked gaze location in such a configuration, which may be termed a “foveated display”.
- An improvement on such configurations may comprise a scanning fiber display with pattern spacing that may be dynamically slaved to tracked eye gaze. For example, with a typical scanning fiber display operating in a spiral pattern, as shown in
FIG. 18 (theleftmost portion 510 of the image inFIG. 18 illustrates a spiral motion pattern of a scannedmulticore fiber 514; therightmost portion 512 of the image inFIG. 18 illustrates a spiral motion pattern of a scannedsingle fiber 516 for comparison), a constant pattern pitch provides for a uniform display resolution. - In a foveated display configuration, a non-uniform scanning pitch may be utilized, with smaller/tighter pitch (and therefore higher resolution) dynamically slaved to the detected gaze location. For example, if the user's gaze was detected as moving toward the edge of the display screen, the spirals may be clustered more densely in such location, which would create a toroid-type scanning pattern for the high-resolution portions, and the rest of the display being in a lower-resolution mode. In a configuration wherein gaps may be created in the portions of the display in a lower-resolution mode, blur could be intentionally dynamically created to smooth out the transitions between scans, as well as between transitions from high-resolution to lower-resolution scan pitch.
- The term lightfield may be used to describe a volumetric 3-D representation of light traveling from an object to a viewer's eye. However, an optical see-through display can only reflect light to the eye, not the absence of light, and ambient light from the real world will add to any light representing a virtual object. That is, if a virtual object presented to the eye contains a black or very dark portion, the ambient light from the real world may pass through that dark portion and obscure that it was intended to be dark.
- It is nonetheless desirable to be able to present a dark virtual object over a bright real background, and for that dark virtual object to appear to occupy a volume at a desired viewing distance; i.e., it is useful to create a “darkfield” representation of that dark virtual object, in which the absence of light is perceived to be located at a particular point in space. With regard to occlusion elements and the presentation of information to the eye of the user so that he or she can perceive darkfield aspects of virtual objects, even in well lighted actual environments, certain aspects of the aforementioned spatial light modulator, or “SLM”, configurations are pertinent. As described above, with a light-sensing system such as the eye, one way to get selective perception of dark field to selectively attenuate light from such portions of the display, because the subject display systems are about manipulation and presentation of light; in other words, darkfield cannot be specifically projected—it's the lack of illumination that may be perceived as darkfield, and thus, configurations for selective attenuation of illumination have been developed.
- Referring back to the discussion of SLM configurations, one way to selectively attenuate for a darkfield perception is to block all of the light coming from one angle, while allowing light from other angles to be transmitted. This may be accomplished with a plurality of SLM planes comprising elements such as liquid crystal (which may not be the most optimal due to its relatively low transparency when in the transmitting state), DMD elements of DLP systems (which have relative high transmission/reflection ratios when in such mode), and MEMS arrays or shutters that are configured to controllably shutter or pass light radiation, as described above.
- With regard to suitable liquid crystal display (“LCD”) configurations, a cholesteric LCD array may be utilized for a controlled occlusion/blocking array. As opposed to the conventional LCD paradigm wherein a polarization state is changed as a function of voltage, with a cholesteric LCD configuration, a pigment is being bound to the liquid crystal molecule, and then the molecule is physically tilted in response to an applied voltage. Such a configuration may be designed to achieve greater transparency when in a transmissive mode than conventional LCD, and a stack of polarizing films is not needed as it is with conventional LCD.
- In another embodiment, a plurality of layers of controllably interrupted patterns may be utilized to controllably block selected presentation of light using moire effects. For example, in one configuration, two arrays of attenuation patterns, each of which may comprise, for example, fine-pitched sine waves printed or painted upon a transparent planar material such as a glass substrate, may be presented to the eye of a user at a distance close enough that when the viewer looks through either of the patterns alone, the view is essentially transparent, but if the viewer looks through both patterns lined up in sequence, the viewer will see a spatial beat frequency moire attenuation pattern, even when the two attenuation patterns are placed in sequence relatively close to the eye of the user.
- The beat frequency is dependent upon the pitch of the patterns on the two attenuation planes, so in one embodiment, an attenuation pattern for selectively blocking certain light transmission for darkfield perception may be created using two sequential patterns, each of which otherwise would be transparent to the user, but which together in series create a spatial beat frequency moire attenuation pattern selected to attenuate in accordance with the darkfield perception desired in the augmented reality system.
- In another embodiment a controlled occlusion paradigm for darkfield effect may be created using a multi-view display style occluder. For example, one configuration may comprise one pin-holed layer that fully occludes with the exception of small apertures or pinholes, along with a selective attenuation layer in series, which may comprise an LCD, DLP system, or other selective attenuation layer configuration, such as those described above. In one scenario, with the pinhole array placed at a typical eyeglasses lens distance from the cornea (about 30 mm), and with a selective attenuation panel located opposite the pinhole array from the eye, a perception of a sharp mechanical edge out in space may be created. In essence, if the configuration will allow certain angles of light to pass, and others to be blocked or occluded, than a perception of a very sharp pattern, such as a sharp edge projection, may be created. In another related embodiment, the pinhole array layer may be replaced with a second dynamic attenuation layer to provide a somewhat similar configuration, but with more controls than the static pinhole array layer (the static pinhole layer could be simulated, but need not be).
- In another related embodiment, the pinholes may be replaced with cylindrical lenses. The same pattern of occlusion as in the pinhole array layer configuration may be achieved, but with cylindrical lenses, the array is not restricted to the very tiny pinhole geometries. To prevent the eye from being presented with distortions due to the lenses when viewing through to the real world, a second lens array may be added on the side of the aperture or lens array opposite of the side nearest the eye to compensate and provide the view-through illumination with basically a zero power telescope configuration.
- In another embodiment, rather than physically blocking light for occlusion and creation of darkfield perception, the light may be bent or bounced, or a polarization of the light may be changed if a liquid crystal layer is utilized. For example, in one variation, each liquid crystal layer may act as a polarization rotator such that if a patterned polarizing material is incorporated on one face of a panel, then the polarization of individual rays coming from the real world may be selectively manipulated so they catch a portion of the patterned polarizer. There are polarizers known in the art that have checkerboard patterns wherein half of the “checker boxes” have vertical polarization and the other half have horizontal polarization. In addition, if a material such as liquid crystal is used in which polarization may be selectively manipulated, light may be selectively attenuated with this.
- As described above, selective reflectors may provide greater transmission efficiency than LCD. In one embodiment, if a lens system is placed such that it takes light coming in from the real world and focuses a plane from the real world onto an image plane, and if a DMD (i.e., DLP technology) is placed at that image plane to reflect light when in an “on” state towards another set of lenses that pass the light to the eye, and those lenses also have the DMD at their focal length, the one may create an attenuation pattern that is in focus for the eye. In other words, DMDs may be used in a selective reflector plane in a zero magnification telescope configuration, such as is shown in
FIG. 19A , to controllably occlude and facilitate creating darkfield perception. - As shown in
FIG. 19A , a lens (518) is taking light from the real world (144) and focusing it down to an image plane (520); if a DMD (or other spatial attenuation device) (522) is placed at the focal length of the lens (i.e., at the image plane 520), the lens (518) is going to take whatever light is coming from optical infinity and focus that onto the image plane (520). Then the spatial attenuator (522) may be utilized to selectively block out things that are to be attenuated.FIG. 19A shows the attenuator DMDs in the transmissive mode wherein they pass the beams shown crossing the device. The image is then placed at the focal length of the second lens (524). Preferably the two lenses (518, 524) have the same focal power so they end up being a zero-power telescope, or a “relay”, that does not magnify views to the real world (144). Such a configuration may be used to present unmagnified views of the world while also allowing selective blocking/attenuation of certain pixels. - In another embodiment, as shown in
FIGS. 19B and 19C , additional DMDs may be added such that light reflects from each of four DMDs (526, 528, 530, 532) before passing to the eye.FIG. 19B shows an embodiment with two lenses preferably with the same focal power (focal length “F”) placed at a 2F relationship from one another (the focal length of the first being conjugate to the focal length of the second) to have the zero-power telescope effect;FIG. 19C shows an embodiment without lenses. The angles of orientation of the four reflective panels (526, 528, 530, 532) in the depicted embodiments ofFIGS. 19B and 19C are shown to be around 45 degrees for simple illustration purposes, but specific relative orientation is required (for example, a typical DMD reflect at about a 12 degree angle). - In another embodiment, the panels may also be ferroelectric, or may be any other kind of reflective or selective attenuator panel or array. In one embodiment similar to those depicted in
FIGS. 19B and 19C , one of the three reflector arrays may be a simple mirror, such that the other 3 are selective attenuators, thus still providing three independent planes to controllably occlude portions of the incoming illumination in furtherance of darkfield perception. By having multiple dynamic reflective attenuators in series, masks at different optical distances relative to the real world may be created. - Alternatively, referring back to
FIG. 19C , one may create a configuration wherein one or more DMDs are placed in a reflective periscope configuration without any lenses. Such a configuration may be driven in lightfield algorithms to selectively attenuate certain rays while others are passed. - In another embodiment, a DMD or similar matrix of controllably movable devices may be created upon a transparent substrate as opposed to a generally opaque substrate, for use in a transmissive configuration such as virtual reality.
- In another embodiment, two LCD panels may be utilized as lightfield occluders. In one variation, they may be thought of as attenuators due to their attenuating capability as described above; alternatively they may be considered polarization rotators with a shared polarizer stack. Suitable LCDs may comprise components such as blue phase liquid crystal, cholesteric liquid crystal, ferroelectric liquid crystal, and/or twisted nematic liquid crystal.
- One embodiment may comprise an array of directionally-selective occlusion elements, such as a MEMS device featuring a set of louvers that can change rotation such that they pass the majority of light that is coming from a particular angle, but are presenting more of a broad face to light that is coming from a different angle (somewhat akin to the manner in which plantation shutters may be utilized with a typical human scale window). The MEMS/louvers configuration may be placed upon an optically transparent substrate, with the louvers substantially opaque. Ideally such a configuration would have a louver pitch fine enough to selectably occlude light on a pixel-by-pixel basis. In another embodiment, two or more layers or stacks of louvers may be combined to provide yet further controls. In another embodiment, rather than selectively blocking light, the louvers may be polarizers configured to change the polarization state of light on a controllably variable basis.
- As described above, another embodiment for selective occlusion may comprise an array of sliding panels in a MEMS device such that the sliding panels may be controllably opened (i.e., by sliding in a planar fashion from a first position to a second position; or by rotating from a first orientation to a second orientation; or, for example, combined rotational reorientation and displacement) to transmit light through a small frame or aperture, and controllably closed to occlude the frame or aperture and prevent transmission. The array may be configured to open or occlude the various frames or apertures such that they maximally attenuate the rays that are to be attenuated, and only minimally attenuate the rays to be transmitted.
- In an embodiment wherein a fixed number of sliding panels can either occupy a first position occluding a first aperture and opening a second aperture, or a second position occluding the second aperture and opening the first aperture, there will always be the same amount of light transmitted overall (because 50% of the apertures are occluded, and the other 50% are open, with such a configuration), but the local position changes of the shutters or doors may create targeted moire or other effects for darkfield perception with the dynamic positioning of the various sliding panels. In one embodiment, the sliding panels may comprise sliding polarizers, and if placed in a stacked configuration with other polarizing elements that are either static or dynamic, may be utilized to selectively attenuate.
- Referring to
FIG. 19D , another configuration providing an opportunity for selective reflection, such as via a DMD style reflector array (534), is shown, such that a stacked set of two waveguides (536, 538) along with a pair of focus elements (540, 542) and a reflector (534; such as a DMD) may be used to capture a portion of incoming light with an entrance reflector (544). The reflected light may be totally internally reflected down the length of the first waveguide (536), into a focusing element (540) to bring the light into focus on a reflector (534) such as a DMD array, after which the DMD may selectively attenuate and reflect a portion of the light back through a focusing lens (542; the lens configured to facilitate injection of the light back into the second waveguide) and into the second waveguide (538) for total internal reflection down to an exit reflector (546) configured to exit the light out of the waveguide and toward the eye (58). - Such a configuration may have a relatively thin shape factor, and is designed to allow light from the real world (144) to be selectively attenuated. As waveguides work most cleanly with collimated light, such a configuration may be well suited for virtual reality configurations wherein focal lengths are in the range of optical infinity. For closer focal lengths, a lightfield display may be used as a layer on top of the silhouette created by the aforementioned selective attenuation/darkfield configuration to provide other cues to the eye of the user that light is coming from another focal distance. An occlusion mask may be out of focus, even nondesirably so, and then in one embodiment, a lightfield on top of the masking layer may be used to hide the fact that the darkfield may be at the wrong focal distance.
- Referring to
FIG. 19E , an embodiment is shown featuring two waveguides (552, 554) each having two angled reflectors (558, 544; 556, 546) for illustrative purposes shown at approximately 45 degrees; in actual configurations the angle may differ depending upon the reflective surface, reflective/refractive properties of the waveguides, etc.) directing a portion of light incoming from the real world down each side of a first waveguide (or down two separate waveguides if the top layer is not monolithic) such that it hits a reflector (548, 550) at each end, such as a DMD which may be used for selective attenuation, after which the reflected light may be injected back into the second waveguide (or into two separate waveguides if the bottom layer is not monolithic) and back toward two angled reflectors (again, they need not be at 45 degrees as shown) for exit out toward the eye (58). - Focusing lenses may also be placed between the reflectors at each end and the waveguides. In another embodiment the reflectors (548, 550) at each end may comprise standard mirrors (such as alumized mirrors). Further, the reflectors may be wavelength selective reflectors, such as dichroic mirrors or film interference filters. Further, the reflectors may be diffractive elements configured to reflect incoming light.
-
FIG. 19F illustrates a configuration wherein four reflective surfaces in a pyramid type configuration are utilized to direct light through two waveguides (560, 562), in which incoming light from the real world may be divided up and reflected to four difference axes. The pyramid-shaped reflector (564) may have more than four facets, and may be resident within the substrate prism, as with the reflectors of the configuration ofFIG. 19E . The configuration ofFIG. 19F is an extension of that ofFIG. 19E . - Referring to
FIG. 19G , a single waveguide (566) may be utilized to capture light from the world (144) with one or more reflective surfaces (574, 576, 578, 580, 582), relay it (570) to a selective attenuator (568; such as a DMD array), and recouple it back into the same waveguide so that it propagates (572) and encounters one or more other reflective surfaces (584, 586, 588, 590, 592) that cause it to at least partially exit (594) the waveguide on a path toward the eye (58) of the user. Preferably the waveguide comprises selective reflectors such that one group (574, 576, 578, 580, 582) may be switched on to capture incoming light and direct it down to the selective attenuator, while separate another group (584, 586, 588, 590, 592) may be switched on to exit light returning from the selective attenuator out toward the eye (58). - For simplicity the selective attenuator is shown oriented substantially perpendicularly to the waveguide; in other embodiments, various optics components, such as refractive or reflective optics, may be utilized to have the selective attenuator at a different and more compact orientation relative to the waveguide.
- Referring to
FIG. 19H , a variation on the configuration described in reference toFIG. 19D is illustrated. This configuration is somewhat analogous to that discussed above in reference toFIG. 5B , wherein a switchable array of reflectors may be embedded within each of a pair of waveguides (602, 604). Referring toFIG. 19H , a controller may be configured to turn the reflectors (598, 600) on and off in sequence, such that multiple reflectors may be operated on a frame sequential basis; then the DMD or other selective attenuator (594) may also be sequentially driven in sync with the different mirrors being turned on and off. - Referring to
FIG. 19I , a pair of wedge-shaped waveguides similar to those described above (for example, in reference toFIGS. 15A-15C ) are shown in side or sectional view to illustrate that the two long surfaces of each wedge-shaped waveguide (610, 612) are not co-planar. A “turning film” (606, 608; such as that available from 3M corporation under the trade name, “TRAF”, which in essence comprises a microprism array), may be utilized on one or more surfaces of the wedge-shaped waveguides to either turn incoming rays at an angle so that they will be captured by total internal reflection, or to turn outgoing rays as they are exiting the waveguide toward an eye or other target. Incoming rays are directed down the first wedge and toward the selective attenuator (614) such as a DMD, LCD (such as a ferroelectric LCD), or an LCD stack to act as a mask). - After the selective attenuator (614), reflected light is coupled back into the second wedge-shaped waveguide which then relays the light by total internal reflection along the wedge. The properties of the wedge-shaped waveguide are intentionally such that each bounce of light causes an angle change; the point at which the angle has changed enough to be the critical angle to escape total internal reflection becomes the exit point from the wedge-shaped waveguide. Typically the exit will be at an oblique angle, so another layer of turning film may be used to “turn” the exiting light toward a targeted object such as the eye (58).
- Referring to
FIG. 19J , several arcuate lenslet arrays (616, 620, 622) are positioned relative to an eye and configured such that a spatial attenuator array (618) is positioned at a focal/image plane so that it may be in focus with the eye (58). The first (616) and second (620) arrays are configured such that in the aggregate, light passing from the real world to the eye is essentially passed through a zero power telescope. The embodiment ofFIG. 19J shows a third array (622) of lenslets which may be utilized for improved optical compensation, but the general case does not require such a third layer. As discussed above, having telescopic lenses that are the diameter of the viewing optic may create an undesirably large form factor (somewhat akin to having a bunch of small sets of binoculars in front of the eyes). - One way to optimize the overall geometry is to reduce the diameter of the lenses by splitting them out into smaller lenslets, as shown in
FIG. 19J (i.e., an array of lenses rather than one single large lens). The lenslet arrays (616, 620, 622) are shown wrapped radially or arcuately around the eye (58) to ensure that beams incoming to the pupil are aligned through the appropriate lenslets (else the system may suffer from optical problems such as dispersion, aliasing, and/or lack of focus). Thus all of the lenslets are oriented “toed in” and pointed at the pupil of the eye (58), and the system facilitates avoidance of scenarios wherein rays are propagated through unintended sets of lenses en route to the pupil. - Referring to
FIGS. 19K-19N , various software approaches may be utilized to assist in the presentation of darkfield in a virtual or augmented reality displace scenario. Referring toFIG. 19K , a typical challenging scenario for augmented reality is depicted (632), with a textured carpet (624) and non-uniform background architectural features (626), both of which are lightly-colored. The black box (628) depicted indicates the region of the display in which one or more augmented reality features are to be presented to the user for three-dimensional perception, and in the black box a robot creature (630) is being presented that may, for example, be part of an augmented reality game in which the user is engaged. In the depicted example, the robot character (630) is darkly-colored, which makes for a challenging presentation in three-dimensional perception, particularly with the background selected for this example scenario. - As discussed briefly above, one of the main challenges for a presenting darkfield augmented reality object is that the system generally cannot add or paint in “darkness”; generally the display is configured to add light. Thus, referring to
FIG. 19L , without any specialized software treatments to enhance darkfield perception, presentation of the robot character in the augmented reality view results in a scene wherein portions of the robot character that are to be essentially flat black in presentation are not visible, and portions of the robot character that are to have some lighting (such as the lightly-pigmented cover of the shoulder gun of the robot character) are only barely visible (634)—they appear almost like a light grayscale disruption to the otherwise normal background image. - Referring to
FIG. 19M , using a software-based global attenuation treatment (akin to digitally putting on a pair of sunglasses) provides enhanced visibility to the robot character because the brightness of the nearly black robot character is effective increased relative to the rest of the space, which now appears more dark (640). Also shown inFIG. 19M is a digitally-added light halo (636) which may be added to enhance and distinguish the now-more-visible robot character shapes (638) from the background. With the halo treatment, even the portions of the robot character that are to be presented as flat black become visible with the contrast to the white halo, or “aura” presented around the robot character. - Preferably the halo may be presented to the user with a perceived focal distance that is behind the focal distance of the robot character in three-dimensional space. In a configuration wherein single panel occlusion techniques such as those described above is being utilized to present darkfield, the light halo may be presented with an intensity gradient to match the dark halo that may accompany the occlusion, minimizing the visibility of either darkfield effect. Further, the halo may be presented with blurring to the background behind the presented halo illumination for further distinguishing effect. A more subtle aura or halo effect may be created by matching, at least in part, the color and/or brightness of a relatively light-colored background.
- Referring to
FIG. 19N , some or all of the black intonations of the robot character may be changed to dark, cool blue colors to provide a further distinguishing effect relative to the background, and relatively good visualization of the robot (642). - Wedge-shaped waveguides have been described above, such as in reference to
FIGS. 15A-15D andFIG. 19I . With a wedge-shaped waveguide, every time a ray bounces off of one of the non-coplanar surfaces, it gets an angle change, which ultimately results in the ray exiting total internal reflection when its approach angle to one of the surfaces goes past the critical angle. Turning films may be used to redirect exiting light so that exiting beams leave with a trajectory that is more or less perpendicular to the exit surface, depending upon the geometric and ergonomic issues at play. - With a series or array of displays injecting image information into a wedge-shaped waveguide, as shown in
FIG. 15C , for example, the wedge-shaped waveguide may be configured to create a fine-pitched array of angle-biased rays emerging from the wedge. Somewhat similarly, it has been discussed above that a lightfield display, or a variable wavefront creating waveguide, both may produce a multiplicity of beamlets or beams to represent a single pixel in space such that wherever the eye is positioned, the eye is hit by a plurality of different beamlets or beams that are unique to that particular eye position in front of the display panel. - As was further discussed above in the context of lightfield displays, a plurality of viewing zones may be created within a given pupil, and each may be used for a different focal distance, with the aggregate producing a perception similar to that of a variable wavefront creating waveguide, or similar to the actual optical physics of reality of the objects viewed were real. Thus a wedge-shaped waveguide with multiple displays may be utilized to generate a lightfield. In an embodiment similar to that of
FIG. 15C with a linear array of displays injecting image information, a fan of exiting rays is created for each pixel. This concept may be extended in an embodiment wherein multiple linear arrays are stacked to all inject image information into the wedge-shaped waveguide (in one variation, one array may inject at one angle relative to the wedge-shaped waveguide face, while the second array may inject at a second angle relative to the wedge-shaped waveguide face), in which case exit beams fan out at two different axes from the wedge. - Thus such a configuration may be utilized to produce pluralities of beams spraying out at lots of different angles, and each beam may be driven separately due to the fact that under such configuration, each beam is driven using a separate display. In another embodiment, one or more arrays or displays may be configured to inject image information into wedge-shaped waveguide through sides or faces of the wedge-shaped waveguide other than that shown in
FIG. 15C , such as by using a diffractive optic to bend injected image information into total an internal reflection configuration relative to the wedge-shaped waveguide. - Various reflectors or reflecting surfaces may also be utilized in concert with such a wedge-shaped waveguide embodiment to outcouple and manage light from the wedge-shaped waveguide. In one embodiment, an entrance aperture to a wedge-shaped waveguide, or injection of image information through a different face other than shown in
FIG. 15C , may be utilized to facilitate staggering (geometric and/or temporal) of different displays and arrays such that a Z-axis delta may also be developed as a means for injecting three-dimensional information into the wedge-shaped waveguide. For a greater than three-dimensions array configuration, various displays may be configured to enter a wedge-shaped waveguide at multiple edges in multiple stacks with staggering to get higher dimensional configurations. - Referring to
FIG. 20A , a configuration similar to that depicted inFIG. 8H is shown wherein a waveguide (646) has a diffractive optical element (648; or “DOE”, as noted above) sandwiched in the middle (alternatively, as described above, the diffractive optical element may reside on the front or back face of the depicted waveguide). A ray may enter the waveguide (646) from the projector or display (644). Once in the waveguide (646), each time the ray intersects the DOE (648), part of it is exited out of the waveguide (646). As described above, the DOE may be designed such that the exit illuminance across the length of the waveguide (646) is somewhat uniform (for example, the first such DOE intersection may be configured to exit about 10% of the light; then the second DOE intersection may be configured to exit about 10% of the remaining light so that 81% is passed on, and so on; in another embodied a DOE may be designed to have a variable diffraction efficiency, such as linearly-decreasing diffraction efficiency, along its length to map out a more uniform exit illuminance across the length of the waveguide). - To further distribute remaining light that reaches an end (and in one embodiment to allow for selection of a relatively low diffraction efficiency DOE which would be favorable from a view-to-the-world transparency perspective), a reflective element (650) at one or both ends may be included. Further, referring to the embodiment of
FIG. 20B , additional distribution and preservation may be achieved by including an elongate reflector (652) across the length of the waveguide as shown (comprising, for example, a thin film dichroic coating that is wavelength-selective); preferably such reflector would be blocking light that accidentally is reflected upward (back toward thereal world 144 for exit in a way that it would not be utilized by the viewer). In some embodiments, such an elongate reflector may contribute to a “ghosting” effect perception by the user. - In one embodiment, this ghosting effect may be eliminated by having a dual-waveguide (646, 654) circulating reflection configuration, such as that shown in
FIG. 20C , which is designed to keep the light moving around until it has been exited toward the eye (58) in a preferably substantially equally distributed manner across the length of the waveguide assembly. Referring toFIG. 20C , light may be injected with a projector or display (644), and as it travels across the DOE (656) of the first waveguide (654), it ejects a preferably substantially uniform pattern of light out toward the eye (58); light that remains in the first waveguide is reflected by a first reflector assembly (660) into the second waveguide (646). In one embodiment, the second waveguide (646) may be configured to not have a DOE, such that it merely transports or recycles the remaining light back to the first waveguide, using the second reflector assembly. - In another embodiment (as shown in
FIG. 20C ) the second waveguide (646) may also have a DOE (648) configured to uniformly eject fractions of travelling light to provide a second plane of focus for three-dimensional perception. Unlike the configurations ofFIGS. 20A and 20B , the configuration ofFIG. 20C is designed for light to travel the waveguide in one direction, which avoids the aforementioned ghosting problem that is related to passing light backwards through a waveguide with a DOE. Referring toFIG. 20D , rather than having a mirror or box style reflector assembly (660) at the ends of a waveguide for recycling the light, an array of smaller retroreflectors (662), or a retroreflective material, may be utilized. - Referring to
FIG. 20E , an embodiment is shown that utilizes some of the light recycling configurations of the embodiment ofFIG. 20C to “snake” the light down through a waveguide (646) having a sandwiched DOE (648) after it has been injected with a display or projector (644) so that it crosses the waveguide (646) many times back and forth before reaching the bottom, at which point it may be recycled back up to the top level for further recycling. Such a configuration not only recycles the light and facilitates use of relatively low diffraction efficiency DOE elements for exiting light toward the eye (58), but also distributes the light, to provide for a large exit pupil configuration akin to that described in reference toFIG. 8K . - Referring to
FIG. 20F , an illustrative configuration similar to that ofFIG. 5A is shown, with incoming light injected along a conventional prism or beamsplitter substrate (104) to a reflector (102) without total internal reflection (i.e., without the prism being considered a waveguide) because the input projection (106), scanning or otherwise, is kept within the bounds of the prism—which means that the geometry of such prism becomes a significant constraint. In another embodiment, a waveguide may be utilized in place of the simple prism ofFIG. 20F , which facilitates the use of total internal reflection to provide more geometric flexibility. - Other configurations describe above are configured to profit from the inclusion of waveguides for similar manipulations and light. For example, referring back to
FIG. 7A , the general concept illustrated therein is that a collimated image injected into a waveguide may be refocused before transfer out toward an eye, in a configuration also designed to facilitate viewing light from the real world. In place of the refractive lens shown inFIG. 7A , a diffractive optical element may be used as a variable focus element. - Referring back to
FIG. 7B , another waveguide configuration is illustrated in the context of having multiple layers stacked upon each other with controllable access toggling between a smaller path (total internal reflection through a waveguide) and a larger path (total internal reflection through a hybrid waveguide comprising the original waveguide and a liquid crystal isolated region with the liquid crystal switched to a mode wherein the refractive indices are substantially matched between the main waveguide and the auxiliary waveguide), so that the controller can tune on a frame-by-frame basis which path is being taken. High-speed switching electro-active materials, such as lithium niobate, facilitate path changes with such a configuration at gigahertz rates, which allows one to change the path of light on a pixel-by-pixel basis. - Referring back to
FIG. 8A , a stack of waveguides paired with weak lenses is illustrated to demonstrate a multifocal configuration wherein the lens and waveguide elements may be static. Each pair of waveguide and lens may be functionally replaced with waveguide having an embedded DOE element (which may be static, in a closer analogy to the configuration ofFIG. 8A , or dynamic), such as that described in reference toFIG. 8I . - Referring to
FIG. 20G , if a transparent prism or block (104; i.e., not a waveguide) is utilized to hold a mirror or reflector (102) in a periscope type of configuration to receive light from other components, such as a lens (662) and projector or display (644), the field of view is limited by the size of that reflector (102; the bigger the reflector, the wider the field of view). Thus to have a larger field of view with such configuration, a thicker substrate may be needed to hold a larger reflector; otherwise, the functionality of an aggregated plurality of reflectors may be utilized to increase the functional field of view, as described in reference toFIGS. 8O , 8P, and 8Q. Referring toFIG. 20H , a stack (664) of planar waveguides (666), each fed with a display or projector (644; or in another embodiment a multiplexing of a single display) and having an exit reflector (668), may be utilized to aggregate toward the function of a larger single reflector. The exit reflectors may be at the same angle in some cases, or not the same angle in other cases, depending upon the positioning of the eye (58) relative to the assembly. -
FIG. 20I illustrates a related configuration, wherein the reflectors (680, 682, 684, 686, 688) in each of the planar waveguides (670, 672, 674, 676, 678) have been offset from each other, and wherein each takes in light from a projector or display (644) which may be sent through a lens (690) to ultimately contribute exiting light to the pupil (45) of the eye (58) by virtue of the reflectors (680, 682, 684, 686, 688) in each of the planar waveguides (670, 672, 674, 676, 678). If one can create a total range of all of the angles that would be expected to be seen in the scene (i.e., preferably without blind spots in the key field of view), then a useful field of view has been achieved. As described above, the eye (58) functions based at least on what angle light rays enter the eye, and this can be simulated. The rays need not pass through the exact same point in space at the pupil—rather the light rays just need to get through the pupil and be sensed by the retina.FIG. 20K illustrates a variation wherein the shaded portion of the optical assembly may be utilized as a compensating lens to functionally pass light from the real world (144) through the assembly as though it has been passed through a zero power telescope. - Referring to
FIG. 20J , each of the aforementioned rays may also be a relative wide beam that is being reflected through the pertinent waveguide (670, 672) by total internal reflection. The reflector (680, 682) facet size will determine what the exiting beam width can be. - Referring to
FIG. 20L , a further discretization of the reflector is shown, wherein a plurality of small straight angular reflectors may form a roughly parabolic reflecting surface (694) in the aggregate through a waveguide or stack thereof (696). Light coming in from the displays (644; or single MUXed display, for example), such as through a lens (690), is all directed toward the same shared focal point at the pupil (45) of the eye (58). - Referring back to
FIG. 13M , a linear array of displays (378) injects light into a shared waveguide (376). In another embodiment a single display may be multiplexed to a series of entry lenses to provide similar functionality as the embodiment ofFIG. 13M , with the entry lenses creating parallel paths of rays running through the waveguide. - In a conventional waveguide approach wherein total internal reflection is relied upon for light propagation, the field of view is restricted because there is only a certain angular range of rays propagating through the waveguide (others may escape out). In one embodiment, if a red/green/blue (or “RGB”) laserline reflector is placed at one or both ends of the planar surfaces, akin to a thin film interference filter that is highly reflective for only certain wavelengths and poorly reflective for other wavelengths, than one can functionally increase the range of angles of light propagation. Windows (without the coating) may be provided for allowing light to exit in predetermined locations. Further, the coating may be selected to have a directional selectivity (somewhat like reflective elements that are only highly reflective for certain angles of incidence). Such a coating may be most relevant for the larger planes/sides of a waveguide.
- Referring back to
FIG. 13E , a variation on a scanning fiber display was discussed, which may be deemed a scanning thin waveguide configuration, such that a plurality of very thin planar waveguides (358) may be oscillated or vibrated such that if a variety of injected beams is coming through with total internal reflection, the configuration functionally would provide a linear array of beams escaping out of the edges of the vibrating elements (358). The depicted configuration has approximately five externally-projecting planar waveguide portions (358) in a host medium or substrate (356) that is transparent, but which preferably has a different refractive index so that the light will stay in total internal reflection within each of the substrate-bound smaller waveguides that ultimately feed (in the depicted embodiment there is a 90 degree turn in each path at which point a planar, curved, or other reflector may be utilized to bounce the light outward) the externally-projecting planar waveguide portions (358). - The externally-projecting planar waveguide portions (358) may be vibrated individually, or as a group along with oscillatory motion of the substrate (356). Such scanning motion may provide horizontal scanning, and for vertical scanning, the input (360) aspect of the assembly (i.e., such as one or more scanning fiber displays scanning in the vertical axis) may be utilized. Thus a variation of the scanning fiber display is presented.
- Referring back to
FIG. 13H , a waveguide (370) may be utilized to create a lightfield. With waveguides working best with collimated beams that may be associated with optical infinity from a perception perspective, all beams staying in focus may cause perception discomfort (i.e., the eye will not make a discernible difference in dioptric blur as a function of accommodation; in other words, the narrow diameter, such as 0.5 mm or less, collimated beamlets may open loop the eye's accommodation/vergence system, causing discomfort). - In one embodiment, a single beam may be fed in with a number of cone beamlets coming out, but if the introduction vector of the entering beam is changed (i.e., laterally shift the beam injection location for the projector/display relative to the waveguide), one may control where the beam exits from the waveguide as it is directed toward the eye. Thus one may use a waveguide to create a lightfield by creating a bunch of narrow diameter collimated beams, and such a configuration is not reliant upon a true variation in a light wavefront to be associated with the desired perception at the eye.
- If a set of angularly and laterally diverse beamlets is injected into a waveguide (for example, by using a multicore fiber and driving each core separately; another configuration may utilize a plurality of fiber scanners coming from different angles; another configuration may utilize a high-resolution panel display with a lenslet array on top of it), a number of exiting beamlets can be created at different exit angles and exit locations. Since the waveguide may scramble the lightfield, the decoding is preferably predetermined.
- Referring to
FIGS. 20M and 20N , a waveguide (646) assembly (696) is shown that comprises stacked waveguide components in the vertical or horizontal axis. Rather than having one monolithic planar waveguide, the notion with these embodiments is to stack a plurality of smaller waveguides (646) immediately adjacent each other such that light introduced into one waveguide, in addition to propagating down (i.e., propagating along a Z axis with total internal reflection in +X,−X) such waveguide by total internal reflection, also totally internally reflects in the perpendicular axis (+y, −Y) as well, such that it is not spilling into other areas. In other words, if total internal reflection is from left to right and back during Z axis propagation, the configuration will be set up to totally internally reflect any light that hits the top or bottom sides as well; each layer may be driven separately without interference from other layers. Each waveguide may have a DOE (648) embedded and configured to eject out light with a predetermined distribution along the length of the waveguide, as described above, with a predetermined focal length configuration (shown inFIG. 20M as ranging from 0.5 meters to optical infinity). - In another variation, a very dense stack of waveguides with embedded DOEs may be produced such that it spans the size of the anatomical pupil of the eye (i.e., such that
multiple layers 698 of the composite waveguide are required to cross the exit pupil, as illustrated inFIG. 20N ). With such a configuration, one may feed a collimated image for one wavelength, and then the portion located the next millimeter down producing a diverging wavefront that represents an object coming from a focal distance of, say, 15 meters away, and so on, with the notion being that an exit pupil is coming from a number of different waveguides as a result of the DOEs and total internal reflection through the waveguides and across the DOEs. Thus rather than creating one uniform exit pupil, such a configuration creates a plurality of stripes that, in the aggregate, facilitate the perception of different focal depths with the eye/brain. - Such a concept may be extended to configurations comprising a waveguide with a switchable/controllable embedded DOE (i.e. that is switchable to different focal distances), such as those described in relation to
FIGS. 8B-8N , which allows more efficient light trapping in the axis across each waveguide. Multiple displays may be coupled into each of the layers, and each waveguide with DOE would emit rays along its own length. In another embodiment, rather than relying on total internal reflection, a laserline reflector may be used to increase angular range. In between layers of the composite waveguide, a completely reflective metallized coating may be utilized, such as aluminum, to ensure total reflection, or alternatively dichroic style or narrow band reflectors may be utilized. - Referring to
FIG. 20O , the whole composite waveguide assembly (696) may be curved concavely toward the eye (58) such that each of the individual waveguides is directed toward the pupil. In other words, the configuration may be designed to more efficiently direct the light toward the location where the pupil is likely to be present. Such a configuration also may be utilized to increase the field of view. - As was discussed above in relation to
FIGS. 8L , 8M, and 8N, a changeable diffraction configuration allows for scanning in one axis, somewhat akin to a scanning light display.FIG. 21A illustrates a waveguide (698) having an embedded (i.e., sandwiched within) DOE (700) with a linear grating term that may be changed to alter the exit angle of exiting light (702) from the waveguide, as shown. A high-frequency switching DOE material such as lithium niobate may be utilized. In one embodiment, such a scanning configuration may be used as the sole mechanism for scanning a beam in one axis; in another embodiment, the scanning configuration may be combined with other scanning axes, and may be used to create a larger field of view (i.e., if a normal field of view is 40 degrees, and by changing the linear diffraction pitch one can steer over another 40 degrees, the effective usable field of view for the system is 80 degrees). - Referring to
FIG. 21B , in a conventional configuration, a waveguide (708) may be placed perpendicular to a panel display (704), such as an LCD or OLED panel, such that beams may be injected from the waveguide (708), through a lens (706), and into the panel (704) in a scanning configuration to provide a viewable display for television or other purposes. Thus the waveguide may be utilized in such configuration as a scanning image source, in contrast to the configurations described in reference toFIG. 21A , wherein a single beam of light may be manipulated by a scanning fiber or other element to sweep through different angular locations, and in addition, another direction may be scanned using the high-frequency diffractive optical element. - In another embodiment, a uniaxial scanning fiber display (say scanning the fast line scan, as the scanning fiber is relatively high frequency) may be used to inject the fast line scan into the waveguide, and then the relatively slow DOE switching (i.e., in the range of 100 Hz) may be used to scan lines in the other axis to form an image.
- In another embodiment, a DOE with a grating of fixed pitch may be combined with an adjacent layer of electro-active material having a dynamic refractive index (such as liquid crystal), so that light may be redirected into the grating at different angles. This is an application of the basic multipath configuration described above in reference to
FIG. 7B , in which an electro-active layer comprising an electro-active material such as liquid crystal or lithium niobate may change its refractive index such that it changes the angle at which a ray emerges from the waveguide. A linear diffraction grating may be added to the configuration ofFIG. 7B (in one embodiment, sandwiched within the glass or other material comprising the larger lower waveguide) such that the diffraction grating may remain at a fixed pitch, but the light is biased before it hits the grating. -
FIG. 21C shows another embodiment featuring two wedge-like waveguide elements (710, 712), wherein one or more of them may be electro-active so that the related refractive index may be changed. The elements may be configured such that when the wedges have matching refractive indices, the light totally internally reflects through the pair (which in the aggregate performs akin to a planar waveguide with both wedges matching) while the wedge interfaces have no effect. Then if one of the refractive indices is changed to create a mismatch, a beam deflection at the wedge interface (714) is caused, and there is total internal reflection from that surface back into the associated wedge. Then a controllable DOE (716) with a linear grating may be coupled along one of the long edges of the wedge to allow light to exit out and reach the eye at a desirable exit angle. - In another embodiment, a DOE such as a Bragg grating, may be configured to change pitch versus time, such as by a mechanical stretching of the grating (for example, if the grating resides on or comprises an elastic material), a moire beat pattern between two gratings on two different planes (the gratings may be the same or different pitches), Z-axis motion (i.e., closer to the eye, or farther away from the eye) of the grating, which functionally is similar in effect to stretching of the grating, or electro-active gratings that may be switched on or off, such as one created using a polymer dispersed liquid crystal approach wherein liquid crystal droplets may be controllably activated to change the refractive index to become an active grating, versus turning the voltage off and allowing a switch back to a refractive index that matches that of the host medium.
- In another embodiment, a time-varying grating may be utilized for field of view expansion by creating a tiled display configuration. Further, a time-varying grating may be utilized to address chromatic aberration (failure to focus all colors/wavelengths at the same focal point). One property of diffraction gratings is that they will deflect a beam as a function of its angle of incidence and wavelength (i.e., a DOE will deflect different wavelengths by different angles: somewhat akin to the manner in which a simple prism will divide out a beam into its wavelength components).
- One may use time-varying grating control to compensate for chromatic aberration in addition to field of view expansion. Thus, for example, in a waveguide with embedded DOE type of configuration as described above, the DOE may be configured to drive the red wavelength to a slightly different place than the green and blue to address unwanted chromatic aberration. The DOE may be time-varied by having a stack of elements that switch on and off (i.e. to get red, green, and blue to be diffracted outbound similarly).
- In another embodiment, a time-varying grating may be utilized for exit pupil expansion. For example, referring to
FIG. 21D , it is possible that a waveguide (718) with embedded DOE (720) may be positioned relative to a target pupil such that none of the beams exiting in a baseline mode actually enter the target pupil (45)—such that the pertinent pixel would be missed by the user. A time-varying configuration may be utilized to fill in the gaps in the outbound exit pattern by shifting the exit pattern laterally (shown in dashed/dotted lines) to effectively scan each of the 5 exiting beams to better ensure that one of them hits the pupil of the eye. In other words, the functional exit pupil of the display system is expanded. - In another embodiment, a time-varying grating may be utilized with a waveguide for one, two, or three axis light scanning. In a manner akin to that described in reference to
FIG. 21A , one may use a term in a grating that is scanning a beam in the vertical axis, as well as a grating that is scanning in the horizontal axis. Further, if radial elements of a grating are incorporated, as is discussed above in relation toFIGS. 8B-8N , one may have scanning of the beam in the Z axis (i.e., toward/away from the eye), all of which may be time sequential scanning. - Notwithstanding the discussions herein regarding specialized treatments and uses of DOEs generally in connection with waveguides, many of these uses of DOE are usable whether or not the DOE is embedded in a waveguide. For example, the output of a waveguide may be separately manipulated using a DOE; or a beam may be manipulated by a DOE before it is injected into a waveguide; further, one or more DOEs, such as a time-varying DOE, may be utilized as an input for freeform optics configurations, as discussed below.
- As discussed above in reference to
FIGS. 8B-8N , an element of a DOE may have a circularly-symmetric term, which may be summed with a linear term to create a controlled exit pattern (i.e., as described above, the same DOE that outcouples light may also focus it). In another embodiment, the circular term of the DOE diffraction grating may be varied such that the focus of the beams representing those pertinent pixels is modulated. In addition, one configuration may have a second/separate circular DOE, obviating the need to have a linear term in the DOE. - Referring to
FIG. 21E , one may have a waveguide (722) outputting collimated light with no DOE element embedded, and a second waveguide that has a circularly-symmetric DOE that can be switched between multiple configurations—in one embodiment by having a stack (724) of such DOE elements (FIG. 21F shows another configuration wherein afunctional stack 728 of DOE elements may comprise a stack of polymer dispersedliquid crystal elements 726, as described above, wherein without a voltage applied, a host medium refraction index matches that of a dispersed molecules of liquid crystal; in another embodiment, molecules of lithium niobate may be dispersed for faster response times; with voltage applied, such as through transparent indium tin oxide layers on either side of the host medium, the dispersed molecules change index of refraction and functionally form a diffraction pattern within the host medium) that can be switched on/off. - In another embodiment, a circular DOE may be layered in front of a waveguide for focus modulation. Referring to
FIG. 21G , the waveguide (722) is outputting collimated light, which will be perceived as associated with a focal depth of optical infinity unless otherwise modified. The collimated light from the waveguide may be input into a diffractive optical element (730) which may be used for dynamic focus modulation (i.e., one may switch on and off different circular DOE patterns to impart various different focuses to the exiting light). In a related embodiment, a static DOE may be used to focus collimated light exiting from a waveguide to a single depth of focus that may be useful for a particular user application. - In another embodiment, multiple stacked circular DOEs may be used for additive power and many focus levels—from a relatively small number of switchable DOE layers. In other words, three different DOE layers may be switched on in various combinations relative to each other; the optical powers of the DOEs that are switched on may be added. In one embodiment wherein a range of up to 4 diopters is desired, for example, a first DOE may be configured to provide half of the total diopter range desired (in this example, 2 diopters of change in focus); a second DOE may be configured to induce a 1 diopter change in focus; then a third DOE may be configured to induce a ½ diopter change in focus. These three DOEs may be mixed and matched to provide ½, 1, 1.5, 2, 2.5, 3, and 3.5 diopters of change in focus. Thus a super large number of DOEs would not be required to get a relatively broad range of control.
- In one embodiment, a matrix of switchable DOE elements may be utilized for scanning, field of view expansion, and/or exit pupil expansion. Generally in the above discussions of DOEs, it has been assume that a typical DOE is either all on or all off. In one variation, a DOE (732) may be subdivided into a plurality of functional subsections (such as the one labeled as
element 734 inFIG. 21H ), each of which preferably is uniquely controllable to be on or off (for example, referring toFIG. 21H , each subsection may be operated by its own set of indium tin oxide, or other control lead material, voltage application leads 736 back to a central controller). Given this level of control over a DOE paradigm, additional configurations are facilitated. - Referring to
FIG. 21I , a waveguide (738) with embedded DOE (740) is viewed from the top down, with the user's eye positioned in front of the waveguide. A given pixel may be represented as a beam coming into the waveguide and totally internally reflecting along until it may be exited by a diffraction pattern to come out of the waveguide as a set of beams. Depending upon the diffraction configuration, the beams may come out parallel/collimated (as shown inFIG. 21I for convenience), or in a diverging fan configuration if representing a focal distance closer than optical infinity. - The depicted set of parallel exiting beams may represent, for example, the farthest left pixel of what the user is seeing in the real world as viewed through the waveguide, and light off to the rightmost extreme will be a different group of parallel exiting beams. Indeed, with modular control of the DOE subsections as described above, one may spend more computing resource or time creating and manipulating the small subset of beams that is likely to be actively addressing the user's pupil (i.e., because the other beams never reach the user's eye and are effectively wasted). Thus, referring to
FIG. 21J , a waveguide (738) configuration is shown wherein only the two subsections (740, 742) of the DOE (744) are deemed to be likely to address the user's pupil (45) are activated. Preferably one subsection may be configured to direct light in one direction simultaneously as another subsection is directing light in a different direction. -
FIG. 21K shows an orthogonal view of two independently controlled subsections (734, 746) of a DOE (732). Referring to the top view ofFIG. 21L , such independent control may be used for scanning or focusing light. In the configuration depicted inFIG. 21K , an assembly (748) of three independently controlled DOE/waveguide subsections (750, 752, 754) may be used to scan, increase the field of view, and/or increase the exit pupil region. Such functionality may arise from a single waveguide with such independently controllable DOE subsections, or a vertical stack of these for additional complexity. - In one embodiment, if a circular DOE may be controllably stretched radially-symmetrically, the diffraction pitch may be modulated, and the DOE may be utilized as a tunable lens with an analog type of control. In another embodiment, a single axis of stretch (for example, to adjust an angle of a linear DOE term) may be utilized for DOE control. Further, in another embodiment a membrane, akin to a drum head, may be vibrated, with oscillatory motion in the Z-axis (i.e., toward/away from the eye) providing Z-axis control and focus change over time.
- Referring to
FIG. 21M , a stack of several DOEs (756) is shown receiving collimated light from a waveguide (722) and refocusing it based upon the additive powers of the activated DOEs. Linear and/or radial terms of DOEs may be modulated over time, such as on a frame sequential basis, to produce a variety of treatments (such as tiled display configurations or expanded field of view) for the light coming from the waveguide and exiting, preferably toward the user's eye. In configurations wherein the DOE or DOEs are embedded within the waveguide, a low diffraction efficiency is desired to maximize transparency for light passed from the real world; in configurations wherein the DOE or DOEs are not embedded, a high diffraction efficiency may be desired, as described above. In one embodiment, both linear and radial DOE terms may be combined outside of the waveguide, in which case high diffraction efficiency would be desired. - Referring to
FIG. 21N , a segmented or parabolic reflector, such as those discussed above inFIG. 8Q , is shown. Rather than executing a segmented reflector by combining a plurality of smaller reflectors, in one embodiment the same functionality may result from a single waveguide with a DOE having different phase profiles for each section of it, such that it is controllable by subsection. In other words, while the entire segmented reflector functionality may be turned on or off together, generally the DOE may be configured to direct light toward the same region in space (i.e., the pupil of the user). - Referring to
FIGS. 22A-22Z , optical configurations known as “freeform optics” may be utilized certain of the aforementioned challenges. The term “freeform” generally is used in reference to arbitrarily curved surfaces that may be utilized in situations wherein a spherical, parabolic, or cylindrical lens does not meet a design complexity such as a geometric constraint. For example, referring toFIG. 22A , one of the common challenges with display (762) configurations when a user is looking through a mirror (and also sometimes a lens 760) is that the field of view is limited by the area subtended by the final lens (760) of the system. - Referring to
FIG. 22B , in more simple terms, if one has a display (762), which may include some lens elements, there is a straightforward geometric relationship such that the field of view cannot be larger than the angle subtended by the display (762). Referring toFIG. 22C , this challenge is exacerbated if the user is trying to have an augmented reality experience wherein light from the real world is also be to passed through the optical system, because in such case, there often is a reflector (764) that leads to a lens (760); by interposing a reflector, the overall path length to get to the lens from the eye is increased, which tightens the angle and reduces the field of view. - Given this, if one wants to increase the field of view, he must increase the size of the lens, but that might mean pushing a physical lens toward the forehead of the user from an ergonomic perspective. Further, the reflector may not catch all of the light from the larger lens. Thus, there is a practical limitation imposed by human head geometry, and it generally is a challenge to get more than a 40-degree field of view using conventional see-through displays and lenses.
- With freeform lenses, rather than having a standard planar reflector as described above, one has a combined reflector and lens with power (i.e., a curved reflector 766), which means that the curved lens geometry determines the field of view. Referring to
FIG. 22D , without the circuitous path length of a conventional paradigm as described above in reference toFIG. 22C , it is possible for a freeform arrangement to realize a significantly larger field of view for a given set of optical requirements. - Referring to
FIG. 22E , a typical freeform optic has three active surfaces. Referring toFIG. 22E , in one typical freeform optic (770) configuration, light may be directed toward the freeform optic from an image plane, such as a flat panel display (768), into the first active surface (772), which typically is a primarily transmissive freeform surface that refracts transmitted light and imparts a focal change (such as an added stigmatism, because the final bounce from the third surface will add a matching/opposite stigmatism and these are desirably canceled). The incoming light may be directed from the first surface to a second surface (774), wherein it may strike with an angle shallow enough to cause the light to be reflected under total internal reflection toward the third surface (776). - The third surface may comprise a half-silvered, arbitrarily-curved surface configured to bounce the light out through the second surface toward the eye, as shown in
FIG. 22E . Thus in the depicted typical freeform configuration, the light enters through the first surface, bounces from the second surface, bounces from the third surface, and is directed out of the second surface. Due to the optimization of the second surface to have the requisite reflective properties on the first pass, as well as refractive properties on the second pass as the light is exited toward the eye, a variety of curved surfaces with higher-order shapes than a simple sphere or parabola are formed into the freeform optic. - Referring to
FIG. 22F , a compensating lens (780) may be added to the freeform optic (770) such that the total thickness of the optic assembly is substantially uniform in thickness, and preferably without magnification, to light incoming from the real world (144) in an augmented reality configuration. - Referring to
FIG. 22G , a freeform optic (770) may be combined with a waveguide (778) configured to facilitate total internal reflection of captured light within certain constraints. For example, as shown inFIG. 22G , light may be directed into the freeform/waveguide assembly from an image plane, such as a flat panel display, and totally internally reflected within the waveguide until it hits the curved freeform surface and escapes toward the eye of the user. Thus the light bounces several times in total internal reflection until it reaches the freeform wedge portion. - One of the main objectives with such an assembly is to try to lengthen the optic assembly while retaining as uniform a thickness as possible (to facilitate transport by total internal reflection, and also viewing of the world through the assembly without further compensation) for a larger field of view.
FIG. 22H depicts a configuration similar to that ofFIG. 22G , with the exception that the configuration ofFIG. 22H also features a compensating lens portion to further extend the thickness uniformity and assist with viewing the world through the assembly without further compensation. - Referring to
FIG. 22I , in another embodiment, a freeform optic (782) is shown with a small flat surface, or fourth face (784), at the lower left corner that is configured to facilitate injection of image information at a different location than is typically used with freeform optics. The input device (786) may comprise, for example, a scanning fiber display, which may be designed to have a very small output geometry. The fourth face may comprise various geometries itself and have its own refractive power, such as by use planar or freeform surface geometries. - Referring to
FIG. 22J , in practice, such a configuration may also feature a reflective coating (788) along the first surface such that it directs light back to the second surface, which then bounces the light to the third surface, which directs the light out across the second surface and to the eye (58). The addition of the fourth small surface for injection of the image information facilitates a more compact configuration. In an embodiment wherein a classical freeform input configuration and a scanning fiber display (790) are utilized, some lenses (792, 794) may be required in order to appropriately form an image plane (796) using the output from the scanning fiber display; these hardware components add extra bulk that may not be desired. - Referring to
FIG. 22K , an embodiment is shown wherein light from a scanning fiber display (790) is passed through an input optics assembly (792, 794) to an image plane (796), and then directed across the first surface of the freeform optic (770) to a total internal reflection bounce off of the second surface, then another total internal reflection bounce from the third surface results in the light exiting across the second surface and being directed toward the eye (58). - An all-total-internal-reflection freeform waveguide may be created such that there are no reflective coatings (i.e., such that total-internal-reflection is being relied upon for propagation of light until a critical angle of incidence with a surface is met, at which point the light exits in a manner akin to the wedge-shaped optics described above). In other words, rather than having two planar surfaces, one may have a surface comprising one or more sub-surfaces from a set of conical curves, such as parabolas, spheres, ellipses, etc.).
- Such a configuration still may produce a shallow-enough angles for total internal reflection within the optic; thus an approach that is somewhat a hybrid between a conventional freeform optic and a wedge-shaped waveguide is presented. One motivation to have such a configuration is to get away from the use of reflective coatings, which do help product reflection, but also are known to prevent transmission of a relatively large portion (such as 50%) of the light transmitting through from the real world (144). Further, such coatings also may block an equivalent amount of the light coming into the freeform optic from the input device. Thus there are reasons to develop designs that do not have reflective coatings.
- As described above, one of the surfaces of a conventional freeform optic may comprise a half-silvered reflective surface. Generally such a reflective surface will be of “neutral density”, meaning that it will generally reflect all wavelengths similarly. In another embodiment, such as one wherein a scanning fiber display is utilized as an input, the conventional reflector paradigm may be replaced with a narrow band reflector that is wavelength sensitive, such as a thin film laserline reflector. Thus in one embodiment, a configuration may reflect particular red/green/blue wavelength ranges and remain passive to other wavelengths, which generally will increase transparency of the optic and therefore be preferred for augmented reality configurations wherein transmission of image information from the real world (144) across the optic also is valued.
- Referring to
FIG. 22L , an embodiment is depicted wherein multiple freeform optics (770) may be stacked in the Z axis (i.e., along an axis substantially aligned with the optical axis of the eye). In one variation, each of the three depicted freeform optics may have a wavelength-selective coating (for example, one highly selective for blue, the next for green, the next for red) so that images may be injected into each to have blue reflected from one surface, green from another, and red from a third surface. Such a configuration may be utilized, for example, to address chromatic aberration issues, to create a lightfield, or to increase the functional exit pupil size. - Referring to
FIG. 22M , an embodiment is shown wherein a single freeform optic (798) has multiple reflective surfaces (800, 802, 804), each of which may be wavelength or polarization selective so that their reflective properties may be individually controlled. - Referring to
FIG. 22N , in one embodiment, multiple microdisplays, such as scanning light displays, (786) may be injected into a single freeform optic to tile images (thereby providing an increased field of view), increase the functional pupil size, or address challenges such as chromatic aberration (i.e., by reflecting one wavelength per display). Each of the depicted displays would inject light that would take a different path through the freeform optic due to the different positioning of the displays relative to the freeform optic, which would provide a larger functional exit pupil output. - In one embodiment, a packet or bundle of scanning fiber displays may be utilized as an input to overcome one of the challenges in operatively coupling a scanning fiber display to a freeform optic. One such challenge with a scanning fiber display configuration is that the output of an individual fiber is emitted with a certain numerical aperture, or “NA”, which is like the projectional angle of light from the fiber; ultimately this angle determines the diameter of the beam that passes through various optics, and ultimately determines the exit functional exit pupil size; thus in order to maximize exit pupil size with a freeform optic configuration, one may either increase the NA of the fiber using optimized refractive relationships, such as between core and cladding, or one may place a lens (i.e., a refractive lens, such as a gradient refractive index lens, or “GRIN” lens) at the end of the fiber or build one into the end of the fiber as described above, or create an array of fibers that is feeding into the freeform optic, in which case all of those NAs in the bundle remain small, and at the exit pupil an array of small exit pupils is produced that in the aggregate forms the functional equivalent of a large exit pupil.
- Alternatively, in another embodiment a more sparse array (i.e., not bundled tightly as a packet) of scanning fiber displays or other displays may be utilized to functionally increase the field of view of the virtual image through the freeform optic. Referring to
FIG. 22O , in another embodiment, a plurality of displays or displays (786) may be injected through the top of a freeform optic (770), as well as another plurality (786) through the lower corner; the display arrays may be two or three dimensional arrays. Referring toFIG. 22P , in another related embodiment, image information also may be injected in from the side (806) of the freeform optic (770) as well. - In an embodiment wherein a plurality of smaller exit pupils is to be aggregated into a functionally larger exit pupil, one may elect to have each of the scanning fibers monochromatic, such that within a given bundle or plurality of projectors or displays, one may have a subgroup of solely red fibers, a subgroup of solely blue fibers, and a subgroup of solely green fibers. Such a configuration facilitates more efficiency in output coupling for bringing light into the optical fibers; for instance, there would be no need in such an embodiment to superimpose red, green, and blue into the same band.
- Referring to
FIGS. 22Q-22V , various freeform optic tiling configurations are depicted. Referring toFIG. 22Q , an embodiment is depicted wherein two freeform optics are tiled side-by-side and a microdisplay, such as a scanning light display, (786) on each side is configured to inject image information from each side, such that one freeform optic wedge represents each half of the field of view. - Referring to
FIG. 22R , a compensator lens (808) may be included to facilitate views of the real world through the optics assembly.FIG. 22S illustrates a configuration wherein freeform optics wedges are tiled side by side to increase the functional field of view while keeping the thickness of such optical assembly relatively uniform. - Referring to
FIG. 22T , a star-shaped assembly comprises a plurality of freeform optics wedges (also shown with a plurality of displays for inputting image information) in a configuration that may provide a larger field of view expansion while also maintaining a relatively thin overall optics assembly thickness. - With a tiled freeform optics assembly, the optics elements may be aggregated to produce a larger field of view; the tiling configurations described above have addressed this notion. For example, in a configuration wherein two freeform waveguides are aimed at the eye such as that depicted in
FIG. 22R , there are several ways to increase the field of view. One option is to “toe in” the freeform waveguides such that their outputs share, or are superimposed in, the space of the pupil (for example, the user may see the left half of the visual field through the left freeform waveguide, and the right half of the visual field through the right freeform waveguide). - With such a configuration, the field of view has been increased with the tiled freeform waveguides, but the exit pupil has not grown in size. Alternatively, the freeform waveguides may be oriented such that they do not toe in as much—so they create exit pupils that are side-by-side at the eye's anatomical pupil. In one example, the anatomical pupil may be 8 mm wide, and each of the side-by-side exit pupils may be 8 mm, such that the functional exit pupil is expanded by about two times. Thus such a configuration provides an enlarged exit pupil, but if the eye is moved around in the “eyebox” defined by that exit pupil, that eye may lose parts of the visual field (i.e., lose either a portion of the left or right incoming light because of the side-by-side nature of such configuration).
- In one embodiment using such an approach for tiling freeform optics, especially in the Z-axis relative to the eye of the user, red wavelengths may be driven through one freeform optic, green through another, and blue through another, such red/green/blue chromatic aberration may be addressed. Multiple freeform optics also may be provided to such a configuration that are stacked up, each of which is configured to address a particular wavelength.
- Referring to
FIG. 22U , two oppositely-oriented freeform optics are shown stacked in the Z-axis (i.e., they are upside down relative to each other). With such a configuration, a compensating lens may not be required to facilitate accurate views of the world through the assembly; in other words, rather than having a compensating lens such as in the embodiment ofFIG. 22F orFIG. 22R , an additional freeform optic may be utilized, which may further assist in routing light to the eye.FIG. 22V shows another similar configuration wherein the assembly of two freeform optics is presented as a vertical stack. - To ensure that one surface is not interfering with another surface in the freeform optics, one may use wavelength or polarization selective reflector surfaces. For example, referring to
FIG. 22V , red, green, and blue wavelengths in the form of 650 nm, 530 nm, and 450 nm may be injected, as well as red, green, and blue wavelengths in the form of 620 nm, 550 nm, and 470 nm; different selective reflectors may be utilized in each of the freeform optics so that they do not interfere with each other. In a configuration wherein polarization filtering is used for a similar purpose, the reflection/transmission selectivity for light that is polarized in a particular axis may be varied (i.e., the images may be pre-polarized before they are sent to each freeform waveguide, to work with reflector selectivity). - Referring to
FIGS. 22W and 22X , configurations are illustrated wherein a plurality of freeform waveguides may be utilized together in series. Referring toFIG. 22W , light may enter from the real world and be directed sequentially through a first freeform optic (770), through an optional lens (812) which may be configured to relay light to a reflector (810) such as a DMD from a DLP system, which may be configured to reflect the light that has been filtered on a pixel by pixel basis (i.e., an occlusion mask may be utilized to block out certain elements of the real world, such as for darkfield perception, as described above; suitable spatial light modulators may be used which comprise DMDs, LCDs, ferroelectric LCOSs, MEMS shutter arrays, and the like, as described above) to another freeform optic (770) that is relaying light to the eye (28) of the user. Such a configuration may be more compact than one using conventional lenses for spatial light modulation. - Referring to
FIG. 22X , in a scenario wherein it is very important to keep overall thickness minimized, a configuration may be utilized that has one surface that is highly-reflective so that it may bounce light straight into another compactly positioned freeform optic. In one embodiment a selective attenuator (814) may be interposed between the two freeform optics (770). - Referring to
FIG. 22Y , an embodiment is depicted wherein a freeform optic (770) may comprise one aspect of a contact lens system. A miniaturized freeform optic is shown engaged against the cornea of a user's eye (58) with a miniaturized compensator lens portion (780), akin to that described in reference toFIG. 22F . Signals may be injected into the miniaturized freeform assembly using a tethered scanning fiber display which may, for example, be coupled between the freeform optic and a tear duct area of the user, or between the freeform optic and another head-mounted display configuration. - Various example embodiments of the invention are described herein. Reference is made to these examples in a non-limiting sense. They are provided to illustrate more broadly applicable aspects of the invention. Various changes may be made to the invention described and equivalents may be substituted without departing from the true spirit and scope of the invention. In addition, many modifications may be made to adapt a particular situation, material, composition of matter, process, process act(s) or step(s) to the objective(s), spirit or scope of the present invention. Further, as will be appreciated by those with skill in the art that each of the individual variations described and illustrated herein has discrete components and features which may be readily separated from or combined with the features of any of the other several embodiments without departing from the scope or spirit of the present inventions. All such modifications are intended to be within the scope of claims associated with this disclosure.
- The invention includes methods that may be performed using the subject devices. The methods may comprise the act of providing such a suitable device. Such provision may be performed by the end user. In other words, the “providing” act merely requires the end user obtain, access, approach, position, set-up, activate, power-up or otherwise act to provide the requisite device in the subject method. Methods recited herein may be carried out in any order of the recited events which is logically possible, as well as in the recited order of events.
- Example aspects of the invention, together with details regarding material selection and manufacture have been set forth above. As for other details of the present invention, these may be appreciated in connection with the above-referenced patents and publications as well as generally known or appreciated by those with skill in the art. The same may hold true with respect to method-based aspects of the invention in terms of additional acts as commonly or logically employed.
- In addition, though the invention has been described in reference to several examples optionally incorporating various features, the invention is not to be limited to that which is described or indicated as contemplated with respect to each variation of the invention. Various changes may be made to the invention described and equivalents (whether recited herein or not included for the sake of some brevity) may be substituted without departing from the true spirit and scope of the invention. In addition, where a range of values is provided, it is understood that every intervening value, between the upper and lower limit of that range and any other stated or intervening value in that stated range, is encompassed within the invention.
- Also, it is contemplated that any optional feature of the inventive variations described may be set forth and claimed independently, or in combination with any one or more of the features described herein. Reference to a singular item, includes the possibility that there are plural of the same items present. More specifically, as used herein and in claims associated hereto, the singular forms “a,” “an,” “said,” and “the” include plural referents unless the specifically stated otherwise. In other words, use of the articles allow for “at least one” of the subject item in the description above as well as claims associated with this disclosure. It is further noted that such claims may be drafted to exclude any optional element. As such, this statement is intended to serve as antecedent basis for use of such exclusive terminology as “solely,” “only” and the like in connection with the recitation of claim elements, or use of a “negative” limitation.
- Without the use of such exclusive terminology, the term “comprising” in claims associated with this disclosure shall allow for the inclusion of any additional element—irrespective of whether a given number of elements are enumerated in such claims, or the addition of a feature could be regarded as transforming the nature of an element set forth in such claims. Except as specifically defined herein, all technical and scientific terms used herein are to be given as broad a commonly understood meaning as possible while maintaining claim validity.
- The breadth of the present invention is not to be limited to the examples provided and/or the subject specification, but rather only by the scope of claim language associated with this disclosure.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/706,596 US20150241696A1 (en) | 2013-11-27 | 2015-05-07 | Inducing phase delays to vary an aggregate wavefront for augmented or virtual reality |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361909774P | 2013-11-27 | 2013-11-27 | |
US14/555,585 US9791700B2 (en) | 2013-11-27 | 2014-11-27 | Virtual and augmented reality systems and methods |
US14/706,596 US20150241696A1 (en) | 2013-11-27 | 2015-05-07 | Inducing phase delays to vary an aggregate wavefront for augmented or virtual reality |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/555,585 Continuation US9791700B2 (en) | 2013-11-27 | 2014-11-27 | Virtual and augmented reality systems and methods |
Publications (1)
Publication Number | Publication Date |
---|---|
US20150241696A1 true US20150241696A1 (en) | 2015-08-27 |
Family
ID=53199737
Family Applications (94)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/555,585 Active US9791700B2 (en) | 2013-11-27 | 2014-11-27 | Virtual and augmented reality systems and methods |
US14/703,168 Active US9846967B2 (en) | 2013-11-27 | 2015-05-04 | Varying a focus through a variable focus element based on user accommodation |
US14/704,519 Abandoned US20150235440A1 (en) | 2013-11-27 | 2015-05-05 | Providing augmented reality using microprojectors |
US14/704,784 Abandoned US20150235443A1 (en) | 2013-11-27 | 2015-05-05 | Selectively blurring a portion of an image based on a user's accommodation |
US14/704,832 Abandoned US20150235446A1 (en) | 2013-11-27 | 2015-05-05 | Driving sub-images based on a user's accommodation |
US14/704,782 Abandoned US20150235442A1 (en) | 2013-11-27 | 2015-05-05 | Using wedge-shaped waveguides for augmented or virtual reality |
US14/704,537 Abandoned US20150234476A1 (en) | 2013-11-27 | 2015-05-05 | Determining user accommodation to display an image through a waveguide assembly |
US14/704,275 Abandoned US20150235436A1 (en) | 2013-11-27 | 2015-05-05 | Delivering light rays associated with virtual images based on user accommodation |
US14/704,803 Abandoned US20150235444A1 (en) | 2013-11-27 | 2015-05-05 | Methods and system for using microprojects for augmented or virtual reality |
US14/704,662 Abandoned US20150234190A1 (en) | 2013-11-27 | 2015-05-05 | Using blurring to create multiple depth planes for augmented or virtual reality |
US14/704,827 Abandoned US20150235445A1 (en) | 2013-11-27 | 2015-05-05 | Modulating a depth of focus of a plurality of pixels displayed to a user |
US14/704,484 Abandoned US20150235439A1 (en) | 2013-11-27 | 2015-05-05 | Combining display elements having different frame rates and bit depths for augmented or virtual reality |
US14/704,438 Abandoned US20150235438A1 (en) | 2013-11-27 | 2015-05-05 | Using a display assembly for augmented or virtual reality |
US14/704,816 Abandoned US20150234205A1 (en) | 2013-11-27 | 2015-05-05 | Contact lens device for displaying augmented or virtual reality |
US14/704,321 Abandoned US20150235437A1 (en) | 2013-11-27 | 2015-05-05 | Determining user accommodation to display an image at a focal plane corresponding to a user's current state of focus |
US14/704,863 Abandoned US20150235448A1 (en) | 2013-11-27 | 2015-05-05 | Using multiple exit pupils to transmit light into a user's pupil for augmented or virtual reality |
US14/704,242 Abandoned US20150235418A1 (en) | 2013-11-27 | 2015-05-05 | Determining user accommodation to display an image at a desired focal distance using freeform optics |
US14/704,983 Active US9846306B2 (en) | 2013-11-27 | 2015-05-06 | Using a plurality of optical fibers for augmented or virtual reality display |
US14/705,648 Abandoned US20150243089A1 (en) | 2013-11-27 | 2015-05-06 | Varying pixel size based on line pitch for augmented or virtual reality |
US14/705,491 Abandoned US20150235461A1 (en) | 2013-11-27 | 2015-05-06 | Using an array of spatial light modulators to generate a lightfield |
US14/705,214 Abandoned US20150235457A1 (en) | 2013-11-27 | 2015-05-06 | Driving light patterns to exit pupils for augmented or virtual reality |
US14/705,237 Abandoned US20150235458A1 (en) | 2013-11-27 | 2015-05-06 | Waveguide assembly having reflective layers for augmented or virtual reality |
US14/705,568 Abandoned US20150235462A1 (en) | 2013-11-27 | 2015-05-06 | Generating a lightfield using a plurality of spatial light modulators |
US14/705,223 Abandoned US20150235421A1 (en) | 2013-11-27 | 2015-05-06 | USING MEMs LOUVERS TO CHANGE AN ANGLE OF LIGHT FOR AUGMENTED OR VIRTUAL REALITY |
US14/705,276 Abandoned US20150248046A1 (en) | 2013-11-27 | 2015-05-06 | Controlling diffractive optical elements for augmented or virtual reality |
US14/705,490 Abandoned US20150234254A1 (en) | 2013-11-27 | 2015-05-06 | Separately addressable diffractive optical elements for augmented or virtual reality |
US14/705,741 Active US9939643B2 (en) | 2013-11-27 | 2015-05-06 | Modifying a focus of virtual images through a variable focus element |
US14/705,769 Active US9804397B2 (en) | 2013-11-27 | 2015-05-06 | Using a freedom reflective and lens optical component for augmented or virtual reality display |
US14/704,990 Abandoned US20150243088A1 (en) | 2013-11-27 | 2015-05-06 | Using a variable focus element coupled to a waveguide to create multiple depth planes |
US14/705,184 Abandoned US20150235455A1 (en) | 2013-11-27 | 2015-05-06 | Using polarization modulators for augmented or virtual reality |
US14/705,667 Abandoned US20150235464A1 (en) | 2013-11-27 | 2015-05-06 | Coupling a lens to an optical fiber for augmented or virtual reality displays |
US14/705,804 Active US9915824B2 (en) | 2013-11-27 | 2015-05-06 | Combining at least one variable focus element with a plurality of stacked waveguides for augmented or virtual reality display |
US14/705,723 Abandoned US20150235466A1 (en) | 2013-11-27 | 2015-05-06 | Using optical fibers to deliver multiple depth planes for augmented or virtual reality |
US14/705,748 Active US9946071B2 (en) | 2013-11-27 | 2015-05-06 | Modifying light of a multicore assembly to produce a plurality of viewing zones |
US14/705,245 Abandoned US20150235459A1 (en) | 2013-11-27 | 2015-05-06 | Using an eye box for augmented or virtual reality |
US14/704,985 Abandoned US20150235454A1 (en) | 2013-11-27 | 2015-05-06 | Providing augmented or virtual reality using transmissive beamsplitters |
US14/704,987 Abandoned US20150235419A1 (en) | 2013-11-27 | 2015-05-06 | Methods and systems for displaying multiple depth planes through a variable focus element |
US14/705,255 Abandoned US20150235460A1 (en) | 2013-11-27 | 2015-05-06 | Diffractive optical elements used for augmented or virtual reality |
US14/705,603 Abandoned US20150235463A1 (en) | 2013-11-27 | 2015-05-06 | Modulating a size of pixels displayed to a user for augmented or virtual reality |
US14/705,197 Abandoned US20150235456A1 (en) | 2013-11-27 | 2015-05-06 | Modulating a polarization of light for augmented or virtual reality |
US14/705,867 Active US9841601B2 (en) | 2013-11-27 | 2015-05-06 | Delivering viewing zones associated with portions of an image for augmented or virtual reality |
US14/704,989 Abandoned US20150235420A1 (en) | 2013-11-27 | 2015-05-06 | Method for displaying multiple depth planes through variable focus elements |
US14/705,675 Abandoned US20150235465A1 (en) | 2013-11-27 | 2015-05-06 | Polishing an array of optical fibers at an angle to deliver augmented or virtual reality images |
US14/705,530 Abandoned US20150234191A1 (en) | 2013-11-27 | 2015-05-06 | Using freeform optical elements to display augmented or virtual reality |
US14/705,825 Abandoned US20150235467A1 (en) | 2013-11-27 | 2015-05-06 | Waveguide assembly to display images at multiple focal planes |
US14/706,783 Abandoned US20150241702A1 (en) | 2013-11-27 | 2015-05-07 | Lens array operatively coupled to a spatial light modulator for augmented or virtual reality |
US14/706,444 Abandoned US20150243092A1 (en) | 2013-11-27 | 2015-05-07 | Pixel size modulation for augmented or virtual reality |
US14/706,196 Abandoned US20150243090A1 (en) | 2013-11-27 | 2015-05-07 | Using polished microprojectors for augmented or virtual reality |
US14/706,551 Abandoned US20150243094A1 (en) | 2013-11-27 | 2015-05-07 | Producing an aggregate wavefront for augmented or virtual reality |
US14/706,763 Abandoned US20150241701A1 (en) | 2013-11-27 | 2015-05-07 | Pinhole array operatively coupled to a spatial light modulator for augmented or virtual reality |
US14/706,232 Abandoned US20150235469A1 (en) | 2013-11-27 | 2015-05-07 | Determining user accommodation to project image data at a desired focal distance |
US14/706,734 Abandoned US20150243097A1 (en) | 2013-11-27 | 2015-05-07 | Selective attenuation of outside light in an augmented or virtual reality device |
US14/706,739 Abandoned US20150243098A1 (en) | 2013-11-27 | 2015-05-07 | Using an array of spatial light modulators for selective attenuation |
US14/706,428 Abandoned US20150243091A1 (en) | 2013-11-27 | 2015-05-07 | Coupling phase modulators to optical fibers for augmented or virtual reality |
US14/706,216 Abandoned US20150235468A1 (en) | 2013-11-27 | 2015-05-07 | Coupling optical elements to an array of microprojectors for augmented or virtual reality |
US14/706,241 Abandoned US20150235470A1 (en) | 2013-11-27 | 2015-05-07 | Coupling a plurality of multicore assemblies polished at an angle for augmented or virtual reality |
US14/706,358 Abandoned US20150235472A1 (en) | 2013-11-27 | 2015-05-07 | Delivering light beams at a plurality of angles for augmented or virtual reality |
US14/706,398 Abandoned US20150235473A1 (en) | 2013-11-27 | 2015-05-07 | Displaying augmented reality or virtual reality through a substrate coupled to the user's eye |
US14/706,635 Abandoned US20150241698A1 (en) | 2013-11-27 | 2015-05-07 | Methods and systems to use multicore fibers for augmented or virtual reality |
US14/706,808 Abandoned US20150241703A1 (en) | 2013-11-27 | 2015-05-07 | Using spatial light modulators to selectively attenuate light from an outside environment for augmented or virtual reality |
US14/706,658 Abandoned US20150243096A1 (en) | 2013-11-27 | 2015-05-07 | Using a fiber scanning display to present a lightfield to a user |
US14/706,840 Abandoned US20150319342A1 (en) | 2013-11-27 | 2015-05-07 | Using a halo to facilitate viewing dark virtual objects in augmented or virtual reality |
US14/706,625 Abandoned US20150241697A1 (en) | 2013-11-27 | 2015-05-07 | Physical actuators coupled to optical fiber cores for augmented or virtual reality |
US14/706,842 Abandoned US20150243099A1 (en) | 2013-11-27 | 2015-05-07 | Rendering a halo around virtual objects for displaying augmented or virtual reality |
US14/706,813 Abandoned US20160109708A1 (en) | 2013-11-27 | 2015-05-07 | Projecting images to a waveguide through microprojectors for augmented or virtual reality |
US14/706,507 Abandoned US20150243093A1 (en) | 2013-11-27 | 2015-05-07 | Determining user accommodation to display an image at a desired focal plane using diffractive optical elements |
US14/706,681 Abandoned US20150241699A1 (en) | 2013-11-27 | 2015-05-07 | Selectively attenuating light from the outside world for augmented or virtual reality |
US14/706,278 Abandoned US20150235471A1 (en) | 2013-11-27 | 2015-05-07 | Delivering light beams through optical fiber cores at a plurality of angles for augmented or virtual reality |
US14/706,586 Abandoned US20150243095A1 (en) | 2013-11-27 | 2015-05-07 | Modulating light associated with image data through phase modulators for augmented or virtual reality |
US14/706,690 Abandoned US20150241700A1 (en) | 2013-11-27 | 2015-05-07 | Attenuating outside light for augmented or virtual reality |
US14/706,596 Abandoned US20150241696A1 (en) | 2013-11-27 | 2015-05-07 | Inducing phase delays to vary an aggregate wavefront for augmented or virtual reality |
US14/707,281 Abandoned US20150248158A1 (en) | 2013-11-27 | 2015-05-08 | Curved waveguides for augmented or virtual reality |
US14/707,813 Abandoned US20150241707A1 (en) | 2013-11-27 | 2015-05-08 | Modifying light using freeform optics for augmented or virtual reality |
US14/707,432 Abandoned US20150241706A1 (en) | 2013-11-27 | 2015-05-08 | Injecting images having an inverse fourier transform to produce a desired wavefront |
US14/707,224 Abandoned US20150248786A1 (en) | 2013-11-27 | 2015-05-08 | Modulating light intensity to enable viewing of dark virtual objects |
US14/707,296 Abandoned US20150243104A1 (en) | 2013-11-27 | 2015-05-08 | Delivering virtual image slices at different depth planes for augmented or virtual reality |
US14/707,302 Abandoned US20150248006A1 (en) | 2013-11-27 | 2015-05-08 | Circular diffractive optical elements for augmented or virtual reality |
US14/707,002 Abandoned US20150243101A1 (en) | 2013-11-27 | 2015-05-08 | Modifying a curvature of light rays to produce multiple depth planes |
US14/707,337 Abandoned US20150248010A1 (en) | 2013-11-27 | 2015-05-08 | Inducing phase delays in a multicore assembly for augmented or virtual reality |
US14/707,332 Abandoned US20150248790A1 (en) | 2013-11-27 | 2015-05-08 | Using circularly-symmetric diffractive optical elements for augmented or virtual reality |
US14/707,779 Abandoned US20150309315A1 (en) | 2013-11-27 | 2015-05-08 | Using freeform optics for augmented or virtual reality |
US14/707,247 Abandoned US20150243102A1 (en) | 2013-11-27 | 2015-05-08 | Rendering visual emphasis proximate to virtual objects for augmented or virtual reality |
US14/707,236 Abandoned US20150248011A1 (en) | 2013-11-27 | 2015-05-08 | Delivering virtual images of different portions of the user's pupil for augmented or virtual reality |
US14/707,265 Abandoned US20150243103A1 (en) | 2013-11-27 | 2015-05-08 | Rendering dark virtual objects as blue to facilitate viewing augmented or virtual reality |
US14/707,519 Abandoned US20150248012A1 (en) | 2013-11-27 | 2015-05-08 | Stacked configuration of freeform optics for augmented or virtual reality |
US14/707,735 Abandoned US20150243107A1 (en) | 2013-11-27 | 2015-05-08 | Displaying augmented or virtual reality through freeform optics |
US14/707,257 Abandoned US20150241704A1 (en) | 2013-11-27 | 2015-05-08 | Using a plurality of waveguides coupled with edge reflectors for augmented or virtual reality |
US15/729,494 Active 2034-12-08 US10643392B2 (en) | 2013-11-27 | 2017-10-10 | Virtual and augmented reality systems and methods |
US15/729,462 Active US10629004B2 (en) | 2013-11-27 | 2017-10-10 | Virtual and augmented reality systems and methods |
US15/970,552 Active 2034-12-25 US10529138B2 (en) | 2013-11-27 | 2018-05-03 | Virtual and augmented reality systems and methods |
US16/814,975 Active US10935806B2 (en) | 2013-11-27 | 2020-03-10 | Virtual and augmented reality systems and methods |
US17/155,412 Active US11237403B2 (en) | 2013-11-27 | 2021-01-22 | Virtual and augmented reality systems and methods |
US17/555,640 Active US11714291B2 (en) | 2013-11-27 | 2021-12-20 | Virtual and augmented reality systems and methods |
US18/329,982 Pending US20230324706A1 (en) | 2013-11-27 | 2023-06-06 | Virtual and augmented reality systems and methods |
Family Applications Before (70)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/555,585 Active US9791700B2 (en) | 2013-11-27 | 2014-11-27 | Virtual and augmented reality systems and methods |
US14/703,168 Active US9846967B2 (en) | 2013-11-27 | 2015-05-04 | Varying a focus through a variable focus element based on user accommodation |
US14/704,519 Abandoned US20150235440A1 (en) | 2013-11-27 | 2015-05-05 | Providing augmented reality using microprojectors |
US14/704,784 Abandoned US20150235443A1 (en) | 2013-11-27 | 2015-05-05 | Selectively blurring a portion of an image based on a user's accommodation |
US14/704,832 Abandoned US20150235446A1 (en) | 2013-11-27 | 2015-05-05 | Driving sub-images based on a user's accommodation |
US14/704,782 Abandoned US20150235442A1 (en) | 2013-11-27 | 2015-05-05 | Using wedge-shaped waveguides for augmented or virtual reality |
US14/704,537 Abandoned US20150234476A1 (en) | 2013-11-27 | 2015-05-05 | Determining user accommodation to display an image through a waveguide assembly |
US14/704,275 Abandoned US20150235436A1 (en) | 2013-11-27 | 2015-05-05 | Delivering light rays associated with virtual images based on user accommodation |
US14/704,803 Abandoned US20150235444A1 (en) | 2013-11-27 | 2015-05-05 | Methods and system for using microprojects for augmented or virtual reality |
US14/704,662 Abandoned US20150234190A1 (en) | 2013-11-27 | 2015-05-05 | Using blurring to create multiple depth planes for augmented or virtual reality |
US14/704,827 Abandoned US20150235445A1 (en) | 2013-11-27 | 2015-05-05 | Modulating a depth of focus of a plurality of pixels displayed to a user |
US14/704,484 Abandoned US20150235439A1 (en) | 2013-11-27 | 2015-05-05 | Combining display elements having different frame rates and bit depths for augmented or virtual reality |
US14/704,438 Abandoned US20150235438A1 (en) | 2013-11-27 | 2015-05-05 | Using a display assembly for augmented or virtual reality |
US14/704,816 Abandoned US20150234205A1 (en) | 2013-11-27 | 2015-05-05 | Contact lens device for displaying augmented or virtual reality |
US14/704,321 Abandoned US20150235437A1 (en) | 2013-11-27 | 2015-05-05 | Determining user accommodation to display an image at a focal plane corresponding to a user's current state of focus |
US14/704,863 Abandoned US20150235448A1 (en) | 2013-11-27 | 2015-05-05 | Using multiple exit pupils to transmit light into a user's pupil for augmented or virtual reality |
US14/704,242 Abandoned US20150235418A1 (en) | 2013-11-27 | 2015-05-05 | Determining user accommodation to display an image at a desired focal distance using freeform optics |
US14/704,983 Active US9846306B2 (en) | 2013-11-27 | 2015-05-06 | Using a plurality of optical fibers for augmented or virtual reality display |
US14/705,648 Abandoned US20150243089A1 (en) | 2013-11-27 | 2015-05-06 | Varying pixel size based on line pitch for augmented or virtual reality |
US14/705,491 Abandoned US20150235461A1 (en) | 2013-11-27 | 2015-05-06 | Using an array of spatial light modulators to generate a lightfield |
US14/705,214 Abandoned US20150235457A1 (en) | 2013-11-27 | 2015-05-06 | Driving light patterns to exit pupils for augmented or virtual reality |
US14/705,237 Abandoned US20150235458A1 (en) | 2013-11-27 | 2015-05-06 | Waveguide assembly having reflective layers for augmented or virtual reality |
US14/705,568 Abandoned US20150235462A1 (en) | 2013-11-27 | 2015-05-06 | Generating a lightfield using a plurality of spatial light modulators |
US14/705,223 Abandoned US20150235421A1 (en) | 2013-11-27 | 2015-05-06 | USING MEMs LOUVERS TO CHANGE AN ANGLE OF LIGHT FOR AUGMENTED OR VIRTUAL REALITY |
US14/705,276 Abandoned US20150248046A1 (en) | 2013-11-27 | 2015-05-06 | Controlling diffractive optical elements for augmented or virtual reality |
US14/705,490 Abandoned US20150234254A1 (en) | 2013-11-27 | 2015-05-06 | Separately addressable diffractive optical elements for augmented or virtual reality |
US14/705,741 Active US9939643B2 (en) | 2013-11-27 | 2015-05-06 | Modifying a focus of virtual images through a variable focus element |
US14/705,769 Active US9804397B2 (en) | 2013-11-27 | 2015-05-06 | Using a freedom reflective and lens optical component for augmented or virtual reality display |
US14/704,990 Abandoned US20150243088A1 (en) | 2013-11-27 | 2015-05-06 | Using a variable focus element coupled to a waveguide to create multiple depth planes |
US14/705,184 Abandoned US20150235455A1 (en) | 2013-11-27 | 2015-05-06 | Using polarization modulators for augmented or virtual reality |
US14/705,667 Abandoned US20150235464A1 (en) | 2013-11-27 | 2015-05-06 | Coupling a lens to an optical fiber for augmented or virtual reality displays |
US14/705,804 Active US9915824B2 (en) | 2013-11-27 | 2015-05-06 | Combining at least one variable focus element with a plurality of stacked waveguides for augmented or virtual reality display |
US14/705,723 Abandoned US20150235466A1 (en) | 2013-11-27 | 2015-05-06 | Using optical fibers to deliver multiple depth planes for augmented or virtual reality |
US14/705,748 Active US9946071B2 (en) | 2013-11-27 | 2015-05-06 | Modifying light of a multicore assembly to produce a plurality of viewing zones |
US14/705,245 Abandoned US20150235459A1 (en) | 2013-11-27 | 2015-05-06 | Using an eye box for augmented or virtual reality |
US14/704,985 Abandoned US20150235454A1 (en) | 2013-11-27 | 2015-05-06 | Providing augmented or virtual reality using transmissive beamsplitters |
US14/704,987 Abandoned US20150235419A1 (en) | 2013-11-27 | 2015-05-06 | Methods and systems for displaying multiple depth planes through a variable focus element |
US14/705,255 Abandoned US20150235460A1 (en) | 2013-11-27 | 2015-05-06 | Diffractive optical elements used for augmented or virtual reality |
US14/705,603 Abandoned US20150235463A1 (en) | 2013-11-27 | 2015-05-06 | Modulating a size of pixels displayed to a user for augmented or virtual reality |
US14/705,197 Abandoned US20150235456A1 (en) | 2013-11-27 | 2015-05-06 | Modulating a polarization of light for augmented or virtual reality |
US14/705,867 Active US9841601B2 (en) | 2013-11-27 | 2015-05-06 | Delivering viewing zones associated with portions of an image for augmented or virtual reality |
US14/704,989 Abandoned US20150235420A1 (en) | 2013-11-27 | 2015-05-06 | Method for displaying multiple depth planes through variable focus elements |
US14/705,675 Abandoned US20150235465A1 (en) | 2013-11-27 | 2015-05-06 | Polishing an array of optical fibers at an angle to deliver augmented or virtual reality images |
US14/705,530 Abandoned US20150234191A1 (en) | 2013-11-27 | 2015-05-06 | Using freeform optical elements to display augmented or virtual reality |
US14/705,825 Abandoned US20150235467A1 (en) | 2013-11-27 | 2015-05-06 | Waveguide assembly to display images at multiple focal planes |
US14/706,783 Abandoned US20150241702A1 (en) | 2013-11-27 | 2015-05-07 | Lens array operatively coupled to a spatial light modulator for augmented or virtual reality |
US14/706,444 Abandoned US20150243092A1 (en) | 2013-11-27 | 2015-05-07 | Pixel size modulation for augmented or virtual reality |
US14/706,196 Abandoned US20150243090A1 (en) | 2013-11-27 | 2015-05-07 | Using polished microprojectors for augmented or virtual reality |
US14/706,551 Abandoned US20150243094A1 (en) | 2013-11-27 | 2015-05-07 | Producing an aggregate wavefront for augmented or virtual reality |
US14/706,763 Abandoned US20150241701A1 (en) | 2013-11-27 | 2015-05-07 | Pinhole array operatively coupled to a spatial light modulator for augmented or virtual reality |
US14/706,232 Abandoned US20150235469A1 (en) | 2013-11-27 | 2015-05-07 | Determining user accommodation to project image data at a desired focal distance |
US14/706,734 Abandoned US20150243097A1 (en) | 2013-11-27 | 2015-05-07 | Selective attenuation of outside light in an augmented or virtual reality device |
US14/706,739 Abandoned US20150243098A1 (en) | 2013-11-27 | 2015-05-07 | Using an array of spatial light modulators for selective attenuation |
US14/706,428 Abandoned US20150243091A1 (en) | 2013-11-27 | 2015-05-07 | Coupling phase modulators to optical fibers for augmented or virtual reality |
US14/706,216 Abandoned US20150235468A1 (en) | 2013-11-27 | 2015-05-07 | Coupling optical elements to an array of microprojectors for augmented or virtual reality |
US14/706,241 Abandoned US20150235470A1 (en) | 2013-11-27 | 2015-05-07 | Coupling a plurality of multicore assemblies polished at an angle for augmented or virtual reality |
US14/706,358 Abandoned US20150235472A1 (en) | 2013-11-27 | 2015-05-07 | Delivering light beams at a plurality of angles for augmented or virtual reality |
US14/706,398 Abandoned US20150235473A1 (en) | 2013-11-27 | 2015-05-07 | Displaying augmented reality or virtual reality through a substrate coupled to the user's eye |
US14/706,635 Abandoned US20150241698A1 (en) | 2013-11-27 | 2015-05-07 | Methods and systems to use multicore fibers for augmented or virtual reality |
US14/706,808 Abandoned US20150241703A1 (en) | 2013-11-27 | 2015-05-07 | Using spatial light modulators to selectively attenuate light from an outside environment for augmented or virtual reality |
US14/706,658 Abandoned US20150243096A1 (en) | 2013-11-27 | 2015-05-07 | Using a fiber scanning display to present a lightfield to a user |
US14/706,840 Abandoned US20150319342A1 (en) | 2013-11-27 | 2015-05-07 | Using a halo to facilitate viewing dark virtual objects in augmented or virtual reality |
US14/706,625 Abandoned US20150241697A1 (en) | 2013-11-27 | 2015-05-07 | Physical actuators coupled to optical fiber cores for augmented or virtual reality |
US14/706,842 Abandoned US20150243099A1 (en) | 2013-11-27 | 2015-05-07 | Rendering a halo around virtual objects for displaying augmented or virtual reality |
US14/706,813 Abandoned US20160109708A1 (en) | 2013-11-27 | 2015-05-07 | Projecting images to a waveguide through microprojectors for augmented or virtual reality |
US14/706,507 Abandoned US20150243093A1 (en) | 2013-11-27 | 2015-05-07 | Determining user accommodation to display an image at a desired focal plane using diffractive optical elements |
US14/706,681 Abandoned US20150241699A1 (en) | 2013-11-27 | 2015-05-07 | Selectively attenuating light from the outside world for augmented or virtual reality |
US14/706,278 Abandoned US20150235471A1 (en) | 2013-11-27 | 2015-05-07 | Delivering light beams through optical fiber cores at a plurality of angles for augmented or virtual reality |
US14/706,586 Abandoned US20150243095A1 (en) | 2013-11-27 | 2015-05-07 | Modulating light associated with image data through phase modulators for augmented or virtual reality |
US14/706,690 Abandoned US20150241700A1 (en) | 2013-11-27 | 2015-05-07 | Attenuating outside light for augmented or virtual reality |
Family Applications After (23)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/707,281 Abandoned US20150248158A1 (en) | 2013-11-27 | 2015-05-08 | Curved waveguides for augmented or virtual reality |
US14/707,813 Abandoned US20150241707A1 (en) | 2013-11-27 | 2015-05-08 | Modifying light using freeform optics for augmented or virtual reality |
US14/707,432 Abandoned US20150241706A1 (en) | 2013-11-27 | 2015-05-08 | Injecting images having an inverse fourier transform to produce a desired wavefront |
US14/707,224 Abandoned US20150248786A1 (en) | 2013-11-27 | 2015-05-08 | Modulating light intensity to enable viewing of dark virtual objects |
US14/707,296 Abandoned US20150243104A1 (en) | 2013-11-27 | 2015-05-08 | Delivering virtual image slices at different depth planes for augmented or virtual reality |
US14/707,302 Abandoned US20150248006A1 (en) | 2013-11-27 | 2015-05-08 | Circular diffractive optical elements for augmented or virtual reality |
US14/707,002 Abandoned US20150243101A1 (en) | 2013-11-27 | 2015-05-08 | Modifying a curvature of light rays to produce multiple depth planes |
US14/707,337 Abandoned US20150248010A1 (en) | 2013-11-27 | 2015-05-08 | Inducing phase delays in a multicore assembly for augmented or virtual reality |
US14/707,332 Abandoned US20150248790A1 (en) | 2013-11-27 | 2015-05-08 | Using circularly-symmetric diffractive optical elements for augmented or virtual reality |
US14/707,779 Abandoned US20150309315A1 (en) | 2013-11-27 | 2015-05-08 | Using freeform optics for augmented or virtual reality |
US14/707,247 Abandoned US20150243102A1 (en) | 2013-11-27 | 2015-05-08 | Rendering visual emphasis proximate to virtual objects for augmented or virtual reality |
US14/707,236 Abandoned US20150248011A1 (en) | 2013-11-27 | 2015-05-08 | Delivering virtual images of different portions of the user's pupil for augmented or virtual reality |
US14/707,265 Abandoned US20150243103A1 (en) | 2013-11-27 | 2015-05-08 | Rendering dark virtual objects as blue to facilitate viewing augmented or virtual reality |
US14/707,519 Abandoned US20150248012A1 (en) | 2013-11-27 | 2015-05-08 | Stacked configuration of freeform optics for augmented or virtual reality |
US14/707,735 Abandoned US20150243107A1 (en) | 2013-11-27 | 2015-05-08 | Displaying augmented or virtual reality through freeform optics |
US14/707,257 Abandoned US20150241704A1 (en) | 2013-11-27 | 2015-05-08 | Using a plurality of waveguides coupled with edge reflectors for augmented or virtual reality |
US15/729,494 Active 2034-12-08 US10643392B2 (en) | 2013-11-27 | 2017-10-10 | Virtual and augmented reality systems and methods |
US15/729,462 Active US10629004B2 (en) | 2013-11-27 | 2017-10-10 | Virtual and augmented reality systems and methods |
US15/970,552 Active 2034-12-25 US10529138B2 (en) | 2013-11-27 | 2018-05-03 | Virtual and augmented reality systems and methods |
US16/814,975 Active US10935806B2 (en) | 2013-11-27 | 2020-03-10 | Virtual and augmented reality systems and methods |
US17/155,412 Active US11237403B2 (en) | 2013-11-27 | 2021-01-22 | Virtual and augmented reality systems and methods |
US17/555,640 Active US11714291B2 (en) | 2013-11-27 | 2021-12-20 | Virtual and augmented reality systems and methods |
US18/329,982 Pending US20230324706A1 (en) | 2013-11-27 | 2023-06-06 | Virtual and augmented reality systems and methods |
Country Status (10)
Country | Link |
---|---|
US (94) | US9791700B2 (en) |
EP (2) | EP4220999A3 (en) |
JP (18) | JP2017500605A (en) |
KR (5) | KR102493498B1 (en) |
CN (15) | CN107329260B (en) |
AU (14) | AU2014354673B2 (en) |
CA (1) | CA2931776A1 (en) |
IL (12) | IL305162B1 (en) |
NZ (2) | NZ755272A (en) |
WO (1) | WO2015081313A2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10338400B2 (en) | 2017-07-03 | 2019-07-02 | Holovisions LLC | Augmented reality eyewear with VAPE or wear technology |
US10859834B2 (en) | 2017-07-03 | 2020-12-08 | Holovisions | Space-efficient optical structures for wide field-of-view augmented reality (AR) eyewear |
Families Citing this family (1108)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
IL166799A (en) | 2005-02-10 | 2014-09-30 | Lumus Ltd | Substrate-guided optical device utilizing beam splitters |
US10073264B2 (en) | 2007-08-03 | 2018-09-11 | Lumus Ltd. | Substrate-guide optical device |
US20070081123A1 (en) * | 2005-10-07 | 2007-04-12 | Lewis Scott W | Digital eyewear |
US10048499B2 (en) | 2005-11-08 | 2018-08-14 | Lumus Ltd. | Polarizing optical system |
GB0522968D0 (en) | 2005-11-11 | 2005-12-21 | Popovich Milan M | Holographic illumination device |
GB0718706D0 (en) | 2007-09-25 | 2007-11-07 | Creative Physics Ltd | Method and apparatus for reducing laser speckle |
US9823737B2 (en) * | 2008-04-07 | 2017-11-21 | Mohammad A Mazed | Augmented reality personal assistant apparatus |
US9865043B2 (en) | 2008-03-26 | 2018-01-09 | Ricoh Company, Ltd. | Adaptive image acquisition and display using multi-focal display |
US9866826B2 (en) * | 2014-11-25 | 2018-01-09 | Ricoh Company, Ltd. | Content-adaptive multi-focal display |
US9229233B2 (en) | 2014-02-11 | 2016-01-05 | Osterhout Group, Inc. | Micro Doppler presentations in head worn computing |
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US20150205111A1 (en) | 2014-01-21 | 2015-07-23 | Osterhout Group, Inc. | Optical configurations for head worn computing |
US9298007B2 (en) | 2014-01-21 | 2016-03-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9952664B2 (en) | 2014-01-21 | 2018-04-24 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9715112B2 (en) | 2014-01-21 | 2017-07-25 | Osterhout Group, Inc. | Suppression of stray light in head worn computing |
US9400390B2 (en) | 2014-01-24 | 2016-07-26 | Osterhout Group, Inc. | Peripheral lighting for head worn computing |
US9335604B2 (en) | 2013-12-11 | 2016-05-10 | Milan Momcilo Popovich | Holographic waveguide display |
US11726332B2 (en) | 2009-04-27 | 2023-08-15 | Digilens Inc. | Diffractive projection apparatus |
US10156722B2 (en) | 2010-12-24 | 2018-12-18 | Magic Leap, Inc. | Methods and systems for displaying stereoscopy with a freeform optical system with addressable focus for virtual and augmented reality |
US9274349B2 (en) | 2011-04-07 | 2016-03-01 | Digilens Inc. | Laser despeckler based on angular diversity |
US20140204455A1 (en) | 2011-08-24 | 2014-07-24 | Milan Momcilo Popovich | Wearable data display |
US10670876B2 (en) | 2011-08-24 | 2020-06-02 | Digilens Inc. | Waveguide laser illuminator incorporating a despeckler |
WO2016020630A2 (en) | 2014-08-08 | 2016-02-11 | Milan Momcilo Popovich | Waveguide laser illuminator incorporating a despeckler |
US20150010265A1 (en) | 2012-01-06 | 2015-01-08 | Milan, Momcilo POPOVICH | Contact image sensor using switchable bragg gratings |
CN106125308B (en) | 2012-04-25 | 2019-10-25 | 罗克韦尔柯林斯公司 | Device and method for displaying images |
WO2013167864A1 (en) | 2012-05-11 | 2013-11-14 | Milan Momcilo Popovich | Apparatus for eye tracking |
IL219907A (en) | 2012-05-21 | 2017-08-31 | Lumus Ltd | Head-mounted display eyeball tracker integrated system |
US9933684B2 (en) * | 2012-11-16 | 2018-04-03 | Rockwell Collins, Inc. | Transparent waveguide display providing upper and lower fields of view having a specific light output aperture configuration |
US10528135B2 (en) | 2013-01-14 | 2020-01-07 | Ctrl-Labs Corporation | Wearable muscle interface systems, devices and methods that interact with content displayed on an electronic display |
US9858721B2 (en) | 2013-01-15 | 2018-01-02 | The University Of North Carolina At Chapel Hill | Methods, systems, and computer readable media for generating an augmented scene display |
US9699433B2 (en) * | 2013-01-24 | 2017-07-04 | Yuchen Zhou | Method and apparatus to produce re-focusable vision with detecting re-focusing event from human eye |
US11504051B2 (en) | 2013-01-25 | 2022-11-22 | Wesley W. O. Krueger | Systems and methods for observing eye and head information to measure ocular parameters and determine human health status |
US10231614B2 (en) | 2014-07-08 | 2019-03-19 | Wesley W. O. Krueger | Systems and methods for using virtual reality, augmented reality, and/or a synthetic 3-dimensional information for the measurement of human ocular performance |
US10602927B2 (en) | 2013-01-25 | 2020-03-31 | Wesley W. O. Krueger | Ocular-performance-based head impact measurement using a faceguard |
US9788714B2 (en) | 2014-07-08 | 2017-10-17 | Iarmourholdings, Inc. | Systems and methods using virtual reality or augmented reality environments for the measurement and/or improvement of human vestibulo-ocular performance |
US11490809B2 (en) | 2013-01-25 | 2022-11-08 | Wesley W. O. Krueger | Ocular parameter-based head impact measurement using a face shield |
US12133567B2 (en) | 2013-01-25 | 2024-11-05 | Wesley W. O. Krueger | Systems and methods for using eye imaging on face protection equipment to assess human health |
US12042294B2 (en) | 2013-01-25 | 2024-07-23 | Wesley W. O. Krueger | Systems and methods to measure ocular parameters and determine neurologic health status |
US10716469B2 (en) | 2013-01-25 | 2020-07-21 | Wesley W. O. Krueger | Ocular-performance-based head impact measurement applied to rotationally-centered impact mitigation systems and methods |
US11389059B2 (en) | 2013-01-25 | 2022-07-19 | Wesley W. O. Krueger | Ocular-performance-based head impact measurement using a faceguard |
US11714487B2 (en) | 2013-03-04 | 2023-08-01 | Tobii Ab | Gaze and smooth pursuit based continuous foveal adjustment |
US10895908B2 (en) | 2013-03-04 | 2021-01-19 | Tobii Ab | Targeting saccade landing prediction using visual history |
US10082870B2 (en) | 2013-03-04 | 2018-09-25 | Tobii Ab | Gaze and saccade based graphical manipulation |
US9898081B2 (en) | 2013-03-04 | 2018-02-20 | Tobii Ab | Gaze and saccade based graphical manipulation |
US9665171B1 (en) | 2013-03-04 | 2017-05-30 | Tobii Ab | Gaze and saccade based graphical manipulation |
WO2014188149A1 (en) | 2013-05-20 | 2014-11-27 | Milan Momcilo Popovich | Holographic waveguide eye tracker |
US9977256B2 (en) * | 2013-05-30 | 2018-05-22 | Johnson & Johnson Vision Care, Inc. | Methods for manufacturing and programming an energizable ophthalmic lens with a programmable media insert |
US9874749B2 (en) | 2013-11-27 | 2018-01-23 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
WO2015012024A1 (en) * | 2013-07-26 | 2015-01-29 | シチズンホールディングス株式会社 | Light source device and projection device |
WO2015015138A1 (en) | 2013-07-31 | 2015-02-05 | Milan Momcilo Popovich | Method and apparatus for contact image sensing |
US11921471B2 (en) | 2013-08-16 | 2024-03-05 | Meta Platforms Technologies, Llc | Systems, articles, and methods for wearable devices having secondary power sources in links of a band for providing secondary power in addition to a primary power source |
US20150124566A1 (en) | 2013-10-04 | 2015-05-07 | Thalmic Labs Inc. | Systems, articles and methods for wearable electronic devices employing contact sensors |
US10042422B2 (en) | 2013-11-12 | 2018-08-07 | Thalmic Labs Inc. | Systems, articles, and methods for capacitive electromyography sensors |
US9143880B2 (en) | 2013-08-23 | 2015-09-22 | Tobii Ab | Systems and methods for providing audio to a user based on gaze input |
US10430150B2 (en) | 2013-08-23 | 2019-10-01 | Tobii Ab | Systems and methods for changing behavior of computer program elements based on gaze input |
US9665172B2 (en) | 2013-09-03 | 2017-05-30 | Tobii Ab | Portable eye tracking device |
US10310597B2 (en) | 2013-09-03 | 2019-06-04 | Tobii Ab | Portable eye tracking device |
US10686972B2 (en) | 2013-09-03 | 2020-06-16 | Tobii Ab | Gaze assisted field of view control |
US9791700B2 (en) | 2013-11-27 | 2017-10-17 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
WO2015081113A1 (en) | 2013-11-27 | 2015-06-04 | Cezar Morun | Systems, articles, and methods for electromyography sensors |
US9915826B2 (en) | 2013-11-27 | 2018-03-13 | Magic Leap, Inc. | Virtual and augmented reality systems and methods having improved diffractive grating structures |
US10677969B2 (en) | 2013-11-27 | 2020-06-09 | Magic Leap, Inc. | Manufacturing for virtual and augmented reality systems and components |
US9857591B2 (en) | 2014-05-30 | 2018-01-02 | Magic Leap, Inc. | Methods and system for creating focal planes in virtual and augmented reality |
US10274731B2 (en) | 2013-12-19 | 2019-04-30 | The University Of North Carolina At Chapel Hill | Optical see-through near-eye display using point light source backlight |
EP3087312B1 (en) * | 2013-12-23 | 2024-09-11 | Essilor International | Head-mounted display with filter function |
US9465237B2 (en) | 2013-12-27 | 2016-10-11 | Intel Corporation | Automatic focus prescription lens eyeglasses |
US20150187115A1 (en) * | 2013-12-27 | 2015-07-02 | Mark A. MacDonald | Dynamically adjustable 3d goggles |
US9841598B2 (en) | 2013-12-31 | 2017-12-12 | 3M Innovative Properties Company | Lens with embedded multilayer optical film for near-eye display systems |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US9299194B2 (en) | 2014-02-14 | 2016-03-29 | Osterhout Group, Inc. | Secure sharing in head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US20150228119A1 (en) | 2014-02-11 | 2015-08-13 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9529195B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | See-through computer display systems |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US9448409B2 (en) | 2014-11-26 | 2016-09-20 | Osterhout Group, Inc. | See-through computer display systems |
US10191279B2 (en) | 2014-03-17 | 2019-01-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US20160019715A1 (en) | 2014-07-15 | 2016-01-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US20150277118A1 (en) | 2014-03-28 | 2015-10-01 | Osterhout Group, Inc. | Sensor dependent content position in head worn computing |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9671613B2 (en) | 2014-09-26 | 2017-06-06 | Osterhout Group, Inc. | See-through computer display systems |
US9594246B2 (en) | 2014-01-21 | 2017-03-14 | Osterhout Group, Inc. | See-through computer display systems |
US9575321B2 (en) | 2014-06-09 | 2017-02-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US20160048019A1 (en) * | 2014-08-12 | 2016-02-18 | Osterhout Group, Inc. | Content presentation in head worn computing |
US9810906B2 (en) | 2014-06-17 | 2017-11-07 | Osterhout Group, Inc. | External user interface for head worn computing |
US9746686B2 (en) | 2014-05-19 | 2017-08-29 | Osterhout Group, Inc. | Content position calibration in head worn computing |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US11227294B2 (en) | 2014-04-03 | 2022-01-18 | Mentor Acquisition One, Llc | Sight information collection in head worn computing |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US9746676B2 (en) | 2014-01-21 | 2017-08-29 | Osterhout Group, Inc. | See-through computer display systems |
US9811159B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US9494800B2 (en) | 2014-01-21 | 2016-11-15 | Osterhout Group, Inc. | See-through computer display systems |
US9532714B2 (en) | 2014-01-21 | 2017-01-03 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9811153B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11892644B2 (en) | 2014-01-21 | 2024-02-06 | Mentor Acquisition One, Llc | See-through computer display systems |
US20150205135A1 (en) | 2014-01-21 | 2015-07-23 | Osterhout Group, Inc. | See-through computer display systems |
US9651784B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US12093453B2 (en) | 2014-01-21 | 2024-09-17 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11737666B2 (en) | 2014-01-21 | 2023-08-29 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US9846308B2 (en) | 2014-01-24 | 2017-12-19 | Osterhout Group, Inc. | Haptic systems for head-worn computers |
CN106233189B (en) | 2014-01-31 | 2020-06-26 | 奇跃公司 | Multi-focus display system and method |
US10317690B2 (en) | 2014-01-31 | 2019-06-11 | Magic Leap, Inc. | Multi-focal display system and method |
US9401540B2 (en) | 2014-02-11 | 2016-07-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US12112089B2 (en) | 2014-02-11 | 2024-10-08 | Mentor Acquisition One, Llc | Spatial location presentation in head worn computing |
US9852545B2 (en) | 2014-02-11 | 2017-12-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US10430985B2 (en) | 2014-03-14 | 2019-10-01 | Magic Leap, Inc. | Augmented reality systems and methods utilizing reflections |
US11138793B2 (en) | 2014-03-14 | 2021-10-05 | Magic Leap, Inc. | Multi-depth plane display system with reduced switching between depth planes |
US10048647B2 (en) | 2014-03-27 | 2018-08-14 | Microsoft Technology Licensing, Llc | Optical waveguide including spatially-varying volume hologram |
US20160187651A1 (en) | 2014-03-28 | 2016-06-30 | Osterhout Group, Inc. | Safety for a vehicle operator with an hmd |
KR20160143748A (en) | 2014-04-09 | 2016-12-14 | 쓰리엠 이노베이티브 프로퍼티즈 컴파니 | Near-eye display system having a pellicle as a combiner |
IL232197B (en) | 2014-04-23 | 2018-04-30 | Lumus Ltd | Compact head-mounted display system |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US9672210B2 (en) | 2014-04-25 | 2017-06-06 | Osterhout Group, Inc. | Language translation with head-worn computing |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
NZ764960A (en) | 2014-05-30 | 2022-05-27 | Magic Leap Inc | Methods and system for creating focal planes in virtual and augmented reality |
CN111856755B (en) | 2014-05-30 | 2022-07-19 | 奇跃公司 | Method and system for displaying stereoscopic vision of virtual and augmented reality |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9880632B2 (en) | 2014-06-19 | 2018-01-30 | Thalmic Labs Inc. | Systems, devices, and methods for gesture identification |
US9874744B2 (en) | 2014-06-25 | 2018-01-23 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays |
US10204530B1 (en) | 2014-07-11 | 2019-02-12 | Shape Matrix Geometric Instruments, LLC | Shape-matrix geometric instrument |
WO2016020632A1 (en) | 2014-08-08 | 2016-02-11 | Milan Momcilo Popovich | Method for holographic mastering and replication |
KR20160029245A (en) * | 2014-09-04 | 2016-03-15 | 삼성디스플레이 주식회사 | Head mounted display apparatus |
WO2016040836A1 (en) * | 2014-09-12 | 2016-03-17 | Eyelock Llc | Methods and apparatus for directing the gaze of a user in an iris recognition system |
WO2016042283A1 (en) | 2014-09-19 | 2016-03-24 | Milan Momcilo Popovich | Method and apparatus for generating input images for holographic waveguide displays |
US10423222B2 (en) | 2014-09-26 | 2019-09-24 | Digilens Inc. | Holographic waveguide optical tracker |
AU2015323940B2 (en) | 2014-09-29 | 2021-05-20 | Magic Leap, Inc. | Architectures and methods for outputting different wavelength light out of waveguides |
EP3212068B1 (en) | 2014-10-31 | 2020-11-25 | Lake Region Medical, Inc. | Fiber bragg grating multi-point pressure sensing guidewire with birefringent component |
IL235642B (en) | 2014-11-11 | 2021-08-31 | Lumus Ltd | Compact head-mounted display system protected by a hyperfine structure |
TWI688789B (en) * | 2014-11-20 | 2020-03-21 | 美商英特爾股份有限公司 | Virtual image generator and method to project a virtual image |
US9864205B2 (en) * | 2014-11-25 | 2018-01-09 | Ricoh Company, Ltd. | Multifocal display |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
KR101844883B1 (en) * | 2014-12-23 | 2018-04-03 | 메타 컴퍼니 | Apparatuses, methods and systems coupling visual accommodation and visual convergence to the same plane at any depth of an object of interest |
NZ759405A (en) * | 2014-12-29 | 2022-01-28 | Magic Leap Inc | Light projector using an acousto-optical control device |
USD751552S1 (en) | 2014-12-31 | 2016-03-15 | Osterhout Group, Inc. | Computer glasses |
EP3241055B1 (en) * | 2014-12-31 | 2020-04-15 | Essilor International | Binocular device comprising a monocular display device |
USD753114S1 (en) | 2015-01-05 | 2016-04-05 | Osterhout Group, Inc. | Air mouse |
US20180275402A1 (en) | 2015-01-12 | 2018-09-27 | Digilens, Inc. | Holographic waveguide light field displays |
US10437064B2 (en) | 2015-01-12 | 2019-10-08 | Digilens Inc. | Environmentally isolated waveguide display |
US9659411B2 (en) * | 2015-01-14 | 2017-05-23 | Oculus Vr, Llc | Passive locators for a virtual reality headset |
EP3248026B1 (en) | 2015-01-20 | 2019-09-04 | DigiLens Inc. | Holographic waveguide lidar |
KR102289923B1 (en) | 2015-01-22 | 2021-08-12 | 매직 립, 인코포레이티드 | Methods and system for creating focal planes using an Alvarez lens |
AU2016211623B2 (en) | 2015-01-26 | 2020-04-16 | Magic Leap, Inc. | Virtual and augmented reality systems and methods having improved diffractive grating structures |
US11726241B2 (en) | 2015-01-26 | 2023-08-15 | Magic Leap, Inc. | Manufacturing for virtual and augmented reality systems and components |
US9632226B2 (en) | 2015-02-12 | 2017-04-25 | Digilens Inc. | Waveguide grating device |
US10878775B2 (en) | 2015-02-17 | 2020-12-29 | Mentor Acquisition One, Llc | See-through computer display systems |
US20160239985A1 (en) | 2015-02-17 | 2016-08-18 | Osterhout Group, Inc. | See-through computer display systems |
US9989764B2 (en) | 2015-02-17 | 2018-06-05 | Thalmic Labs Inc. | Systems, devices, and methods for eyebox expansion in wearable heads-up displays |
US11468639B2 (en) * | 2015-02-20 | 2022-10-11 | Microsoft Technology Licensing, Llc | Selective occlusion system for augmented reality devices |
EP3062142B1 (en) | 2015-02-26 | 2018-10-03 | Nokia Technologies OY | Apparatus for a near-eye display |
AU2016227316B2 (en) | 2015-03-01 | 2020-06-25 | Novasight Ltd. | A system and method for measuring ocular motility |
CN112764536A (en) | 2015-03-05 | 2021-05-07 | 奇跃公司 | System and method for augmented reality |
US10838207B2 (en) | 2015-03-05 | 2020-11-17 | Magic Leap, Inc. | Systems and methods for augmented reality |
US10180734B2 (en) | 2015-03-05 | 2019-01-15 | Magic Leap, Inc. | Systems and methods for augmented reality |
WO2016146963A1 (en) | 2015-03-16 | 2016-09-22 | Popovich, Milan, Momcilo | Waveguide device incorporating a light pipe |
CA2979687A1 (en) | 2015-03-16 | 2016-09-22 | Nicole Elizabeth SAMEC | Methods and systems for diagnosing and treating health ailments |
US20160274365A1 (en) * | 2015-03-17 | 2016-09-22 | Thalmic Labs Inc. | Systems, devices, and methods for wearable heads-up displays with heterogeneous display quality |
JP6528498B2 (en) * | 2015-03-25 | 2019-06-12 | セイコーエプソン株式会社 | Head mounted display |
WO2016156776A1 (en) | 2015-03-31 | 2016-10-06 | Milan Momcilo Popovich | Method and apparatus for contact image sensing |
WO2016161463A1 (en) * | 2015-04-03 | 2016-10-06 | David Markus | Method and apparatus for an imaging lens |
JP2016212177A (en) * | 2015-05-01 | 2016-12-15 | セイコーエプソン株式会社 | Transmission type display device |
US10197805B2 (en) | 2015-05-04 | 2019-02-05 | North Inc. | Systems, devices, and methods for eyeboxes with heterogeneous exit pupils |
WO2016183020A1 (en) | 2015-05-11 | 2016-11-17 | Magic Leap, Inc. | Devices, methods and systems for biometric user recognition utilizing neural networks |
CN107710048A (en) | 2015-05-28 | 2018-02-16 | 赛尔米克实验室公司 | The system, apparatus and method of eye tracks and scanning laser projection are integrated in wearable head-up display |
EP3310043B1 (en) * | 2015-06-10 | 2024-08-07 | Sony Interactive Entertainment Inc. | Head-mounted display and display control method |
WO2016205256A1 (en) * | 2015-06-15 | 2016-12-22 | Magic Leap, Inc. | Display system with optical elements for in-coupling multiplexed light streams |
US10210844B2 (en) | 2015-06-29 | 2019-02-19 | Microsoft Technology Licensing, Llc | Holographic near-eye display |
WO2017015162A1 (en) * | 2015-07-17 | 2017-01-26 | Magic Leap, Inc. | Virtual/augmented reality system having dynamic region resolution |
JP6792607B2 (en) | 2015-07-20 | 2020-11-25 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Designed to collimate fiber scanners using inward angles in virtual / augmented reality systems |
US10307246B2 (en) | 2015-07-23 | 2019-06-04 | Elwha Llc | Intraocular lens devices, systems, and related methods |
US9877824B2 (en) * | 2015-07-23 | 2018-01-30 | Elwha Llc | Intraocular lens systems and related methods |
US10376357B2 (en) * | 2015-07-23 | 2019-08-13 | Elwha Llc | Intraocular lens systems and related methods |
US10154897B2 (en) | 2015-07-23 | 2018-12-18 | Elwha Llc | Intraocular lens systems and related methods |
US10007115B2 (en) * | 2015-08-12 | 2018-06-26 | Daqri, Llc | Placement of a computer generated display with focal plane at finite distance using optical devices and a see-through head-mounted display incorporating the same |
EP3338251B1 (en) | 2015-08-18 | 2023-01-25 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
AU2016310452B2 (en) | 2015-08-21 | 2021-04-22 | Magic Leap, Inc. | Eyelid shape estimation |
JP6918781B2 (en) | 2015-08-21 | 2021-08-11 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Eyelid shape estimation using eye pose measurement |
JP6367166B2 (en) | 2015-09-01 | 2018-08-01 | 株式会社東芝 | Electronic apparatus and method |
JP2018526278A (en) * | 2015-09-02 | 2018-09-13 | ラークテイル ピー・ティー・ワイ リミテッド | Folding stroller |
US9555589B1 (en) | 2015-09-03 | 2017-01-31 | 3M Innovative Properties Company | Method of making optical films and stacks |
CN108474873A (en) | 2015-09-04 | 2018-08-31 | 赛尔米克实验室公司 | System, product and method for combining holographic optical elements (HOE) and eyeglass |
WO2017131816A1 (en) | 2016-01-30 | 2017-08-03 | Leia Inc. | Privacy display and dual-mode privacy display system |
CA2996925C (en) * | 2015-09-05 | 2022-06-28 | Leia Inc. | Light concentrating backlight and near-eye display system using same |
WO2017048713A1 (en) | 2015-09-16 | 2017-03-23 | Magic Leap, Inc. | Head pose mixing of audio files |
WO2017049106A1 (en) * | 2015-09-17 | 2017-03-23 | Lumii, Inc. | Multi-view displays and associated systems and methods |
US9978183B2 (en) * | 2015-09-18 | 2018-05-22 | Fove, Inc. | Video system, video generating method, video distribution method, video generating program, and video distribution program |
WO2017053382A1 (en) * | 2015-09-23 | 2017-03-30 | Magic Leap, Inc. | Eye imaging with an off-axis imager |
EP3357175B1 (en) * | 2015-09-29 | 2021-05-05 | Newracom, Inc. | Resource allocation indication for multi-user multiple-input-multiple-output (mu-mimo) orthogonal frequency division multiple access (ofdma) communication |
US20170097753A1 (en) | 2015-10-01 | 2017-04-06 | Thalmic Labs Inc. | Systems, devices, and methods for interacting with content displayed on head-mounted displays |
CN113759555B (en) * | 2015-10-05 | 2024-09-20 | 迪吉伦斯公司 | Waveguide display |
EP3360002A4 (en) | 2015-10-05 | 2018-10-31 | Magic Leap, Inc. | Microlens collimator for scanning optical fiber in virtual/augmented reality system |
NZ775329A (en) | 2015-10-06 | 2022-07-01 | Magic Leap Inc | Virtual/augmented reality system having reverse angle diffraction grating |
CA2994231C (en) | 2015-10-16 | 2022-06-21 | Leia Inc. | Multibeam diffraction grating-based near-eye display |
CN108369653B (en) | 2015-10-16 | 2021-12-14 | 奇跃公司 | Eye pose recognition using eye features |
US11609427B2 (en) | 2015-10-16 | 2023-03-21 | Ostendo Technologies, Inc. | Dual-mode augmented/virtual reality (AR/VR) near-eye wearable displays |
US10088685B1 (en) | 2015-10-19 | 2018-10-02 | Meta Company | Apparatuses, methods and systems for multiple focal distance display |
EP3365724B1 (en) | 2015-10-20 | 2021-05-05 | Magic Leap, Inc. | Selecting virtual objects in a three-dimensional space |
US9904051B2 (en) | 2015-10-23 | 2018-02-27 | Thalmic Labs Inc. | Systems, devices, and methods for laser eye tracking |
US9842868B2 (en) * | 2015-10-26 | 2017-12-12 | Sensors Unlimited, Inc. | Quantum efficiency (QE) restricted infrared focal plane arrays |
US11106273B2 (en) | 2015-10-30 | 2021-08-31 | Ostendo Technologies, Inc. | System and methods for on-body gestural interfaces and projection displays |
US10281744B2 (en) | 2015-11-02 | 2019-05-07 | Focure Inc. | Continuous autofocusing eyewear using structured light |
WO2017079342A1 (en) | 2015-11-02 | 2017-05-11 | Focure, Inc. | Continuous autofocusing eyewear |
US10260864B2 (en) | 2015-11-04 | 2019-04-16 | Magic Leap, Inc. | Dynamic display calibration based on eye-tracking |
US11231544B2 (en) | 2015-11-06 | 2022-01-25 | Magic Leap, Inc. | Metasurfaces for redirecting light and methods for fabricating |
US10204451B2 (en) | 2015-11-30 | 2019-02-12 | Microsoft Technology Licensing, Llc | Multi-optical surface optical design |
CA3006556A1 (en) * | 2015-12-02 | 2017-06-08 | Abl Ip Holding Llc | Projection and/or waveguide arrangements for a software configurable lighting device |
WO2017096241A1 (en) | 2015-12-02 | 2017-06-08 | Augmenteum, Inc. | System for and method of projecting augmentation imagery in a head-mounted display |
AU2016365422A1 (en) | 2015-12-04 | 2018-06-28 | Magic Leap, Inc. | Relocalization systems and methods |
US10445860B2 (en) | 2015-12-08 | 2019-10-15 | Facebook Technologies, Llc | Autofocus virtual reality headset |
KR102436809B1 (en) * | 2015-12-15 | 2022-08-29 | 삼성디스플레이 주식회사 | Window display apparatus |
US10802190B2 (en) | 2015-12-17 | 2020-10-13 | Covestro Llc | Systems, devices, and methods for curved holographic optical elements |
US10345594B2 (en) | 2015-12-18 | 2019-07-09 | Ostendo Technologies, Inc. | Systems and methods for augmented near-eye wearable displays |
CN108139801B (en) * | 2015-12-22 | 2021-03-16 | 谷歌有限责任公司 | System and method for performing electronic display stabilization via preserving light field rendering |
WO2017112958A1 (en) * | 2015-12-24 | 2017-06-29 | Meta Company | Optical engine for creating wide-field of view fovea-based display |
US20200301150A1 (en) * | 2015-12-28 | 2020-09-24 | Intelligent Technologies International, Inc. | Secure testing device with liquid crystal shutter |
US10764503B2 (en) | 2015-12-28 | 2020-09-01 | Nec Corporation | Information processing apparatus, control method, and program for outputting a guide for correcting a field of view of a camera |
WO2017115587A1 (en) * | 2015-12-28 | 2017-07-06 | 日本電気株式会社 | Information processing device, control method, and program |
US10578882B2 (en) | 2015-12-28 | 2020-03-03 | Ostendo Technologies, Inc. | Non-telecentric emissive micro-pixel array light modulators and methods of fabrication thereof |
US11030443B2 (en) | 2015-12-28 | 2021-06-08 | Nec Corporation | Information processing apparatus, control method, and program |
US9964925B2 (en) * | 2015-12-29 | 2018-05-08 | Oculus Vr, Llc | Holographic display architecture |
US11281003B2 (en) * | 2015-12-30 | 2022-03-22 | Dualitas Ltd | Near eye dynamic holography |
WO2017120271A1 (en) | 2016-01-04 | 2017-07-13 | Meta Company | Apparatuses, methods and systems for application of forces within a 3d virtual environment |
WO2017120552A1 (en) | 2016-01-06 | 2017-07-13 | Meta Company | Apparatuses, methods and systems for pre-warping images for a display system with a distorting optical component |
JP6681042B2 (en) * | 2016-02-17 | 2020-04-15 | 株式会社リコー | Light guide and virtual image display device |
JP6701559B2 (en) * | 2016-02-17 | 2020-05-27 | 株式会社リコー | Light guide and virtual image display device |
JP6608065B2 (en) * | 2016-01-06 | 2019-11-20 | ビュージックス コーポレーション | Double-sided imaging light guide with embedded dichroic filter |
IL260280B2 (en) | 2016-01-07 | 2023-09-01 | Magic Leap Inc | Virtual and augmented reality systems and methods having unequal numbers of component color images distributed across depth planes |
US10042166B2 (en) * | 2016-01-12 | 2018-08-07 | Magic Leap, Inc. | Beam angle sensor in virtual/augmented reality system |
US10681328B1 (en) | 2016-01-13 | 2020-06-09 | Apple Inc. | Dynamic focus 3D display |
US11262580B1 (en) | 2016-01-13 | 2022-03-01 | Apple Inc. | Virtual reality system |
US11209898B2 (en) | 2016-01-19 | 2021-12-28 | Magic Leap, Inc. | Eye image collection |
KR20180104051A (en) | 2016-01-19 | 2018-09-19 | 매직 립, 인코포레이티드 | Augmented reality systems and methods that utilize reflections |
US10303246B2 (en) | 2016-01-20 | 2019-05-28 | North Inc. | Systems, devices, and methods for proximity-based eye tracking |
JP6907218B2 (en) * | 2016-01-20 | 2021-07-21 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Polarization maintenance optical fiber in virtual / augmented reality systems |
US10110935B2 (en) | 2016-01-29 | 2018-10-23 | Cable Television Laboratories, Inc | Systems and methods for video delivery based upon saccadic eye motion |
WO2017131770A1 (en) * | 2016-01-29 | 2017-08-03 | Hewlett-Packard Development Company, L.P | Viewing device adjustment based on eye accommodation in relation to a display |
IL292790B2 (en) | 2016-01-29 | 2024-09-01 | Magic Leap Inc | Display for three-dimensional image |
US10151926B2 (en) | 2016-01-29 | 2018-12-11 | North Inc. | Systems, devices, and methods for preventing eyebox degradation in a wearable heads-up display |
US11284109B2 (en) | 2016-01-29 | 2022-03-22 | Cable Television Laboratories, Inc. | Visual coding for sensitivities to light, color and spatial resolution in human visual system |
WO2017134412A1 (en) | 2016-02-04 | 2017-08-10 | Milan Momcilo Popovich | Holographic waveguide optical tracker |
EP3205512B1 (en) | 2016-02-09 | 2018-06-13 | CSEM Centre Suisse d'Electronique et de Microtechnique SA - Recherche et Développement | Optical security device |
KR102587841B1 (en) * | 2016-02-11 | 2023-10-10 | 매직 립, 인코포레이티드 | Multi-depth plane display system with reduced switching between depth planes |
JP6908872B2 (en) * | 2016-02-17 | 2021-07-28 | 株式会社リコー | Light guide, virtual image display device and light guide member |
US10591728B2 (en) | 2016-03-02 | 2020-03-17 | Mentor Acquisition One, Llc | Optical systems for head-worn computers |
EP3420387A4 (en) * | 2016-02-24 | 2019-10-30 | Magic Leap, Inc. | Polarizing beam splitter with low light leakage |
US11157072B1 (en) * | 2016-02-24 | 2021-10-26 | Apple Inc. | Direct retinal projector |
JP6991981B2 (en) | 2016-02-24 | 2022-01-13 | マジック リープ, インコーポレイテッド | Thin interconnect for light emitters |
NZ745294A (en) | 2016-02-26 | 2020-01-31 | Magic Leap Inc | Light output system with reflector and lens for highly spatially uniform light output |
WO2017147534A1 (en) * | 2016-02-26 | 2017-08-31 | Magic Leap, Inc. | Display system having a plurality of light pipes for a plurality of light emitters |
US10667981B2 (en) | 2016-02-29 | 2020-06-02 | Mentor Acquisition One, Llc | Reading assistance system for visually impaired |
CN108700712B (en) * | 2016-02-29 | 2020-10-13 | 奇跃公司 | Virtual and augmented reality systems and methods |
WO2017151671A1 (en) | 2016-03-01 | 2017-09-08 | Magic Leap, Inc. | Depth sensing systems and methods |
CA3015643A1 (en) * | 2016-03-01 | 2017-09-08 | Magic Leap, Inc. | Reflective switching device for inputting different wavelengths of light into waveguides |
KR102079181B1 (en) * | 2016-03-04 | 2020-02-19 | 주식회사 고영테크놀러지 | Pattern lighting appartus and method thereof |
WO2017151974A1 (en) | 2016-03-04 | 2017-09-08 | Magic Leap, Inc. | Current drain reduction in ar/vr display systems |
EP4451237A2 (en) | 2016-03-07 | 2024-10-23 | Magic Leap, Inc. | Blue light adjustment for biometric security |
US11106276B2 (en) | 2016-03-11 | 2021-08-31 | Facebook Technologies, Llc | Focus adjusting headset |
CA3015658A1 (en) | 2016-03-11 | 2017-09-14 | Magic Leap, Inc. | Structure learning in convolutional neural networks |
CN111329551A (en) * | 2016-03-12 | 2020-06-26 | P·K·朗 | Augmented reality guidance for spinal and joint surgery |
US9886742B2 (en) * | 2016-03-17 | 2018-02-06 | Google Llc | Electro-optic beam steering for super-resolution/lightfield imagery |
CN105589202A (en) * | 2016-03-18 | 2016-05-18 | 京东方科技集团股份有限公司 | Display device, display method, and display system |
CN115032795A (en) | 2016-03-22 | 2022-09-09 | 奇跃公司 | Head-mounted display system configured to exchange biometric information |
WO2017164573A1 (en) * | 2016-03-23 | 2017-09-28 | Samsung Electronics Co., Ltd. | Near-eye display apparatus and near-eye display method |
EP3223063A1 (en) | 2016-03-24 | 2017-09-27 | Thomson Licensing | Device for forming a field intensity pattern in the near zone, from incident electromagnetic waves |
EP3223062A1 (en) | 2016-03-24 | 2017-09-27 | Thomson Licensing | Device for forming at least one focused beam in the near zone, from incident electromagnetic waves |
US10859768B2 (en) | 2016-03-24 | 2020-12-08 | Digilens Inc. | Method and apparatus for providing a polarization selective holographic waveguide device |
KR20180122726A (en) * | 2016-03-25 | 2018-11-13 | 매직 립, 인코포레이티드 | Virtual and augmented reality systems and methods |
KR101788452B1 (en) * | 2016-03-30 | 2017-11-15 | 연세대학교 산학협력단 | Apparatus and method for replaying contents using eye tracking of users |
CA3018758A1 (en) * | 2016-03-31 | 2017-10-05 | Magic Leap, Inc. | Interactions with 3d virtual objects using poses and multiple-dof controllers |
US10815145B2 (en) * | 2016-03-31 | 2020-10-27 | Corning Incorporated | High index glass and devices incorporating such |
US10317679B2 (en) | 2016-04-04 | 2019-06-11 | Akonia Holographics, Llc | Light homogenization |
US10353203B2 (en) | 2016-04-05 | 2019-07-16 | Ostendo Technologies, Inc. | Augmented/virtual reality near-eye displays with edge imaging lens comprising a plurality of display devices |
CN109073819A (en) | 2016-04-07 | 2018-12-21 | 奇跃公司 | System and method for augmented reality |
US10379356B2 (en) | 2016-04-07 | 2019-08-13 | Facebook Technologies, Llc | Accommodation based optical correction |
US9897811B2 (en) * | 2016-04-07 | 2018-02-20 | Google Llc | Curved eyepiece with color correction for head wearable display |
AU2017246901B2 (en) * | 2016-04-08 | 2022-06-02 | Magic Leap, Inc. | Augmented reality systems and methods with variable focus lens elements |
WO2017178781A1 (en) | 2016-04-11 | 2017-10-19 | GRANT, Alastair, John | Holographic waveguide apparatus for structured light projection |
US10178378B2 (en) * | 2016-04-12 | 2019-01-08 | Microsoft Technology Licensing, Llc | Binocular image alignment for near-eye display |
KR20180132854A (en) | 2016-04-13 | 2018-12-12 | 탈믹 랩스 인크 | System, apparatus and method for focusing a laser projector |
IL262279B2 (en) * | 2016-04-21 | 2023-04-01 | Magic Leap Inc | Visual aura around field of view |
US11009714B1 (en) * | 2016-04-22 | 2021-05-18 | Holochip Corporation | Interactive virtual reality display providing accommodation depth cues |
WO2017189450A1 (en) | 2016-04-26 | 2017-11-02 | Magic Leap, Inc. | Electromagnetic tracking with augmented reality systems |
CN105759447A (en) * | 2016-04-27 | 2016-07-13 | 江苏卡罗卡国际动漫城有限公司 | Augmented reality glasses |
US10453431B2 (en) | 2016-04-28 | 2019-10-22 | Ostendo Technologies, Inc. | Integrated near-far light field display systems |
CN105788390A (en) * | 2016-04-29 | 2016-07-20 | 吉林医药学院 | Medical anatomy auxiliary teaching system based on augmented reality |
US20170315347A1 (en) * | 2016-04-29 | 2017-11-02 | Mikko Antton Juhola | Exit Pupil Expander for Laser-Scanner and Waveguide Based Augmented-Reality Displays |
US10522106B2 (en) | 2016-05-05 | 2019-12-31 | Ostendo Technologies, Inc. | Methods and apparatus for active transparency modulation |
KR102230642B1 (en) | 2016-05-06 | 2021-03-19 | 매직 립, 인코포레이티드 | Metasurfaces with asymmetric gratings for redirecting light and methods for manufacturing |
CN109414164B (en) | 2016-05-09 | 2022-06-14 | 奇跃公司 | Augmented reality system and method for user health analysis |
CN105898276A (en) * | 2016-05-10 | 2016-08-24 | 北京理工大学 | Near-to-eye three-dimensional display system based on non-periodic holographic microlens array |
AU2017264780B2 (en) | 2016-05-12 | 2022-05-12 | Magic Leap, Inc. | Distributed light manipulation over imaging waveguide |
US10215986B2 (en) | 2016-05-16 | 2019-02-26 | Microsoft Technology Licensing, Llc | Wedges for light transformation |
WO2017199232A1 (en) * | 2016-05-18 | 2017-11-23 | Lumus Ltd. | Head-mounted imaging device |
EP3459071B1 (en) | 2016-05-20 | 2022-05-11 | Magic Leap, Inc. | Contextual awareness of user interface menus |
US10765314B2 (en) | 2016-05-29 | 2020-09-08 | Novasight Ltd. | Display system and method |
JP6850817B2 (en) | 2016-06-03 | 2021-03-31 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Augmented reality identification verification |
US10429647B2 (en) | 2016-06-10 | 2019-10-01 | Facebook Technologies, Llc | Focus adjusting virtual reality headset |
CA3026872A1 (en) | 2016-06-10 | 2017-12-14 | Magic Leap, Inc. | Integrating point source for texture projecting bulb |
US10684479B2 (en) | 2016-06-15 | 2020-06-16 | Vrvaorigin Vision Technology Corp. Ltd. | Head-mounted personal multimedia systems and visual assistance devices thereof |
CN107526165B (en) * | 2016-06-15 | 2022-08-26 | 威亚视觉科技股份有限公司 | Head-mounted personal multimedia system, visual auxiliary device and related glasses |
JP7385993B2 (en) | 2016-06-20 | 2023-11-24 | マジック リープ, インコーポレイテッド | Augmented reality display system for evaluation and correction of neurological disorders, including disorders of visual processing and perception |
US10444509B2 (en) * | 2016-06-27 | 2019-10-15 | Daqri, Llc | Near eye diffractive holographic projection method |
US10366536B2 (en) | 2016-06-28 | 2019-07-30 | Microsoft Technology Licensing, Llc | Infinite far-field depth perception for near-field objects in virtual environments |
AU2017291131B2 (en) | 2016-06-30 | 2022-03-31 | Magic Leap, Inc. | Estimating pose in 3D space |
CN107561697B (en) * | 2016-07-01 | 2019-04-30 | 成都理想境界科技有限公司 | Near-eye display system, virtual reality device and augmented reality equipment |
CN107561698A (en) * | 2016-07-01 | 2018-01-09 | 成都理想境界科技有限公司 | A kind of near-eye display system, virtual reality device and augmented reality equipment |
CN107561700A (en) * | 2016-07-01 | 2018-01-09 | 成都理想境界科技有限公司 | A kind of near-eye display system, virtual reality device and augmented reality equipment |
CN107562181B (en) * | 2016-07-01 | 2020-01-31 | 成都理想境界科技有限公司 | Near-to-eye display system, virtual reality equipment and augmented reality equipment |
US10649209B2 (en) | 2016-07-08 | 2020-05-12 | Daqri Llc | Optical combiner apparatus |
CN109690387B (en) * | 2016-07-13 | 2022-11-01 | 视瑞尔技术公司 | Display device |
JP6917444B2 (en) | 2016-07-14 | 2021-08-11 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Iris boundary estimation using corneal curvature |
KR102648770B1 (en) | 2016-07-14 | 2024-03-15 | 매직 립, 인코포레이티드 | Deep neural network for iris identification |
US10877210B2 (en) * | 2016-07-15 | 2020-12-29 | Light Field Lab, Inc. | Energy propagation and transverse anderson localization with two-dimensional, light field and holographic relays |
US10152122B2 (en) | 2016-07-18 | 2018-12-11 | Tobii Ab | Foveated rendering |
CN106201213A (en) * | 2016-07-19 | 2016-12-07 | 深圳市金立通信设备有限公司 | The control method of a kind of virtual reality focus and terminal |
KR102461953B1 (en) * | 2016-07-25 | 2022-10-31 | 매직 립, 인코포레이티드 | Imaging modification, display and visualization using augmented reality and virtual reality eyewear |
JP6992046B2 (en) * | 2016-07-25 | 2022-01-13 | マジック リープ, インコーポレイテッド | Lightfield processor system |
US10277874B2 (en) | 2016-07-27 | 2019-04-30 | North Inc. | Systems, devices, and methods for laser projectors |
US10241244B2 (en) | 2016-07-29 | 2019-03-26 | Lumentum Operations Llc | Thin film total internal reflection diffraction grating for single polarization or dual polarization |
IL294128B2 (en) | 2016-07-29 | 2024-04-01 | Magic Leap Inc | Secure exchange of cryptographically signed records |
AU2017305249B2 (en) | 2016-08-01 | 2021-07-22 | Magic Leap, Inc. | Mixed reality system with spatialized audio |
KR102626821B1 (en) | 2016-08-02 | 2024-01-18 | 매직 립, 인코포레이티드 | Fixed-distance virtual and augmented reality systems and methods |
EP3497676B1 (en) | 2016-08-11 | 2024-07-17 | Magic Leap, Inc. | Automatic placement of a virtual object in a three-dimensional space |
US10185153B2 (en) | 2016-08-12 | 2019-01-22 | Avegant Corp. | Orthogonal optical path length extender |
US10057488B2 (en) | 2016-08-12 | 2018-08-21 | Avegant Corp. | Image capture with digital light path length modulation |
US10401639B2 (en) | 2016-08-12 | 2019-09-03 | Avegant Corp. | Method and apparatus for an optical path length extender |
US10379388B2 (en) | 2016-08-12 | 2019-08-13 | Avegant Corp. | Digital light path length modulation systems |
US10516879B2 (en) | 2016-08-12 | 2019-12-24 | Avegant Corp. | Binocular display with digital light path length modulation |
US10187634B2 (en) | 2016-08-12 | 2019-01-22 | Avegant Corp. | Near-eye display system including a modulation stack |
US10607609B2 (en) | 2016-08-12 | 2020-03-31 | Magic Leap, Inc. | Word flow annotation |
US10809546B2 (en) | 2016-08-12 | 2020-10-20 | Avegant Corp. | Digital light path length modulation |
WO2018027326A1 (en) | 2016-08-12 | 2018-02-15 | Thalmic Labs Inc. | Systems, devices, and methods for variable luminance in wearable heads-up displays |
WO2018039269A1 (en) | 2016-08-22 | 2018-03-01 | Magic Leap, Inc. | Augmented reality display device with deep learning sensors |
JP6813666B2 (en) * | 2016-08-22 | 2021-01-13 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Multi-layer diffraction eyepiece |
KR102634343B1 (en) | 2016-08-22 | 2024-02-05 | 매직 립, 인코포레이티드 | Virtual, augmented, and mixed reality systems and methods |
US10255714B2 (en) | 2016-08-24 | 2019-04-09 | Disney Enterprises, Inc. | System and method of gaze predictive rendering of a focal area of an animation |
US20180061084A1 (en) * | 2016-08-24 | 2018-03-01 | Disney Enterprises, Inc. | System and method of bandwidth-sensitive rendering of a focal area of an animation |
KR101894555B1 (en) * | 2016-09-07 | 2018-10-04 | 주식회사 레티널 | Reflecting lens module |
CN113467093B (en) | 2016-09-07 | 2023-09-22 | 奇跃公司 | Virtual reality, augmented reality, and mixed reality systems including thick media and related methods |
US9958749B2 (en) * | 2016-09-12 | 2018-05-01 | Disney Enterprises, Inc. | Optical assemblies that are both brightly backlit and transparent for use in costumed characters |
JP6798010B2 (en) | 2016-09-13 | 2020-12-09 | マジック リープ, インコーポレイテッドMagic Leap,Inc. | Sensory eyewear |
CN109691141B (en) | 2016-09-14 | 2022-04-29 | 奇跃公司 | Spatialization audio system and method for rendering spatialization audio |
CN107835288A (en) * | 2016-09-16 | 2018-03-23 | 天津思博科科技发展有限公司 | The interdynamic recreational apparatus realized using intelligent terminal |
WO2018057660A2 (en) | 2016-09-20 | 2018-03-29 | Apple Inc. | Augmented reality system |
CA3037044A1 (en) * | 2016-09-21 | 2018-03-29 | Magic Leap, Inc. | Systems and methods for optical systems with exit pupil expander |
WO2018057910A1 (en) | 2016-09-22 | 2018-03-29 | Apple, Inc. | Postponing the state change of an information affecting the graphical user interface until during the conditon of inattentiveness |
US10330935B2 (en) | 2016-09-22 | 2019-06-25 | Apple Inc. | Predictive, foveated virtual reality system |
KR20210072157A (en) | 2016-09-22 | 2021-06-16 | 매직 립, 인코포레이티드 | Augmented reality spectroscopy |
JP2020509790A (en) | 2016-09-23 | 2020-04-02 | ノバサイト リミテッド | Screening device and method |
EP3516485A4 (en) | 2016-09-26 | 2020-06-03 | Magic Leap, Inc. | Calibration of magnetic and optical sensors in a virtual reality or augmented reality display system |
GB2554416A (en) | 2016-09-26 | 2018-04-04 | Design Led Ltd | Illuminated eyewear device |
US10976549B2 (en) | 2016-09-28 | 2021-04-13 | Magic Leap, Inc. | Face model capture by a wearable device |
RU2016138608A (en) | 2016-09-29 | 2018-03-30 | Мэджик Лип, Инк. | NEURAL NETWORK FOR SEGMENTING THE EYE IMAGE AND ASSESSING THE QUALITY OF THE IMAGE |
US10425636B2 (en) | 2016-10-03 | 2019-09-24 | Microsoft Technology Licensing, Llc | Automatic detection and correction of binocular misalignment in a display device |
WO2018067603A1 (en) | 2016-10-04 | 2018-04-12 | Magic Leap, Inc. | Efficient data layouts for convolutional neural networks |
IL295059A (en) | 2016-10-05 | 2022-09-01 | Magic Leap Inc | Periocular test for mixed reality calibration |
WO2018067801A1 (en) | 2016-10-05 | 2018-04-12 | Magic Leap, Inc. | Surface modeling systems and methods |
US10466479B2 (en) | 2016-10-07 | 2019-11-05 | Coretronic Corporation | Head-mounted display apparatus and optical system |
CA2992213C (en) | 2016-10-09 | 2023-08-29 | Yochay Danziger | Aperture multiplier using a rectangular waveguide |
JP6209662B1 (en) * | 2016-10-13 | 2017-10-04 | 株式会社Qdレーザ | Image projection device |
EP4333428A3 (en) | 2016-10-21 | 2024-05-22 | Magic Leap, Inc. | System and method for presenting image content on multiple depth planes by providing multiple intra-pupil parallax views |
EP3312646A1 (en) | 2016-10-21 | 2018-04-25 | Thomson Licensing | Device and method for shielding at least one sub-wavelength-scale object from an incident electromagnetic wave |
EP3312660A1 (en) | 2016-10-21 | 2018-04-25 | Thomson Licensing | Device for forming at least one tilted focused beam in the near zone, from incident electromagnetic waves |
CN106371218B (en) * | 2016-10-28 | 2019-05-24 | 苏州苏大维格光电科技股份有限公司 | A kind of wear-type three-dimensional display apparatus |
CN108024187A (en) * | 2016-10-31 | 2018-05-11 | 苏州乐听电子科技有限公司 | It is a kind of independently to debug the intelligent hearing aid tested and matched somebody with somebody |
US10254542B2 (en) | 2016-11-01 | 2019-04-09 | Microsoft Technology Licensing, Llc | Holographic projector for a waveguide display |
CN108007386B (en) * | 2016-11-02 | 2021-04-20 | 光宝电子(广州)有限公司 | Three-dimensional scanning method based on structured light and device and system thereof |
US10274732B2 (en) | 2016-11-04 | 2019-04-30 | Microsoft Technology Licensing, Llc | Hologram focus accommodation |
BR112018014673A2 (en) | 2016-11-08 | 2018-12-11 | Lumus Ltd. | optical cutting edge light guide device and corresponding production methods |
US10215987B2 (en) | 2016-11-10 | 2019-02-26 | North Inc. | Systems, devices, and methods for astigmatism compensation in a wearable heads-up display |
KR102555925B1 (en) | 2016-11-10 | 2023-07-13 | 매직 립, 인코포레이티드 | Method and system for eye tracking using speckle patterns |
EP3538946B1 (en) | 2016-11-11 | 2023-02-15 | Magic Leap, Inc. | Periocular and audio synthesis of a full face image |
US11960083B2 (en) | 2016-11-15 | 2024-04-16 | Creal Sa | Near-eye sequential light-field projector with correct monocular depth cues |
EP3542250A4 (en) | 2016-11-15 | 2020-08-26 | Magic Leap, Inc. | Deep learning system for cuboid detection |
CA3042554C (en) | 2016-11-16 | 2023-07-18 | Magic Leap, Inc. | Multi-resolution display assembly for head-mounted display systems |
IL294413B2 (en) | 2016-11-16 | 2024-07-01 | Magic Leap Inc | Thermal management systems for wearable components |
CN106657976B (en) * | 2016-11-17 | 2019-06-11 | 宇龙计算机通信科技(深圳)有限公司 | A kind of visual range extension method, device and virtual reality glasses |
JP7237830B2 (en) * | 2016-11-18 | 2023-03-13 | マジック リープ, インコーポレイテッド | Waveguide optical multiplexer using cross gratings |
CN115685626A (en) | 2016-11-18 | 2023-02-03 | 奇跃公司 | Multilayer liquid crystal diffraction grating for redirecting light with a wide range of incident angles |
US11067860B2 (en) * | 2016-11-18 | 2021-07-20 | Magic Leap, Inc. | Liquid crystal diffractive devices with nano-scale pattern and methods of manufacturing the same |
EP4152085A1 (en) * | 2016-11-18 | 2023-03-22 | Magic Leap, Inc. | Spatially variable liquid crystal diffraction gratings |
WO2018098436A1 (en) | 2016-11-28 | 2018-05-31 | Spy Eye, Llc | Unobtrusive eye mounted display |
US10509153B2 (en) | 2016-11-29 | 2019-12-17 | Akonia Holographics Llc | Input coupling |
CA3045046A1 (en) | 2016-11-30 | 2018-06-07 | Magic Leap, Inc. | Method and system for high resolution digitized display |
WO2018098579A1 (en) | 2016-11-30 | 2018-06-07 | Thalmic Labs Inc. | Systems, devices, and methods for laser eye tracking in wearable heads-up displays |
CN106780297B (en) * | 2016-11-30 | 2019-10-25 | 天津大学 | Image high registration accuracy method under scene and Varying Illumination |
US20190278090A1 (en) * | 2016-11-30 | 2019-09-12 | Nova-Sight Ltd. | Methods and devices for displaying image with changed field of view |
CA3045014A1 (en) | 2016-12-01 | 2018-06-07 | Magic Leap, Inc. | Projector with scanning array light engine |
EP3548939A4 (en) | 2016-12-02 | 2020-11-25 | DigiLens Inc. | Waveguide device with uniform output illumination |
US10531220B2 (en) | 2016-12-05 | 2020-01-07 | Magic Leap, Inc. | Distributed audio capturing techniques for virtual reality (VR), augmented reality (AR), and mixed reality (MR) systems |
CN116883628A (en) | 2016-12-05 | 2023-10-13 | 奇跃公司 | Wearable system and method for providing virtual remote control in mixed reality environment |
US10310268B2 (en) | 2016-12-06 | 2019-06-04 | Microsoft Technology Licensing, Llc | Waveguides with peripheral side geometries to recycle light |
KR20230144109A (en) | 2016-12-08 | 2023-10-13 | 매직 립, 인코포레이티드 | Diffractive devices based on cholesteric liquid crystal |
US9946075B1 (en) * | 2016-12-08 | 2018-04-17 | Darwin Hu | See-through display glasses for virtual reality and augmented reality applications |
US10353213B2 (en) * | 2016-12-08 | 2019-07-16 | Darwin Hu | See-through display glasses for viewing 3D multimedia |
JP7071363B2 (en) | 2016-12-13 | 2022-05-18 | マジック リープ, インコーポレイテッド | Augmented reality and virtual reality eyewear, systems, and methods for delivering polarization and determining glucose levels. |
JP7038713B2 (en) | 2016-12-13 | 2022-03-18 | マジック リープ, インコーポレイテッド | 3D object rendering with detected features |
KR102170123B1 (en) | 2016-12-14 | 2020-10-26 | 주식회사 엘지화학 | Waveguide having light shielding and manufacturing method for the same |
US10895784B2 (en) * | 2016-12-14 | 2021-01-19 | Magic Leap, Inc. | Patterning of liquid crystals using soft-imprint replication of surface alignment patterns |
CN110959131A (en) * | 2016-12-15 | 2020-04-03 | 株式会社Ntt都科摩 | Wearable display optics using laser beam scanner |
US11209586B2 (en) * | 2016-12-15 | 2021-12-28 | Fusao Ishii | Ghost image elimination of doe using fourier optics method |
US9977248B1 (en) | 2016-12-21 | 2018-05-22 | PhantaField, Inc. | Augmented reality display system |
US10930082B2 (en) | 2016-12-21 | 2021-02-23 | Pcms Holdings, Inc. | Systems and methods for selecting spheres of relevance for presenting augmented reality information |
US10371896B2 (en) | 2016-12-22 | 2019-08-06 | Magic Leap, Inc. | Color separation in planar waveguides using dichroic filters |
CN106599893B (en) * | 2016-12-22 | 2020-01-24 | 深圳大学 | Processing method and device for object deviating from recognition graph based on augmented reality |
KR102524010B1 (en) | 2016-12-22 | 2023-04-19 | 매직 립, 인코포레이티드 | Systems and methods for manipulating light from ambient light sources |
US10663732B2 (en) | 2016-12-23 | 2020-05-26 | North Inc. | Systems, devices, and methods for beam combining in wearable heads-up displays |
WO2018122902A1 (en) * | 2016-12-26 | 2018-07-05 | マクセル株式会社 | Image display device and image display method |
US10746999B2 (en) | 2016-12-28 | 2020-08-18 | Magic Leap, Inc. | Dual depth exit pupil expander |
CN108254918B (en) * | 2016-12-28 | 2021-10-26 | 精工爱普生株式会社 | Optical element and display device |
KR102630774B1 (en) | 2016-12-29 | 2024-01-30 | 매직 립, 인코포레이티드 | Automatic control of wearable display device based on external conditions |
US10650552B2 (en) | 2016-12-29 | 2020-05-12 | Magic Leap, Inc. | Systems and methods for augmented reality |
US10209520B2 (en) * | 2016-12-30 | 2019-02-19 | Microsoft Technology Licensing, Llc | Near eye display multi-component dimming system |
CN106713882A (en) * | 2016-12-30 | 2017-05-24 | 中国科学院苏州生物医学工程技术研究所 | Photostimulation visual restoration device and photostimulation visual imaging method |
EP3343267B1 (en) | 2016-12-30 | 2024-01-24 | Magic Leap, Inc. | Polychromatic light out-coupling apparatus, near-eye displays comprising the same, and method of out-coupling polychromatic light |
US11022939B2 (en) * | 2017-01-03 | 2021-06-01 | Microsoft Technology Licensing, Llc | Reduced bandwidth holographic near-eye display |
USD864959S1 (en) | 2017-01-04 | 2019-10-29 | Mentor Acquisition One, Llc | Computer glasses |
EP4122897A1 (en) | 2017-01-05 | 2023-01-25 | Magic Leap, Inc. | Patterning of high refractive index glasses by plasma etching |
WO2018129398A1 (en) | 2017-01-05 | 2018-07-12 | Digilens, Inc. | Wearable heads up displays |
KR101894955B1 (en) | 2017-01-05 | 2018-09-05 | 주식회사 미디어프론트 | Live social media system for using virtual human awareness and real-time synthesis technology, server for augmented synthesis |
US10481678B2 (en) | 2017-01-11 | 2019-11-19 | Daqri Llc | Interface-based modeling and design of three dimensional spaces using two dimensional representations |
US20180197624A1 (en) | 2017-01-11 | 2018-07-12 | Magic Leap, Inc. | Medical assistant |
CN108303804A (en) * | 2017-01-12 | 2018-07-20 | 北京维信诺光电技术有限公司 | 3d glasses |
TWI629506B (en) * | 2017-01-16 | 2018-07-11 | 國立台灣大學 | Stereoscopic video see-through augmented reality device with vergence control and gaze stabilization, head-mounted display and method for near-field augmented reality application |
US10409066B2 (en) | 2017-01-19 | 2019-09-10 | Coretronic Corporation | Head-mounted display device with waveguide elements |
CN110431470B (en) * | 2017-01-19 | 2022-03-01 | 脸谱科技有限责任公司 | Focal plane display |
IL268135B2 (en) | 2017-01-23 | 2024-03-01 | Magic Leap Inc | Eyepiece for virtual, augmented, or mixed reality systems |
US10812936B2 (en) | 2017-01-23 | 2020-10-20 | Magic Leap, Inc. | Localization determination for mixed reality systems |
US10718951B2 (en) | 2017-01-25 | 2020-07-21 | North Inc. | Systems, devices, and methods for beam combining in laser projectors |
WO2018140502A1 (en) | 2017-01-27 | 2018-08-02 | Magic Leap, Inc. | Antireflection coatings for metasurfaces |
JP7142015B2 (en) * | 2017-01-27 | 2022-09-26 | マジック リープ, インコーポレイテッド | Diffraction gratings formed by metasurfaces with differently directed nanobeams |
EP3574360B1 (en) | 2017-01-28 | 2024-07-10 | Lumus Ltd. | Augmented reality imaging system |
US10303211B2 (en) * | 2017-02-01 | 2019-05-28 | Facebook Technologies, Llc | Two part cone display using flexible substrates |
US9983412B1 (en) | 2017-02-02 | 2018-05-29 | The University Of North Carolina At Chapel Hill | Wide field of view augmented reality see through head mountable display with distance accommodation |
US10410566B1 (en) * | 2017-02-06 | 2019-09-10 | Andrew Kerdemelidis | Head mounted virtual reality display system and method |
US10904514B2 (en) | 2017-02-09 | 2021-01-26 | Facebook Technologies, Llc | Polarization illumination using acousto-optic structured light in 3D depth sensing |
WO2018147811A1 (en) * | 2017-02-12 | 2018-08-16 | Lemnis Technologies Pte. Ltd. | Methods, devices and systems for focus adjustment of displays |
US11016292B2 (en) | 2017-02-15 | 2021-05-25 | Magic Leap, Inc. | Projector architecture incorporating artifact mitigation |
US11347054B2 (en) | 2017-02-16 | 2022-05-31 | Magic Leap, Inc. | Systems and methods for augmented reality |
CN110383139A (en) * | 2017-02-17 | 2019-10-25 | 康尼柴克斯 | Method and system for displaying images |
US10485420B2 (en) * | 2017-02-17 | 2019-11-26 | Analog Devices Global Unlimited Company | Eye gaze tracking |
CN110268448B (en) | 2017-02-20 | 2023-11-24 | 交互数字Vc控股公司 | Dynamically presenting augmented reality information to reduce peak cognitive demands |
JP7027856B2 (en) * | 2017-02-21 | 2022-03-02 | 株式会社リコー | Display devices and equipment |
WO2018154576A1 (en) | 2017-02-22 | 2018-08-30 | Lumus Ltd. | Light guide optical assembly |
JP7158395B2 (en) | 2017-02-23 | 2022-10-21 | マジック リープ, インコーポレイテッド | Variable focus imaging device based on polarization conversion |
CN106980983A (en) | 2017-02-23 | 2017-07-25 | 阿里巴巴集团控股有限公司 | Service authentication method and device based on virtual reality scenario |
JP6774603B2 (en) * | 2017-03-06 | 2020-10-28 | 株式会社Jvcケンウッド | Laser light irradiation detection device, laser light irradiation detection method, laser light irradiation detection system |
EP3593183B1 (en) | 2017-03-07 | 2022-02-16 | Apple Inc. | Head-mounted display system |
AU2018231083B2 (en) * | 2017-03-09 | 2023-03-09 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Head-mounted light field display with integral imaging and waveguide prism |
US12078802B2 (en) | 2017-03-09 | 2024-09-03 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Head-mounted light field display with integral imaging and relay optics |
CN107452031B (en) * | 2017-03-09 | 2020-06-26 | 叠境数字科技(上海)有限公司 | Virtual ray tracking method and light field dynamic refocusing display system |
CN110402350B (en) | 2017-03-14 | 2022-05-24 | 奇跃公司 | Waveguide having light absorbing film and method of forming the same |
US10502948B2 (en) | 2017-03-15 | 2019-12-10 | Magic Leap, Inc. | Techniques for improving a fiber scanning system |
IL251189A0 (en) | 2017-03-15 | 2017-06-29 | Ophir Yoav | Gradual transitioning between two-dimensional and theree-dimensional augmented reality images |
CN110392844B (en) * | 2017-03-16 | 2024-03-12 | 特里纳米克斯股份有限公司 | Detector for optically detecting at least one object |
AU2018234921B2 (en) | 2017-03-17 | 2021-10-07 | Magic Leap, Inc. | Mixed reality system with color virtual content warping and method of generating virtual content using same |
CN110419061B (en) | 2017-03-17 | 2023-09-29 | 奇跃公司 | Mixed reality system and method for generating virtual content using the same |
US10769752B2 (en) | 2017-03-17 | 2020-09-08 | Magic Leap, Inc. | Mixed reality system with virtual content warping and method of generating virtual content using same |
CA3055597A1 (en) | 2017-03-17 | 2018-09-20 | Magic Leap, Inc. | Room layout estimation methods and techniques |
EP3602173A4 (en) | 2017-03-21 | 2021-01-13 | Magic Leap, Inc. | Stacked waveguides having different diffraction gratings for combined field of view |
US11079603B2 (en) * | 2017-03-21 | 2021-08-03 | Magic Leap, Inc. | Display system with spatial light modulator illumination for divided pupils |
CN113568508B (en) | 2017-03-21 | 2024-07-26 | 奇跃公司 | Depth sensing techniques for virtual, augmented and mixed reality systems |
AU2018239525A1 (en) | 2017-03-21 | 2019-10-17 | Magic Leap, Inc. | Methods, devices, and systems for illuminating spatial light modulators |
CN114326278A (en) * | 2017-03-21 | 2022-04-12 | 奇跃公司 | Method and system for fiber scanning projector |
IL301719B1 (en) * | 2017-03-21 | 2024-08-01 | Magic Leap Inc | Method and system for tracking eye movement in conjunction with a light scanning projector |
KR102576133B1 (en) | 2017-03-21 | 2023-09-07 | 매직 립, 인코포레이티드 | Eye-imaging device using diffractive optical elements |
IL269289B2 (en) * | 2017-03-21 | 2024-10-01 | Magic Leap Inc | Low-profile beam splitter |
TWI751262B (en) | 2017-03-22 | 2022-01-01 | 以色列商魯姆斯有限公司 | Overlapping facets |
IL300511A (en) | 2017-03-22 | 2023-04-01 | Magic Leap Inc | Depth based foveated rendering for display systems |
IL269103B2 (en) * | 2017-03-22 | 2024-07-01 | Magic Leap Inc | Dynamic field of view variable focus display system |
US10748333B2 (en) | 2017-03-23 | 2020-08-18 | Nvidia Corporation | Finite aperture omni-directional stereo light transport |
US10277943B2 (en) | 2017-03-27 | 2019-04-30 | Microsoft Technology Licensing, Llc | Selective rendering of sparse peripheral displays based on user movements |
US10216260B2 (en) | 2017-03-27 | 2019-02-26 | Microsoft Technology Licensing, Llc | Selective rendering of sparse peripheral displays based on element saliency |
JP7175281B2 (en) * | 2017-03-28 | 2022-11-18 | マジック リープ, インコーポレイテッド | Augmented reality system with spatialized audio associated with user-scanned virtual objects |
CN106959514B (en) * | 2017-03-29 | 2021-09-14 | 联想(北京)有限公司 | Head-mounted equipment |
EP3602182A1 (en) | 2017-03-31 | 2020-02-05 | Universiteit Gent | Integrated near-eye display |
CN106933022A (en) * | 2017-04-01 | 2017-07-07 | 深圳优立全息科技有限公司 | A kind of virtual reality Interactive Experience device |
DE102017107346A1 (en) * | 2017-04-05 | 2018-10-11 | Carl Zeiss Ag | Device for power supply of and / or communication with an eye implant by means of illumination radiation |
IL251645B (en) | 2017-04-06 | 2018-08-30 | Lumus Ltd | Light-guide optical element and method of its manufacture |
WO2018184718A1 (en) * | 2017-04-06 | 2018-10-11 | Konstantin Roggatz | Augmented reality (ar) glasses and method for mixing virtual images into an image visible to a wearer of the glasses through at least one glasses lens |
EP3385219B1 (en) | 2017-04-07 | 2021-07-14 | InterDigital CE Patent Holdings | Method for manufacturing a device for forming at least one focused beam in a near zone |
KR102629577B1 (en) | 2017-04-18 | 2024-01-24 | 매직 립, 인코포레이티드 | Waveguides having reflective layers formed by reflective flowable materials |
CN113608617A (en) | 2017-04-19 | 2021-11-05 | 奇跃公司 | Multi-modal task execution and text editing for wearable systems |
CN107123096B (en) * | 2017-04-20 | 2018-11-06 | 腾讯科技(深圳)有限公司 | Method for displaying image and device in VR equipment and VR equipment |
US11474354B2 (en) * | 2017-04-25 | 2022-10-18 | Ati Technologies Ulc | Display pacing in multi-head mounted display virtual reality configurations |
IL310727A (en) | 2017-04-27 | 2024-04-01 | Magic Leap Inc | Light-emitting user input device |
CN110573992B (en) * | 2017-04-27 | 2023-07-18 | 西门子股份公司 | Editing augmented reality experiences using augmented reality and virtual reality |
US10854146B1 (en) * | 2017-04-28 | 2020-12-01 | Cloudminds (Shenzhen) Robotics Systems Co., Ltd. | Directional optical waveguide, directional backlight module, and display device |
US10409074B2 (en) | 2017-05-03 | 2019-09-10 | Microsoft Technology Licensing, Llc | Near-to-eye display with steerable phased arrays |
US10412378B2 (en) | 2017-05-08 | 2019-09-10 | Microsoft Technology Licensing, Llc | Resonating optical waveguide using multiple diffractive optical elements |
US10386923B2 (en) | 2017-05-08 | 2019-08-20 | International Business Machines Corporation | Authenticating users and improving virtual reality experiences via ocular scans and pupillometry |
US20200201038A1 (en) * | 2017-05-15 | 2020-06-25 | Real View Imaging Ltd. | System with multiple displays and methods of use |
CN108873326A (en) | 2017-05-16 | 2018-11-23 | 中强光电股份有限公司 | Head-mounted display apparatus |
IL270459B2 (en) * | 2017-05-16 | 2024-04-01 | Magic Leap Inc | Systems and methods for mixed reality |
EP3625658B1 (en) | 2017-05-19 | 2024-10-09 | Magic Leap, Inc. | Keyboards for virtual, augmented, and mixed reality display systems |
WO2018211074A1 (en) * | 2017-05-19 | 2018-11-22 | Seereal Technologies S.A. | Display device comprising a light guide |
IL294143B2 (en) | 2017-05-22 | 2023-12-01 | Magic Leap Inc | Pairing with companion device |
US10855977B2 (en) * | 2017-05-26 | 2020-12-01 | Google Llc | Near-eye display with extended accommodation range adjustment |
WO2018215834A1 (en) * | 2017-05-26 | 2018-11-29 | Spectrum Optix, Inc. | Reflective truncated ball imaging system |
US10764552B2 (en) | 2017-05-26 | 2020-09-01 | Google Llc | Near-eye display with sparse sampling super-resolution |
US10222615B2 (en) | 2017-05-26 | 2019-03-05 | Microsoft Technology Licensing, Llc | Optical waveguide with coherent light source |
EP3631559A1 (en) * | 2017-05-26 | 2020-04-08 | Google LLC | Near-eye display with sparse sampling super-resolution |
US10869517B1 (en) | 2017-05-28 | 2020-12-22 | Nexus House LLC | Folding hat with integrated display system |
WO2018222618A1 (en) | 2017-05-30 | 2018-12-06 | Magic Leap, Inc. | Power supply assembly with fan assembly for electronic device |
EP3631567B1 (en) | 2017-05-31 | 2022-09-21 | Magic Leap, Inc. | Eye tracking calibration techniques |
CN108984075B (en) * | 2017-05-31 | 2021-09-07 | 华为技术有限公司 | Display mode switching method and device and terminal |
US10613413B1 (en) | 2017-05-31 | 2020-04-07 | Facebook Technologies, Llc | Ultra-wide field-of-view scanning devices for depth sensing |
US11119353B2 (en) | 2017-06-01 | 2021-09-14 | E-Vision Smart Optics, Inc. | Switchable micro-lens array for augmented reality and mixed reality |
US10921613B2 (en) * | 2017-06-01 | 2021-02-16 | NewSight Reality, Inc. | Near eye display and related computer-implemented software and firmware |
US10634921B2 (en) * | 2017-06-01 | 2020-04-28 | NewSight Reality, Inc. | See-through near eye optical display |
JP6812579B2 (en) * | 2017-06-02 | 2021-01-13 | アップル インコーポレイテッドApple Inc. | Methods and devices for detecting planes and / or quadtrees for use as virtual substrates |
US10409001B2 (en) | 2017-06-05 | 2019-09-10 | Applied Materials, Inc. | Waveguide fabrication with sacrificial sidewall spacers |
EP3639510A4 (en) * | 2017-06-12 | 2021-03-24 | Magic Leap, Inc. | Augmented reality display having multi-element adaptive lens for changing depth planes |
KR102365726B1 (en) * | 2017-06-13 | 2022-02-22 | 한국전자통신연구원 | Method for providing composite image based on optical see-through and apparatus using the same |
US10712567B2 (en) | 2017-06-15 | 2020-07-14 | Microsoft Technology Licensing, Llc | Holographic display system |
CN107121787A (en) * | 2017-06-16 | 2017-09-01 | 北京灵犀微光科技有限公司 | three-dimensional imaging display device and method |
EP3639042A1 (en) * | 2017-06-16 | 2020-04-22 | Tektronix, Inc. | Test and measurement devices, systems and methods associated with augmented reality |
CN107065196B (en) * | 2017-06-16 | 2019-03-15 | 京东方科技集团股份有限公司 | A kind of augmented reality display device and augmented reality display methods |
US11598971B2 (en) * | 2017-06-21 | 2023-03-07 | Fusao Ishii | Image device with a compact homogenizer |
US10181200B1 (en) | 2017-06-28 | 2019-01-15 | Facebook Technologies, Llc | Circularly polarized illumination and detection for depth sensing |
KR102314789B1 (en) * | 2017-06-29 | 2021-10-20 | 에스케이텔레콤 주식회사 | Apparatus for displaying augmented reality contents |
US10908680B1 (en) | 2017-07-12 | 2021-02-02 | Magic Leap, Inc. | Pose estimation using electromagnetic tracking |
CN107219629B (en) * | 2017-07-14 | 2021-01-08 | 惠州Tcl移动通信有限公司 | Method, storage medium and device for preventing VR device dispersion by RGB superposition |
CN107277496B (en) * | 2017-07-17 | 2019-05-10 | 京东方科技集团股份有限公司 | Nearly eye field display system and control circuit |
RU2698919C2 (en) * | 2017-07-18 | 2019-09-02 | Святослав Иванович АРСЕНИЧ | Stereo display (embodiments), video camera for stereoscopic shooting and method for stereoscopic images computer formation for such stereo display |
TWI770234B (en) | 2017-07-19 | 2022-07-11 | 以色列商魯姆斯有限公司 | Lcos illumination via loe |
US20190025602A1 (en) * | 2017-07-20 | 2019-01-24 | Google Llc | Compact near-eye display optics for augmented reality |
CN111133294B (en) * | 2017-07-24 | 2024-05-24 | 宽腾矽公司 | Photonic structure for light suppression |
KR102461253B1 (en) * | 2017-07-24 | 2022-10-31 | 삼성전자주식회사 | Projection display apparatus including eye tracker |
CN110914790A (en) | 2017-07-26 | 2020-03-24 | 奇跃公司 | Training neural networks using representations of user interface devices |
US10578870B2 (en) | 2017-07-26 | 2020-03-03 | Magic Leap, Inc. | Exit pupil expander |
WO2019023489A1 (en) | 2017-07-28 | 2019-01-31 | Magic Leap, Inc. | Fan assembly for displaying an image |
US11122256B1 (en) | 2017-08-07 | 2021-09-14 | Apple Inc. | Mixed reality system |
EP3665013B1 (en) | 2017-08-09 | 2021-12-29 | Fathom Optics Inc. | Manufacturing light field prints |
CN107817471B (en) * | 2017-08-11 | 2021-07-20 | 北京圣威特科技有限公司 | Optical tracking method, device and system |
CN107479705B (en) * | 2017-08-14 | 2020-06-02 | 中国电子科技集团公司第二十八研究所 | Command institute collaborative operation electronic sand table system based on HoloLens |
CN107390365A (en) * | 2017-08-18 | 2017-11-24 | 联想(北京)有限公司 | A kind of imaging device, augmented reality display device and imaging method |
JP7231546B2 (en) | 2017-08-23 | 2023-03-01 | ソニーセミコンダクタソリューションズ株式会社 | Imaging element and imaging device |
CN107644443B (en) * | 2017-09-01 | 2020-07-28 | 北京七鑫易维信息技术有限公司 | Parameter setting method and device in sight tracking equipment |
US10521661B2 (en) | 2017-09-01 | 2019-12-31 | Magic Leap, Inc. | Detailed eye shape model for robust biometric applications |
US10574973B2 (en) | 2017-09-06 | 2020-02-25 | Facebook Technologies, Llc | Non-mechanical beam steering for depth sensing |
WO2019060283A1 (en) | 2017-09-20 | 2019-03-28 | Magic Leap, Inc. | Personalized neural network for eye tracking |
IL273397B2 (en) | 2017-09-21 | 2024-09-01 | Magic Leap Inc | Augmented reality display with waveguide configured to capture images of eye and/or environment |
US11857260B2 (en) * | 2017-09-21 | 2024-01-02 | Verily Life Sciences Llc | Retinal cameras having movable optical stops |
KR102481884B1 (en) | 2017-09-22 | 2022-12-28 | 삼성전자주식회사 | Method and apparatus for displaying a virtual image |
CN111051938B (en) | 2017-09-26 | 2022-11-25 | 苹果公司 | Display with volume phase grating |
EP3460561A1 (en) | 2017-09-26 | 2019-03-27 | Thomson Licensing | Device for deviating and focusing light |
US10890767B1 (en) | 2017-09-27 | 2021-01-12 | United Services Automobile Association (Usaa) | System and method for automatic vision correction in near-to-eye displays |
US11231581B2 (en) | 2017-09-27 | 2022-01-25 | Magic Leap, Inc. | Near eye 3D display with separate phase and amplitude modulators |
IL255049B (en) * | 2017-10-16 | 2022-08-01 | Oorym Optics Ltd | Highly efficient compact head-mounted display system |
US10867368B1 (en) | 2017-09-29 | 2020-12-15 | Apple Inc. | Foveated image capture for power efficient video see-through |
WO2019068304A1 (en) | 2017-10-02 | 2019-04-11 | CSEM Centre Suisse d'Electronique et de Microtechnique SA - Recherche et Développement | Resonant waveguide grating and applications thereof |
US10930709B2 (en) | 2017-10-03 | 2021-02-23 | Lockheed Martin Corporation | Stacked transparent pixel structures for image sensors |
US10788677B2 (en) * | 2017-10-03 | 2020-09-29 | Facebook Technologies, Llc | Fresnel assembly for light redirection in eye tracking systems |
WO2019072363A1 (en) | 2017-10-09 | 2019-04-18 | Huawei Technologies Co., Ltd. | Focus tunable optical system and multi-focal display device |
KR20240056672A (en) | 2017-10-11 | 2024-04-30 | 매직 립, 인코포레이티드 | Augmented reality display comprising eyepiece having a transparent emissive display |
IL273734B2 (en) | 2017-10-16 | 2024-11-01 | Oorym Optics Ltd | Highly Efficient Compact Head-Mounted Display System |
US10551625B2 (en) | 2017-10-16 | 2020-02-04 | Palo Alto Research Center Incorporated | Laser homogenizing and beam shaping illumination optical system and method |
CN116149058A (en) | 2017-10-16 | 2023-05-23 | 迪吉伦斯公司 | System and method for multiplying image resolution of pixellated display |
EP3697297A4 (en) | 2017-10-19 | 2020-12-16 | Facebook Technologies, Inc. | Systems and methods for identifying biological structures associated with neuromuscular source signals |
US11300788B2 (en) | 2017-10-23 | 2022-04-12 | Google Llc | Free space multiple laser diode modules |
WO2019084189A1 (en) | 2017-10-26 | 2019-05-02 | Magic Leap, Inc. | Gradient normalization systems and methods for adaptive loss balancing in deep multitask networks |
JP7273034B2 (en) | 2017-10-26 | 2023-05-12 | マジック リープ, インコーポレイテッド | Augmented reality display with liquid crystal variable focus element and roll-to-roll method and apparatus for forming same |
KR102698364B1 (en) | 2017-10-26 | 2024-08-23 | 매직 립, 인코포레이티드 | Wideband adaptive lens assembly for augmented reality displays |
IL310847A (en) | 2017-10-27 | 2024-04-01 | Magic Leap Inc | Virtual reticle for augmented reality systems |
TW201917447A (en) * | 2017-10-27 | 2019-05-01 | 廣達電腦股份有限公司 | Head-mounted display devices and methods for increasing color difference |
KR102507626B1 (en) * | 2017-10-31 | 2023-03-07 | 엘지디스플레이 주식회사 | Volumetric type 3-dimension display device |
CN109344677B (en) * | 2017-11-07 | 2021-01-15 | 长城汽车股份有限公司 | Method, device, vehicle and storage medium for recognizing three-dimensional object |
US10510812B2 (en) | 2017-11-09 | 2019-12-17 | Lockheed Martin Corporation | Display-integrated infrared emitter and sensor structures |
EP3710990A4 (en) | 2017-11-14 | 2021-10-27 | Magic Leap, Inc. | Meta-learning for multi-task learning for neural networks |
CN207965356U (en) * | 2017-11-14 | 2018-10-12 | 塔普翊海(上海)智能科技有限公司 | A kind of aobvious optical system of the see-through head of nearly eye |
US20190146152A1 (en) * | 2017-11-15 | 2019-05-16 | Source Photonics (Chengdu) Company , Ltd. | Waveguide Array Module and Receiver Optical Sub-Assembly |
US10867407B2 (en) * | 2017-11-15 | 2020-12-15 | Magic Leap, Inc. | System and methods for extrinsic calibration of cameras and diffractive optical elements |
IL274484B2 (en) * | 2017-11-22 | 2024-07-01 | Magic Leap Inc | Thermally actuated cantilevered beam optical scanner |
KR102182768B1 (en) | 2017-11-24 | 2020-11-25 | 주식회사 엘지화학 | Waveguide having light shielding and manufacturing method for the same |
WO2019104046A1 (en) * | 2017-11-27 | 2019-05-31 | University Of Central Florida Research | Optical display system, method, and applications |
KR102467882B1 (en) * | 2017-11-28 | 2022-11-16 | 엘지디스플레이 주식회사 | Personal immersion display device and driving method thereof |
CN109839742A (en) * | 2017-11-29 | 2019-06-04 | 深圳市掌网科技股份有限公司 | A kind of augmented reality device based on Eye-controlling focus |
CN108012139B (en) * | 2017-12-01 | 2019-11-29 | 北京理工大学 | The image generating method and device shown applied to the nearly eye of the sense of reality |
KR102005508B1 (en) * | 2017-12-01 | 2019-07-30 | 김태경 | Image display optical apparatus and image generation method thereof |
KR102436730B1 (en) * | 2017-12-06 | 2022-08-26 | 삼성전자주식회사 | Method and apparatus for estimating parameter of virtual screen |
KR102596429B1 (en) | 2017-12-10 | 2023-10-30 | 매직 립, 인코포레이티드 | Anti-reflection coatings on optical waveguides |
KR20240152954A (en) | 2017-12-11 | 2024-10-22 | 매직 립, 인코포레이티드 | Waveguide illuminator |
KR102601622B1 (en) | 2017-12-14 | 2023-11-10 | 매직 립, 인코포레이티드 | Contextual rendering of virtual avatars |
EP3724713A4 (en) | 2017-12-15 | 2021-08-25 | Magic Leap, Inc. | Enhanced pose determination for display device |
JP7407111B2 (en) | 2017-12-15 | 2023-12-28 | マジック リープ, インコーポレイテッド | Eyepiece for augmented reality display system |
DE102017130344A1 (en) | 2017-12-18 | 2019-06-19 | Carl Zeiss Ag | Optical system for transmitting a source image |
WO2019126331A1 (en) | 2017-12-20 | 2019-06-27 | Magic Leap, Inc. | Insert for augmented reality viewing device |
US10768431B2 (en) | 2017-12-20 | 2020-09-08 | Aperture In Motion, LLC | Light control devices and methods for regional variation of visual information and sampling |
US10175490B1 (en) | 2017-12-20 | 2019-01-08 | Aperture In Motion, LLC | Light control devices and methods for regional variation of visual information and sampling |
US10845594B1 (en) | 2017-12-21 | 2020-11-24 | Facebook Technologies, Llc | Prism based light redirection system for eye tracking systems |
CN108267856A (en) * | 2017-12-21 | 2018-07-10 | 成都理想境界科技有限公司 | A kind of augmented reality wears display equipment |
CN108072978A (en) * | 2017-12-21 | 2018-05-25 | 成都理想境界科技有限公司 | A kind of augmented reality wears display device |
CN108294739B (en) * | 2017-12-27 | 2021-02-09 | 苏州创捷传媒展览股份有限公司 | Method and device for testing user experience |
US10506220B2 (en) | 2018-01-02 | 2019-12-10 | Lumus Ltd. | Augmented reality displays with active alignment and corresponding methods |
KR102689506B1 (en) | 2018-01-04 | 2024-07-29 | 매직 립, 인코포레이티드 | Optical elements based on polymer structures containing inorganic materials |
US10914950B2 (en) | 2018-01-08 | 2021-02-09 | Digilens Inc. | Waveguide architectures and related methods of manufacturing |
KR20200108030A (en) | 2018-01-08 | 2020-09-16 | 디지렌즈 인코포레이티드. | System and method for high throughput recording of holographic gratings in waveguide cells |
WO2019135837A1 (en) | 2018-01-08 | 2019-07-11 | Digilens, Inc. | Systems and methods for manufacturing waveguide cells |
WO2019143844A1 (en) | 2018-01-17 | 2019-07-25 | Magic Leap, Inc. | Eye center of rotation determination, depth plane selection, and render camera positioning in display systems |
EP4339692A3 (en) * | 2018-01-17 | 2024-06-19 | Magic Leap, Inc. | Display systems and methods for determining registration between a display and a user's eyes |
US10477186B2 (en) * | 2018-01-17 | 2019-11-12 | Nextvr Inc. | Methods and apparatus for calibrating and/or adjusting the arrangement of cameras in a camera pair |
CN111869205B (en) * | 2018-01-19 | 2022-06-10 | Pcms控股公司 | Multiple focal planes with varying positions |
US10551544B2 (en) | 2018-01-21 | 2020-02-04 | Lumus Ltd. | Light-guide optical element with multiple-axis internal aperture expansion |
US10942355B2 (en) | 2018-01-22 | 2021-03-09 | Facebook Technologies, Llc | Systems, devices, and methods for tiled multi-monochromatic displays |
US10739595B2 (en) * | 2018-01-22 | 2020-08-11 | Facebook Technologies, Llc | Application specific integrated circuit for waveguide display |
US11907423B2 (en) | 2019-11-25 | 2024-02-20 | Meta Platforms Technologies, Llc | Systems and methods for contextualized interactions with an environment |
US11961494B1 (en) | 2019-03-29 | 2024-04-16 | Meta Platforms Technologies, Llc | Electromagnetic interference reduction in extended reality environments |
US11567627B2 (en) | 2018-01-30 | 2023-01-31 | Magic Leap, Inc. | Eclipse cursor for virtual content in mixed reality displays |
US10540941B2 (en) | 2018-01-30 | 2020-01-21 | Magic Leap, Inc. | Eclipse cursor for mixed reality displays |
JP7100333B2 (en) * | 2018-02-02 | 2022-07-13 | Fairy Devices株式会社 | Optical scanning image display device |
US10673414B2 (en) | 2018-02-05 | 2020-06-02 | Tectus Corporation | Adaptive tuning of a contact lens |
CN108366250B (en) * | 2018-02-06 | 2020-03-17 | 深圳市鹰硕技术有限公司 | Image display system, method and digital glasses |
KR102689931B1 (en) | 2018-02-06 | 2024-07-29 | 매직 립, 인코포레이티드 | Systems and methods for augmented reality |
US10652529B2 (en) | 2018-02-07 | 2020-05-12 | Lockheed Martin Corporation | In-layer Signal processing |
US10951883B2 (en) | 2018-02-07 | 2021-03-16 | Lockheed Martin Corporation | Distributed multi-screen array for high density display |
US10838250B2 (en) | 2018-02-07 | 2020-11-17 | Lockheed Martin Corporation | Display assemblies with electronically emulated transparency |
US10594951B2 (en) | 2018-02-07 | 2020-03-17 | Lockheed Martin Corporation | Distributed multi-aperture camera array |
US10690910B2 (en) | 2018-02-07 | 2020-06-23 | Lockheed Martin Corporation | Plenoptic cellular vision correction |
US10979699B2 (en) | 2018-02-07 | 2021-04-13 | Lockheed Martin Corporation | Plenoptic cellular imaging system |
US11616941B2 (en) | 2018-02-07 | 2023-03-28 | Lockheed Martin Corporation | Direct camera-to-display system |
US10129984B1 (en) | 2018-02-07 | 2018-11-13 | Lockheed Martin Corporation | Three-dimensional electronics distribution by geodesic faceting |
EP3729176A4 (en) | 2018-02-09 | 2021-09-22 | Vuzix Corporation | Image light guide with circular polarizer |
US10488666B2 (en) | 2018-02-10 | 2019-11-26 | Daqri, Llc | Optical waveguide devices, methods and systems incorporating same |
US10735649B2 (en) | 2018-02-22 | 2020-08-04 | Magic Leap, Inc. | Virtual and augmented reality systems and methods using display system control information embedded in image data |
CN108537111A (en) * | 2018-02-26 | 2018-09-14 | 阿里巴巴集团控股有限公司 | A kind of method, apparatus and equipment of In vivo detection |
JP7342008B2 (en) | 2018-02-27 | 2023-09-11 | マジック リープ, インコーポレイテッド | Mesh mates for virtual avatars |
JP7355748B2 (en) | 2018-02-28 | 2023-10-03 | マジック リープ, インコーポレイテッド | Head scan alignment using eye alignment |
US10866426B2 (en) | 2018-02-28 | 2020-12-15 | Apple Inc. | Scanning mirror display devices |
US20210364817A1 (en) * | 2018-03-05 | 2021-11-25 | Carnegie Mellon University | Display system for rendering a scene with multiple focal planes |
CN111819487A (en) | 2018-03-05 | 2020-10-23 | 奇跃公司 | Display system with low-delay pupil tracker |
US10802285B2 (en) | 2018-03-05 | 2020-10-13 | Invensas Corporation | Remote optical engine for virtual reality or augmented reality headsets |
WO2019173524A1 (en) | 2018-03-07 | 2019-09-12 | Magic Leap, Inc. | Visual tracking of peripheral devices |
CA3090818A1 (en) | 2018-03-07 | 2019-09-12 | Magic Leap, Inc. | Adaptive lens assemblies including polarization-selective lens stacks for augmented reality display |
AU2019236460B2 (en) | 2018-03-12 | 2024-10-03 | Magic Leap, Inc. | Tilting array based display |
US10878620B2 (en) | 2018-03-14 | 2020-12-29 | Magic Leap, Inc. | Display systems and methods for clipping content to increase viewing comfort |
KR102486664B1 (en) | 2018-03-14 | 2023-01-10 | 주식회사 엘지화학 | Module of diffractive light guide plate |
WO2019177870A1 (en) | 2018-03-15 | 2019-09-19 | Magic Leap, Inc. | Animating virtual avatar facial movements |
US10755676B2 (en) | 2018-03-15 | 2020-08-25 | Magic Leap, Inc. | Image correction due to deformation of components of a viewing device |
JP7235146B2 (en) * | 2018-03-16 | 2023-03-08 | 株式会社リコー | Head-mounted display and display system |
USD878420S1 (en) * | 2018-03-16 | 2020-03-17 | Magic Leap, Inc. | Display panel or portion thereof with a transitional mixed reality graphical user interface |
CN112136094A (en) | 2018-03-16 | 2020-12-25 | 奇跃公司 | Depth-based foveated rendering for display systems |
JP7027987B2 (en) * | 2018-03-16 | 2022-03-02 | 株式会社リコー | Head-mounted display device and display system |
WO2019178614A1 (en) | 2018-03-16 | 2019-09-19 | Digilens Inc. | Holographic waveguides incorporating birefringence control and methods for their fabrication |
WO2019177869A1 (en) | 2018-03-16 | 2019-09-19 | Magic Leap, Inc. | Facial expressions from eye-tracking cameras |
US11480467B2 (en) | 2018-03-21 | 2022-10-25 | Magic Leap, Inc. | Augmented reality system and method for spectroscopic analysis |
CN110297324B (en) * | 2018-03-21 | 2021-08-03 | 京东方科技集团股份有限公司 | Display device and vehicle |
RU2748645C1 (en) * | 2018-03-22 | 2021-05-28 | Сэн-Гобэн Гласс Франс | Projection system for windshield indicator (wsi) with areas of p-polarized light |
US11477434B2 (en) | 2018-03-23 | 2022-10-18 | Pcms Holdings, Inc. | Multifocal plane based method to produce stereoscopic viewpoints in a DIBR system (MFP-DIBR) |
US20210041693A1 (en) * | 2018-03-26 | 2021-02-11 | Sony Corporation | Information detection apparatus, video projection apparatus, information detection method, and video projection method |
CN108810519A (en) * | 2018-03-26 | 2018-11-13 | 成都理想境界科技有限公司 | A kind of 3D rendering display equipment |
CN112166367A (en) | 2018-03-26 | 2021-01-01 | Adlens有限公司 | Improvements in or relating to augmented reality display units and augmented reality headsets including the same |
FI128552B (en) * | 2018-03-28 | 2020-07-31 | Dispelix Oy | Waveguide display element with reflector surface |
JPWO2019187592A1 (en) * | 2018-03-29 | 2021-05-20 | ソニーグループ株式会社 | Information processing equipment, information processing methods and programs |
US20190045174A1 (en) * | 2018-03-29 | 2019-02-07 | Intel Corporation | Extended depth of focus integral displays |
CN108398791B (en) * | 2018-03-29 | 2022-11-25 | 陈超平 | Near-to-eye display device based on polarized contact lenses |
EP3776029A4 (en) | 2018-04-02 | 2022-06-29 | Magic Leap, Inc. | Hybrid polymer waveguide and methods for making the same |
JP7349445B2 (en) * | 2018-04-02 | 2023-09-22 | マジック リープ, インコーポレイテッド | Waveguides with integrated optical elements and methods for making the same |
WO2019195193A1 (en) | 2018-04-02 | 2019-10-10 | Magic Leap, Inc. | Waveguides having integrated spacers, waveguides having edge absorbers, and methods for making the same |
WO2019195390A1 (en) | 2018-04-03 | 2019-10-10 | Magic Leap, Inc. | Waveguide display with cantilevered light scanner |
EP3762762B1 (en) | 2018-04-03 | 2023-05-31 | Huawei Technologies Co., Ltd. | Display device for head-mounting and display method |
US10690922B2 (en) * | 2018-04-13 | 2020-06-23 | Facebook Technologies, Llc | Super-resolution scanning display for near-eye displays |
US11048082B1 (en) | 2018-04-13 | 2021-06-29 | Apple Inc. | Wireless bandwidth reduction with display data interleaving |
WO2019204164A1 (en) | 2018-04-16 | 2019-10-24 | Magic Leap, Inc. | Systems and methods for cross-application authoring, transfer, and evaluation of rigging control systems for virtual characters |
US11067805B2 (en) | 2018-04-19 | 2021-07-20 | Magic Leap, Inc. | Systems and methods for operating a display system based on user perceptibility |
FI129306B (en) | 2018-04-19 | 2021-11-30 | Dispelix Oy | Diffractive exit pupil expander arrangement for display applications |
CN108333781B (en) * | 2018-04-20 | 2023-10-27 | 深圳创维新世界科技有限公司 | Near-to-eye display system |
CN108333780A (en) * | 2018-04-20 | 2018-07-27 | 深圳创维新世界科技有限公司 | Near-eye display system |
US10505394B2 (en) | 2018-04-21 | 2019-12-10 | Tectus Corporation | Power generation necklaces that mitigate energy absorption in the human body |
US10318811B1 (en) * | 2018-04-22 | 2019-06-11 | Bubbler International Llc | Methods and systems for detecting objects by non-visible radio frequencies and displaying associated augmented reality effects |
WO2019209431A1 (en) | 2018-04-23 | 2019-10-31 | Magic Leap, Inc. | Avatar facial expression representation in multidimensional space |
US10838239B2 (en) | 2018-04-30 | 2020-11-17 | Tectus Corporation | Multi-coil field generation in an electronic contact lens system |
US10895762B2 (en) | 2018-04-30 | 2021-01-19 | Tectus Corporation | Multi-coil field generation in an electronic contact lens system |
WO2019212698A1 (en) | 2018-05-01 | 2019-11-07 | Magic Leap, Inc. | Avatar animation using markov decision process policies |
US10295723B1 (en) * | 2018-05-01 | 2019-05-21 | Facebook Technologies, Llc | 2D pupil expander using holographic Bragg grating |
US11308673B2 (en) | 2018-05-03 | 2022-04-19 | Magic Leap, Inc. | Using three-dimensional scans of a physical subject to determine positions and/or orientations of skeletal joints in the rigging for a virtual character |
US10783230B2 (en) | 2018-05-09 | 2020-09-22 | Shape Matrix Geometric Instruments, LLC | Methods and apparatus for encoding passwords or other information |
US10747309B2 (en) * | 2018-05-10 | 2020-08-18 | Microsoft Technology Licensing, Llc | Reconfigurable optics for switching between near-to-eye display modes |
EP3625617B1 (en) | 2018-05-14 | 2023-09-06 | Lumus Ltd. | Projector configuration with subdivided optical aperture for near-eye displays, and corresponding optical systems |
WO2019222215A1 (en) * | 2018-05-14 | 2019-11-21 | The Trustees Of Columbia University In The City Of New York | Micromachined waveguide and methods of making and using |
CN112236711A (en) * | 2018-05-17 | 2021-01-15 | 诺基亚技术有限公司 | Apparatus and method for image display |
IL278511B1 (en) | 2018-05-17 | 2024-09-01 | Lumus Ltd | Near-eye display having overlapping projector assemblies |
JP7328993B2 (en) | 2018-05-17 | 2023-08-17 | マジック リープ, インコーポレイテッド | Gradient Adversarial Training of Neural Networks |
CN108508616B (en) * | 2018-05-17 | 2024-04-16 | 成都工业学院 | 3D display system and 3D display device |
JP7079146B2 (en) * | 2018-05-18 | 2022-06-01 | シャープ株式会社 | 3D display device |
US10790700B2 (en) | 2018-05-18 | 2020-09-29 | Tectus Corporation | Power generation necklaces with field shaping systems |
US11282255B2 (en) | 2018-05-21 | 2022-03-22 | Magic Leap, Inc. | Generating textured polygon strip hair from strand-based hair for a virtual character |
US11210835B2 (en) | 2018-05-22 | 2021-12-28 | Magic Leap, Inc. | Computer generated hair groom transfer tool |
IL259518B2 (en) | 2018-05-22 | 2023-04-01 | Lumus Ltd | Optical system and method for improvement of light field uniformity |
JP7516260B2 (en) | 2018-05-22 | 2024-07-16 | マジック リープ, インコーポレイテッド | A skeletal system for animating virtual avatars |
CN112424727A (en) | 2018-05-22 | 2021-02-26 | 奇跃公司 | Cross-modal input fusion for wearable systems |
AU2019274687B2 (en) | 2018-05-23 | 2023-05-11 | Lumus Ltd. | Optical system including light-guide optical element with partially-reflective internal surfaces |
US11625864B2 (en) | 2018-05-25 | 2023-04-11 | Magic Leap, Inc. | Compression of dynamic unstructured point clouds |
KR102558106B1 (en) * | 2018-05-30 | 2023-07-21 | 엘지디스플레이 주식회사 | Display panel, display device and virtual reality/augmented reality device |
CN112236713B (en) * | 2018-05-30 | 2023-01-24 | 奇跃公司 | Compact variable focus configuration |
CN110554593B (en) * | 2018-05-31 | 2021-01-26 | 京东方科技集团股份有限公司 | Holographic optical element, manufacturing method thereof, image reconstruction method and augmented reality glasses |
CN112601975B (en) | 2018-05-31 | 2024-09-06 | 奇跃公司 | Radar head pose positioning |
WO2019232075A1 (en) | 2018-06-01 | 2019-12-05 | Magic Leap, Inc. | Compression of dynamic unstructured point clouds |
US10825424B2 (en) | 2018-06-05 | 2020-11-03 | Magic Leap, Inc. | Homography transformation matrices based temperature calibration of a viewing system |
US11157159B2 (en) | 2018-06-07 | 2021-10-26 | Magic Leap, Inc. | Augmented reality scrollbar |
US11092812B2 (en) | 2018-06-08 | 2021-08-17 | Magic Leap, Inc. | Augmented reality viewer with automated surface selection placement and content orientation placement |
JP7336470B2 (en) | 2018-06-15 | 2023-08-31 | マジック リープ, インコーポレイテッド | Wide-field polarization switch using liquid crystal optical element with pretilt angle |
CN112272789B (en) * | 2018-06-15 | 2022-10-04 | 大陆汽车有限责任公司 | Device for generating virtual images with variable projection distance |
US12007652B2 (en) | 2018-06-15 | 2024-06-11 | Magic Leap, Inc. | Wide field-of-view polarization switches and methods of fabricating liquid crystal optical elements with pretilt |
WO2019246129A2 (en) | 2018-06-18 | 2019-12-26 | Magic Leap, Inc. | Augmented reality display with frame modulation functionality |
WO2019246058A1 (en) | 2018-06-18 | 2019-12-26 | Magic Leap, Inc. | Systems and methods for temporarily disabling user control interfaces during attachment of an electronic device |
WO2019246044A1 (en) | 2018-06-18 | 2019-12-26 | Magic Leap, Inc. | Head-mounted display systems with power saving functionality |
WO2020005757A1 (en) | 2018-06-26 | 2020-01-02 | Magic Leap, Inc. | Waypoint creation in map detection |
US11415812B2 (en) | 2018-06-26 | 2022-08-16 | Lumus Ltd. | Compact collimating optical device and system |
CN108932058B (en) * | 2018-06-29 | 2021-05-18 | 联想(北京)有限公司 | Display method and device and electronic equipment |
WO2020010097A1 (en) * | 2018-07-02 | 2020-01-09 | Magic Leap, Inc. | Pixel intensity modulation using modifying gain values |
CN112602090A (en) | 2018-07-02 | 2021-04-02 | 奇跃公司 | Method and system for interpolating different inputs |
WO2020010226A1 (en) | 2018-07-03 | 2020-01-09 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality |
US11856479B2 (en) | 2018-07-03 | 2023-12-26 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality along a route with markers |
EP4462289A2 (en) | 2018-07-05 | 2024-11-13 | Magic Leap, Inc. | Waveguide-based illumination for head mounted display system |
CN118158376A (en) | 2018-07-05 | 2024-06-07 | Pcms控股公司 | Method and system for near-eye focal plane overlay |
CN109001907A (en) * | 2018-07-06 | 2018-12-14 | 成都理想境界科技有限公司 | A kind of high-resolution display module |
US11302156B1 (en) * | 2018-07-06 | 2022-04-12 | Amazon Technologies, Inc. | User interfaces associated with device applications |
US11624937B2 (en) | 2018-07-07 | 2023-04-11 | Acucela Inc. | Device to prevent retinal hypoxia |
WO2020014038A2 (en) * | 2018-07-11 | 2020-01-16 | Pure Depth Inc. | Ghost multi-layer and single layer display systems |
WO2020014700A1 (en) | 2018-07-13 | 2020-01-16 | Magic Leap, Inc. | Systems and methods for display binocular deformation compensation |
US11137622B2 (en) | 2018-07-15 | 2021-10-05 | Tectus Corporation | Eye-mounted displays including embedded conductive coils |
WO2020018938A1 (en) | 2018-07-19 | 2020-01-23 | Magic Leap, Inc. | Content interaction driven by eye metrics |
EP3827300A4 (en) | 2018-07-23 | 2021-09-15 | Magic Leap, Inc. | Method and system for resolving hemisphere ambiguity using a position vector |
KR102578653B1 (en) * | 2018-07-23 | 2023-09-15 | 삼성전자 주식회사 | Electronic device capable of providing multi focus to light of display |
WO2020023383A1 (en) | 2018-07-23 | 2020-01-30 | Magic Leap, Inc. | Mixed reality system with virtual content warping and method of generating virtual content using same |
JP7562504B2 (en) | 2018-07-23 | 2024-10-07 | マジック リープ, インコーポレイテッド | Deep Predictor Recurrent Neural Networks for Head Pose Prediction |
US11145127B2 (en) | 2018-07-23 | 2021-10-12 | Magic Leap, Inc. | System and method for mapping |
EP3827584A4 (en) | 2018-07-23 | 2021-09-08 | Magic Leap, Inc. | Intra-field sub code timing in field sequential displays |
US11627587B2 (en) | 2018-07-23 | 2023-04-11 | Magic Leap, Inc. | Coexistence interference avoidance between two different radios operating in the same band |
US11624929B2 (en) | 2018-07-24 | 2023-04-11 | Magic Leap, Inc. | Viewing device with dust seal integration |
CN112689869A (en) | 2018-07-24 | 2021-04-20 | 奇跃公司 | Display system and method for determining registration between a display and an eye of a user |
USD930614S1 (en) | 2018-07-24 | 2021-09-14 | Magic Leap, Inc. | Totem controller having an illumination region |
WO2020023582A1 (en) | 2018-07-24 | 2020-01-30 | Magic Leap, Inc. | Methods and apparatuses for determining and/or evaluating localizing maps of image display devices |
EP3827224B1 (en) | 2018-07-24 | 2023-09-06 | Magic Leap, Inc. | Temperature dependent calibration of movement detection devices |
USD924204S1 (en) | 2018-07-24 | 2021-07-06 | Magic Leap, Inc. | Totem controller having an illumination region |
USD918176S1 (en) | 2018-07-24 | 2021-05-04 | Magic Leap, Inc. | Totem controller having an illumination region |
JP7456995B2 (en) | 2018-07-24 | 2024-03-27 | マジック リープ, インコーポレイテッド | Display system and method for determining vertical alignment between left and right displays and a user's eyes |
US12099386B2 (en) | 2018-07-24 | 2024-09-24 | Magic Leap, Inc. | Thermal management system for electronic device |
EP3827294A4 (en) | 2018-07-24 | 2022-04-20 | Magic Leap, Inc. | Diffractive optical elements with mitigation of rebounce-induced light loss and related systems and methods |
WO2020023531A1 (en) | 2018-07-24 | 2020-01-30 | Magic Leap, Inc. | Methods and apparatuses for corner detection |
WO2020023404A1 (en) | 2018-07-24 | 2020-01-30 | Magic Leap, Inc. | Flicker mitigation when toggling eyepiece display illumination in augmented reality systems |
CN108919492B (en) * | 2018-07-25 | 2021-05-07 | 京东方科技集团股份有限公司 | Near-to-eye display device, system and display method |
US11402801B2 (en) | 2018-07-25 | 2022-08-02 | Digilens Inc. | Systems and methods for fabricating a multilayer optical structure |
US10950024B2 (en) | 2018-07-27 | 2021-03-16 | Magic Leap, Inc. | Pose space dimensionality reduction for pose space deformation of a virtual character |
CN109116577B (en) * | 2018-07-30 | 2020-10-20 | 杭州光粒科技有限公司 | Holographic contact lens and application thereof |
EP3830636A4 (en) | 2018-07-30 | 2022-04-13 | Acucela Inc. | Optical designs of electronic contact lens to decrease myopia progression |
US11353767B2 (en) * | 2018-07-30 | 2022-06-07 | Facebook Technologies, Llc | Varifocal system using hybrid tunable liquid crystal lenses |
US10712837B1 (en) * | 2018-07-30 | 2020-07-14 | David Douglas | Using geo-registered tools to manipulate three-dimensional medical images |
JP7401519B2 (en) | 2018-08-02 | 2023-12-19 | マジック リープ, インコーポレイテッド | Visual recognition system with interpupillary distance compensation based on head motion |
US11468640B2 (en) | 2018-08-03 | 2022-10-11 | Magic Leap, Inc. | Depth plane selection for multi-depth plane display systems by user categorization |
EP3830631A4 (en) | 2018-08-03 | 2021-10-27 | Magic Leap, Inc. | Unfused pose-based drift correction of a fused pose of a totem in a user interaction system |
US10955677B1 (en) | 2018-08-06 | 2021-03-23 | Apple Inc. | Scene camera |
CN108965857A (en) * | 2018-08-09 | 2018-12-07 | 张家港康得新光电材料有限公司 | A kind of stereo display method and device, wearable stereoscopic display |
CN110825280A (en) * | 2018-08-09 | 2020-02-21 | 北京微播视界科技有限公司 | Method, apparatus and computer-readable storage medium for controlling position movement of virtual object |
US10778963B2 (en) | 2018-08-10 | 2020-09-15 | Valve Corporation | Head-mounted display (HMD) with spatially-varying retarder optics |
US10996463B2 (en) | 2018-08-10 | 2021-05-04 | Valve Corporation | Head-mounted display (HMD) with spatially-varying retarder optics |
US10957112B2 (en) | 2018-08-13 | 2021-03-23 | Magic Leap, Inc. | Cross reality system |
US11227435B2 (en) | 2018-08-13 | 2022-01-18 | Magic Leap, Inc. | Cross reality system |
KR102605397B1 (en) * | 2018-08-20 | 2023-11-24 | 삼성디스플레이 주식회사 | Device for providing augmented reality |
EP3840645A4 (en) | 2018-08-22 | 2021-10-20 | Magic Leap, Inc. | Patient viewing system |
EP4212929A1 (en) | 2018-08-26 | 2023-07-19 | Lumus Ltd. | Reflection suppression in near eye displays |
US10699383B2 (en) | 2018-08-27 | 2020-06-30 | Nvidia Corp. | Computational blur for varifocal displays |
GB2576738B (en) * | 2018-08-29 | 2020-08-19 | Envisics Ltd | Head-up display |
CN117238224A (en) | 2018-08-31 | 2023-12-15 | 奇跃公司 | Spatially resolved dynamic dimming for augmented reality devices |
US11914148B2 (en) * | 2018-09-07 | 2024-02-27 | Adeia Semiconductor Inc. | Stacked optical waveguides |
DE102018215272A1 (en) * | 2018-09-07 | 2020-03-12 | Bayerische Motoren Werke Aktiengesellschaft | Method for operating a field of view display device for a motor vehicle |
WO2020049542A1 (en) | 2018-09-09 | 2020-03-12 | Lumus Ltd. | Optical systems including light-guide optical elements with two-dimensional expansion |
US10529107B1 (en) | 2018-09-11 | 2020-01-07 | Tectus Corporation | Projector alignment in a contact lens |
US11141645B2 (en) | 2018-09-11 | 2021-10-12 | Real Shot Inc. | Athletic ball game using smart glasses |
US11103763B2 (en) | 2018-09-11 | 2021-08-31 | Real Shot Inc. | Basketball shooting game using smart glasses |
JP2020042212A (en) * | 2018-09-12 | 2020-03-19 | ソニー株式会社 | Display unit, display control method, and recording medium |
US11428942B2 (en) | 2018-09-14 | 2022-08-30 | Magic Leap, Inc. | Systems and methods for external light management |
USD934873S1 (en) | 2018-09-18 | 2021-11-02 | Magic Leap, Inc. | Mobile computing support system having an illumination region |
USD955396S1 (en) | 2018-09-18 | 2022-06-21 | Magic Leap, Inc. | Mobile computing support system having an illumination region |
USD934872S1 (en) | 2018-09-18 | 2021-11-02 | Magic Leap, Inc. | Mobile computing support system having an illumination region |
USD950567S1 (en) | 2018-09-18 | 2022-05-03 | Magic Leap, Inc. | Mobile computing support system having an illumination region |
KR20210043670A (en) | 2018-09-25 | 2021-04-21 | 애플 인크. | Camera lens system |
US11520044B2 (en) | 2018-09-25 | 2022-12-06 | Waymo Llc | Waveguide diffusers for LIDARs |
US11733523B2 (en) | 2018-09-26 | 2023-08-22 | Magic Leap, Inc. | Diffractive optical elements with optical power |
US12050318B2 (en) | 2018-09-26 | 2024-07-30 | Magic Leap, Inc. | Eyewear with pinhole and slit cameras |
JP7086392B2 (en) * | 2018-10-01 | 2022-06-20 | 株式会社Qdレーザ | Image projection device and image projection method |
EP3633438A1 (en) | 2018-10-01 | 2020-04-08 | InterDigital CE Patent Holdings | Inhomogeneous microlens device for near-field focusing, beam forming, and high-efficiency far-field device implementation |
CN113196209A (en) | 2018-10-05 | 2021-07-30 | 奇跃公司 | Rendering location-specific virtual content at any location |
CN111007589A (en) * | 2018-10-08 | 2020-04-14 | 成都理想境界科技有限公司 | Waveguide module, display module based on waveguide and near-to-eye display equipment |
CN111077670B (en) * | 2018-10-18 | 2022-02-18 | 中强光电股份有限公司 | Light transmission module and head-mounted display device |
JP7455121B2 (en) | 2018-10-26 | 2024-03-25 | マジック リープ, インコーポレイテッド | Ambient electromagnetic distortion correction for electromagnetic tracking |
WO2020090487A1 (en) * | 2018-10-30 | 2020-05-07 | 国立大学法人横浜国立大学 | Prism lens, light deflection device, and lidar apparatus |
US11262585B2 (en) * | 2018-11-01 | 2022-03-01 | Google Llc | Optical combiner lens with spacers between lens and lightguide |
KR102099785B1 (en) | 2018-11-06 | 2020-04-10 | 주식회사 레티널 | Optical device for augmented reality |
KR102140733B1 (en) | 2018-11-06 | 2020-08-03 | 주식회사 레티널 | Optical device for augmented reality |
WO2020102030A1 (en) | 2018-11-12 | 2020-05-22 | Magic Leap, Inc. | Multi-depth exit pupil expander |
DE102018219474A1 (en) * | 2018-11-15 | 2020-05-20 | Robert Bosch Gmbh | Method and arrangement for performing a virtual retinal display |
EP3881232A4 (en) | 2018-11-15 | 2022-08-10 | Magic Leap, Inc. | Deep neural network pose estimation system |
CN109348210A (en) * | 2018-11-16 | 2019-02-15 | 成都理想境界科技有限公司 | Image source mould group, near-eye display system, control method and near-eye display device |
EP3881279A4 (en) | 2018-11-16 | 2022-08-17 | Magic Leap, Inc. | Image size triggered clarification to maintain image sharpness |
US11237393B2 (en) | 2018-11-20 | 2022-02-01 | Magic Leap, Inc. | Eyepieces for augmented reality display system |
US10838232B2 (en) | 2018-11-26 | 2020-11-17 | Tectus Corporation | Eye-mounted displays including embedded solenoids |
EP3886693A4 (en) | 2018-11-27 | 2022-06-08 | Facebook Technologies, LLC. | Methods and apparatus for autocalibration of a wearable electrode sensor system |
CN111240145B (en) * | 2018-11-29 | 2022-04-15 | 青岛海信激光显示股份有限公司 | Light valve driving control method and projection equipment |
EP3887887A4 (en) | 2018-11-30 | 2022-01-19 | Magic Leap, Inc. | Methods and systems for high efficiency eyepiece in augmented reality devices |
US11199912B2 (en) | 2018-11-30 | 2021-12-14 | Magic Leap, Inc. | Multi-modal hand location and orientation for avatar movement |
US10866413B2 (en) | 2018-12-03 | 2020-12-15 | Lockheed Martin Corporation | Eccentric incident luminance pupil tracking |
JP7330901B2 (en) * | 2018-12-04 | 2023-08-22 | 京東方科技集團股▲ふん▼有限公司 | Display panel, display device and display method |
WO2020123561A1 (en) | 2018-12-10 | 2020-06-18 | Daqri, Llc | Adaptive viewports for hypervocal viewport (hvp) displays |
US11125993B2 (en) | 2018-12-10 | 2021-09-21 | Facebook Technologies, Llc | Optical hyperfocal reflective systems and methods, and augmented reality and/or virtual reality displays incorporating same |
JP7153087B2 (en) * | 2018-12-11 | 2022-10-13 | 富士フイルム株式会社 | Light guide element, image display device and sensing device |
US11107972B2 (en) | 2018-12-11 | 2021-08-31 | Facebook Technologies, Llc | Nanovoided tunable optics |
US11409240B2 (en) * | 2018-12-17 | 2022-08-09 | Meta Platforms Technologies, Llc | Holographic pattern generation for head-mounted display (HMD) eye tracking using a diffractive optical element |
US10644543B1 (en) | 2018-12-20 | 2020-05-05 | Tectus Corporation | Eye-mounted display system including a head wearable object |
JP7161934B2 (en) * | 2018-12-21 | 2022-10-27 | 株式会社日立エルジーデータストレージ | Video display device and video display system |
CN113454507B (en) | 2018-12-21 | 2024-05-07 | 奇跃公司 | Cavitation structure for promoting total internal reflection within a waveguide |
KR102041261B1 (en) | 2018-12-28 | 2019-11-07 | 한국광기술원 | Reactive Multifocal Optical System and Augmented Reality Device Using the Same |
WO2020139752A1 (en) * | 2018-12-28 | 2020-07-02 | Magic Leap, Inc. | Variable pixel density display system with mechanically-actuated image projector |
JP7539386B2 (en) * | 2018-12-28 | 2024-08-23 | マジック リープ, インコーポレイテッド | Augmented and virtual reality display system with shared displays for left and right eyes - Patents.com |
KR102706916B1 (en) * | 2019-01-02 | 2024-09-20 | 삼성디스플레이 주식회사 | Device for providing augmented reality |
EP3908878A4 (en) | 2019-01-09 | 2022-04-06 | Facebook Technologies, LLC | Non-uniform sub-pupil reflectors and methods in optical waveguides for ar, hmd and hud applications |
CN113544570A (en) | 2019-01-11 | 2021-10-22 | 奇跃公司 | Time multiplexed display of virtual content at various depths |
CN113260889B (en) | 2019-01-15 | 2023-04-18 | 鲁姆斯有限公司 | Method for manufacturing symmetrical light guide optical element |
JP7259341B2 (en) * | 2019-01-18 | 2023-04-18 | セイコーエプソン株式会社 | PROJECTION DEVICE, PROJECTION SYSTEM, AND PROJECTION DEVICE CONTROL METHOD |
US11099406B2 (en) * | 2019-01-24 | 2021-08-24 | International Business Machines Corporation | Delivering augmented reality via embedded contact lens |
US11061254B2 (en) | 2019-01-24 | 2021-07-13 | International Business Machines Corporation | Adjusting contact lens prescription while wearing lens |
AU2020211092B2 (en) | 2019-01-24 | 2023-05-11 | Lumus Ltd. | Optical systems including LOE with three stage expansion |
EP3914997A4 (en) | 2019-01-25 | 2022-10-12 | Magic Leap, Inc. | Eye-tracking using images having different exposure times |
CN113383260B (en) * | 2019-01-31 | 2023-08-25 | 见真实股份有限公司 | Light field mixed reality system with correct monocular depth cues for viewers |
US12105289B2 (en) | 2019-02-01 | 2024-10-01 | Magic Leap, Inc. | Inline in-coupling optical elements |
CN113518961B (en) | 2019-02-06 | 2024-09-24 | 奇跃公司 | Clock speed determination and adjustment based on target intent |
US11237389B1 (en) * | 2019-02-11 | 2022-02-01 | Facebook Technologies, Llc | Wedge combiner for eye-tracking |
US20200264378A1 (en) | 2019-02-15 | 2020-08-20 | Digilens Inc. | Methods and Apparatuses for Providing a Holographic Waveguide Display Using Integrated Gratings |
KR101982098B1 (en) * | 2019-02-27 | 2019-05-24 | 주식회사 두리번 | Motion Sensor With Optical Fiber And Virtual/Augmented Reality System |
US12124050B2 (en) | 2019-02-28 | 2024-10-22 | Lumus Ltd. | Compact collimated image projector |
JP7518844B2 (en) | 2019-02-28 | 2024-07-18 | マジック リープ, インコーポレイテッド | Display system and method for providing variable accommodation cues using multiple intrapupillary parallax views formed by a light emitter array - Patents.com |
US11624906B2 (en) * | 2019-03-04 | 2023-04-11 | Microsoft Technology Licensing, Llc | IR illumination module for MEMS-based eye tracking |
US11360269B2 (en) * | 2019-03-04 | 2022-06-14 | Lumentum Operations Llc | High-power all fiber telescope |
JP7398131B2 (en) | 2019-03-12 | 2023-12-14 | ルムス エルティーディー. | image projector |
EP3939030A4 (en) | 2019-03-12 | 2022-11-30 | Magic Leap, Inc. | Registration of local content between first and second augmented reality viewers |
US11982813B2 (en) | 2019-03-12 | 2024-05-14 | Magic Leap, Inc. | Waveguides with high index materials and methods of fabrication thereof |
EP3938821A4 (en) | 2019-03-12 | 2023-04-26 | Digilens Inc. | Holographic waveguide backlight and related methods of manufacturing |
US12117630B2 (en) | 2019-03-12 | 2024-10-15 | Magic Leap, Inc. | Method of fabricating display device having patterned lithium-based transition metal oxide |
US20200301239A1 (en) * | 2019-03-18 | 2020-09-24 | Microsoft Technology Licensing, Llc | Varifocal display with fixed-focus lens |
US11846778B2 (en) * | 2019-03-20 | 2023-12-19 | Magic Leap, Inc. | System for providing illumination of the eye |
WO2020191224A1 (en) | 2019-03-20 | 2020-09-24 | Magic Leap, Inc. | System for collecting light |
KR101984616B1 (en) * | 2019-03-20 | 2019-06-03 | (주)락앤크리에이티브 | System for providing contents using images |
US11221487B2 (en) * | 2019-03-26 | 2022-01-11 | Kevin Chew Figueroa | Method and device of field sequential imaging for large field of view augmented/virtual reality |
CN111751987B (en) * | 2019-03-29 | 2023-04-14 | 托比股份公司 | Holographic eye imaging apparatus |
US10698201B1 (en) | 2019-04-02 | 2020-06-30 | Lockheed Martin Corporation | Plenoptic cellular axis redirection |
US10867543B2 (en) | 2019-04-09 | 2020-12-15 | Facebook Technologies, Llc | Resolution reduction of color channels of display devices |
US10861369B2 (en) * | 2019-04-09 | 2020-12-08 | Facebook Technologies, Llc | Resolution reduction of color channels of display devices |
US11016305B2 (en) | 2019-04-15 | 2021-05-25 | Magic Leap, Inc. | Sensor fusion for electromagnetic tracking |
EP3956604A4 (en) | 2019-04-15 | 2022-06-08 | Lumus Ltd. | Method of fabricating a light-guide optical element |
CN110069310B (en) * | 2019-04-23 | 2022-04-22 | 北京小米移动软件有限公司 | Method and device for switching desktop wallpaper and storage medium |
US10659772B1 (en) * | 2019-04-23 | 2020-05-19 | Disney Enterprises, Inc. | Augmented reality system for layering depth on head-mounted displays using external stereo screens |
KR102302159B1 (en) * | 2019-04-26 | 2021-09-14 | 주식회사 레티널 | Optical device for augmented reality preventing light leakage |
CN114127837A (en) | 2019-05-01 | 2022-03-01 | 奇跃公司 | Content providing system and method |
JP2022532688A (en) | 2019-05-06 | 2022-07-19 | ルーマス リミテッド | A transparent light guide for viewing scenes and near-eye displays |
JP7284292B2 (en) | 2019-05-10 | 2023-05-30 | トゥエンティ・トゥエンティ・セラピューティクス・エルエルシイ | A natural physio-optical user interface for intraocular microdisplays |
US11778856B2 (en) | 2019-05-15 | 2023-10-03 | Apple Inc. | Electronic device having emissive display with light recycling |
CN114127610A (en) | 2019-05-20 | 2022-03-01 | 奇跃公司 | Systems and techniques for estimating eye pose |
CN113874883A (en) | 2019-05-21 | 2021-12-31 | 奇跃公司 | Hand pose estimation |
US20200371472A1 (en) * | 2019-05-21 | 2020-11-26 | Light Field Lab, Inc. | Light Field Display System Based Commercial System |
WO2020242533A1 (en) | 2019-05-24 | 2020-12-03 | Arno Leon Konings | Annular axial flux motors |
JP2022532931A (en) | 2019-05-24 | 2022-07-20 | マジック リープ, インコーポレイテッド | Varifocal assembly |
CN110187506B (en) * | 2019-05-28 | 2021-12-17 | 京东方科技集团股份有限公司 | Optical display system and augmented reality device |
CN114174463A (en) | 2019-05-28 | 2022-03-11 | 奇跃公司 | Thermal management system for portable electronic devices |
USD962981S1 (en) | 2019-05-29 | 2022-09-06 | Magic Leap, Inc. | Display screen or portion thereof with animated scrollbar graphical user interface |
US20200386947A1 (en) | 2019-06-07 | 2020-12-10 | Digilens Inc. | Waveguides Incorporating Transmissive and Reflective Gratings and Related Methods of Manufacturing |
WO2020257469A1 (en) | 2019-06-20 | 2020-12-24 | Magic Leap, Inc. | Eyepieces for augmented reality display system |
US11803628B2 (en) | 2019-06-21 | 2023-10-31 | Magic Leap, Inc. | Secure authorization via modal window |
KR20220021904A (en) | 2019-06-23 | 2022-02-22 | 루머스 리미티드 | Display by Foveated Optical Correction |
US11049302B2 (en) | 2019-06-24 | 2021-06-29 | Realwear, Inc. | Photo redaction security system and related methods |
US11726317B2 (en) | 2019-06-24 | 2023-08-15 | Magic Leap, Inc. | Waveguides having integral spacers and related systems and methods |
JP7270481B2 (en) | 2019-06-25 | 2023-05-10 | 株式会社日立製作所 | vehicle control system |
TWI690745B (en) * | 2019-06-26 | 2020-04-11 | 點晶科技股份有限公司 | Multifunctional eyeglasses |
GB201909179D0 (en) * | 2019-06-26 | 2019-08-07 | Wave Optics Ltd | Pupil relay system |
JP2022538957A (en) | 2019-06-27 | 2022-09-07 | ルーマス リミテッド | Apparatus and method for eye tracking based on imaging the eye through light guide optics |
US11151794B1 (en) | 2019-06-28 | 2021-10-19 | Snap Inc. | Messaging system with augmented reality messages |
IL309979A (en) | 2019-07-04 | 2024-03-01 | Lumus Ltd | Image waveguide with symmetric beam multiplication |
US11029805B2 (en) | 2019-07-10 | 2021-06-08 | Magic Leap, Inc. | Real-time preview of connectable objects in a physically-modeled virtual space |
CN114096910B (en) * | 2019-07-12 | 2024-09-10 | 奇跃公司 | Method and system for an augmented reality display with dynamic field of view |
USD907037S1 (en) | 2019-07-12 | 2021-01-05 | Magic Leap, Inc. | Clip-on accessory |
USD980840S1 (en) | 2019-07-12 | 2023-03-14 | Magic Leap, Inc. | Clip-on accessory |
JP2022540675A (en) | 2019-07-16 | 2022-09-16 | マジック リープ, インコーポレイテッド | Determination of Eye Rotation Center Using One or More Eye Tracking Cameras |
CN110349383A (en) * | 2019-07-18 | 2019-10-18 | 浙江师范大学 | A kind of intelligent eyeshield device and method |
EP3999884A4 (en) | 2019-07-19 | 2023-08-30 | Magic Leap, Inc. | Display device having diffraction gratings with reduced polarization sensitivity |
WO2021016028A1 (en) | 2019-07-19 | 2021-01-28 | Magic Leap, Inc. | Method of fabricating diffraction gratings |
US11907417B2 (en) | 2019-07-25 | 2024-02-20 | Tectus Corporation | Glance and reveal within a virtual environment |
WO2021021670A1 (en) * | 2019-07-26 | 2021-02-04 | Magic Leap, Inc. | Systems and methods for augmented reality |
CN112305776B (en) * | 2019-07-26 | 2022-06-07 | 驻景(广州)科技有限公司 | Light field display system based on light waveguide coupling light exit pupil segmentation-combination control |
WO2021021926A1 (en) | 2019-07-29 | 2021-02-04 | Digilens Inc. | Methods and apparatus for multiplying the image resolution and field-of-view of a pixelated display |
KR20210014816A (en) * | 2019-07-30 | 2021-02-10 | 삼성디스플레이 주식회사 | Optical device |
CN114467279A (en) | 2019-07-31 | 2022-05-10 | 奇跃公司 | User data management for augmented reality using distributed ledgers |
CN110426853B (en) * | 2019-07-31 | 2020-10-16 | 华为技术有限公司 | Lens and head-mounted display device |
EP4003250A4 (en) | 2019-07-31 | 2023-08-23 | Acucela Inc. | Device for projecting images on the retina |
US10944290B2 (en) | 2019-08-02 | 2021-03-09 | Tectus Corporation | Headgear providing inductive coupling to a contact lens |
US11579425B1 (en) | 2019-08-05 | 2023-02-14 | Meta Platforms Technologies, Llc | Narrow-band peripheral see-through pancake lens assembly and display device with same |
US11586024B1 (en) | 2019-08-05 | 2023-02-21 | Meta Platforms Technologies, Llc | Peripheral see-through pancake lens assembly and display device with same |
US11822083B2 (en) | 2019-08-13 | 2023-11-21 | Apple Inc. | Display system with time interleaving |
CN112399157A (en) * | 2019-08-15 | 2021-02-23 | 中强光电股份有限公司 | Projector and projection method |
US11300791B2 (en) | 2019-08-21 | 2022-04-12 | Magic Leap, Inc. | Flat spectral response gratings using high index materials |
KR102386259B1 (en) * | 2019-08-21 | 2022-04-18 | 주식회사 레티널 | Optical device for augmented reality having visual acuity correction function |
CN112823307A (en) * | 2019-08-26 | 2021-05-18 | 京东方科技集团股份有限公司 | Optical display system and method, display device |
KR20220054386A (en) | 2019-08-29 | 2022-05-02 | 디지렌즈 인코포레이티드. | Vacuum Bragg grating and manufacturing method thereof |
WO2021040096A1 (en) * | 2019-08-30 | 2021-03-04 | 엘지전자 주식회사 | Head-wearable electronic device |
US11391948B2 (en) | 2019-09-10 | 2022-07-19 | Facebook Technologies, Llc | Display illumination using a grating |
US11467332B2 (en) | 2019-09-10 | 2022-10-11 | Meta Platforms Technologies, Llc | Display with switchable retarder array |
US11726336B2 (en) | 2019-09-10 | 2023-08-15 | Meta Platforms Technologies, Llc | Active zonal display illumination using a chopped lightguide |
EP4028354A4 (en) | 2019-09-11 | 2023-11-22 | Magic Leap, Inc. | Display device with diffraction grating having reduced polarization sensitivity |
US11245880B2 (en) * | 2019-09-12 | 2022-02-08 | Universal City Studios Llc | Techniques for spatial data projection |
WO2021049740A1 (en) | 2019-09-12 | 2021-03-18 | Samsung Electronics Co., Ltd. | Eye accommodation distance measuring device and method, and head-mounted display |
US11733545B2 (en) | 2019-09-16 | 2023-08-22 | Acucela Inc. | Assembly process for an electronic soft contact lens designed to inhibit progression of myopia |
US11835722B2 (en) * | 2019-09-17 | 2023-12-05 | Meta Platforms Technologies, Llc | Display device with transparent emissive display and see-through lens assembly |
JP7357976B2 (en) * | 2019-09-18 | 2023-10-10 | レティノル カンパニー リミテッド | Augmented reality optical device with improved light efficiency |
US11137608B2 (en) * | 2019-09-25 | 2021-10-05 | Electronics And Telecommunications Research Institute | Slim immersive display device, slim visualization device, and user eye-tracking device |
US11933949B2 (en) | 2019-09-27 | 2024-03-19 | Apple Inc. | Freeform folded optical system |
US11039113B2 (en) * | 2019-09-30 | 2021-06-15 | Snap Inc. | Multi-dimensional rendering |
JP2021056369A (en) | 2019-09-30 | 2021-04-08 | セイコーエプソン株式会社 | Head-mounted display |
US11175509B2 (en) * | 2019-09-30 | 2021-11-16 | Microsoft Technology Licensing, Llc | Tuned waveguides |
US11176757B2 (en) | 2019-10-02 | 2021-11-16 | Magic Leap, Inc. | Mission driven virtual character for user interaction |
US11276246B2 (en) | 2019-10-02 | 2022-03-15 | Magic Leap, Inc. | Color space mapping for intuitive surface normal visualization |
JP2022551734A (en) | 2019-10-15 | 2022-12-13 | マジック リープ, インコーポレイテッド | Cross-reality system that supports multiple device types |
JP2022551733A (en) | 2019-10-15 | 2022-12-13 | マジック リープ, インコーポレイテッド | Cross-reality system with localization service |
US11632679B2 (en) | 2019-10-15 | 2023-04-18 | Magic Leap, Inc. | Cross reality system with wireless fingerprints |
EP4045964A4 (en) | 2019-10-17 | 2023-11-08 | Magic Leap, Inc. | Attenuation of light transmission artifacts in wearable displays |
KR20220074893A (en) | 2019-10-23 | 2022-06-03 | 루머스 리미티드 | Display using astigmatism optics and aberration compensation |
US10901505B1 (en) | 2019-10-24 | 2021-01-26 | Tectus Corporation | Eye-based activation and tool selection systems and methods |
US11662807B2 (en) | 2020-01-06 | 2023-05-30 | Tectus Corporation | Eye-tracking user interface for virtual tool control |
CN110866895B (en) * | 2019-10-25 | 2023-09-08 | 广西电网有限责任公司电力科学研究院 | Method for detecting quality of hot galvanizing layer of power transmission and transformation steel framework |
CN114616509A (en) | 2019-10-31 | 2022-06-10 | 奇跃公司 | Cross-reality system with quality information about persistent coordinate frames |
KR102282422B1 (en) * | 2019-11-01 | 2021-07-27 | 주식회사 레티널 | Compact type optical device for augmented reality which can prevent ghost images with wide field of view |
KR102200144B1 (en) * | 2019-11-01 | 2021-01-08 | 주식회사 레티널 | Compact type optical device for augmented reality which can prevent ghost images with wide field of view |
KR102248606B1 (en) * | 2019-12-26 | 2021-05-06 | 주식회사 레티널 | Compact type optical device for augmented reality having reflective means arranged in curved line |
JP7291441B2 (en) | 2019-11-01 | 2023-06-15 | レティノル カンパニー リミテッド | Compact augmented reality optical device with ghost image blocking function and wide viewing angle |
CN110727115A (en) * | 2019-11-05 | 2020-01-24 | 华东交通大学 | Super-multi-viewpoint near-to-eye display device based on diffractive optics |
US11493989B2 (en) | 2019-11-08 | 2022-11-08 | Magic Leap, Inc. | Modes of user interaction |
EP4055423A4 (en) | 2019-11-08 | 2024-01-10 | Magic Leap, Inc. | Metasurfaces with light-redirecting structures including multiple materials and methods for fabricating |
USD982593S1 (en) | 2019-11-08 | 2023-04-04 | Magic Leap, Inc. | Portion of a display screen with animated ray |
WO2021096931A1 (en) | 2019-11-12 | 2021-05-20 | Magic Leap, Inc. | Cross reality system with localization service and shared location-based content |
WO2021097102A1 (en) * | 2019-11-13 | 2021-05-20 | Magic Leap, Inc. | Ambient light management systems and methods for wearable devices |
EP4058936A4 (en) | 2019-11-14 | 2023-05-03 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality |
JP2023502927A (en) | 2019-11-15 | 2023-01-26 | マジック リープ, インコーポレイテッド | Visualization system for use in a surgical environment |
WO2021101844A1 (en) | 2019-11-18 | 2021-05-27 | Magic Leap, Inc. | Mapping and localization of a passable world |
US11282288B2 (en) | 2019-11-20 | 2022-03-22 | Shape Matrix Geometric Instruments, LLC | Methods and apparatus for encoding data in notched shapes |
JP7494297B2 (en) | 2019-11-22 | 2024-06-03 | マジック リープ, インコーポレイテッド | Method and system for patterning a liquid crystal layer - Patents.com |
KR102244445B1 (en) * | 2019-11-22 | 2021-04-26 | 인하대학교 산학협력단 | Apparatus and method for occlusion capable near-eye display for augmented reality using single dmd |
US12094139B2 (en) | 2019-11-22 | 2024-09-17 | Magic Leap, Inc. | Systems and methods for enhanced depth determination using projection spots |
JP7467623B2 (en) | 2019-11-26 | 2024-04-15 | マジック リープ, インコーポレイテッド | Improved eye tracking for augmented or virtual reality display systems |
EP4066042A4 (en) * | 2019-11-26 | 2023-11-08 | Telefonaktiebolaget LM Ericsson (publ) | Supply of multi-layer extended reality images to a user |
CN110873971B (en) * | 2019-11-29 | 2021-08-31 | 维沃移动通信有限公司 | Wearable device and light filtering method thereof |
GB2589575B (en) | 2019-12-02 | 2022-01-12 | Envisics Ltd | Pupil expander |
US11561335B2 (en) | 2019-12-05 | 2023-01-24 | Lumus Ltd. | Light-guide optical element employing complementary coated partial reflectors, and light-guide optical element having reduced light scattering |
US11442593B2 (en) | 2019-12-06 | 2022-09-13 | Magic Leap, Inc. | Dynamic browser stage |
CN114788251A (en) | 2019-12-06 | 2022-07-22 | 奇跃公司 | Encoding stereoscopic splash screens in still images |
JP7497079B2 (en) | 2019-12-08 | 2024-06-10 | ルーマス リミテッド | Optical system with a compact image projector |
USD941353S1 (en) | 2019-12-09 | 2022-01-18 | Magic Leap, Inc. | Portion of a display screen with transitional graphical user interface for guiding graphics |
USD940748S1 (en) | 2019-12-09 | 2022-01-11 | Magic Leap, Inc. | Portion of a display screen with transitional graphical user interface for guiding graphics |
USD940749S1 (en) | 2019-12-09 | 2022-01-11 | Magic Leap, Inc. | Portion of a display screen with transitional graphical user interface for guiding graphics |
USD941307S1 (en) | 2019-12-09 | 2022-01-18 | Magic Leap, Inc. | Portion of a display screen with graphical user interface for guiding graphics |
US20210173144A1 (en) * | 2019-12-09 | 2021-06-10 | KDH Design Co., Ltd. | Guiding apparatus for an imaging light path and guiding method thereof |
EP4073763A4 (en) | 2019-12-09 | 2023-12-27 | Magic Leap, Inc. | Cross reality system with simplified programming of virtual content |
USD940189S1 (en) | 2019-12-09 | 2022-01-04 | Magic Leap, Inc. | Portion of a display screen with transitional graphical user interface for guiding graphics |
US12135446B2 (en) * | 2019-12-09 | 2024-11-05 | KDH Advanced Research Pty Ltd. | Light guiding apparatus and guiding method thereof |
US11269181B2 (en) | 2019-12-09 | 2022-03-08 | Magic Leap, Inc. | Systems and methods for operating a head-mounted display system based on user identity |
USD952673S1 (en) | 2019-12-09 | 2022-05-24 | Magic Leap, Inc. | Portion of a display screen with transitional graphical user interface for guiding graphics |
US11288876B2 (en) | 2019-12-13 | 2022-03-29 | Magic Leap, Inc. | Enhanced techniques for volumetric stage mapping based on calibration object |
CN113007618B (en) | 2019-12-19 | 2023-11-28 | 隆达电子股份有限公司 | Optical element and light-emitting device |
WO2021124287A1 (en) * | 2019-12-20 | 2021-06-24 | Niantic, Inc. | Sharded storage of geolocated data with predictable query response times |
CN110989172B (en) * | 2019-12-24 | 2021-08-06 | 平行现实(杭州)科技有限公司 | Waveguide display device with ultra-large field angle |
KR102310804B1 (en) * | 2019-12-26 | 2021-10-08 | 주식회사 레티널 | Compact type optical device for augmented reality having reflective means arranged in curved line |
WO2021137228A1 (en) | 2019-12-30 | 2021-07-08 | Lumus Ltd. | Optical systems including light-guide optical elements with two-dimensional expansion |
TW202125039A (en) * | 2019-12-30 | 2021-07-01 | 宏碁股份有限公司 | Wearable display device |
US11372478B2 (en) * | 2020-01-14 | 2022-06-28 | Htc Corporation | Head mounted display |
KR20210091580A (en) | 2020-01-14 | 2021-07-22 | 삼성전자주식회사 | Device and method for 3-dimensional image display |
CN111175976B (en) * | 2020-01-17 | 2022-02-22 | 歌尔股份有限公司 | Optical waveguide component, display system, augmented reality device and display method |
CN111221126B (en) * | 2020-01-17 | 2022-01-14 | 歌尔股份有限公司 | Imaging system, imaging method and virtual reality equipment |
US11360308B2 (en) | 2020-01-22 | 2022-06-14 | Facebook Technologies, Llc | Optical assembly with holographic optics for folded optical path |
US11340695B2 (en) | 2020-01-24 | 2022-05-24 | Magic Leap, Inc. | Converting a 2D positional input into a 3D point in space |
CN115380236A (en) | 2020-01-24 | 2022-11-22 | 奇跃公司 | Content movement and interaction using a single controller |
JP7490784B2 (en) | 2020-01-27 | 2024-05-27 | マジック リープ, インコーポレイテッド | Augmented Reality Map Curation |
USD949200S1 (en) | 2020-01-27 | 2022-04-19 | Magic Leap, Inc. | Portion of a display screen with a set of avatars |
USD948562S1 (en) | 2020-01-27 | 2022-04-12 | Magic Leap, Inc. | Portion of a display screen with avatar |
JP2023511311A (en) | 2020-01-27 | 2023-03-17 | マジック リープ, インコーポレイテッド | Gaze timer-based extension of user input device functionality |
USD948574S1 (en) | 2020-01-27 | 2022-04-12 | Magic Leap, Inc. | Portion of a display screen with a set of avatars |
JP2023511107A (en) | 2020-01-27 | 2023-03-16 | マジック リープ, インコーポレイテッド | neutral avatar |
US11756509B2 (en) | 2020-01-27 | 2023-09-12 | Magic Leap, Inc. | Enhanced state control for anchor-based cross reality applications |
USD936704S1 (en) | 2020-01-27 | 2021-11-23 | Magic Leap, Inc. | Portion of a display screen with avatar |
US10856253B1 (en) | 2020-01-31 | 2020-12-01 | Dell Products, Lp | System and method for beamsteering acquisition and optimization in an enhanced reality environment |
US11487356B2 (en) | 2020-01-31 | 2022-11-01 | Magic Leap, Inc. | Augmented and virtual reality display systems for oculometric assessments |
US20230072559A1 (en) * | 2020-01-31 | 2023-03-09 | University Of Virginia Patent Foundation | System and method for visual field rapid assessment |
US11709363B1 (en) | 2020-02-10 | 2023-07-25 | Avegant Corp. | Waveguide illumination of a spatial light modulator |
JP7455985B2 (en) | 2020-02-10 | 2024-03-26 | マジック リープ, インコーポレイテッド | Body-centered content positioning for 3D containers in mixed reality environments |
JP2023514207A (en) * | 2020-02-13 | 2023-04-05 | マジック リープ, インコーポレイテッド | Cross-reality system with prioritization of geolocation information for localization |
JP2023514208A (en) | 2020-02-13 | 2023-04-05 | マジック リープ, インコーポレイテッド | Cross-reality system with map processing using multi-resolution frame descriptors |
EP4103910A4 (en) | 2020-02-13 | 2024-03-06 | Magic Leap, Inc. | Cross reality system with accurate shared maps |
JP2023517281A (en) | 2020-02-14 | 2023-04-25 | マジック リープ, インコーポレイテッド | Virtual Object Movement Velocity Curves for Virtual and Augmented Reality Display Systems |
EP4107820A4 (en) | 2020-02-21 | 2024-04-17 | Acucela Inc. | Charging case for electronic contact lens |
CN115461787A (en) | 2020-02-26 | 2022-12-09 | 奇跃公司 | Cross reality system with quick positioning |
EP4111133A4 (en) | 2020-02-26 | 2024-11-06 | Magic Leap Inc | Procedural electron beam lithography |
US11840034B2 (en) | 2020-02-28 | 2023-12-12 | Magic Leap, Inc. | Method of fabricating molds for forming eyepieces with integrated spacers |
US11187858B2 (en) * | 2020-02-28 | 2021-11-30 | International Business Machines Corporation | Electrically-controlled fiber-optic switching system |
WO2021178727A1 (en) * | 2020-03-04 | 2021-09-10 | The Regents Of The University Of California | Evanescent coupler mode converters |
CN115244447A (en) | 2020-03-06 | 2022-10-25 | 奇跃公司 | Angle-selective attenuation of light transmission artifacts in wearable displays |
US11262588B2 (en) | 2020-03-10 | 2022-03-01 | Magic Leap, Inc. | Spectator view of virtual and physical objects |
CN111476104B (en) * | 2020-03-17 | 2022-07-01 | 重庆邮电大学 | AR-HUD image distortion correction method, device and system under dynamic eye position |
JP2023518421A (en) | 2020-03-20 | 2023-05-01 | マジック リープ, インコーポレイテッド | Systems and methods for retinal imaging and tracking |
KR20220152387A (en) | 2020-03-23 | 2022-11-15 | 루머스 리미티드 | Optical device to mitigate ghost images |
WO2021195283A1 (en) | 2020-03-25 | 2021-09-30 | Magic Leap, Inc. | Optical device with one-way mirror |
JP2023520461A (en) | 2020-04-03 | 2023-05-17 | マジック リープ, インコーポレイテッド | Wearable display system with nanowire LED microdisplay |
US11348300B2 (en) | 2020-04-03 | 2022-05-31 | Magic Leap, Inc. | Avatar customization for optimal gaze discrimination |
CN111474813B (en) * | 2020-04-29 | 2021-09-28 | Oppo广东移动通信有限公司 | Projection optical machine and electronic equipment |
JP2023524446A (en) | 2020-04-29 | 2023-06-12 | マジック リープ, インコーポレイテッド | Cross-reality system for large-scale environments |
US11994687B2 (en) * | 2020-05-13 | 2024-05-28 | President And Fellows Of Harvard College | Meta-optics for virtual reality and augmented reality systems |
AU2021270577A1 (en) | 2020-05-13 | 2022-12-08 | Acucela Inc. | Electro-switchable spectacles for myopia treatment |
KR102284743B1 (en) * | 2020-05-14 | 2021-08-03 | 한국과학기술연구원 | Extended dof image display apparatus and method for controlling thereof |
JP2023526430A (en) | 2020-05-22 | 2023-06-21 | マジック リープ, インコーポレイテッド | Method and system for dual projector waveguide display with wide field of view |
CN116194821A (en) | 2020-05-22 | 2023-05-30 | 奇跃公司 | Augmented and virtual reality display system with associated in-and out-coupling optical zones |
CN115552171B (en) | 2020-05-24 | 2024-01-23 | 鲁姆斯有限公司 | Method for manufacturing composite light guide optical element and optical structure |
EP4085287A4 (en) | 2020-05-24 | 2023-07-12 | Lumus Ltd. | Compound light-guide optical elements |
US11195490B1 (en) | 2020-05-29 | 2021-12-07 | International Business Machines Corporation | Smart contact lens with adjustable light transmittance |
US11514649B2 (en) | 2020-05-29 | 2022-11-29 | Microsoft Technology Licensing, Llc | Camera for augmented reality display |
WO2021247435A1 (en) | 2020-06-05 | 2021-12-09 | Magic Leap, Inc. | Enhanced eye tracking techniques based on neural network analysis of images |
AU2021288457A1 (en) | 2020-06-08 | 2023-01-05 | Acucela Inc. | Lens with asymmetric projection to treat astigmatism |
CA3179557A1 (en) | 2020-06-08 | 2021-12-16 | Ryo Kubota | Stick on devices using peripheral defocus to treat progressive refractive error |
CA3174148A1 (en) | 2020-06-08 | 2021-12-16 | Acucela Inc. | Projection of defocused images on the peripheral retina to treat refractive error |
US11281022B2 (en) | 2020-06-10 | 2022-03-22 | Acucela Inc. | Apparatus and methods for the treatment of refractive error using active stimulation |
EP4165360A4 (en) * | 2020-06-16 | 2024-10-30 | Marsupial Holdings Inc | Diffractive optic reflex sight |
EP4172684A4 (en) * | 2020-06-25 | 2024-07-17 | Magic Leap Inc | Tunable attenuation of light transmission artifacts in wearable displays |
US20220004148A1 (en) * | 2020-07-06 | 2022-01-06 | Grimaldi, Inc. | Apparatus and method of reproduction of a diffractive pattern |
KR102489272B1 (en) * | 2020-07-07 | 2023-01-17 | 한국과학기술연구원 | Near eye display apparatus |
US11360310B2 (en) * | 2020-07-09 | 2022-06-14 | Trimble Inc. | Augmented reality technology as a controller for a total station |
US11512956B2 (en) | 2020-07-09 | 2022-11-29 | Trimble Inc. | Construction layout using augmented reality |
US11852817B2 (en) * | 2020-07-14 | 2023-12-26 | Mercury Mission Systems, Llc | Curved waveguide for slim head up displays |
WO2022015847A1 (en) | 2020-07-15 | 2022-01-20 | Magic Leap, Inc. | Eye tracking using aspheric cornea model |
KR102423857B1 (en) * | 2020-07-17 | 2022-07-21 | 주식회사 레티널 | Compact type optical device for augmented reality using total internal reflection |
US11662511B2 (en) | 2020-07-22 | 2023-05-30 | Samsung Electronics Co., Ltd. | Beam expander and method of operating the same |
CN111650745B (en) * | 2020-07-24 | 2022-07-19 | 中国科学院光电技术研究所 | Scanning system based on micro-lens array group and self-adaptive optical fiber collimator |
US12105299B2 (en) | 2020-07-31 | 2024-10-01 | University Of Utah Research Foundation | Broadband diffractive optical element |
WO2022032198A1 (en) | 2020-08-07 | 2022-02-10 | Magic Leap, Inc. | Tunable cylindrical lenses and head-mounted display including the same |
CN112087575B (en) * | 2020-08-24 | 2022-03-08 | 广州启量信息科技有限公司 | Virtual camera control method |
WO2022045707A1 (en) | 2020-08-25 | 2022-03-03 | Samsung Electronics Co., Ltd. | Augmented reality device based on waveguide with holographic diffractive grating structure and apparatus for recording the holographic diffractive grating structure |
WO2022051123A1 (en) * | 2020-09-02 | 2022-03-10 | The Board Of Trustees Of The Leland Stanford Junior University | Doubly resonant micromechanical beam scanners |
KR102436597B1 (en) | 2020-09-09 | 2022-08-26 | 주식회사 레티널 | Optical device for augmented reality having optical structure arranged in straight line and manufacturing method for optical means |
WO2022066431A1 (en) | 2020-09-22 | 2022-03-31 | Sterling Labs Llc | Multiple gaze dependent illumination sources for retinal eye tracking |
WO2022065658A1 (en) | 2020-09-22 | 2022-03-31 | Samsung Electronics Co., Ltd. | Holographic waveguide, method of producing the same, and display device including the holographic waveguide |
US11860366B2 (en) | 2020-09-29 | 2024-01-02 | Avegant Corp. | Architecture to illuminate a display panel |
JP7481225B2 (en) | 2020-10-08 | 2024-05-10 | エルジー ディスプレイ カンパニー リミテッド | Head-mounted display |
KR102425375B1 (en) * | 2020-10-15 | 2022-07-27 | 주식회사 레티널 | Compact type optical device for augmented reality having optical structure arranged in straight line and manufacturing method for optical means |
US20220128756A1 (en) * | 2020-10-27 | 2022-04-28 | Lightspace Technologies, SIA | Display system for generating three-dimensional image and method therefor |
US11747621B2 (en) * | 2020-11-07 | 2023-09-05 | Microsoft Technology Licensing, Llc | Dichroic coatings to improve display uniformity and light security in an optical combiner |
CN114545624A (en) * | 2020-11-25 | 2022-05-27 | 华为技术有限公司 | Near-eye display system and near-eye display method |
TWI793912B (en) * | 2020-12-14 | 2023-02-21 | 瑞士商艾姆微體電子 馬林公司 | Method for sensing a displacement of a pointing device |
CN112702537B (en) * | 2020-12-25 | 2022-06-28 | 上海科技大学 | High dynamic range environment light dynamic collection system based on albedo difference |
CN112817117B (en) * | 2020-12-28 | 2022-10-21 | 西南技术物理研究所 | Parabolic reflector auxiliary device with auto-collimation adjusting function |
US20240151964A1 (en) * | 2021-01-11 | 2024-05-09 | Magic Leap, Inc. | Actuated pupil steering for head-mounted display systems |
US11323691B1 (en) * | 2021-01-14 | 2022-05-03 | Lightspace Technologies, SIA | Display system for displaying three-dimensional image and method therefor |
KR102591589B1 (en) * | 2021-01-30 | 2023-10-20 | 주식회사 피앤씨솔루션 | Wearable computing system of ar glasses using external computing device |
KR20240006707A (en) | 2021-02-25 | 2024-01-15 | 루머스 리미티드 | Optical aperture multipliers having a rectangular waveguide |
JP7558842B2 (en) | 2021-02-25 | 2024-10-01 | 株式会社小糸製作所 | Image Projection Device |
JP7558841B2 (en) | 2021-02-25 | 2024-10-01 | 株式会社小糸製作所 | Image Projection Device |
WO2022181346A1 (en) * | 2021-02-25 | 2022-09-01 | 株式会社小糸製作所 | Image projection device |
USD962290S1 (en) * | 2021-03-03 | 2022-08-30 | Johnson & Johnson Consumer Inc. | Display screen or portion thereof with icon |
KR102470650B1 (en) * | 2021-03-04 | 2022-11-25 | 주식회사 레티널 | Compact type optical device for augmented reality having reflective means arranged in curved line |
KR20220131720A (en) * | 2021-03-22 | 2022-09-29 | 삼성전자주식회사 | Display apparatus including combiner having asymmetric magnification |
US20240168278A1 (en) * | 2021-03-22 | 2024-05-23 | Sony Semiconductor Solutions Corporation | Light source device and image display device |
CN115128737B (en) * | 2021-03-24 | 2023-04-18 | 华为技术有限公司 | Diffractive optical waveguide and electronic device |
JP7093591B1 (en) | 2021-03-24 | 2022-06-30 | 株式会社Qdレーザ | Image projection device |
WO2022213190A1 (en) * | 2021-04-06 | 2022-10-13 | Vuereal Inc. | Ar system with hybrid display |
US11209672B1 (en) | 2021-04-06 | 2021-12-28 | Acucela Inc. | Supporting pillars for encapsulating a flexible PCB within a soft hydrogel contact lens |
US11868531B1 (en) | 2021-04-08 | 2024-01-09 | Meta Platforms Technologies, Llc | Wearable device providing for thumb-to-finger-based input gestures detected based on neuromuscular signals, and systems and methods of use thereof |
CN112950791A (en) * | 2021-04-08 | 2021-06-11 | 腾讯科技(深圳)有限公司 | Display method and related device |
US11741863B2 (en) | 2021-04-16 | 2023-08-29 | Tectus Corporation | Eyeglass-integrated display device using multiple embedded projectors and display windows |
KR20220145668A (en) | 2021-04-22 | 2022-10-31 | 삼성전자주식회사 | Display apparatus including free-formed surface and operating method of the same |
US11366341B1 (en) | 2021-05-04 | 2022-06-21 | Acucela Inc. | Electronic case for electronic spectacles |
CN114675418B (en) * | 2021-05-08 | 2024-08-20 | 胡大文 | Ultra-light wearable display device and method for the same |
TW202307518A (en) | 2021-07-04 | 2023-02-16 | 以色列商魯姆斯有限公司 | Display with stacked light-guide elements providing different parts of field of view |
US20230143223A1 (en) * | 2021-08-18 | 2023-05-11 | Advanced Neuromodulation Systems, Inc. | Systems and methods for providing digital health services |
CN117651892A (en) | 2021-08-23 | 2024-03-05 | 鲁姆斯有限公司 | Method for manufacturing a composite light guide optical element with an embedded in-coupling reflector |
US12135471B2 (en) | 2021-09-10 | 2024-11-05 | Tectus Corporation | Control of an electronic contact lens using eye gestures |
US20230084541A1 (en) * | 2021-09-16 | 2023-03-16 | Meta Platforms Technologies, Llc | Compact imaging optics using spatially located, free form optical components for distortion compensation and image clarity enhancement |
US11592899B1 (en) | 2021-10-28 | 2023-02-28 | Tectus Corporation | Button activation within an eye-controlled user interface |
US11863730B2 (en) | 2021-12-07 | 2024-01-02 | Snap Inc. | Optical waveguide combiner systems and methods |
TWI806293B (en) * | 2021-12-17 | 2023-06-21 | 宏碁股份有限公司 | Augmented reality glasses |
GB2610251B (en) * | 2021-12-21 | 2023-10-18 | Envisics Ltd | Compact head-up display and waveguide therefor |
WO2023121290A1 (en) * | 2021-12-24 | 2023-06-29 | 삼성전자 주식회사 | Display device and projector using holographic optical element |
KR20230103379A (en) | 2021-12-31 | 2023-07-07 | 삼성전자주식회사 | Method and apparatus for processing augmented reality |
US11619994B1 (en) | 2022-01-14 | 2023-04-04 | Tectus Corporation | Control of an electronic contact lens using pitch-based eye gestures |
US20230236415A1 (en) * | 2022-01-26 | 2023-07-27 | Meta Platforms Technologies, Llc | Image generation and delivery in a display system utilizing a two-dimensional (2d) field of view expander |
US11741861B1 (en) | 2022-02-08 | 2023-08-29 | Lumus Ltd. | Optical system including selectively activatable facets |
CN118830239A (en) * | 2022-03-07 | 2024-10-22 | Uab研究基金会 | Myopia prevention visual display therapy using simulated myopia blur |
US11556010B1 (en) * | 2022-04-01 | 2023-01-17 | Wen-Tsun Wu | Mini display device |
US20230350203A1 (en) * | 2022-04-29 | 2023-11-02 | Snap Inc. | Ar/vr enabled contact lens |
US11874961B2 (en) | 2022-05-09 | 2024-01-16 | Tectus Corporation | Managing display of an icon in an eye tracking augmented reality device |
TWI828150B (en) * | 2022-05-20 | 2024-01-01 | 宇力電通數位整合有限公司 | Stereoscopic vision glasses |
WO2023239417A1 (en) * | 2022-06-10 | 2023-12-14 | Magic Leap, Inc. | Compensating thickness variations in substrates for optical devices |
WO2023244811A1 (en) * | 2022-06-17 | 2023-12-21 | Magic Leap, Inc. | Method and system for performing eye tracking in augmented reality devices |
US20240087238A1 (en) * | 2022-06-21 | 2024-03-14 | T-Mobile Innovations Llc | System and method for extended reality processing in a local wireless environment with reduced latency |
CN115128829B (en) * | 2022-08-25 | 2023-01-31 | 惠科股份有限公司 | Display device |
KR102474532B1 (en) * | 2022-09-13 | 2022-12-06 | 한국광기술원 | Contact Lens Type Holographic Display Apparatus and System with Optical Waveguide Structure |
WO2024071692A1 (en) * | 2022-09-29 | 2024-04-04 | 삼성전자 주식회사 | Electronic device and operation method thereof |
US20240177398A1 (en) * | 2022-11-29 | 2024-05-30 | Rovi Guides, Inc. | Systems and methods for forming folded focal planes |
KR20240080686A (en) | 2022-11-30 | 2024-06-07 | (주)라이언로켓 | Apparatus and method for processing pseudonyms of video contents using one-shot face swap |
KR20240080687A (en) | 2022-11-30 | 2024-06-07 | (주)라이언로켓 | Real-time face special makeup device and method through one-shot face swap |
CN115840295B (en) * | 2023-02-23 | 2023-05-02 | 北京数字光芯集成电路设计有限公司 | Linear array micro LED scans AR equipment |
JP7475751B1 (en) | 2023-10-11 | 2024-04-30 | アルディーテック株式会社 | Collimating contact lenses and XR glasses |
CN118506260A (en) * | 2024-04-23 | 2024-08-16 | 南京脉期尔电气有限公司 | Intelligent identification system for different target contact scenes of power distribution network |
Family Cites Families (283)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3466021A (en) * | 1967-09-14 | 1969-09-09 | Falconbridge Nickel Mines Ltd | Thermal treatments in fluidized beds |
US3566021A (en) | 1968-11-29 | 1971-02-23 | Bell Telephone Labor Inc | Real time three dimensional television system |
US4173391A (en) * | 1978-04-06 | 1979-11-06 | New York Institute Of Technology | Three dimensional display |
JPS56153315A (en) * | 1980-04-30 | 1981-11-27 | Nippon Telegr & Teleph Corp <Ntt> | Light beam deflector |
US5061049A (en) | 1984-08-31 | 1991-10-29 | Texas Instruments Incorporated | Spatial light modulator and method |
US4786150A (en) | 1985-09-19 | 1988-11-22 | Canon Kabushiki Kaisha | Zoom lens with beam splitter |
JPS6388529A (en) * | 1986-10-01 | 1988-04-19 | Univ Osaka | Multifunctional optical signal processing system |
US4848884A (en) * | 1987-02-05 | 1989-07-18 | Ricoh Company, Ltd. | Variable focusing lens barrel |
JPS63254884A (en) * | 1987-04-10 | 1988-10-21 | Toshio Hiji | Image pickup, transmitting, reproducing device for picture |
US5003300A (en) * | 1987-07-27 | 1991-03-26 | Reflection Technology, Inc. | Head mounted display for miniature video display system |
GB8727212D0 (en) | 1987-11-20 | 1987-12-23 | Secr Defence | Optical beam steering device |
EP0365406B1 (en) * | 1988-10-21 | 1993-09-29 | Thomson-Csf | Optical collimating system for a helmet visual |
US5081615A (en) * | 1988-12-16 | 1992-01-14 | Fuji Photo Film Co., Ltd. | Method of coupling external light into an optical waveguide and a guided wave from an optical waveguide and optical pickup employing an optical waveguide |
JPH0384516A (en) * | 1989-08-29 | 1991-04-10 | Fujitsu Ltd | Three-dimensional display device |
JP2874208B2 (en) | 1989-09-08 | 1999-03-24 | ブラザー工業株式会社 | Image display device |
US8730129B2 (en) | 1990-12-07 | 2014-05-20 | Dennis J Solomon | Advanced immersive visual display system |
US5334333A (en) * | 1990-12-17 | 1994-08-02 | Elf Atochem North America, Inc. | Electroactive compounds, compositions and devices and methods of making the same |
JPH04336523A (en) * | 1991-05-14 | 1992-11-24 | Sumitomo Electric Ind Ltd | Multicore type optical isolator |
JP2971626B2 (en) * | 1991-07-10 | 1999-11-08 | 旭化成工業株式会社 | Image fiber |
JP3438732B2 (en) | 1991-07-19 | 2003-08-18 | セイコーエプソン株式会社 | Optical device and display device |
JPH0546161A (en) | 1991-08-12 | 1993-02-26 | Casio Comput Co Ltd | Virtual reality display device |
US5233673A (en) | 1991-10-09 | 1993-08-03 | Hughes Aircraft Company | Output steerable optical phased array |
US5493426A (en) * | 1991-11-14 | 1996-02-20 | University Of Colorado Foundation, Inc. | Lateral electrode smectic liquid crystal devices |
US5377287A (en) | 1991-12-19 | 1994-12-27 | Hughes Aircraft Company | Fiber optic corporate power divider/combiner and method |
US5596339A (en) * | 1992-10-22 | 1997-01-21 | University Of Washington | Virtual retinal display with fiber optic point source |
US6008781A (en) | 1992-10-22 | 1999-12-28 | Board Of Regents Of The University Of Washington | Virtual retinal display |
US5285509A (en) * | 1992-12-18 | 1994-02-08 | Hughes Aircraft Company | Coupler for waveguides of differing cross section |
US5564959A (en) | 1993-09-08 | 1996-10-15 | Silicon Video Corporation | Use of charged-particle tracks in fabricating gated electron-emitting devices |
JPH0795622A (en) * | 1993-09-21 | 1995-04-07 | Olympus Optical Co Ltd | Stereo image photographing device, stereoscopic image display device and stereoscopic image recording and/or reproducing |
JP3341795B2 (en) * | 1994-01-13 | 2002-11-05 | 富士写真フイルム株式会社 | Laser scanning oscillator |
US5416876A (en) * | 1994-01-28 | 1995-05-16 | Hughes Training, Inc. | Fiber optic ribbon subminiature display for head/helmet mounted display |
JPH07261112A (en) * | 1994-03-22 | 1995-10-13 | Hitachi Ltd | On-head type display device |
JP3335248B2 (en) * | 1994-03-31 | 2002-10-15 | セントラル硝子株式会社 | Holographic display device |
JP3847799B2 (en) * | 1994-08-05 | 2006-11-22 | キヤノン株式会社 | Display device having gaze detection system |
JP3298082B2 (en) * | 1994-12-13 | 2002-07-02 | 日本電信電話株式会社 | Head mount display device |
US5768025A (en) | 1995-08-21 | 1998-06-16 | Olympus Optical Co., Ltd. | Optical system and image display apparatus |
EP1072934A3 (en) * | 1995-08-25 | 2002-05-29 | Massachusetts Institute Of Technology | VLSI visual display |
JP3556389B2 (en) * | 1996-05-01 | 2004-08-18 | 日本電信電話株式会社 | Head mounted display device |
JPH09211374A (en) * | 1996-01-31 | 1997-08-15 | Nikon Corp | Head mounted display device |
JPH09219832A (en) | 1996-02-13 | 1997-08-19 | Olympus Optical Co Ltd | Image display |
JPH09243955A (en) * | 1996-03-11 | 1997-09-19 | Seiko Epson Corp | Mounting-on-head type liquid crystal display device |
US5701132A (en) | 1996-03-29 | 1997-12-23 | University Of Washington | Virtual retinal display with expanded exit pupil |
US6243350B1 (en) * | 1996-05-01 | 2001-06-05 | Terastor Corporation | Optical storage systems with flying optical heads for near-field recording and reading |
US5885822A (en) | 1996-08-20 | 1999-03-23 | David A. Paslin | Method and system for growing molluscum contagiosum in xenografts to immunocompromised hosts |
US5886822A (en) * | 1996-10-08 | 1999-03-23 | The Microoptical Corporation | Image combining system for eyeglasses and face masks |
US6204974B1 (en) * | 1996-10-08 | 2001-03-20 | The Microoptical Corporation | Compact image display system for eyeglasses or other head-borne frames |
JP3787939B2 (en) * | 1997-02-27 | 2006-06-21 | コニカミノルタホールディングス株式会社 | 3D image display device |
JPH10257411A (en) * | 1997-03-14 | 1998-09-25 | Minolta Co Ltd | Video observation system |
US6046720A (en) | 1997-05-07 | 2000-04-04 | University Of Washington | Point source scanning apparatus and method |
US6415087B1 (en) | 1997-06-04 | 2002-07-02 | Corning Laserton, Inc. | Polished fused optical fiber endface |
JP3865906B2 (en) | 1997-06-27 | 2007-01-10 | オリンパス株式会社 | Image display device |
JP2001517329A (en) * | 1997-12-16 | 2001-10-02 | ゴスダルストベンニ ナウチニ ツェントル ロシイスコイ フェデラツィイ“ニオピク”(ゲーエヌテーエス エルエフ“ニオピク”) | Polarizer and liquid crystal display device |
US6043799A (en) | 1998-02-20 | 2000-03-28 | University Of Washington | Virtual retinal display with scanner array for generating multiple exit pupils |
JP3279265B2 (en) | 1998-03-26 | 2002-04-30 | 株式会社エム・アール・システム研究所 | Image display device |
FR2777359B1 (en) * | 1998-04-09 | 2000-07-07 | Corning Inc | CONNECTION OF OPTICAL FIBER AND OPTICAL WAVEGUIDE BY MERGER |
JP3583613B2 (en) * | 1998-04-15 | 2004-11-04 | 日本電信電話株式会社 | Stereoscopic display method and apparatus |
US6100862A (en) | 1998-04-20 | 2000-08-08 | Dimensional Media Associates, Inc. | Multi-planar volumetric display system and method of operation |
JPH11352325A (en) * | 1998-06-05 | 1999-12-24 | Shimadzu Corp | Optical interference filter and head up display using optical interference filter |
US6215532B1 (en) | 1998-07-27 | 2001-04-10 | Mixed Reality Systems Laboratory Inc. | Image observing apparatus for observing outside information superposed with a display image |
JP2000105348A (en) * | 1998-07-27 | 2000-04-11 | Mr System Kenkyusho:Kk | Picture observation device |
JP4232231B2 (en) * | 1998-09-30 | 2009-03-04 | ソニー株式会社 | Information processing apparatus and method, and recording medium |
JP2000131640A (en) * | 1998-10-23 | 2000-05-12 | Sony Corp | Picture display device |
US6281862B1 (en) | 1998-11-09 | 2001-08-28 | University Of Washington | Scanned beam display with adjustable accommodation |
CN1362868A (en) * | 1999-02-05 | 2002-08-07 | 成象诊断系统公司 | CCD array as a multiple-detector in an optical imaging apparatus |
CA2363012A1 (en) * | 1999-02-17 | 2000-08-24 | University Of Washington | A halo display system generating a panoramic virtual image surrounding the use |
JP2000249974A (en) * | 1999-03-02 | 2000-09-14 | Canon Inc | Display device and stereoscopic display device |
US6480337B2 (en) | 1999-03-04 | 2002-11-12 | Mixed Reality Systems Laboratory Inc. | Image display apparatus |
JP3891723B2 (en) | 1999-03-04 | 2007-03-14 | 富士フイルム株式会社 | Laser deflection scanning device |
JP2000249969A (en) * | 1999-03-04 | 2000-09-14 | Mr System Kenkyusho:Kk | Picture display optical system and picture display device using the same |
JP2000295637A (en) * | 1999-04-12 | 2000-10-20 | Mr System Kenkyusho:Kk | Stereoscopic image display device |
JP3453086B2 (en) * | 1999-05-18 | 2003-10-06 | 日本電信電話株式会社 | Three-dimensional display method and head-mounted display device |
JP4372891B2 (en) | 1999-06-22 | 2009-11-25 | オリンパス株式会社 | Video display device |
JP2001021831A (en) * | 1999-07-09 | 2001-01-26 | Shimadzu Corp | Head mounted type display device |
JP2001066504A (en) * | 1999-08-30 | 2001-03-16 | Canon Inc | Optical device and photographing device using the optical device |
JP3854763B2 (en) * | 1999-11-19 | 2006-12-06 | キヤノン株式会社 | Image display device |
JP4921634B2 (en) | 2000-01-31 | 2012-04-25 | グーグル インコーポレイテッド | Display device |
JP5059937B2 (en) * | 2000-01-31 | 2012-10-31 | グーグル インコーポレイテッド | Display device |
US6987911B2 (en) | 2000-03-16 | 2006-01-17 | Lightsmyth Technologies, Inc. | Multimode planar waveguide spectral filter |
JP2001265314A (en) * | 2000-03-17 | 2001-09-28 | Toshiba Corp | Display system, data display method, shadowed character font generating method, and recording medium |
JP2001290101A (en) | 2000-04-06 | 2001-10-19 | Tomohiko Hattori | System for detecting will to adjust visual point in length direction and method for driving the will and spectacles for automatic correction of perspective |
CA2411442C (en) * | 2000-06-05 | 2010-07-13 | Lumus Ltd. | Substrate-guided optical beam expander |
AU2001267074A1 (en) * | 2000-06-08 | 2001-12-17 | Interactive Imaging Systems, Inc. | Two stage optical magnification and image correction system |
US6975898B2 (en) | 2000-06-19 | 2005-12-13 | University Of Washington | Medical imaging, diagnosis, and therapy using a scanning single optical fiber system |
JP2002296626A (en) * | 2000-10-23 | 2002-10-09 | Sony Corp | Optical switch and display apparatus |
US6856712B2 (en) | 2000-11-27 | 2005-02-15 | University Of Washington | Micro-fabricated optical waveguide for use in scanning fiber displays and scanned fiber image acquisition |
TW522256B (en) * | 2000-12-15 | 2003-03-01 | Samsung Electronics Co Ltd | Wearable display system |
US7405884B2 (en) | 2000-12-21 | 2008-07-29 | Olympus Corporation | Optical apparatus |
JP2002214545A (en) * | 2001-01-18 | 2002-07-31 | Olympus Optical Co Ltd | Optical device |
US6542665B2 (en) | 2001-02-17 | 2003-04-01 | Lucent Technologies Inc. | GRIN fiber lenses |
US6654070B1 (en) | 2001-03-23 | 2003-11-25 | Michael Edward Rofe | Interactive heads up display (IHUD) |
GB0108838D0 (en) | 2001-04-07 | 2001-05-30 | Cambridge 3D Display Ltd | Far field display |
KR20020083737A (en) * | 2001-04-30 | 2002-11-04 | 삼성전자 주식회사 | Wearable display system |
US7616986B2 (en) * | 2001-05-07 | 2009-11-10 | University Of Washington | Optical fiber scanner for performing multimodal optical imaging |
US6574043B2 (en) | 2001-11-07 | 2003-06-03 | Eastman Kodak Company | Method for enhanced bit depth in an imaging apparatus using a spatial light modulator |
US7012756B2 (en) | 2001-11-14 | 2006-03-14 | Canon Kabushiki Kaisha | Display optical system, image display apparatus, image taking optical system, and image taking apparatus |
JP4083684B2 (en) * | 2002-01-23 | 2008-04-30 | 道彦 庄司 | Image processing system and image processing apparatus |
US7497574B2 (en) | 2002-02-20 | 2009-03-03 | Brother Kogyo Kabushiki Kaisha | Retinal image display device |
JP4107102B2 (en) * | 2002-02-20 | 2008-06-25 | ブラザー工業株式会社 | Image display device |
US6702442B2 (en) * | 2002-03-08 | 2004-03-09 | Eastman Kodak Company | Monocentric autostereoscopic optical apparatus using resonant fiber-optic image generation |
US7428001B2 (en) * | 2002-03-15 | 2008-09-23 | University Of Washington | Materials and methods for simulating focal shifts in viewers using large depth of focus displays |
KR20030088218A (en) * | 2002-05-13 | 2003-11-19 | 삼성전자주식회사 | Wearable color-display system |
DE10392754T5 (en) | 2002-06-17 | 2005-08-25 | Zygo Corp., Middlefield | An interferometric optical system and methods providing an optical path length and a focus that are scanned simultaneously |
US7031579B2 (en) * | 2002-06-26 | 2006-04-18 | L-3 Communications Corporation | High resolution display component, system and method |
JP2004038012A (en) * | 2002-07-05 | 2004-02-05 | Minolta Co Ltd | Image display device |
US7463783B1 (en) | 2002-09-20 | 2008-12-09 | Lockheed Martin Corporation | Constant magnification imaging method and system |
GB0222244D0 (en) | 2002-09-25 | 2002-10-30 | Sau Anthony L T | Improved 3D imaging system using reflectors |
JP2004289108A (en) * | 2002-09-27 | 2004-10-14 | Mitsubishi Electric Corp | Semiconductor optical element |
EP1592992B1 (en) * | 2003-01-24 | 2012-05-30 | University of Washington | Optical beam scanning system for compact image display or image acquisition |
JP2006516761A (en) | 2003-01-24 | 2006-07-06 | モンタナ ステート ユニバーシティー−ボーズマン | Off-axis variable focus and aberration control mirror and method |
CN2611927Y (en) * | 2003-02-09 | 2004-04-14 | 广辉电子股份有限公司 | Optical fiber display |
CN100416340C (en) * | 2003-04-25 | 2008-09-03 | 微型光学公司 | Binocular viewing system |
US7146084B2 (en) * | 2003-06-16 | 2006-12-05 | Cmc Electronics, Inc. | Fiber optic light source for display devices |
IL157837A (en) | 2003-09-10 | 2012-12-31 | Yaakov Amitai | Substrate-guided optical device particularly for three-dimensional displays |
US20050073471A1 (en) * | 2003-10-03 | 2005-04-07 | Uni-Pixel Displays, Inc. | Z-axis redundant display/multilayer display |
JP2005227324A (en) | 2004-02-10 | 2005-08-25 | Matsushita Electric Ind Co Ltd | Display element and display apparatus |
US7077523B2 (en) | 2004-02-13 | 2006-07-18 | Angstorm Inc. | Three-dimensional display using variable focusing lens |
US7274835B2 (en) | 2004-02-18 | 2007-09-25 | Cornell Research Foundation, Inc. | Optical waveguide displacement sensor |
US7088743B2 (en) | 2004-03-15 | 2006-08-08 | Northrop Grumman Corp. | Laser source comprising amplifier and adaptive wavefront/polarization driver |
CN100410727C (en) * | 2004-03-29 | 2008-08-13 | 索尼株式会社 | Optical device and virtual image display device |
JP4636808B2 (en) * | 2004-03-31 | 2011-02-23 | キヤノン株式会社 | Image display device |
JP4138690B2 (en) * | 2004-03-31 | 2008-08-27 | 株式会社東芝 | Display device |
EP1757974B1 (en) | 2004-05-17 | 2012-08-29 | Olympus Corporation | Head-mounted type image display device |
JP4766913B2 (en) | 2004-05-17 | 2011-09-07 | オリンパス株式会社 | Head-mounted image display device |
JP4609160B2 (en) * | 2004-05-17 | 2011-01-12 | 株式会社ニコン | Optical element, combiner optical system, and information display device |
US7764428B2 (en) * | 2004-06-23 | 2010-07-27 | Nikon Corporation | Optical element, optical system, and waveguide |
WO2006017771A1 (en) | 2004-08-06 | 2006-02-16 | University Of Washington | Variable fixation viewing distance scanned light displays |
JP4637532B2 (en) * | 2004-08-30 | 2011-02-23 | オリンパス株式会社 | Decentered optical system and optical system using it |
CN100359362C (en) * | 2004-09-13 | 2008-01-02 | 宋义 | LED optical fiber display |
WO2006054518A1 (en) * | 2004-11-18 | 2006-05-26 | Pioneer Corporation | 3d display device |
JP2006154041A (en) | 2004-11-26 | 2006-06-15 | Konica Minolta Opto Inc | Projection optical system |
JP2006154437A (en) * | 2004-11-30 | 2006-06-15 | Konica Minolta Photo Imaging Inc | Video display device |
WO2006064301A1 (en) * | 2004-12-13 | 2006-06-22 | Nokia Corporation | System and method for beam expansion with near focus in a display device |
US7206107B2 (en) * | 2004-12-13 | 2007-04-17 | Nokia Corporation | Method and system for beam expansion in a display device |
US7499174B2 (en) | 2005-01-12 | 2009-03-03 | John Farah | Lensless imaging with reduced aperture |
JP2006195084A (en) | 2005-01-12 | 2006-07-27 | Sharp Corp | Display apparatus |
KR101210804B1 (en) | 2005-03-22 | 2012-12-10 | 혼하이 프리시젼 인더스트리 컴퍼니 리미티드 | optical system using total internal reflection images |
EP1893074B2 (en) * | 2005-05-18 | 2017-06-14 | Visual Physics, LLC | Image presentation and micro-optic security system |
JP4655771B2 (en) * | 2005-06-17 | 2011-03-23 | ソニー株式会社 | Optical device and virtual image display device |
JP2007010830A (en) * | 2005-06-29 | 2007-01-18 | Nikon Corp | Image display optical system and image display apparatus |
JP4776285B2 (en) | 2005-07-01 | 2011-09-21 | ソニー株式会社 | Illumination optical device and virtual image display device using the same |
JP4662256B2 (en) | 2005-08-03 | 2011-03-30 | サミー株式会社 | Display device and display program |
JP4768367B2 (en) * | 2005-09-02 | 2011-09-07 | 日本放送協会 | Stereoscopic image pickup apparatus and stereoscopic image display apparatus |
KR100725895B1 (en) * | 2005-09-15 | 2007-06-08 | 주식회사 아이캔텍 | Apparatus for optical data input using optical fiber |
US7324732B2 (en) | 2005-10-21 | 2008-01-29 | Fanasys, Llc | Lithium niobate coated optical fiber apparatus and method |
US20070171328A1 (en) | 2005-11-21 | 2007-07-26 | Freeman Mark O | Substrate-guided display |
WO2007067163A1 (en) | 2005-11-23 | 2007-06-14 | University Of Washington | Scanning beam with variable sequential framing using interrupted scanning resonance |
US8526096B2 (en) | 2006-02-23 | 2013-09-03 | Pixtronix, Inc. | Mechanical light modulators with stressed beams |
US20080144174A1 (en) * | 2006-03-15 | 2008-06-19 | Zebra Imaging, Inc. | Dynamic autostereoscopic displays |
DE102006047777A1 (en) * | 2006-03-17 | 2007-09-20 | Daimlerchrysler Ag | Virtual spotlight for marking objects of interest in image data |
EP1996688B1 (en) | 2006-03-22 | 2011-10-26 | The Procter & Gamble Company | Liquid treatment composition |
CA2649411C (en) * | 2006-04-19 | 2019-02-12 | Setred As | Bandwidth improvement for 3d display |
CN101086608A (en) * | 2006-06-05 | 2007-12-12 | 中国科学院物理研究所 | Projection display device |
KR101258584B1 (en) | 2006-06-21 | 2013-05-02 | 엘지디스플레이 주식회사 | Volumetric type 3-Dimension Image Display Device |
FR2903786B1 (en) * | 2006-07-11 | 2008-09-05 | Thales Sa | HELMET VISUALIZATION SYSTEM WITH INTERCHANGEABLE OPTICAL MODULES |
JP2008046253A (en) * | 2006-08-11 | 2008-02-28 | Canon Inc | Image display device |
EP2059838A4 (en) | 2006-08-21 | 2009-12-30 | Univ Washington | Optical fiber scope with both non-resonant illumination and resonant collection/imaging for multiple modes of operation |
JP2008070604A (en) | 2006-09-14 | 2008-03-27 | Canon Inc | Image display device |
JP5023632B2 (en) * | 2006-09-15 | 2012-09-12 | ブラザー工業株式会社 | Head mounted display |
JP4893200B2 (en) * | 2006-09-28 | 2012-03-07 | ブラザー工業株式会社 | Optical system for light beam transfer, and retinal scanning display using the same |
DE102007024237B4 (en) * | 2007-05-21 | 2009-01-29 | Seereal Technologies S.A. | Holographic reconstruction system with optical waveguide tracking |
JP2008197242A (en) * | 2007-02-09 | 2008-08-28 | Sony Corp | Image reproduction method, image reproduction device, and three-dimensional image display apparatus |
EP2115519A4 (en) * | 2007-02-23 | 2012-12-05 | Pixeloptics Inc | Ophthalmic dynamic aperture |
JP2008268846A (en) | 2007-03-22 | 2008-11-06 | Citizen Holdings Co Ltd | Spectacles with electronic image display function |
JP2008304580A (en) | 2007-06-06 | 2008-12-18 | Sharp Corp | Image display device |
US7700908B2 (en) * | 2007-06-08 | 2010-04-20 | University Of Washington | Two dimensional optical scanning image system |
WO2009032813A2 (en) * | 2007-09-06 | 2009-03-12 | 3M Innovative Properties Company | Lightguides having light extraction structures providing regional control of light output |
EP2187259B1 (en) * | 2007-09-14 | 2014-05-14 | Panasonic Corporation | Projector |
JP2010541001A (en) * | 2007-09-25 | 2010-12-24 | エクスプレイ エルティーディー. | Micro projector |
JP4739304B2 (en) | 2007-09-28 | 2011-08-03 | 株式会社エヌ・ティ・ティ・ドコモ | Light wavefront display device and light wavefront display method |
JP2009092810A (en) * | 2007-10-05 | 2009-04-30 | Nikon Corp | Head-mounted display device |
CN101414425B (en) * | 2007-10-16 | 2013-07-17 | 宋学锋 | Display device and display method |
WO2009066465A1 (en) * | 2007-11-20 | 2009-05-28 | Panasonic Corporation | Image display device, display method thereof, program, integrated circuit, glasses type head mounted display, automobile, binoculars, and desktop type display |
WO2009100113A1 (en) | 2008-02-07 | 2009-08-13 | Imra America, Inc. | High power parallel fiber arrays |
US20100149073A1 (en) * | 2008-11-02 | 2010-06-17 | David Chaum | Near to Eye Display System and Appliance |
US20090295683A1 (en) * | 2008-05-27 | 2009-12-03 | Randall Pugh | Head mounted display with variable focal length lens |
JP5493144B2 (en) * | 2008-05-30 | 2014-05-14 | 国立大学法人東北大学 | Optical scanning device |
EP2138886A3 (en) | 2008-06-25 | 2011-10-05 | Samsung Electronics Co., Ltd. | Compact virtual display |
RU2427015C2 (en) * | 2008-06-25 | 2011-08-20 | Корпорация "САМСУНГ ЭЛЕКТРОНИКС Ко., Лтд." | Compact virtual display |
GB2461294B (en) | 2008-06-26 | 2011-04-06 | Light Blue Optics Ltd | Holographic image display systems |
JP2010008948A (en) * | 2008-06-30 | 2010-01-14 | Shinko Electric Ind Co Ltd | Scanning type optical projector |
JP2010020025A (en) * | 2008-07-09 | 2010-01-28 | Ricoh Co Ltd | Optical scanner and image forming apparatus |
JP5062432B2 (en) * | 2008-07-22 | 2012-10-31 | 大日本印刷株式会社 | Head mounted display |
US8809758B2 (en) * | 2008-07-25 | 2014-08-19 | Cornell University | Light field image sensor with an angle-sensitive pixel (ASP) device |
EP2311007B1 (en) * | 2008-08-08 | 2016-12-28 | Thomson Licensing | Method and apparatus for banding artifact detection |
WO2010062481A1 (en) * | 2008-11-02 | 2010-06-03 | David Chaum | Near to eye display system and appliance |
US10274660B2 (en) | 2008-11-17 | 2019-04-30 | Luminit, Llc | Holographic substrate-guided wave-based see-through display |
GB2465786A (en) * | 2008-11-28 | 2010-06-02 | Sharp Kk | An optical system for varying the perceived shape of a display surface |
JP5491833B2 (en) | 2008-12-05 | 2014-05-14 | 株式会社半導体エネルギー研究所 | Semiconductor device |
WO2010067114A1 (en) * | 2008-12-12 | 2010-06-17 | Bae Systems Plc | Improvements in or relating to waveguides |
JP4674634B2 (en) * | 2008-12-19 | 2011-04-20 | ソニー株式会社 | Head-mounted display |
JP5136442B2 (en) * | 2009-01-27 | 2013-02-06 | ブラザー工業株式会社 | Head mounted display |
DE102009000724A1 (en) * | 2009-02-09 | 2010-08-12 | Robert Bosch Gmbh | Device for deflecting light rays |
JP5133925B2 (en) * | 2009-03-25 | 2013-01-30 | オリンパス株式会社 | Head-mounted image display device |
US20120081800A1 (en) * | 2009-04-20 | 2012-04-05 | Dewen Cheng | Optical see-through free-form head-mounted display |
JP2010271526A (en) * | 2009-05-21 | 2010-12-02 | Konica Minolta Opto Inc | Video display device, head mounted display, and head-up display |
JP2010271565A (en) * | 2009-05-22 | 2010-12-02 | Seiko Epson Corp | Head-mounted display device |
JP5104820B2 (en) * | 2009-07-10 | 2012-12-19 | 株式会社島津製作所 | Display device |
JP5104823B2 (en) * | 2009-07-29 | 2012-12-19 | 株式会社島津製作所 | Display device |
DE102009037835B4 (en) * | 2009-08-18 | 2012-12-06 | Metaio Gmbh | Method for displaying virtual information in a real environment |
US20110075257A1 (en) * | 2009-09-14 | 2011-03-31 | The Arizona Board Of Regents On Behalf Of The University Of Arizona | 3-Dimensional electro-optical see-through displays |
CN101661163A (en) * | 2009-09-27 | 2010-03-03 | 合肥工业大学 | Three-dimensional helmet display of augmented reality system |
US20110083742A1 (en) * | 2009-10-01 | 2011-04-14 | RNY Solar | Multiconverter system comprising spectral separating reflector assembly and methods thereof |
US9195115B2 (en) * | 2009-10-15 | 2015-11-24 | Nec Corporation | Image projection device, image protection method, distance measuring device and distance measuring method |
JP5370071B2 (en) * | 2009-10-26 | 2013-12-18 | 株式会社島津製作所 | Display device |
EP2494388B1 (en) * | 2009-10-27 | 2018-11-21 | DigiLens Inc. | Compact holographic eyeglass display |
JP2011145607A (en) * | 2010-01-18 | 2011-07-28 | Sony Corp | Head mount display |
JP5218438B2 (en) | 2010-01-25 | 2013-06-26 | 株式会社島津製作所 | Display device |
JP5240214B2 (en) | 2010-02-15 | 2013-07-17 | 株式会社島津製作所 | Display device |
US20130278631A1 (en) * | 2010-02-28 | 2013-10-24 | Osterhout Group, Inc. | 3d positioning of augmented reality information |
US8488246B2 (en) * | 2010-02-28 | 2013-07-16 | Osterhout Group, Inc. | See-through near-eye display glasses including a curved polarizing film in the image source, a partially reflective, partially transmitting optical element and an optically flat film |
CN102906623A (en) | 2010-02-28 | 2013-01-30 | 奥斯特豪特集团有限公司 | Local advertising content on an interactive head-mounted eyepiece |
US20120194553A1 (en) * | 2010-02-28 | 2012-08-02 | Osterhout Group, Inc. | Ar glasses with sensor and user action based control of external devices with feedback |
WO2011110728A1 (en) * | 2010-03-12 | 2011-09-15 | Nokia Corporation | Light-guiding structures |
KR101772153B1 (en) * | 2010-03-17 | 2017-08-29 | 삼성디스플레이 주식회사 | Display device using diffractive lens |
US8134719B2 (en) | 2010-03-19 | 2012-03-13 | Carestream Health, Inc. | 3-D imaging using telecentric defocus |
JP5361800B2 (en) * | 2010-05-21 | 2013-12-04 | 三菱電機株式会社 | Optical waveguide type Q switch element and Q switch laser device |
US8649099B2 (en) | 2010-09-13 | 2014-02-11 | Vuzix Corporation | Prismatic multiple waveguide for near-eye display |
JP5646263B2 (en) * | 2010-09-27 | 2014-12-24 | 任天堂株式会社 | Image processing program, image processing apparatus, image processing system, and image processing method |
US9223137B2 (en) * | 2010-10-08 | 2015-12-29 | Seiko Epson Corporation | Virtual image display apparatus |
US8884984B2 (en) * | 2010-10-15 | 2014-11-11 | Microsoft Corporation | Fusing virtual content into real content |
CN102033319B (en) * | 2010-10-25 | 2015-07-15 | 北京理工大学 | Oxyopter type display device using holographic elements |
US8582209B1 (en) * | 2010-11-03 | 2013-11-12 | Google Inc. | Curved near-to-eye display |
US9292973B2 (en) * | 2010-11-08 | 2016-03-22 | Microsoft Technology Licensing, Llc | Automatic variable virtual focus for augmented reality displays |
US9304319B2 (en) * | 2010-11-18 | 2016-04-05 | Microsoft Technology Licensing, Llc | Automatic focus improvement for augmented reality displays |
JP2012119867A (en) | 2010-11-30 | 2012-06-21 | Nintendo Co Ltd | Program, device, system, and method for image processing |
US8988765B2 (en) * | 2010-12-07 | 2015-03-24 | Laser Light Engines, Inc. | Laser projection system with improved bit depth |
US8988463B2 (en) * | 2010-12-08 | 2015-03-24 | Microsoft Technology Licensing, Llc | Sympathetic optic adaptation for see-through display |
US9690099B2 (en) * | 2010-12-17 | 2017-06-27 | Microsoft Technology Licensing, Llc | Optimized focal area for augmented reality displays |
KR101876944B1 (en) * | 2010-12-22 | 2018-08-09 | 시리얼 테크놀로지즈 에스.에이. | Combined light modulation device for tracking users |
AU2011348122A1 (en) | 2010-12-24 | 2013-07-11 | Magic Leap Inc. | An ergonomic head mounted display device and optical system |
CN103261944A (en) * | 2010-12-28 | 2013-08-21 | 洛克希德马丁公司 | Head-mounted display apparatus employing one or more reflective optical surfaces |
US8432542B2 (en) | 2011-01-10 | 2013-04-30 | Eric T. Marple | Fiber optic probes utilizing GRIN lenses for spatially precise optical spectroscopy |
US8939579B2 (en) * | 2011-01-28 | 2015-01-27 | Light Prescriptions Innovators, Llc | Autofocusing eyewear, especially for presbyopia correction |
WO2012115768A1 (en) | 2011-02-27 | 2012-08-30 | Dolby Laboratories Licensing Corporation | Multiview projector system |
TWI465768B (en) * | 2011-03-21 | 2014-12-21 | Nat Univ Tsing Hua | Head-up display (hud) |
CN202041739U (en) * | 2011-05-14 | 2011-11-16 | 马建栋 | Glasses for treating myopia |
US20120306850A1 (en) | 2011-06-02 | 2012-12-06 | Microsoft Corporation | Distributed asynchronous localization and mapping for augmented reality |
JP2013013450A (en) * | 2011-06-30 | 2013-01-24 | Namco Bandai Games Inc | Program, information storage medium, terminal and server |
CA2750287C (en) * | 2011-08-29 | 2012-07-03 | Microsoft Corporation | Gaze detection in a see-through, near-eye, mixed reality display |
US9025252B2 (en) * | 2011-08-30 | 2015-05-05 | Microsoft Technology Licensing, Llc | Adjustment of a mixed reality display for inter-pupillary distance alignment |
US9323325B2 (en) | 2011-08-30 | 2016-04-26 | Microsoft Technology Licensing, Llc | Enhancing an object of interest in a see-through, mixed reality display device |
US8998414B2 (en) * | 2011-09-26 | 2015-04-07 | Microsoft Technology Licensing, Llc | Integrated eye tracking and display system |
WO2013051051A1 (en) | 2011-10-03 | 2013-04-11 | 古野電気株式会社 | Touch-panel-equipped device, radar device, plotter device, shipboard network system, information display method, and information display program |
US20130108229A1 (en) * | 2011-10-28 | 2013-05-02 | Google Inc. | Heads-up display including ambient light control |
US8752963B2 (en) * | 2011-11-04 | 2014-06-17 | Microsoft Corporation | See-through display brightness control |
CA2858208C (en) | 2011-11-23 | 2019-01-15 | Magic Leap, Inc. | Three dimensional virtual and augmented reality display system |
US20130135359A1 (en) | 2011-11-30 | 2013-05-30 | Qualcomm Mems Technologies, Inc. | Display systems including illumination and optical touchscreen |
CN103135233B (en) * | 2011-12-01 | 2015-10-14 | 财团法人车辆研究测试中心 | Head-up display device |
US20130147686A1 (en) | 2011-12-12 | 2013-06-13 | John Clavin | Connecting Head Mounted Displays To External Displays And Other Communication Networks |
US9223138B2 (en) * | 2011-12-23 | 2015-12-29 | Microsoft Technology Licensing, Llc | Pixel opacity for augmented reality |
US8917453B2 (en) * | 2011-12-23 | 2014-12-23 | Microsoft Corporation | Reflective array waveguide |
US20130194304A1 (en) * | 2012-02-01 | 2013-08-01 | Stephen Latta | Coordinate-system sharing for augmented reality |
JP6111635B2 (en) | 2012-02-24 | 2017-04-12 | セイコーエプソン株式会社 | Virtual image display device |
CN103293675A (en) * | 2012-02-24 | 2013-09-11 | 精工爱普生株式会社 | Virtual image display device |
KR20130097429A (en) | 2012-02-24 | 2013-09-03 | 삼성전자주식회사 | Focusing apparatus for optical device |
JP5919885B2 (en) * | 2012-02-28 | 2016-05-18 | セイコーエプソン株式会社 | Virtual image display device |
JP2013178639A (en) * | 2012-02-28 | 2013-09-09 | Seiko Epson Corp | Head mounted display device and image display system |
US20130229712A1 (en) * | 2012-03-02 | 2013-09-05 | Google Inc. | Sandwiched diffractive optical combiner |
JP5957972B2 (en) * | 2012-03-07 | 2016-07-27 | セイコーエプソン株式会社 | Virtual image display device |
JP5919899B2 (en) * | 2012-03-08 | 2016-05-18 | セイコーエプソン株式会社 | Virtual image display device and method for adjusting position of virtual image display device |
JP6035793B2 (en) * | 2012-03-14 | 2016-11-30 | ソニー株式会社 | Image display device and image generation device |
US9578318B2 (en) * | 2012-03-14 | 2017-02-21 | Microsoft Technology Licensing, Llc | Imaging structure emitter calibration |
US8985803B2 (en) * | 2012-03-21 | 2015-03-24 | Microsoft Technology Licensing, Llc | Freeform-prism eyepiece with illumination waveguide |
US9116337B1 (en) * | 2012-03-21 | 2015-08-25 | Google Inc. | Increasing effective eyebox size of an HMD |
US9274338B2 (en) | 2012-03-21 | 2016-03-01 | Microsoft Technology Licensing, Llc | Increasing field of view of reflective waveguide |
US8736963B2 (en) * | 2012-03-21 | 2014-05-27 | Microsoft Corporation | Two-dimensional exit-pupil expansion |
JP2013200474A (en) * | 2012-03-26 | 2013-10-03 | Jvc Kenwood Corp | Image display device and control method for image display device |
US20130285885A1 (en) * | 2012-04-25 | 2013-10-31 | Andreas G. Nowatzyk | Head-mounted light-field display |
JP6145966B2 (en) * | 2012-05-09 | 2017-06-14 | ソニー株式会社 | Display device |
US20130300635A1 (en) | 2012-05-09 | 2013-11-14 | Nokia Corporation | Method and apparatus for providing focus correction of displayed information |
US20130300634A1 (en) * | 2012-05-09 | 2013-11-14 | Nokia Corporation | Method and apparatus for determining representations of displayed information based on focus distance |
EP4235270A3 (en) * | 2012-05-18 | 2023-11-08 | RealD Spark, LLC | Directionally illuminated waveguide arrangement |
US10502876B2 (en) * | 2012-05-22 | 2019-12-10 | Microsoft Technology Licensing, Llc | Waveguide optics focus elements |
US9116666B2 (en) | 2012-06-01 | 2015-08-25 | Microsoft Technology Licensing, Llc | Gesture based region identification for holograms |
US9250445B2 (en) | 2012-08-08 | 2016-02-02 | Carol Ann Tosaya | Multiple-pixel-beam retinal displays |
US9151887B2 (en) | 2012-09-04 | 2015-10-06 | Corning Incorporated | Multi-core optical fibers with single mode and multimode core elements |
US20140118360A1 (en) * | 2012-10-30 | 2014-05-01 | Pixtronix, Inc. | Thinfilm stacks for light modulating displays |
CN202975477U (en) * | 2012-11-28 | 2013-06-05 | 联想(北京)有限公司 | Head-mounted electronic equipment |
CN102998799A (en) * | 2012-12-04 | 2013-03-27 | 深圳市长江力伟股份有限公司 | Near-to-eye display system for blending virtual with real scenes |
AU2014207545B2 (en) * | 2013-01-15 | 2018-03-15 | Magic Leap, Inc. | Ultra-high resolution scanning fiber display |
WO2015006784A2 (en) | 2013-07-12 | 2015-01-15 | Magic Leap, Inc. | Planar waveguide apparatus with diffraction element(s) and system employing same |
TWI541543B (en) * | 2013-10-21 | 2016-07-11 | 財團法人工業技術研究院 | Beam splitting module and projector device using the same |
US9791700B2 (en) | 2013-11-27 | 2017-10-17 | Magic Leap, Inc. | Virtual and augmented reality systems and methods |
-
2014
- 2014-11-27 US US14/555,585 patent/US9791700B2/en active Active
- 2014-11-27 EP EP23161107.0A patent/EP4220999A3/en active Pending
- 2014-11-27 KR KR1020227009156A patent/KR102493498B1/en active IP Right Grant
- 2014-11-27 CN CN201710242150.7A patent/CN107329260B/en active Active
- 2014-11-27 CN CN201480074051.7A patent/CN105934902B/en active Active
- 2014-11-27 CN CN201811242425.8A patent/CN109445095B/en active Active
- 2014-11-27 NZ NZ755272A patent/NZ755272A/en unknown
- 2014-11-27 CN CN201910628010.2A patent/CN110542938B/en active Active
- 2014-11-27 CN CN201710242158.3A patent/CN107315249B/en active Active
- 2014-11-27 CN CN201710242141.8A patent/CN109298526B/en active Active
- 2014-11-27 CN CN201710242115.5A patent/CN107329259B/en active Active
- 2014-11-27 KR KR1020167017170A patent/KR102268462B1/en active IP Right Grant
- 2014-11-27 AU AU2014354673A patent/AU2014354673B2/en active Active
- 2014-11-27 IL IL305162A patent/IL305162B1/en unknown
- 2014-11-27 CN CN201710242045.3A patent/CN107193126B/en active Active
- 2014-11-27 CN CN202110867733.5A patent/CN113568175B/en active Active
- 2014-11-27 KR KR1020247009378A patent/KR20240042677A/en not_active Application Discontinuation
- 2014-11-27 KR KR1020237002914A patent/KR102651578B1/en active IP Right Grant
- 2014-11-27 EP EP14866595.3A patent/EP3075090B1/en active Active
- 2014-11-27 IL IL291010A patent/IL291010B2/en unknown
- 2014-11-27 CN CN201710242095.1A patent/CN107219628B/en active Active
- 2014-11-27 CN CN201710242109.XA patent/CN107272199B/en active Active
- 2014-11-27 WO PCT/US2014/067791 patent/WO2015081313A2/en active Application Filing
- 2014-11-27 KR KR1020217018584A patent/KR102378457B1/en active IP Right Grant
- 2014-11-27 IL IL313875A patent/IL313875A/en unknown
- 2014-11-27 NZ NZ720610A patent/NZ720610A/en unknown
- 2014-11-27 CA CA2931776A patent/CA2931776A1/en active Pending
- 2014-11-27 CN CN201710242061.2A patent/CN107300769B/en active Active
- 2014-11-27 CN CN201811242449.3A patent/CN109597202B/en active Active
- 2014-11-27 CN CN201710242143.7A patent/CN107203045B/en active Active
- 2014-11-27 JP JP2016534677A patent/JP2017500605A/en not_active Withdrawn
- 2014-11-27 CN CN202110760710.4A patent/CN113433700B/en active Active
-
2015
- 2015-05-04 US US14/703,168 patent/US9846967B2/en active Active
- 2015-05-05 US US14/704,519 patent/US20150235440A1/en not_active Abandoned
- 2015-05-05 US US14/704,784 patent/US20150235443A1/en not_active Abandoned
- 2015-05-05 US US14/704,832 patent/US20150235446A1/en not_active Abandoned
- 2015-05-05 US US14/704,782 patent/US20150235442A1/en not_active Abandoned
- 2015-05-05 US US14/704,537 patent/US20150234476A1/en not_active Abandoned
- 2015-05-05 US US14/704,275 patent/US20150235436A1/en not_active Abandoned
- 2015-05-05 US US14/704,803 patent/US20150235444A1/en not_active Abandoned
- 2015-05-05 US US14/704,662 patent/US20150234190A1/en not_active Abandoned
- 2015-05-05 US US14/704,827 patent/US20150235445A1/en not_active Abandoned
- 2015-05-05 US US14/704,484 patent/US20150235439A1/en not_active Abandoned
- 2015-05-05 US US14/704,438 patent/US20150235438A1/en not_active Abandoned
- 2015-05-05 US US14/704,816 patent/US20150234205A1/en not_active Abandoned
- 2015-05-05 US US14/704,321 patent/US20150235437A1/en not_active Abandoned
- 2015-05-05 US US14/704,863 patent/US20150235448A1/en not_active Abandoned
- 2015-05-05 US US14/704,242 patent/US20150235418A1/en not_active Abandoned
- 2015-05-06 US US14/704,983 patent/US9846306B2/en active Active
- 2015-05-06 US US14/705,648 patent/US20150243089A1/en not_active Abandoned
- 2015-05-06 US US14/705,491 patent/US20150235461A1/en not_active Abandoned
- 2015-05-06 US US14/705,214 patent/US20150235457A1/en not_active Abandoned
- 2015-05-06 US US14/705,237 patent/US20150235458A1/en not_active Abandoned
- 2015-05-06 US US14/705,568 patent/US20150235462A1/en not_active Abandoned
- 2015-05-06 US US14/705,223 patent/US20150235421A1/en not_active Abandoned
- 2015-05-06 US US14/705,276 patent/US20150248046A1/en not_active Abandoned
- 2015-05-06 US US14/705,490 patent/US20150234254A1/en not_active Abandoned
- 2015-05-06 US US14/705,741 patent/US9939643B2/en active Active
- 2015-05-06 US US14/705,769 patent/US9804397B2/en active Active
- 2015-05-06 US US14/704,990 patent/US20150243088A1/en not_active Abandoned
- 2015-05-06 US US14/705,184 patent/US20150235455A1/en not_active Abandoned
- 2015-05-06 US US14/705,667 patent/US20150235464A1/en not_active Abandoned
- 2015-05-06 US US14/705,804 patent/US9915824B2/en active Active
- 2015-05-06 US US14/705,723 patent/US20150235466A1/en not_active Abandoned
- 2015-05-06 US US14/705,748 patent/US9946071B2/en active Active
- 2015-05-06 US US14/705,245 patent/US20150235459A1/en not_active Abandoned
- 2015-05-06 US US14/704,985 patent/US20150235454A1/en not_active Abandoned
- 2015-05-06 US US14/704,987 patent/US20150235419A1/en not_active Abandoned
- 2015-05-06 US US14/705,255 patent/US20150235460A1/en not_active Abandoned
- 2015-05-06 US US14/705,603 patent/US20150235463A1/en not_active Abandoned
- 2015-05-06 US US14/705,197 patent/US20150235456A1/en not_active Abandoned
- 2015-05-06 US US14/705,867 patent/US9841601B2/en active Active
- 2015-05-06 US US14/704,989 patent/US20150235420A1/en not_active Abandoned
- 2015-05-06 US US14/705,675 patent/US20150235465A1/en not_active Abandoned
- 2015-05-06 US US14/705,530 patent/US20150234191A1/en not_active Abandoned
- 2015-05-06 US US14/705,825 patent/US20150235467A1/en not_active Abandoned
- 2015-05-07 US US14/706,783 patent/US20150241702A1/en not_active Abandoned
- 2015-05-07 US US14/706,444 patent/US20150243092A1/en not_active Abandoned
- 2015-05-07 US US14/706,196 patent/US20150243090A1/en not_active Abandoned
- 2015-05-07 US US14/706,551 patent/US20150243094A1/en not_active Abandoned
- 2015-05-07 US US14/706,763 patent/US20150241701A1/en not_active Abandoned
- 2015-05-07 US US14/706,232 patent/US20150235469A1/en not_active Abandoned
- 2015-05-07 US US14/706,734 patent/US20150243097A1/en not_active Abandoned
- 2015-05-07 US US14/706,739 patent/US20150243098A1/en not_active Abandoned
- 2015-05-07 US US14/706,428 patent/US20150243091A1/en not_active Abandoned
- 2015-05-07 US US14/706,216 patent/US20150235468A1/en not_active Abandoned
- 2015-05-07 US US14/706,241 patent/US20150235470A1/en not_active Abandoned
- 2015-05-07 US US14/706,358 patent/US20150235472A1/en not_active Abandoned
- 2015-05-07 US US14/706,398 patent/US20150235473A1/en not_active Abandoned
- 2015-05-07 US US14/706,635 patent/US20150241698A1/en not_active Abandoned
- 2015-05-07 US US14/706,808 patent/US20150241703A1/en not_active Abandoned
- 2015-05-07 US US14/706,658 patent/US20150243096A1/en not_active Abandoned
- 2015-05-07 US US14/706,840 patent/US20150319342A1/en not_active Abandoned
- 2015-05-07 US US14/706,625 patent/US20150241697A1/en not_active Abandoned
- 2015-05-07 US US14/706,842 patent/US20150243099A1/en not_active Abandoned
- 2015-05-07 US US14/706,813 patent/US20160109708A1/en not_active Abandoned
- 2015-05-07 US US14/706,507 patent/US20150243093A1/en not_active Abandoned
- 2015-05-07 US US14/706,681 patent/US20150241699A1/en not_active Abandoned
- 2015-05-07 US US14/706,278 patent/US20150235471A1/en not_active Abandoned
- 2015-05-07 US US14/706,586 patent/US20150243095A1/en not_active Abandoned
- 2015-05-07 US US14/706,690 patent/US20150241700A1/en not_active Abandoned
- 2015-05-07 US US14/706,596 patent/US20150241696A1/en not_active Abandoned
- 2015-05-08 US US14/707,281 patent/US20150248158A1/en not_active Abandoned
- 2015-05-08 US US14/707,813 patent/US20150241707A1/en not_active Abandoned
- 2015-05-08 US US14/707,432 patent/US20150241706A1/en not_active Abandoned
- 2015-05-08 US US14/707,224 patent/US20150248786A1/en not_active Abandoned
- 2015-05-08 US US14/707,296 patent/US20150243104A1/en not_active Abandoned
- 2015-05-08 US US14/707,302 patent/US20150248006A1/en not_active Abandoned
- 2015-05-08 US US14/707,002 patent/US20150243101A1/en not_active Abandoned
- 2015-05-08 US US14/707,337 patent/US20150248010A1/en not_active Abandoned
- 2015-05-08 US US14/707,332 patent/US20150248790A1/en not_active Abandoned
- 2015-05-08 US US14/707,779 patent/US20150309315A1/en not_active Abandoned
- 2015-05-08 US US14/707,247 patent/US20150243102A1/en not_active Abandoned
- 2015-05-08 US US14/707,236 patent/US20150248011A1/en not_active Abandoned
- 2015-05-08 US US14/707,265 patent/US20150243103A1/en not_active Abandoned
- 2015-05-08 US US14/707,519 patent/US20150248012A1/en not_active Abandoned
- 2015-05-08 US US14/707,735 patent/US20150243107A1/en not_active Abandoned
- 2015-05-08 US US14/707,257 patent/US20150241704A1/en not_active Abandoned
-
2016
- 2016-05-26 IL IL245878A patent/IL245878B/en active IP Right Grant
-
2017
- 2017-10-10 US US15/729,494 patent/US10643392B2/en active Active
- 2017-10-10 US US15/729,462 patent/US10629004B2/en active Active
- 2017-10-30 AU AU2017254813A patent/AU2017254813B2/en active Active
- 2017-10-30 AU AU2017254807A patent/AU2017254807B2/en active Active
- 2017-10-30 AU AU2017254811A patent/AU2017254811B2/en active Active
- 2017-10-30 AU AU2017254801A patent/AU2017254801B2/en active Active
- 2017-10-30 AU AU2017254803A patent/AU2017254803B2/en active Active
- 2017-10-30 AU AU2017254800A patent/AU2017254800B2/en active Active
- 2017-10-30 AU AU2017254798A patent/AU2017254798B2/en active Active
- 2017-11-24 JP JP2017225727A patent/JP6529143B2/en active Active
- 2017-11-24 JP JP2017225735A patent/JP6510016B2/en active Active
- 2017-11-24 JP JP2017225726A patent/JP6510012B2/en active Active
- 2017-11-24 JP JP2017225733A patent/JP6513167B2/en active Active
- 2017-11-24 JP JP2017225736A patent/JP6600675B2/en active Active
- 2017-11-24 JP JP2017225729A patent/JP6510014B2/en active Active
- 2017-11-24 JP JP2017225732A patent/JP6514302B2/en active Active
- 2017-11-24 JP JP2017225730A patent/JP2018060212A/en not_active Withdrawn
- 2017-11-24 JP JP2017225734A patent/JP2018060214A/en active Pending
- 2017-11-24 JP JP2017225731A patent/JP6510015B2/en active Active
- 2017-11-24 JP JP2017225728A patent/JP6510013B2/en active Active
-
2018
- 2018-05-03 US US15/970,552 patent/US10529138B2/en active Active
- 2018-07-30 IL IL260865A patent/IL260865B/en active IP Right Grant
- 2018-07-30 IL IL260870A patent/IL260870B/en active IP Right Grant
- 2018-07-30 IL IL260871A patent/IL260871B/en active IP Right Grant
- 2018-07-30 IL IL260869A patent/IL260869B/en active IP Right Grant
- 2018-07-30 IL IL260872A patent/IL260872B/en unknown
- 2018-07-30 IL IL26086618A patent/IL260866B/en active IP Right Grant
- 2018-07-30 IL IL260867A patent/IL260867B/en active IP Right Grant
- 2018-10-22 AU AU2018250534A patent/AU2018250534A1/en not_active Abandoned
-
2019
- 2019-02-28 JP JP2019035288A patent/JP6720368B2/en active Active
- 2019-08-05 JP JP2019143520A patent/JP6971281B2/en active Active
- 2019-08-05 JP JP2019143519A patent/JP6763070B2/en active Active
- 2019-11-06 AU AU2019261725A patent/AU2019261725B2/en active Active
-
2020
- 2020-03-10 US US16/814,975 patent/US10935806B2/en active Active
- 2020-06-18 AU AU2020204085A patent/AU2020204085B2/en active Active
-
2021
- 2021-01-22 US US17/155,412 patent/US11237403B2/en active Active
- 2021-02-07 IL IL280685A patent/IL280685B/en unknown
- 2021-11-01 JP JP2021178713A patent/JP7179140B2/en active Active
- 2021-12-20 US US17/555,640 patent/US11714291B2/en active Active
-
2022
- 2022-10-05 AU AU2022246404A patent/AU2022246404B2/en active Active
- 2022-11-15 JP JP2022182285A patent/JP7432687B2/en active Active
-
2023
- 2023-06-06 US US18/329,982 patent/US20230324706A1/en active Pending
- 2023-06-21 AU AU2023203920A patent/AU2023203920B2/en active Active
-
2024
- 2024-02-05 JP JP2024015576A patent/JP2024040250A/en active Pending
- 2024-02-23 AU AU2024201233A patent/AU2024201233B2/en active Active
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10338400B2 (en) | 2017-07-03 | 2019-07-02 | Holovisions LLC | Augmented reality eyewear with VAPE or wear technology |
US10859834B2 (en) | 2017-07-03 | 2020-12-08 | Holovisions | Space-efficient optical structures for wide field-of-view augmented reality (AR) eyewear |
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9804397B2 (en) | Using a freedom reflective and lens optical component for augmented or virtual reality display |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STCB | Information on status: application discontinuation |
Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION |
|
AS | Assignment |
Owner name: JP MORGAN CHASE BANK, N.A., NEW YORK Free format text: PATENT SECURITY AGREEMENT;ASSIGNORS:MAGIC LEAP, INC.;MOLECULAR IMPRINTS, INC.;MENTOR ACQUISITION ONE, LLC;REEL/FRAME:050138/0287 Effective date: 20190820 |
|
AS | Assignment |
Owner name: CITIBANK, N.A., NEW YORK Free format text: ASSIGNMENT OF SECURITY INTEREST IN PATENTS;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:050967/0138 Effective date: 20191106 |