US20120326948A1 - Environmental-light filter for see-through head-mounted display device - Google Patents
Environmental-light filter for see-through head-mounted display device Download PDFInfo
- Publication number
- US20120326948A1 US20120326948A1 US13/166,346 US201113166346A US2012326948A1 US 20120326948 A1 US20120326948 A1 US 20120326948A1 US 201113166346 A US201113166346 A US 201113166346A US 2012326948 A1 US2012326948 A1 US 2012326948A1
- Authority
- US
- United States
- Prior art keywords
- light
- user
- environmental
- augmented
- head
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 230000007613 environmental effect Effects 0.000 claims abstract description 24
- 230000003287 optical effect Effects 0.000 claims abstract description 20
- 230000008878 coupling Effects 0.000 claims abstract description 9
- 238000010168 coupling process Methods 0.000 claims abstract description 9
- 238000005859 coupling reaction Methods 0.000 claims abstract description 9
- 230000003190 augmentative effect Effects 0.000 claims description 23
- 238000001914 filtration Methods 0.000 claims description 6
- 238000000576 coating method Methods 0.000 claims description 4
- 239000011521 glass Substances 0.000 claims description 4
- 239000002131 composite material Substances 0.000 claims description 2
- 239000004744 fabric Substances 0.000 claims description 2
- 239000002184 metal Substances 0.000 claims description 2
- 239000004033 plastic Substances 0.000 claims description 2
- 238000001429 visible spectrum Methods 0.000 claims description 2
- 238000000034 method Methods 0.000 description 10
- 210000003128 head Anatomy 0.000 description 8
- 230000033001 locomotion Effects 0.000 description 7
- 238000013459 approach Methods 0.000 description 5
- 230000015654 memory Effects 0.000 description 5
- 241001481833 Coryphaena hippurus Species 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 210000001747 pupil Anatomy 0.000 description 4
- 239000007787 solid Substances 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 230000000981 bystander Effects 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 210000004279 orbit Anatomy 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- CDBYLPFSWZWCQE-UHFFFAOYSA-L Sodium Carbonate Chemical compound [Na+].[Na+].[O-]C([O-])=O CDBYLPFSWZWCQE-UHFFFAOYSA-L 0.000 description 1
- 239000000853 adhesive Substances 0.000 description 1
- 230000001070 adhesive effect Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 210000004087 cornea Anatomy 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 238000010438 heat treatment Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 230000013011 mating Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 239000000049 pigment Substances 0.000 description 1
- 230000019612 pigmentation Effects 0.000 description 1
- 230000010287 polarization Effects 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 239000005060 rubber Substances 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B5/00—Optical elements other than lenses
- G02B5/20—Filters
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B27/0172—Head mounted characterised by optical features
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0112—Head-up displays characterised by optical features comprising device for genereting colour display
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/0101—Head-up displays characterised by optical features
- G02B2027/0118—Head-up displays characterised by optical features comprising devices for improving the contrast of the display / brillance control visibility
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
- G02B2027/0178—Eyeglass type
Definitions
- Head-mounted displays can be used in various applications, including military, aviation, medicine, video gaming, entertainment, sports, and so forth. See-through head-mounted displays allow the user to observe the physical world around him or her, while optical elements add light from one or more small micro-displays into the user's visual path, to provide an augmented reality image.
- the augmented-reality image may relate to a real-world scene that represents an environment in which a user is located.
- various challenges exist in providing an augmented-reality image that is realistic and that can represent a full range of colors and intensities.
- this disclosure describes, among other things, an environmental-light filter for an optical see-through head-mounted display device.
- the head-mounted display device uses the environmental-light filter to increase the primacy of an augmented-reality image as seen by a user.
- the filter may block light from a real-world scene to avoid the augmented-reality image appearing transparent.
- an environmental-light filter is configured to be removably coupled to an optical see-through head-mounted display (HMD) device.
- the HMD device includes a see-through lens extending between a user's eye and a real-world scene when the display device is worn by the user.
- the see-through lens also has a display component that, in combination with an augmented reality emitter, such as a micro-display, provides an augmented-reality image to a user's eye.
- the environmental-light filter may be selected from a group of filters with varied opacity, from a minimum opacity level which allows a substantial amount of light to pass, to a maximum opacity level which allows little or no light to pass. As such, the environmental-light filter enhances the appearance of the augmented-reality image by reducing or eliminating environmental light that reaches the user's eye.
- FIG. 1 is perspective view depicting an environmental-light filter and a head-mounted display unit worn on a user's head in accordance with an embodiment of the invention
- FIG. 2 is a perspective view depicting an environmental-light filter removably coupled to a head-mounted display unit in accordance with an embodiment of the invention
- FIG. 3 is an elevational side view of an environmental-light filter removably coupled to a head-mounted display unit in accordance with an embodiment of the invention
- FIG. 4 is a block diagram depicting an environmental-light filter removably coupled to a head-mounted display unit in accordance with an embodiment of the invention
- FIG. 5 is a block diagram of a head-mounted display unit in accordance with and embodiment of the invention.
- FIG. 6 is a flow diagram depicting a method for providing an augmented-reality image in accordance with an embodiment of the invention
- FIG. 7A is an exemplary image of a real-world scene in accordance with an embodiment of the invention.
- FIG. 7B is an exemplary image of an augmented-reality image in accordance with an embodiment of the invention.
- FIG. 7C is an exemplary image of a combined image without an environmental-light filter in accordance with an embodiment of the invention.
- FIG. 7D is an exemplary image of a combined image with an environmental-light filter having a light transmissivity of about 70% in accordance with an embodiment of the invention.
- FIG. 7E is an exemplary image of a combined image with an environmental-light filter having a light transmissivity of about 50% in accordance with an embodiment of the invention.
- FIG. 7F is an exemplary image of a combined image with an environmental-light filter having a light transmissivity of about 10% in accordance with an embodiment of the invention.
- HMDs head-mounted displays
- optical elements such as minors, prisms, and holographic lenses to add light from one or more small micro-displays into a user's visual path.
- these elements can only add light, but cannot remove light.
- Augmented-reality images as described herein comprise any image, graphic, or other output provided to the user's eye by the augmented-reality emitter.
- Augmented-reality images provided by an HMD device generally appear superimposed on a background and may appear to interact with or be integral with the background.
- the background is comprised of a real-world scene, e.g. a scene that a user would perceive without any augmented-reality image emitted by the HMD device.
- the real-world scene may be similar to what the user would see without the HMD device or a filter lens may be provided between the user's eye and the real-world scene to filter at least a portion of the environmental light from the real-world scene as describe herein.
- the augmented-reality image may need to be provided at a sufficiently high intensity that is brighter than the corresponding portion of a real-world scene, for the augmented-reality image to be distinct and not transparent.
- a lens of an HMD device can be provided with an environmental-light filter of a selected opacity or light transmissivity to block a desired amount of environmental light.
- Environmental-light filter lenses are described herein with respect to their opacity or light transmissivity.
- Opacity and light transmissivity each refer to an amount of light that is allowed to pass through a lens.
- An opacity of 100% or a light transmissivity of 0% indicates that no light passes through a lens while an opacity of 0% or light transmissivity of 100% is indicative of a perfectly clear lens.
- a lens having decreasing transmissivity from about 70% to about 10% exhibits an increasing tinted appearance while a lens having less that about 10-15% transmissivity appears substantially opaque.
- an environmental-light filter lens for a head-mounted display device is disclosed.
- the filter lens is configured to at least partially filter environmental light received by a user's eye and is removeably coupled to a head-mounted display device having a see-through lens extending between the user's eye and a real-world scene when the head-mounted device is worn by the user.
- the head-mounted display device includes a display component and an augmented-reality emitter which emits light to the user's eye using the display component to provide an augmented reality image.
- the filter lens causes the augmented-reality image to appear less transparent than when the filter lens is not coupled to the head-mounted device.
- a head-mounted display device in another embodiment, includes see-through lens extending between a user's eye and a real-world scene when the head-mounted device is worn by the user.
- the see-through lens includes a display component that receives light comprising an augmented-reality image and directs the light toward the user's eye.
- An augmented-reality emitter that emits light to provide the augmented reality image is also provided.
- the device further includes a frame configured to carry the see-through lens and the augmented reality emitter for wearing on a user's head.
- a filter lens that has a light transmissivity between 0 and 70% to at least partially filter environmental light received by a user's eye is removeably coupled to the frame.
- the frame includes the augmented-reality emitter and the filter lens is substantially opaque.
- the environmental-light filter 100 includes one or more filter lenses 104 , a frame 106 , and a plurality of coupling features 108 .
- a pair of filter lenses 104 is configured to correspond to a pair of see-through lenses 110 of the HMD device 102 .
- the filter lenses 104 comprise a single lens that extends across both see-through lenses 110 of the HMD device 102 .
- the lenses 104 may extend along sides or temples 112 of the HMD device 102 to provide filtering of environmental light from a user's periphery.
- additional filter lenses can be provided along the user's periphery and may be coupled to the filter lenses 104 or to the temple 112 of the HMD device 102 directly.
- the filter lenses 104 are comprised of any available materials including, for example and not limitation, glass, plastic, metal, rubber, fabrics, composites, or other suitable materials configured to provide a desired level of light transmissivity.
- the filter lenses 104 might include one or more features such as pigments, tints, colorations, coatings (e.g. anti-glare coatings), or filter layers (e.g. polarizing filters) disposed on surfaces thereof or formed integral therewith.
- the features increase the visibility and/or quality of an augmented-reality image.
- the features block, absorb, or filter light that is reflected off of a user's eye 114 such that the reflected light is not visible to bystanders viewing the user. Such may avoid an undesired appearance of a user's eyes 114 being illuminated and may secure any data or images viewed by a user from being visible to bystanders.
- the coatings, filter layers, and any other pigmentation or filtration elements of the lenses 104 are uniform across the entire lens 104 or can include gradients or variations throughout the lens 104 . Such might be employed to provide predetermined areas of the lens 104 with increased opacity where portions of an augmented-reality image are commonly displayed.
- the filter lenses 104 are configured to filter, block, or absorb a desired amount and/or desired spectrum of wavelengths of environmental light that is incident thereon.
- Environmental light includes any radiation that a user might ordinarily encounter (e.g., visible light, ultraviolet light, and infrared light) from sources such as the sun, light bulbs, heating elements, and the like.
- the filter lenses 104 are configured to filter, block, or absorb substantially all ultraviolet light and infrared light and at least a portion of light received in the visible spectrum. Filtration of infrared light may avoid interference with infrared sensors employed by the HMD device 102 .
- the frame 106 is configured similarly to a frame employed for eyeglasses.
- the frame 106 is constructed from any desired materials known in the art and retains the filter lenses 104 in an orientation comparable to that of the see-through lenses 110 of the HMD device 102 .
- the frame 106 is integral with the filter lenses 104 or is not employed with the filter lenses 104 .
- the frame 106 is bendable and/or includes one or more hinged joints to enable folding of the environmental-light filter 100 for storage.
- the frame 106 may also extend along the temples 112 of the HMD device 102 to retain peripheral environmental-light filters (not shown).
- the coupling features 108 include any component useable to removably couple the filter lenses 104 and/or the frame 106 to the HMD device 102 .
- the features 108 might include one or more clips, clasps, hooks, tabs, flanges, latches, lugs, or the like.
- the features might include a plurality of tabs 116 extending from the frame 106 with an orthogonally-extending flange 118 at a distal end thereof.
- the tabs 116 extend the thickness of the HMD device 102 and the flange 118 engages a backside 120 thereof to removeably retain the frame 106 against or adjacent to a front surface 122 of the HMD device 102 .
- the coupling features 108 may provide a coupling by snap-fit, friction-fit, adhesion, or mechanical interlocking, among others.
- a spring-biased clip is mounted on the frame 106 between the filter lenses 104 and is coupled to the HMD device 102 by a user (not shown).
- a plurality of magnetic elements are included on the frame 106 and magnetically couple to mating elements on the HMD device 102 (not shown).
- features such as suction cups, static cling, adhesives, or the like might also be employed.
- the coupling features 108 may also be configure for quick and easy removal of the filter 100 from the HMD device 102 such that a user's normal vision can be easily restored in the event of an emergency.
- the HMD device 102 is depicted as worn on a user's head 124 .
- the HMD device 102 includes a frame 126 similar to a conventional eyeglasses frame and can be worn with a similar comfort level.
- a face shield which is mounted to the user's head by a helmet, strap or other means.
- the frame 126 includes a frame front 122 and the temples 112 .
- the frame front 122 holds the see-through lens 110 L for the user's left eye and a see-through lens 110 R for the user's right eye.
- the left and right orientations are from the user's perspective.
- the left-side see-through lens 110 L includes a light-transmissive environmental-light filter lens 104 L removably coupled in association therewith and a light-transmissive optical display component 128 L such as a beam splitter which mixes an augmented-reality image 402 with light from a real-world scene 404 for viewing by the left eye 114 L, as depicted in FIG. 4 .
- An opening (not shown) in the environmental-light filter lens 104 L can be provided to allow an eye tracking component 132 to image the left eye 114 L, including the pupil thereof.
- the opening can be, e.g., a hole in the filter lens 104 L, or a region of the filter lens 104 L in which the filtering is reduced or not provided.
- the eye tracking component 132 employs infrared light and at least a portion of the filter lens 104 L has a high light transmissivity for infrared light.
- the eye tracking component 132 includes an infrared (IR) emitter 134 that emits IR light 136 and an IR sensor 138 that senses reflected IR light 140 .
- the eye tracking component 132 is mounted directly on, and inside, the frame 126 . In this implementation, the eye tracking component 132 does not need to project infrared light through the filter lens 104 or the see-through lens 110 . In another embodiment, the eye tracking component 132 can be mounted to the frame via an arm (not show) that extends from the frame 126 .
- the right-side see-through lens 11 OR includes a light-transmissive environmental-light filter 104 R removably coupled in association therewith and an optical display component 128 , such as a beam splitter that mixes an augmented-reality image 402 with light from a real-world scene 404 for viewing by the right eye 114 R.
- a right-side augmented-reality emitter 130 R is mounted to the frame 126 via an arm 142 R, and a left-side augmented-reality emitter 130 L is mounted to the frame 126 via an arm 142 L.
- the display device is shown from a perspective of the user looking forward, so that the right-side lens 11 OR and the left-side lens 110 L are depicted.
- the right-side augmented-reality emitter 130 R includes a light emitting portion (not shown), such as a grid of pixels, and a portion which may include circuitry for controlling the light-emitting portion.
- the left-side augmented-reality emitter 130 L includes a light-emitting portion and a portion with circuitry for controlling the light emitting portion.
- each of the optical components 128 L and 128 R may have the same dimensions.
- the right-side optical component 128 R includes a top surface 148 R through which light enters from the right-side augmented-reality emitter 130 R, an angled half-mirrored surface 150 R within the optical component 128 R, and a face 152 R.
- Light from the right-side augmented-reality emitter 130 R and from portions of a real-world scene (represented by ray 406 in FIG. 4 ) that are not blocked by the environmental-light filter 104 R pass through the face 152 R and enter the user's right-side eye 114 R.
- the left-side optical component 128 L includes a top surface 148 L through which light enters from the left-side augmented-reality emitter 130 L, an angled half-mirrored surface 150 L within the optical component 128 L, and a face 152 L.
- Light from the left-side augmented-reality emitter 130 L and from portions of the real-world scene which are not blocked by the environmental-light filter 104 L pass through the face 152 L and enter the user's left-side eye 114 L.
- Each of the environmental-light filters 104 may have the same dimensions.
- the same augmented-reality image is provided to both eyes 114 , although it is possible to provide a separate image to each eye 114 , such as for a stereoscopic effect.
- only one augmented-reality emitter 130 is routed by appropriate optical components to both eyes 114 .
- the display device 102 includes the see-through lens 110 which is placed in front of a user's eye 114 , similarly to an eyeglass lens. Typically, a pair of see-through lenses 110 is provided, one for each eye 114 .
- the lens 110 includes an optical display component 128 , such as a beam splitter, e.g., a half-silvered mirror or other light-transmissive minor. Then environmental-light filter 100 is removeably coupled in front of the lens 110 .
- Light from the real-world scene 404 such as a light ray 406 , reaches the lens 110 and is partially or completely blocked by the environmental-light filter 100 .
- the light from the real-world scene 404 that passes through the environmental-light filter 100 also passes through the display component 128 .
- An augmented-reality emitter 130 emits a 2-D array of light representing an augmented-reality image 402 and exemplified by a light ray 408 . Additional optics are typically used to refocus the augmented-reality image 402 so that it appears to originate from several feet away from the eye 114 rather than from about one inch away, where the display component 128 actually is.
- the augmented-reality image 402 is reflected by the display component 128 toward a user's eye 114 , as exemplified by a light ray 410 , so that the user sees an image 412 .
- a portion of the real-world scene 404 such as a grove of trees, is visible, along with the entire augmented-reality image 402 , such as a flying dolphin.
- the user therefore sees a fanciful image 412 in which a dolphin flies past trees, in this entertainment-oriented example.
- an augmented-reality image might appear as a can of soda on a user's desk. Many other applications are possible.
- the user can wear the HMD device 102 anywhere, including indoors or outdoors.
- Various pieces of information can be obtained to determine what type of augmented-reality image 402 is appropriate and where it should be provided on the display component 128 .
- the location of the user, the direction in which the user is looking, and the location of floors, walls and perhaps furniture, when the user is indoors can be used to decide where to place the augmented-reality image 402 in an appropriate location in the real world scene 404 when combined into the image 412 .
- the direction in which the user is looking can be determined by tracking a position of the user's head using a combination of motion tracking techniques and an inertial measure unit which is attached to the user's head, such as via the HMD device 102 .
- Motion tracking techniques use a depth sensing camera to obtain a 3-D model of the user.
- a depth sensing camera can similarly be used to obtain the location of floors, walls, and other aspects of the user's environment. See, e.g., U.S. Patent Publication No. 2010/0197399, published Aug. 5, 2010, titled “Visual Target Tracking,” U.S. Patent Publication No. 2010/0194872, published Aug. 5, 2010, titled “Body Scan,” and U.S. Pat. No. 7,717,173, issued Apr. 7, 2009, titled “Head Pose Tracking System,” each of which is hereby incorporated herein in its entirety by reference.
- the tracking camera 132 can be used to identify a location of the user's eye 114 with respect to a frame 126 on which the HMD device 102 is mounted.
- the frame 126 can be similar to conventional eyeglass frames, in one approach, as depicted in FIGS. 1-3 .
- a frame 126 can move slightly on the user's head when worn, e.g., due to motions of the user, slipping of the bridge of the frame on the user's nose, and so forth.
- the augmented-reality emitter 130 can adjust its image 402 , accordingly.
- the augmented-reality image 402 can be made to appear more stable.
- the tracking camera 132 includes an infrared (IR) emitter 134 which emits IR light 136 toward the eye 114 , and an IR sensor 138 which senses reflected IR light 140 .
- IR infrared
- the position of the pupil can be identified by known imaging techniques such as detecting the reflection of the cornea. See, for example, U.S. Pat. No. 7,401,920, titled “Head Mounted Eye Tracking and Display System,” issued Jul. 22, 2008 to Ophir et al., incorporated herein by reference in its entirety. Such techniques can locate a position of the center of the eye 114 relative to the tracking camera 132 .
- eye tracking involves obtaining an image of the eye 114 and using computer vision techniques to determine the location of the pupil within the eye socket.
- Other eye tracking techniques can use arrays of photo detectors and LEDs.
- the location of the eye 114 with respect to any other location that is fixed relative to the frame 126 can be determined.
- the rotation of the eyes e.g., the movement of the pupil within the eye socket
- the tracking camera 132 images the eye 114 from a side position on the frame 126 that is independent from the environmental-light filter 100 and optical display component 128 .
- light used by the tracking camera 132 could be carried via the display component 128 or otherwise integrated into the lens 110 .
- the HMD device 102 provides passive stereoscopic vision. Since the environmental-light filters 100 may be polarized, right and left lenses 104 can be oriented so that the polarization is different by 90 degrees. As such, the HMD device 102 equipped with the environmental-light filters 104 can be used with a comparably equipped 3-D display to view images in 3-D.
- FIG. 5 depicts a system diagram of the HMD device 102 of FIGS. 1-3 .
- the system includes the eye tracking camera 132 and the augmented-reality emitter 130 which can communicate with one another via a bus 502 or other communication paths.
- the eye tracking camera 132 includes a processor 504 , a memory 506 , the IR emitter 134 , the IR sensor 138 , and an interface 508 .
- the memory 506 can contain instructions which are executed by the processor 504 to enable the eye tracking camera 132 to perform its functions as described herein.
- the interface 508 allows the eye tracking camera 132 to communicate data to the augmented-reality emitter 130 that indicates the relative location of the user's eye 114 with respect to the frame 126 .
- the augmented-reality emitter 130 includes a processor 510 , a memory 512 , a light emitter 514 that emits visible light, and an interface 516 .
- the memory 512 can contain instructions which are executed by the processor 510 to enable the augmented-reality emitter 130 to perform its functions as described herein.
- the light emitter 514 can be a micro-display such as an LCD which emits a 2-D color image in a small area such as one quarter inch square.
- the interface 516 may be used to communicate with the eye tracking camera 132 .
- One of more of the processors 504 and 510 can be considered to be control circuits.
- one or more of the memories 506 and 512 can be considered to be a tangible computer-readable storage having computer-readable software embodied thereon for programming at least one processor or control circuit to perform a method for use in an optical see-through HMD device 102 as described herein.
- the system may further include components, discussed previously, such as for determining a direction in which the user is looking, the location of floors, walls and other aspects of the user's environment.
- FIG. 6 depicts a process 600 for providing an augmented-reality image in an HMD device, such as the HMD device 102 with the environmental-light filter 100 coupled thereto.
- an eye tracking component provides data regarding the relative location of a user's eye(s). Generally, this can be performed several times per second. The data can indicate an offset of the eye from a default location, such as when the eye is looking straight ahead. The location data can be based on the data regarding the relative location of the eye.
- An augmented-reality image is an image which is set based on the needs of an application in which it is used. For instance, the previous example of a flying dolphin is provided for an entertainment application.
- the augmented-reality emitter emits the augmented reality image, so that it reaches the user's eye via one or more optical display components.
- the process is repeated starting at step 602 . When there is no next augmented reality image, the process ends at step 608 .
- the next augmented-reality image can refer to the same augmented-reality image as previously provided, but in a different location, as seen by the user, such as when the previous augmented-reality image is moved to a slightly different location to depict movement of the augmented reality image.
- the next augmented-reality image can also refer to a new type of image, such as switching from a dolphin to another type of object.
- the next augmented-reality image can also refer to adding a new object while a previously displayed object continues to be displayed.
- the augmented-reality emitter emits video images at a fixed frame rate.
- static images are emitted and persisted for a period of time which is greater than a typical video frame period.
- Step 610 optionally provides a gradual fade in the augmented reality image, such as when it is near a boundary of an augmented reality display region of a field of view.
- the augmented reality display region can be defined by the maximum angular extent (vertically and horizontally) in the user's field of view in which the augmented-reality image is constrained, due to limitations of the augmented-reality emitter and/or optical components 128 .
- the augmented-reality image can appear in any portion of the augmented reality display region, but not outside the augmented reality display region.
- the field of view of a user is the angular extent of the observable world, vertically and horizontally, that is seen at any given moment. Humans have an almost 180-degree forward-facing field of view. However, the ability to perceive color is greater in the center of the field of view, while the ability to perceive shapes and motion is greater in the periphery of the field of view.
- the augmented-reality image is constrained to being provided in a subset region of the user's field of view.
- the augmented-reality image is provided in the center of the field of view over an angular extent of about 20 degrees, which lines up with the fovea of the eye. This is the augmented reality display region of the field of view.
- the augmented-reality image is constrained by factors such as the size of the optical components used to route the augmented-reality image to the user's eye.
- FIGS. 7A-F exemplary images provided by environmental-light filters removably coupled to an HMD device are described in accordance with an embodiment of the invention.
- the images of the FIGS. 7A-F are depicted as viewed by a user wearing an HMD device and under identical environmental lighting conditions.
- FIG. 7A depicts a real-world scene 700 that might be viewed by a user wearing the HMD device without the environmental filter coupled thereto.
- An exemplary augmented-reality image 702 is depicted in FIG. 6B and a combined image 704 comprised of the real-world scene 700 and the augmented-reality image 702 is depicted in FIG. 7C .
- the augmented-reality image 702 appears generally transparent and the real-world scene 700 is visible through the augmented reality image 702 .
- This combined image 704 might be the result of viewing when environmental lighting conditions are bright and/or when the brightness of the augmented-reality image 702 is too dim with respect to the real-world scene 700 .
- the brightness of the augmented-reality image 702 may be increased, however the brightness required to sufficiently overcome environmental lighting may exceed the HMD device's capabilities, increase power consumption, and result in an uncomfortable viewing experience for a user.
- the transparent effect might also be overcome by applying an environmental-light filter, such as the environmental-light filter 100 .
- FIG. 7D depicts a combined image 706 that might be viewed by a user with a mild environmental-light filter removably coupled to an HMD device.
- the mild environmental-light filter has a light transmissivity between about 70 and about 50%%, however any light transmissivity from about 100 to about 50% might be employed.
- the augmented-reality image 702 appears more opaque or solid, however the real-world scene 700 is still, at least partially visible through the augmented-reality image 702 .
- FIG. 7E depicts a combined image 708 that might be viewed by a user with a strong environmental-light filter removably coupled to an HMD device.
- the strong environmental-light filter has a light transmissivity between about 15 and about 50%.
- the augmented-reality image 702 appears more opaque or solid, and the real-world scene 700 is not clearly visible through the augmented-reality image 702 .
- the real-world scene 700 is at least partially visible outside the augmented-reality image 702 .
- FIG. 7F depicts a combined image 710 that might be viewed by a user with a substantially opaque environmental-light filter removably coupled to an HMD device.
- the substantially opaque environmental-light filter has a light transmissivity between about 15 and about 0%.
- the augmented-reality image 702 appears opaque or solid and the real-world scene 700 is only visible to a very small extent or is not visible.
- environmental-light filters are selectable based on a level of light transmissivity to provide a desired appearance in a combined image.
- a user might select a mild filter to enhance an augmented-reality image to a small extent, such as when the augmented-reality image is only being viewed periodically or nonchalantly, e.g. the augmented-reality image may provide a heads-up display of the current time and date which the user only periodically focuses his or her attention on.
- the user might select a mild filter on a cloudy day or when in low environmental lighting conditions because greater filtration of the environmental light is not needed.
- Such, may provide the user with a desired viewing experience of the augmented-reality image while also not obstructing the user's ability to view the real-world scene.
- a user might desire to view an augmented-reality image without the real-world scene.
- the user might select a substantially opaque environmental-light filter. For instance, a user playing a video game, reading a book, or watching a movie via the augmented-reality image while sunbathing on a beach would likely desire to block out most of the environmental light to provide sufficient viewability of the augmented-reality image.
- environmental-light filters might be provided in a kit that includes a plurality of filters each having a different light transmissivity. As such, a user can select an environmental-light filter based on environmental conditions and a desired viewing experience.
Abstract
An environmental-light filter removably coupled to an optical see-through head-mounted display (HMD) device is disclosed. The environmental-light filter couples to the HMD device between a display component and a real-world scene. Coupling features are provided to allow the filter to be easily and removably attached to the HMD device when desired by a user. The filter increases the primacy of a provided augmented-reality image with respect to a real-world scene and reduces brightness and power consumption requirements for presenting the augmented-reality image. A plurality of filters of varied light transmissivity may be provided from which to select a desired filter based on environmental lighting conditions and user preference. The light transmissivity of the filter may be about 70% light transmissive to substantially or completely opaque.
Description
- Head-mounted displays can be used in various applications, including military, aviation, medicine, video gaming, entertainment, sports, and so forth. See-through head-mounted displays allow the user to observe the physical world around him or her, while optical elements add light from one or more small micro-displays into the user's visual path, to provide an augmented reality image. The augmented-reality image may relate to a real-world scene that represents an environment in which a user is located. However, various challenges exist in providing an augmented-reality image that is realistic and that can represent a full range of colors and intensities.
- Embodiments of the invention are defined by the claims below, not this summary. A high-level overview of various aspects of the invention are provided here for that reason, to provide an overview of the disclosure, and to introduce a selection of concepts that are further described below in the detailed-description section below. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in isolation to determine the scope of the claimed subject matter. In brief and at a high level, this disclosure describes, among other things, an environmental-light filter for an optical see-through head-mounted display device. The head-mounted display device uses the environmental-light filter to increase the primacy of an augmented-reality image as seen by a user. For example, the filter may block light from a real-world scene to avoid the augmented-reality image appearing transparent.
- In one embodiment, an environmental-light filter is configured to be removably coupled to an optical see-through head-mounted display (HMD) device. The HMD device includes a see-through lens extending between a user's eye and a real-world scene when the display device is worn by the user. The see-through lens also has a display component that, in combination with an augmented reality emitter, such as a micro-display, provides an augmented-reality image to a user's eye. The environmental-light filter may be selected from a group of filters with varied opacity, from a minimum opacity level which allows a substantial amount of light to pass, to a maximum opacity level which allows little or no light to pass. As such, the environmental-light filter enhances the appearance of the augmented-reality image by reducing or eliminating environmental light that reaches the user's eye.
- Illustrative embodiments of the invention are described in detail below with reference to the attached drawing figures, and wherein:
-
FIG. 1 is perspective view depicting an environmental-light filter and a head-mounted display unit worn on a user's head in accordance with an embodiment of the invention; -
FIG. 2 is a perspective view depicting an environmental-light filter removably coupled to a head-mounted display unit in accordance with an embodiment of the invention; -
FIG. 3 is an elevational side view of an environmental-light filter removably coupled to a head-mounted display unit in accordance with an embodiment of the invention; -
FIG. 4 is a block diagram depicting an environmental-light filter removably coupled to a head-mounted display unit in accordance with an embodiment of the invention; -
FIG. 5 is a block diagram of a head-mounted display unit in accordance with and embodiment of the invention; -
FIG. 6 is a flow diagram depicting a method for providing an augmented-reality image in accordance with an embodiment of the invention; -
FIG. 7A is an exemplary image of a real-world scene in accordance with an embodiment of the invention; -
FIG. 7B is an exemplary image of an augmented-reality image in accordance with an embodiment of the invention; -
FIG. 7C is an exemplary image of a combined image without an environmental-light filter in accordance with an embodiment of the invention; -
FIG. 7D is an exemplary image of a combined image with an environmental-light filter having a light transmissivity of about 70% in accordance with an embodiment of the invention; -
FIG. 7E is an exemplary image of a combined image with an environmental-light filter having a light transmissivity of about 50% in accordance with an embodiment of the invention; and -
FIG. 7F is an exemplary image of a combined image with an environmental-light filter having a light transmissivity of about 10% in accordance with an embodiment of the invention. - The subject matter of select embodiments of the invention is described with specificity herein to meet statutory requirements. But the description itself is not intended to necessarily limit the scope of claims. Rather, the claimed subject matter might be embodied in other ways to include different steps or combinations of steps similar to the ones described in this document, in conjunction with other present or future technologies. Terms should not be interpreted as implying any particular order among or between various steps herein disclosed unless and except when the order of individual steps is explicitly described.
- See-through head-mounted displays (HMDs) most often use optical elements such as minors, prisms, and holographic lenses to add light from one or more small micro-displays into a user's visual path. By their very nature, these elements can only add light, but cannot remove light. This means a virtual display cannot display darker colors—they tend towards transparent in the case of pure black—and virtual objects such as augmented reality images seem translucent or ghosted.
- For compelling augmented-reality or other mixed-reality scenarios, it is desirable to have the ability to selectively remove natural or environmental light from the view so that virtual color imagery can both represent the full range of colors and intensities, while making that imagery seem more solid or real. Moreover, removal of environmental light reduces power consumption by the augmented-reality emitter because the augmented-reality image can be provided at a lower intensity.
- Augmented-reality images as described herein comprise any image, graphic, or other output provided to the user's eye by the augmented-reality emitter. Augmented-reality images provided by an HMD device generally appear superimposed on a background and may appear to interact with or be integral with the background. The background is comprised of a real-world scene, e.g. a scene that a user would perceive without any augmented-reality image emitted by the HMD device. The real-world scene may be similar to what the user would see without the HMD device or a filter lens may be provided between the user's eye and the real-world scene to filter at least a portion of the environmental light from the real-world scene as describe herein.
- Without the environmental-light filter, the augmented-reality image may need to be provided at a sufficiently high intensity that is brighter than the corresponding portion of a real-world scene, for the augmented-reality image to be distinct and not transparent. To achieve this goal, a lens of an HMD device can be provided with an environmental-light filter of a selected opacity or light transmissivity to block a desired amount of environmental light.
- Environmental-light filter lenses are described herein with respect to their opacity or light transmissivity. Opacity and light transmissivity each refer to an amount of light that is allowed to pass through a lens. An opacity of 100% or a light transmissivity of 0% indicates that no light passes through a lens while an opacity of 0% or light transmissivity of 100% is indicative of a perfectly clear lens. For example, a lens having decreasing transmissivity from about 70% to about 10% exhibits an increasing tinted appearance while a lens having less that about 10-15% transmissivity appears substantially opaque.
- In an embodiment, an environmental-light filter lens for a head-mounted display device is disclosed. The filter lens is configured to at least partially filter environmental light received by a user's eye and is removeably coupled to a head-mounted display device having a see-through lens extending between the user's eye and a real-world scene when the head-mounted device is worn by the user. The head-mounted display device includes a display component and an augmented-reality emitter which emits light to the user's eye using the display component to provide an augmented reality image. The filter lens causes the augmented-reality image to appear less transparent than when the filter lens is not coupled to the head-mounted device.
- In another embodiment, a head-mounted display device is disclosed. The device includes see-through lens extending between a user's eye and a real-world scene when the head-mounted device is worn by the user. The see-through lens includes a display component that receives light comprising an augmented-reality image and directs the light toward the user's eye. An augmented-reality emitter that emits light to provide the augmented reality image is also provided. The device further includes a frame configured to carry the see-through lens and the augmented reality emitter for wearing on a user's head. A filter lens that has a light transmissivity between 0 and 70% to at least partially filter environmental light received by a user's eye is removeably coupled to the frame. In another embodiment, the frame includes the augmented-reality emitter and the filter lens is substantially opaque.
- With reference to
FIGS. 1-3 , an environmental-light filter 100 configured to removeably couple to anHMD device 102 is described in accordance with an embodiment of the invention. The environmental-light filter 100 includes one or more filter lenses 104, aframe 106, and a plurality of coupling features 108. A pair of filter lenses 104 is configured to correspond to a pair of see-through lenses 110 of theHMD device 102. However, in an embodiment, the filter lenses 104 comprise a single lens that extends across both see-through lenses 110 of theHMD device 102. Additionally, the lenses 104 may extend along sides or temples 112 of theHMD device 102 to provide filtering of environmental light from a user's periphery. Alternatively, additional filter lenses (not show) can be provided along the user's periphery and may be coupled to the filter lenses 104 or to the temple 112 of theHMD device 102 directly. - The filter lenses 104 are comprised of any available materials including, for example and not limitation, glass, plastic, metal, rubber, fabrics, composites, or other suitable materials configured to provide a desired level of light transmissivity. The filter lenses 104 might include one or more features such as pigments, tints, colorations, coatings (e.g. anti-glare coatings), or filter layers (e.g. polarizing filters) disposed on surfaces thereof or formed integral therewith. In an embodiment, the features increase the visibility and/or quality of an augmented-reality image. In another embodiment, the features block, absorb, or filter light that is reflected off of a user's eye 114 such that the reflected light is not visible to bystanders viewing the user. Such may avoid an undesired appearance of a user's eyes 114 being illuminated and may secure any data or images viewed by a user from being visible to bystanders.
- Additionally, the coatings, filter layers, and any other pigmentation or filtration elements of the lenses 104 are uniform across the entire lens 104 or can include gradients or variations throughout the lens 104. Such might be employed to provide predetermined areas of the lens 104 with increased opacity where portions of an augmented-reality image are commonly displayed.
- The filter lenses 104 are configured to filter, block, or absorb a desired amount and/or desired spectrum of wavelengths of environmental light that is incident thereon. Environmental light includes any radiation that a user might ordinarily encounter (e.g., visible light, ultraviolet light, and infrared light) from sources such as the sun, light bulbs, heating elements, and the like. In an embodiment, the filter lenses 104 are configured to filter, block, or absorb substantially all ultraviolet light and infrared light and at least a portion of light received in the visible spectrum. Filtration of infrared light may avoid interference with infrared sensors employed by the
HMD device 102. - The
frame 106 is configured similarly to a frame employed for eyeglasses. Theframe 106 is constructed from any desired materials known in the art and retains the filter lenses 104 in an orientation comparable to that of the see-through lenses 110 of theHMD device 102. In an embodiment, theframe 106 is integral with the filter lenses 104 or is not employed with the filter lenses 104. In another embodiment, theframe 106 is bendable and/or includes one or more hinged joints to enable folding of the environmental-light filter 100 for storage. Theframe 106 may also extend along the temples 112 of theHMD device 102 to retain peripheral environmental-light filters (not shown). - The coupling features 108 include any component useable to removably couple the filter lenses 104 and/or the
frame 106 to theHMD device 102. Thefeatures 108 might include one or more clips, clasps, hooks, tabs, flanges, latches, lugs, or the like. For example, as depicted inFIGS. 1 and 2 , the features might include a plurality oftabs 116 extending from theframe 106 with an orthogonally-extendingflange 118 at a distal end thereof. Thetabs 116 extend the thickness of theHMD device 102 and theflange 118 engages abackside 120 thereof to removeably retain theframe 106 against or adjacent to afront surface 122 of theHMD device 102. The coupling features 108 may provide a coupling by snap-fit, friction-fit, adhesion, or mechanical interlocking, among others. - In an embodiment, a spring-biased clip is mounted on the
frame 106 between the filter lenses 104 and is coupled to theHMD device 102 by a user (not shown). In yet another embodiment, a plurality of magnetic elements are included on theframe 106 and magnetically couple to mating elements on the HMD device 102 (not shown). Further, features such as suction cups, static cling, adhesives, or the like might also be employed. The coupling features 108 may also be configure for quick and easy removal of thefilter 100 from theHMD device 102 such that a user's normal vision can be easily restored in the event of an emergency. - Although, embodiments of the environmental-
light filter 100 are described herein with respect to selected configurations, it is to be understood that the environmental-light filter 100 may be configurable in a variety of ways with additional or different features without departing from the scope described herein. Such configurations are understood as being disclosed herein. - With continued reference to
FIG. 1 , theHMD device 102 is depicted as worn on a user'shead 124. In this example, theHMD device 102 includes aframe 126 similar to a conventional eyeglasses frame and can be worn with a similar comfort level. However, other implementations are possible, such as a face shield which is mounted to the user's head by a helmet, strap or other means. Theframe 126 includes aframe front 122 and the temples 112. Theframe front 122 holds the see-throughlens 110L for the user's left eye and a see-throughlens 110R for the user's right eye. - Many components of the environmental-
light filter 100 and theHMD device 102 are provided in pairs, one for each eye 114 of a user. Such components are indicated herein by similar reference numbers having an “L” or an “R” appended thereto to indicate a left or right side component, respectively. Or those components are generally referred to by the reference numeral alone. - The left and right orientations are from the user's perspective. The left-side see-through
lens 110L includes a light-transmissive environmental-light filter lens 104L removably coupled in association therewith and a light-transmissiveoptical display component 128L such as a beam splitter which mixes an augmented-reality image 402 with light from a real-world scene 404 for viewing by theleft eye 114L, as depicted inFIG. 4 . An opening (not shown) in the environmental-light filter lens 104L can be provided to allow aneye tracking component 132 to image theleft eye 114L, including the pupil thereof. The opening can be, e.g., a hole in thefilter lens 104L, or a region of thefilter lens 104L in which the filtering is reduced or not provided. In an embodiment, theeye tracking component 132 employs infrared light and at least a portion of thefilter lens 104L has a high light transmissivity for infrared light. - In an embodiment, the
eye tracking component 132 includes an infrared (IR) emitter 134 that emits IR light 136 and anIR sensor 138 that senses reflectedIR light 140. Theeye tracking component 132 is mounted directly on, and inside, theframe 126. In this implementation, theeye tracking component 132 does not need to project infrared light through the filter lens 104 or the see-through lens 110. In another embodiment, theeye tracking component 132 can be mounted to the frame via an arm (not show) that extends from theframe 126. - The right-side see-through lens 11OR includes a light-transmissive environmental-
light filter 104R removably coupled in association therewith and anoptical display component 128, such as a beam splitter that mixes an augmented-reality image 402 with light from a real-world scene 404 for viewing by the right eye 114R. A right-side augmented-reality emitter 130R is mounted to theframe 126 via anarm 142R, and a left-side augmented-reality emitter 130L is mounted to theframe 126 via anarm 142L. - Referring now to
FIG. 2 , the display device is shown from a perspective of the user looking forward, so that the right-side lens 11OR and the left-side lens 110L are depicted. The right-side augmented-reality emitter 130R includes a light emitting portion (not shown), such as a grid of pixels, and a portion which may include circuitry for controlling the light-emitting portion. Similarly, the left-side augmented-reality emitter 130L includes a light-emitting portion and a portion with circuitry for controlling the light emitting portion. In one approach, each of theoptical components optical component 128R includes atop surface 148R through which light enters from the right-side augmented-reality emitter 130R, an angled half-mirroredsurface 150R within theoptical component 128R, and aface 152R. Light from the right-side augmented-reality emitter 130R and from portions of a real-world scene (represented byray 406 inFIG. 4 ) that are not blocked by the environmental-light filter 104R pass through theface 152R and enter the user's right-side eye 114R. - Similarly, the left-side
optical component 128L includes atop surface 148L through which light enters from the left-side augmented-reality emitter 130L, an angled half-mirroredsurface 150L within theoptical component 128L, and aface 152L. Light from the left-side augmented-reality emitter 130L and from portions of the real-world scene which are not blocked by the environmental-light filter 104L pass through theface 152L and enter the user's left-side eye 114L. Each of the environmental-light filters 104 may have the same dimensions. - Typically, the same augmented-reality image is provided to both eyes 114, although it is possible to provide a separate image to each eye 114, such as for a stereoscopic effect. In an alternative implementation, only one augmented-
reality emitter 130 is routed by appropriate optical components to both eyes 114. - With additional reference now to
FIG. 4 , the operation of the environmental-light filter 100 and theHMD device 102 is described in accordance with an embodiment of the invention. Thedisplay device 102 includes the see-through lens 110 which is placed in front of a user's eye 114, similarly to an eyeglass lens. Typically, a pair of see-through lenses 110 is provided, one for each eye 114. The lens 110 includes anoptical display component 128, such as a beam splitter, e.g., a half-silvered mirror or other light-transmissive minor. Then environmental-light filter 100 is removeably coupled in front of the lens 110. Light from the real-world scene 404, such as alight ray 406, reaches the lens 110 and is partially or completely blocked by the environmental-light filter 100. The light from the real-world scene 404 that passes through the environmental-light filter 100 also passes through thedisplay component 128. - An augmented-
reality emitter 130 emits a 2-D array of light representing an augmented-reality image 402 and exemplified by alight ray 408. Additional optics are typically used to refocus the augmented-reality image 402 so that it appears to originate from several feet away from the eye 114 rather than from about one inch away, where thedisplay component 128 actually is. - The augmented-
reality image 402 is reflected by thedisplay component 128 toward a user's eye 114, as exemplified by alight ray 410, so that the user sees animage 412. In theimage 412, a portion of the real-world scene 404, such as a grove of trees, is visible, along with the entire augmented-reality image 402, such as a flying dolphin. The user therefore sees afanciful image 412 in which a dolphin flies past trees, in this entertainment-oriented example. In an advertising-oriented example, an augmented-reality image might appear as a can of soda on a user's desk. Many other applications are possible. - Generally, the user can wear the
HMD device 102 anywhere, including indoors or outdoors. Various pieces of information can be obtained to determine what type of augmented-reality image 402 is appropriate and where it should be provided on thedisplay component 128. For example, the location of the user, the direction in which the user is looking, and the location of floors, walls and perhaps furniture, when the user is indoors, can be used to decide where to place the augmented-reality image 402 in an appropriate location in thereal world scene 404 when combined into theimage 412. - The direction in which the user is looking can be determined by tracking a position of the user's head using a combination of motion tracking techniques and an inertial measure unit which is attached to the user's head, such as via the
HMD device 102. Motion tracking techniques use a depth sensing camera to obtain a 3-D model of the user. A depth sensing camera can similarly be used to obtain the location of floors, walls, and other aspects of the user's environment. See, e.g., U.S. Patent Publication No. 2010/0197399, published Aug. 5, 2010, titled “Visual Target Tracking,” U.S. Patent Publication No. 2010/0194872, published Aug. 5, 2010, titled “Body Scan,” and U.S. Pat. No. 7,717,173, issued Apr. 7, 2009, titled “Head Pose Tracking System,” each of which is hereby incorporated herein in its entirety by reference. - The tracking
camera 132 can be used to identify a location of the user's eye 114 with respect to aframe 126 on which theHMD device 102 is mounted. Theframe 126 can be similar to conventional eyeglass frames, in one approach, as depicted inFIGS. 1-3 . Typically, such aframe 126 can move slightly on the user's head when worn, e.g., due to motions of the user, slipping of the bridge of the frame on the user's nose, and so forth. By providing real-time information regarding the location of the eye 114 with respect to theframe 126, the augmented-reality emitter 130 can adjust itsimage 402, accordingly. For example, the augmented-reality image 402 can be made to appear more stable. As depicted inFIG. 1 , in an embodiment, the trackingcamera 132 includes an infrared (IR) emitter 134 which emits IR light 136 toward the eye 114, and anIR sensor 138 which senses reflectedIR light 140. - The position of the pupil can be identified by known imaging techniques such as detecting the reflection of the cornea. See, for example, U.S. Pat. No. 7,401,920, titled “Head Mounted Eye Tracking and Display System,” issued Jul. 22, 2008 to Ophir et al., incorporated herein by reference in its entirety. Such techniques can locate a position of the center of the eye 114 relative to the
tracking camera 132. - Generally, eye tracking involves obtaining an image of the eye 114 and using computer vision techniques to determine the location of the pupil within the eye socket. Other eye tracking techniques can use arrays of photo detectors and LEDs. With a known mounting location of the tracking
camera 132 on theframe 126, the location of the eye 114 with respect to any other location that is fixed relative to theframe 126, such as the environmental-light filter 100 and thedisplay component 128, can be determined. Typically it is sufficient to track the location of one of the user's eyes 114 since the eyes 114 move in unison. However, it is also possible to track each eye 114 separately and use the location of each eye 114 to determine the location of the augmented-reality image 402 for the associated see-through lens 110. In most cases, it is sufficient to know the displacement of the augmented reality glasses relative to the eyes as the glasses bounce around during motion. The rotation of the eyes (e.g., the movement of the pupil within the eye socket) is often less consequential. - In the example depicted in
FIG. 1 , the trackingcamera 132 images the eye 114 from a side position on theframe 126 that is independent from the environmental-light filter 100 andoptical display component 128. However, other approaches are possible. For example, light used by the trackingcamera 132 could be carried via thedisplay component 128 or otherwise integrated into the lens 110. - In another embodiment, the
HMD device 102 provides passive stereoscopic vision. Since the environmental-light filters 100 may be polarized, right and left lenses 104 can be oriented so that the polarization is different by 90 degrees. As such, theHMD device 102 equipped with the environmental-light filters 104 can be used with a comparably equipped 3-D display to view images in 3-D. -
FIG. 5 depicts a system diagram of theHMD device 102 ofFIGS. 1-3 . The system includes theeye tracking camera 132 and the augmented-reality emitter 130 which can communicate with one another via abus 502 or other communication paths. Theeye tracking camera 132 includes aprocessor 504, amemory 506, theIR emitter 134, theIR sensor 138, and aninterface 508. Thememory 506 can contain instructions which are executed by theprocessor 504 to enable theeye tracking camera 132 to perform its functions as described herein. Theinterface 508 allows theeye tracking camera 132 to communicate data to the augmented-reality emitter 130 that indicates the relative location of the user's eye 114 with respect to theframe 126. - The augmented-
reality emitter 130 includes aprocessor 510, amemory 512, alight emitter 514 that emits visible light, and aninterface 516. Thememory 512 can contain instructions which are executed by theprocessor 510 to enable the augmented-reality emitter 130 to perform its functions as described herein. Thelight emitter 514 can be a micro-display such as an LCD which emits a 2-D color image in a small area such as one quarter inch square. Theinterface 516 may be used to communicate with theeye tracking camera 132. - One of more of the
processors memories HMD device 102 as described herein. The system may further include components, discussed previously, such as for determining a direction in which the user is looking, the location of floors, walls and other aspects of the user's environment. -
FIG. 6 depicts aprocess 600 for providing an augmented-reality image in an HMD device, such as theHMD device 102 with the environmental-light filter 100 coupled thereto. At astep 602, an eye tracking component provides data regarding the relative location of a user's eye(s). Generally, this can be performed several times per second. The data can indicate an offset of the eye from a default location, such as when the eye is looking straight ahead. The location data can be based on the data regarding the relative location of the eye. - An augmented-reality image is an image which is set based on the needs of an application in which it is used. For instance, the previous example of a flying dolphin is provided for an entertainment application. At a
step 604, the augmented-reality emitter emits the augmented reality image, so that it reaches the user's eye via one or more optical display components. At adecision step 606, when there is a next augmented reality image, the process is repeated starting atstep 602. When there is no next augmented reality image, the process ends atstep 608. - The next augmented-reality image can refer to the same augmented-reality image as previously provided, but in a different location, as seen by the user, such as when the previous augmented-reality image is moved to a slightly different location to depict movement of the augmented reality image. The next augmented-reality image can also refer to a new type of image, such as switching from a dolphin to another type of object. The next augmented-reality image can also refer to adding a new object while a previously displayed object continues to be displayed. In one approach, the augmented-reality emitter emits video images at a fixed frame rate. In another approach, static images are emitted and persisted for a period of time which is greater than a typical video frame period.
- Step 610 optionally provides a gradual fade in the augmented reality image, such as when it is near a boundary of an augmented reality display region of a field of view. The augmented reality display region can be defined by the maximum angular extent (vertically and horizontally) in the user's field of view in which the augmented-reality image is constrained, due to limitations of the augmented-reality emitter and/or
optical components 128. Thus, the augmented-reality image can appear in any portion of the augmented reality display region, but not outside the augmented reality display region. - Generally, the field of view of a user is the angular extent of the observable world, vertically and horizontally, that is seen at any given moment. Humans have an almost 180-degree forward-facing field of view. However, the ability to perceive color is greater in the center of the field of view, while the ability to perceive shapes and motion is greater in the periphery of the field of view.
- Furthermore, as mentioned, the augmented-reality image is constrained to being provided in a subset region of the user's field of view. In an exemplary implementation, the augmented-reality image is provided in the center of the field of view over an angular extent of about 20 degrees, which lines up with the fovea of the eye. This is the augmented reality display region of the field of view. The augmented-reality image is constrained by factors such as the size of the optical components used to route the augmented-reality image to the user's eye.
- Referring now to
FIGS. 7A-F , exemplary images provided by environmental-light filters removably coupled to an HMD device are described in accordance with an embodiment of the invention. The images of theFIGS. 7A-F are depicted as viewed by a user wearing an HMD device and under identical environmental lighting conditions.FIG. 7A depicts a real-world scene 700 that might be viewed by a user wearing the HMD device without the environmental filter coupled thereto. An exemplary augmented-reality image 702 is depicted inFIG. 6B and a combinedimage 704 comprised of the real-world scene 700 and the augmented-reality image 702 is depicted inFIG. 7C . - As depicted in
FIG. 7C , the augmented-reality image 702 appears generally transparent and the real-world scene 700 is visible through theaugmented reality image 702. Thiscombined image 704 might be the result of viewing when environmental lighting conditions are bright and/or when the brightness of the augmented-reality image 702 is too dim with respect to the real-world scene 700. To overcome the transparent effect, the brightness of the augmented-reality image 702 may be increased, however the brightness required to sufficiently overcome environmental lighting may exceed the HMD device's capabilities, increase power consumption, and result in an uncomfortable viewing experience for a user. The transparent effect might also be overcome by applying an environmental-light filter, such as the environmental-light filter 100. -
FIG. 7D depicts a combinedimage 706 that might be viewed by a user with a mild environmental-light filter removably coupled to an HMD device. In an embodiment, the mild environmental-light filter has a light transmissivity between about 70 and about 50%%, however any light transmissivity from about 100 to about 50% might be employed. As such, the augmented-reality image 702 appears more opaque or solid, however the real-world scene 700 is still, at least partially visible through the augmented-reality image 702. -
FIG. 7E depicts a combinedimage 708 that might be viewed by a user with a strong environmental-light filter removably coupled to an HMD device. In an embodiment, the strong environmental-light filter has a light transmissivity between about 15 and about 50%. As such, the augmented-reality image 702 appears more opaque or solid, and the real-world scene 700 is not clearly visible through the augmented-reality image 702. The real-world scene 700 is at least partially visible outside the augmented-reality image 702. -
FIG. 7F depicts a combinedimage 710 that might be viewed by a user with a substantially opaque environmental-light filter removably coupled to an HMD device. In an embodiment, the substantially opaque environmental-light filter has a light transmissivity between about 15 and about 0%. As such, the augmented-reality image 702 appears opaque or solid and the real-world scene 700 is only visible to a very small extent or is not visible. - As depicted in
FIGS. 7A-F , environmental-light filters are selectable based on a level of light transmissivity to provide a desired appearance in a combined image. For example, a user might select a mild filter to enhance an augmented-reality image to a small extent, such as when the augmented-reality image is only being viewed periodically or nonchalantly, e.g. the augmented-reality image may provide a heads-up display of the current time and date which the user only periodically focuses his or her attention on. Or the user might select a mild filter on a cloudy day or when in low environmental lighting conditions because greater filtration of the environmental light is not needed. Such, may provide the user with a desired viewing experience of the augmented-reality image while also not obstructing the user's ability to view the real-world scene. - Alternatively, a user might desire to view an augmented-reality image without the real-world scene. Thus, the user might select a substantially opaque environmental-light filter. For instance, a user playing a video game, reading a book, or watching a movie via the augmented-reality image while sunbathing on a beach would likely desire to block out most of the environmental light to provide sufficient viewability of the augmented-reality image.
- Accordingly, environmental-light filters might be provided in a kit that includes a plurality of filters each having a different light transmissivity. As such, a user can select an environmental-light filter based on environmental conditions and a desired viewing experience.
- Many different arrangements of the various components depicted, as well as components not shown, are possible without departing from the scope of the claims below. Embodiments of the technology have been described with the intent to be illustrative rather than restrictive. Alternative embodiments will become apparent to readers of this disclosure after and because of reading it. Alternative means of implementing the aforementioned can be completed without departing from the scope of the claims below. Certain features and subcombinations are of utility and may be employed without reference to other features and subcombinations and are contemplated within the scope of the claims.
Claims (20)
1. An environmental-light filter lens for a head-mounted display device, comprising:
a filter lens configured to at least partially filter environmental light received by a user's eye, and removeably coupled to a head-mounted display device to cause an augmented-reality image to appear less transparent than when the filter lens is not coupled to the head-mounted device, the head mounted device including a see-through lens extending between the user's eye and a real-world scene when the head-mounted device is worn by the user, a display component, and an augmented-reality emitter which emits light to the user's eye using the display component to provide the augmented reality image.
2. The environmental-light filter lens of claim 1 , wherein the head-mounted display device includes a frame and the filter lens includes one or more features configured to couple to the frame.
3. The environmental-light filter lens of claim 2 , wherein the features include one or more of clips, clasps, hooks, tabs, flanges, latches, or lugs for removeably coupling the filter lens to the frame.
4. The environmental-light filter lens of claim 1 , wherein the filter lens has a light transmissivity between 0 and 70%.
5. The environmental-light filter lens of claim 4 , wherein the filter lens is substantially opaque.
6. The environmental-light filter lens of claim 1 , wherein the filter lens includes one or more substantially opaque regions.
7. The environmental-light filter lens of claim 1 , wherein the filter lens blocks ultraviolet light, infrared light, and at least a portion of light in the visible spectrum.
8. The environmental-light filter lens of claim 1 , wherein the filter lens includes one or more coatings configured to absorb light from the augmented-reality image that is reflected off of the user.
9. The environmental-light filter lens of claim 2 , wherein the head-mounted display device further comprises a tracking component that tracks a location of the user's eye relative to the frame.
10. The environmental-light filter lens of claim 1 , wherein the display component comprises at least one optical component that combines light from the real-world scene and light representing the augmented reality image, the display component being between the filter lens and the user's eye.
11. A head-mounted display device, comprising:
a see-through lens extending between a user's eye and a real-world scene when the head-mounted device is worn by the user, the see-though lens including a display component that receives light comprising an augmented-reality image and directs the light toward the user's eye;
an augmented-reality emitter that emits light to provide the augmented reality image;
a frame configured to carry the see-through lens and the augmented reality emitter for wearing on a user's head; and
a filter lens that has a light transmissivity between 0 and 70% to at least partially filter environmental light received by a user's eye, and that is removeably coupled to the frame.
12. The head-mounted display device of claim 11 , wherein the filter lens causes the augmented-reality image to appear less transparent than when the filter lens is not coupled to the head-mounted device.
13. The head-mounted display device of claim 11 , wherein the filter lens includes one or more features configured to couple to the frame.
14. The head-mounted display device of claim 11 , wherein the features include one or more of clips, clasps, hooks, tabs, flanges, latches, or lugs for removeably coupling the filter lens to the frame.
15. The head-mounted display device of claim 11 , wherein the filter lens is made from one or more of a glass, plastic, metal, fabric, and a composite.
16. The head-mounted display device of claim 11 , wherein the filter lens substantially blocks ultraviolet light and infrared light in the environmental light from being received by the user's eye.
17. A head-mounted display device, comprising:
a see-through lens extending between a user's eye and a real-world scene when the head-mounted device is worn by the user, the see-though lens including a display component that receives light comprising an augmented-reality image and directs the light toward the user's eye;
a frame configured to carry the see-through lens for wearing on a user's head and including an augmented-reality emitter that emits light to provide the augmented reality image; and
a filter lens that is substantially opaque and substantially filters environmental light received by a user's eye, the filter lens being removeably coupled to the frame.
18. The head-mounted display device of claim 17 , wherein filtering of the environmental light by the filter lens causes the augmented-reality image to be a primary image seen by a user.
19. The head-mounted display device of claim 17 , wherein the filter lens includes one or more features that provide one or more of a snap-fit, a friction fit, or magnetic coupling between the filter lens and the frame.
20. The head-mounted display device of claim 17 , wherein filter lens has a light transmissivity between 0 and 10%.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/166,346 US20120326948A1 (en) | 2011-06-22 | 2011-06-22 | Environmental-light filter for see-through head-mounted display device |
US13/692,861 US8537075B2 (en) | 2011-06-22 | 2012-12-03 | Environmental-light filter for see-through head-mounted display device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/166,346 US20120326948A1 (en) | 2011-06-22 | 2011-06-22 | Environmental-light filter for see-through head-mounted display device |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/692,861 Continuation US8537075B2 (en) | 2011-06-22 | 2012-12-03 | Environmental-light filter for see-through head-mounted display device |
Publications (1)
Publication Number | Publication Date |
---|---|
US20120326948A1 true US20120326948A1 (en) | 2012-12-27 |
Family
ID=47361352
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/166,346 Abandoned US20120326948A1 (en) | 2011-06-22 | 2011-06-22 | Environmental-light filter for see-through head-mounted display device |
US13/692,861 Active US8537075B2 (en) | 2011-06-22 | 2012-12-03 | Environmental-light filter for see-through head-mounted display device |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/692,861 Active US8537075B2 (en) | 2011-06-22 | 2012-12-03 | Environmental-light filter for see-through head-mounted display device |
Country Status (1)
Country | Link |
---|---|
US (2) | US20120326948A1 (en) |
Cited By (121)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120188179A1 (en) * | 2010-12-10 | 2012-07-26 | Sony Ericsson Mobile Communications Ab | Touch sensitive display |
US20160004085A1 (en) * | 2014-07-02 | 2016-01-07 | Christian Stroetmann | Head-mounted display device with air conditioning device and control approaches |
US9401540B2 (en) | 2014-02-11 | 2016-07-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9423612B2 (en) | 2014-03-28 | 2016-08-23 | Osterhout Group, Inc. | Sensor dependent content position in head worn computing |
US20160247319A1 (en) * | 2015-02-20 | 2016-08-25 | Andreas G. Nowatzyk | Selective occlusion system for augmented reality devices |
US9436006B2 (en) | 2014-01-21 | 2016-09-06 | Osterhout Group, Inc. | See-through computer display systems |
US9494800B2 (en) | 2014-01-21 | 2016-11-15 | Osterhout Group, Inc. | See-through computer display systems |
US9523856B2 (en) | 2014-01-21 | 2016-12-20 | Osterhout Group, Inc. | See-through computer display systems |
US9529192B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9529195B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | See-through computer display systems |
US20170011557A1 (en) * | 2015-07-06 | 2017-01-12 | Samsung Electronics Co., Ltd | Method for providing augmented reality and virtual reality and electronic device using the same |
US9547465B2 (en) | 2014-02-14 | 2017-01-17 | Osterhout Group, Inc. | Object shadowing in head worn computing |
US9575321B2 (en) | 2014-06-09 | 2017-02-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US9607409B2 (en) | 2013-12-23 | 2017-03-28 | Empire Technology Development Llc | Suppression of real features in see-through display |
US9615742B2 (en) | 2014-01-21 | 2017-04-11 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9625723B2 (en) | 2013-06-25 | 2017-04-18 | Microsoft Technology Licensing, Llc | Eye-tracking system using a freeform prism |
US9632315B2 (en) | 2010-10-21 | 2017-04-25 | Lockheed Martin Corporation | Head-mounted display apparatus employing one or more fresnel lenses |
WO2017070226A1 (en) * | 2015-10-20 | 2017-04-27 | Lockheed Martin Corporation | Multiple-eye, single-display, ultrawide-field-of-view optical see-through augmented reality system |
US9651784B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US9672210B2 (en) | 2014-04-25 | 2017-06-06 | Osterhout Group, Inc. | Language translation with head-worn computing |
US9671613B2 (en) | 2014-09-26 | 2017-06-06 | Osterhout Group, Inc. | See-through computer display systems |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
US9696552B1 (en) * | 2014-01-10 | 2017-07-04 | Lockheed Martin Corporation | System and method for providing an augmented reality lightweight clip-on wearable device |
USD792400S1 (en) | 2014-12-31 | 2017-07-18 | Osterhout Group, Inc. | Computer glasses |
US9715112B2 (en) | 2014-01-21 | 2017-07-25 | Osterhout Group, Inc. | Suppression of stray light in head worn computing |
US9720234B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US9720228B2 (en) | 2010-12-16 | 2017-08-01 | Lockheed Martin Corporation | Collimating display with pixel lenses |
US9733478B2 (en) | 2014-04-03 | 2017-08-15 | Thomson Licensing | Ambient light management for virtual reality glasses |
USD794637S1 (en) | 2015-01-05 | 2017-08-15 | Osterhout Group, Inc. | Air mouse |
US9740280B2 (en) | 2014-01-21 | 2017-08-22 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9746686B2 (en) | 2014-05-19 | 2017-08-29 | Osterhout Group, Inc. | Content position calibration in head worn computing |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US9784973B2 (en) | 2014-02-11 | 2017-10-10 | Osterhout Group, Inc. | Micro doppler presentations in head worn computing |
US9811152B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9810906B2 (en) | 2014-06-17 | 2017-11-07 | Osterhout Group, Inc. | External user interface for head worn computing |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US20170363874A1 (en) * | 2016-06-21 | 2017-12-21 | Symbol Technologies, Llc | Modular heads up display assemblies |
CN107643597A (en) * | 2016-07-22 | 2018-01-30 | 蔡清来 | Integration reality border wisdom glasses eyeshield photochopper |
US9897822B2 (en) | 2014-04-25 | 2018-02-20 | Osterhout Group, Inc. | Temple and ear horn assembly for headworn computer |
US9939650B2 (en) | 2015-03-02 | 2018-04-10 | Lockheed Martin Corporation | Wearable display system |
US9939646B2 (en) | 2014-01-24 | 2018-04-10 | Osterhout Group, Inc. | Stray light suppression for head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US9952664B2 (en) | 2014-01-21 | 2018-04-24 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9995936B1 (en) | 2016-04-29 | 2018-06-12 | Lockheed Martin Corporation | Augmented reality systems having a virtual image overlaying an infrared portion of a live scene |
US10062182B2 (en) | 2015-02-17 | 2018-08-28 | Osterhout Group, Inc. | See-through computer display systems |
US10073269B2 (en) | 2013-09-11 | 2018-09-11 | Industrial Technology Research Institute | Virtual image display apparatus |
US10078917B1 (en) | 2015-06-26 | 2018-09-18 | Lucasfilm Entertainment Company Ltd. | Augmented reality simulation |
US20180284454A1 (en) * | 2017-04-03 | 2018-10-04 | Mira Labs, Inc. | Reflective Lens Headset |
US10146067B2 (en) * | 2016-09-09 | 2018-12-04 | Ching-Lai Tsai | Eye-protective shade for augmented reality smart glasses |
EP3413116A1 (en) * | 2017-06-07 | 2018-12-12 | Ching-Lai Tsai | Eye-protective shade for augmented reality smart glasses |
US10191279B2 (en) | 2014-03-17 | 2019-01-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US10194860B2 (en) | 2013-09-11 | 2019-02-05 | Industrial Technology Research Institute | Virtual image display system |
USD840395S1 (en) | 2016-10-17 | 2019-02-12 | Osterhout Group, Inc. | Head-worn computer |
US10228561B2 (en) | 2013-06-25 | 2019-03-12 | Microsoft Technology Licensing, Llc | Eye-tracking system using a freeform prism and gaze-detection light |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US20190187490A1 (en) * | 2017-12-20 | 2019-06-20 | Magic Leap, Inc. | Insert for augmented reality viewing device |
US10338400B2 (en) | 2017-07-03 | 2019-07-02 | Holovisions LLC | Augmented reality eyewear with VAPE or wear technology |
US10359545B2 (en) | 2010-10-21 | 2019-07-23 | Lockheed Martin Corporation | Fresnel lens with reduced draft facet visibility |
US10388199B2 (en) | 2014-10-23 | 2019-08-20 | Signify Holding B.V. | Illumination perception augmentation method, computer program products, head-mountable computing device and lighting system that adjusts a light output of a light source based on a desired light condition |
USD864959S1 (en) | 2017-01-04 | 2019-10-29 | Mentor Acquisition One, Llc | Computer glasses |
US10466491B2 (en) | 2016-06-01 | 2019-11-05 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US10466492B2 (en) | 2014-04-25 | 2019-11-05 | Mentor Acquisition One, Llc | Ear horn assembly for headworn computer |
US10474230B2 (en) * | 2016-12-15 | 2019-11-12 | Tectus Corporation | Brightness control for an augmented reality eye-mounted display |
US10482666B2 (en) | 2015-05-04 | 2019-11-19 | Beijing Zhigu Rui Tuo Tech Co., Ltd. | Display control methods and apparatuses |
US10484824B1 (en) * | 2015-06-26 | 2019-11-19 | Lucasfilm Entertainment Company Ltd. | Content presentation and layering across multiple devices |
US20190385372A1 (en) * | 2018-06-15 | 2019-12-19 | Microsoft Technology Licensing, Llc | Positioning a virtual reality passthrough region at a known distance |
US10520996B2 (en) | 2014-09-18 | 2019-12-31 | Mentor Acquisition One, Llc | Thermal management for head-worn computer |
US10558050B2 (en) | 2014-01-24 | 2020-02-11 | Mentor Acquisition One, Llc | Haptic systems for head-worn computers |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US10684478B2 (en) | 2016-05-09 | 2020-06-16 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10684476B2 (en) | 2014-10-17 | 2020-06-16 | Lockheed Martin Corporation | Head-wearable ultra-wide field of view display device |
US10690936B2 (en) | 2016-08-29 | 2020-06-23 | Mentor Acquisition One, Llc | Adjustable nose bridge assembly for headworn computer |
US10757495B2 (en) | 2016-08-22 | 2020-08-25 | Mentor Acquisition One, Llc | Speaker systems for head-worn computer systems |
US10768500B2 (en) | 2016-09-08 | 2020-09-08 | Mentor Acquisition One, Llc | Electrochromic systems for head-worn computer systems |
US10824253B2 (en) | 2016-05-09 | 2020-11-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US10859834B2 (en) | 2017-07-03 | 2020-12-08 | Holovisions | Space-efficient optical structures for wide field-of-view augmented reality (AR) eyewear |
US10878235B2 (en) | 2015-02-26 | 2020-12-29 | Magic Leap, Inc. | Apparatus for a near-eye display |
US10877274B1 (en) * | 2017-01-27 | 2020-12-29 | Facebook Technologies, Llc | Composite optical element for eye tracking having beam splitter formed by coupling of junction surfaces |
US10914949B2 (en) | 2018-11-16 | 2021-02-09 | Magic Leap, Inc. | Image size triggered clarification to maintain image sharpness |
US20210231951A1 (en) * | 2018-07-23 | 2021-07-29 | Magic Leap, Inc. | Systems and methods for external light management |
US11092812B2 (en) | 2018-06-08 | 2021-08-17 | Magic Leap, Inc. | Augmented reality viewer with automated surface selection placement and content orientation placement |
WO2021168449A1 (en) * | 2020-02-21 | 2021-08-26 | Raytrx, Llc | All-digital multi-option 3d surgery visualization system and control |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11104272B2 (en) | 2014-03-28 | 2021-08-31 | Mentor Acquisition One, Llc | System for assisted operator safety using an HMD |
US11112862B2 (en) | 2018-08-02 | 2021-09-07 | Magic Leap, Inc. | Viewing system with interpupillary distance compensation based on head motion |
US11138301B1 (en) * | 2017-11-20 | 2021-10-05 | Snap Inc. | Eye scanner for user identification and security in an eyewear device |
US11189252B2 (en) | 2018-03-15 | 2021-11-30 | Magic Leap, Inc. | Image correction due to deformation of components of a viewing device |
WO2021242338A1 (en) * | 2020-05-29 | 2021-12-02 | Microsoft Technology Licensing, Llc | Systems and methods for providing mixed-reality experiences under low light conditions |
US20210382315A1 (en) * | 2020-06-03 | 2021-12-09 | Universal City Studios Llc | Interface device with three-dimensional (3-d) viewing functionality |
US11199713B2 (en) | 2016-12-30 | 2021-12-14 | Magic Leap, Inc. | Polychromatic light out-coupling apparatus, near-eye displays comprising the same, and method of out-coupling polychromatic light |
US11200870B2 (en) | 2018-06-05 | 2021-12-14 | Magic Leap, Inc. | Homography transformation matrices based temperature calibration of a viewing system |
US11204491B2 (en) | 2018-05-30 | 2021-12-21 | Magic Leap, Inc. | Compact variable focus configurations |
US11210808B2 (en) | 2016-12-29 | 2021-12-28 | Magic Leap, Inc. | Systems and methods for augmented reality |
US11216086B2 (en) | 2018-08-03 | 2022-01-04 | Magic Leap, Inc. | Unfused pose-based drift correction of a fused pose of a totem in a user interaction system |
US11227294B2 (en) | 2014-04-03 | 2022-01-18 | Mentor Acquisition One, Llc | Sight information collection in head worn computing |
US11269182B2 (en) | 2014-07-15 | 2022-03-08 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11280937B2 (en) | 2017-12-10 | 2022-03-22 | Magic Leap, Inc. | Anti-reflective coatings on optical waveguides |
US11425189B2 (en) | 2019-02-06 | 2022-08-23 | Magic Leap, Inc. | Target intent-based clock speed determination and adjustment to limit total heat generated by multiple processors |
US11445232B2 (en) | 2019-05-01 | 2022-09-13 | Magic Leap, Inc. | Content provisioning system and method |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US11510027B2 (en) | 2018-07-03 | 2022-11-22 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality |
US11514673B2 (en) | 2019-07-26 | 2022-11-29 | Magic Leap, Inc. | Systems and methods for augmented reality |
US11567324B2 (en) | 2017-07-26 | 2023-01-31 | Magic Leap, Inc. | Exit pupil expander |
US11579441B2 (en) | 2018-07-02 | 2023-02-14 | Magic Leap, Inc. | Pixel intensity modulation using modifying gain values |
US11598651B2 (en) | 2018-07-24 | 2023-03-07 | Magic Leap, Inc. | Temperature dependent calibration of movement detection devices |
US11624929B2 (en) | 2018-07-24 | 2023-04-11 | Magic Leap, Inc. | Viewing device with dust seal integration |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11737666B2 (en) | 2014-01-21 | 2023-08-29 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US11737832B2 (en) | 2019-11-15 | 2023-08-29 | Magic Leap, Inc. | Viewing system for use in a surgical environment |
US11762623B2 (en) | 2019-03-12 | 2023-09-19 | Magic Leap, Inc. | Registration of local content between first and second augmented reality viewers |
US11856479B2 (en) | 2018-07-03 | 2023-12-26 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality along a route with markers |
US11885871B2 (en) | 2018-05-31 | 2024-01-30 | Magic Leap, Inc. | Radar head pose localization |
US11892644B2 (en) | 2014-01-21 | 2024-02-06 | Mentor Acquisition One, Llc | See-through computer display systems |
Families Citing this family (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9001030B2 (en) * | 2012-02-15 | 2015-04-07 | Google Inc. | Heads up display |
US9239415B2 (en) | 2012-03-08 | 2016-01-19 | Google Inc. | Near-to-eye display with an integrated out-looking camera |
US9858721B2 (en) | 2013-01-15 | 2018-01-02 | The University Of North Carolina At Chapel Hill | Methods, systems, and computer readable media for generating an augmented scene display |
KR102153599B1 (en) | 2013-11-18 | 2020-09-08 | 삼성전자주식회사 | Head mounted display apparatus and method for changing a light transmittance |
US10274731B2 (en) * | 2013-12-19 | 2019-04-30 | The University Of North Carolina At Chapel Hill | Optical see-through near-eye display using point light source backlight |
US9514573B2 (en) * | 2013-12-26 | 2016-12-06 | Dassault Systemes | Diminished reality |
US9760898B2 (en) * | 2014-01-06 | 2017-09-12 | The Nielsen Company (Us), Llc | Methods and apparatus to detect engagement with media presented on wearable media devices |
US20160178905A1 (en) * | 2014-12-19 | 2016-06-23 | Intel Corporation | Facilitating improved viewing capabitlies for glass displays |
EP3251098A4 (en) * | 2015-01-30 | 2018-07-11 | Ent. Services Development Corporation LP | Dynamic modulation for near eye display |
US9779554B2 (en) | 2015-04-10 | 2017-10-03 | Sony Interactive Entertainment Inc. | Filtering and parental control methods for restricting visual activity on a head mounted display |
US11609427B2 (en) | 2015-10-16 | 2023-03-21 | Ostendo Technologies, Inc. | Dual-mode augmented/virtual reality (AR/VR) near-eye wearable displays |
US11106273B2 (en) | 2015-10-30 | 2021-08-31 | Ostendo Technologies, Inc. | System and methods for on-body gestural interfaces and projection displays |
CN105450929A (en) * | 2015-12-09 | 2016-03-30 | 安徽海聚信息科技有限责任公司 | Information identification and communication method for smart wearable devices |
US10345594B2 (en) | 2015-12-18 | 2019-07-09 | Ostendo Technologies, Inc. | Systems and methods for augmented near-eye wearable displays |
US10578882B2 (en) | 2015-12-28 | 2020-03-03 | Ostendo Technologies, Inc. | Non-telecentric emissive micro-pixel array light modulators and methods of fabrication thereof |
CN108700743A (en) | 2016-01-22 | 2018-10-23 | 康宁股份有限公司 | Wide visual field individual's display |
US10353203B2 (en) | 2016-04-05 | 2019-07-16 | Ostendo Technologies, Inc. | Augmented/virtual reality near-eye displays with edge imaging lens comprising a plurality of display devices |
US10453431B2 (en) | 2016-04-28 | 2019-10-22 | Ostendo Technologies, Inc. | Integrated near-far light field display systems |
US10522106B2 (en) | 2016-05-05 | 2019-12-31 | Ostendo Technologies, Inc. | Methods and apparatus for active transparency modulation |
US9983412B1 (en) | 2017-02-02 | 2018-05-29 | The University Of North Carolina At Chapel Hill | Wide field of view augmented reality see through head mountable display with distance accommodation |
US11307417B2 (en) | 2017-07-14 | 2022-04-19 | Hewlett-Packard Development Company, L.P. | Panel to attenuate light for electronic eyewear device |
US10976551B2 (en) | 2017-08-30 | 2021-04-13 | Corning Incorporated | Wide field personal display device |
KR102607861B1 (en) | 2018-08-08 | 2023-11-29 | 삼성전자주식회사 | See-through type display device |
KR20200107027A (en) | 2019-03-05 | 2020-09-16 | 삼성디스플레이 주식회사 | Device for providing augmented reality |
KR20200110543A (en) | 2019-03-14 | 2020-09-24 | 삼성디스플레이 주식회사 | Augmented reality providing device and method for driving the same |
Family Cites Families (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH09113842A (en) | 1995-10-17 | 1997-05-02 | Olympus Optical Co Ltd | Head or face mount type video display device |
US6064749A (en) * | 1996-08-02 | 2000-05-16 | Hirota; Gentaro | Hybrid tracking for augmented reality using both camera motion detection and landmark tracking |
US7717173B2 (en) * | 1998-07-06 | 2010-05-18 | Ecycling, LLC | Methods of improving oil or gas production with recycled, increased sodium water |
JP2002538542A (en) * | 1999-03-02 | 2002-11-12 | シーメンス アクチエンゲゼルシヤフト | Enhanced Reality System for Supporting User-Industry Equipment Dialogue According to Context |
US7401920B1 (en) * | 2003-05-20 | 2008-07-22 | Elbit Systems Ltd. | Head mounted eye tracking and display system |
JP4364002B2 (en) | 2004-02-06 | 2009-11-11 | オリンパス株式会社 | Head-mounted camera and photographing method using head-mounted camera |
TWI263831B (en) | 2004-09-30 | 2006-10-11 | Himax Tech Inc | Head mounted display |
US7369317B2 (en) | 2005-03-07 | 2008-05-06 | Himax Technologies, Inc. | Head-mounted display utilizing an LCOS panel with a color filter attached thereon |
US7405881B2 (en) | 2005-05-30 | 2008-07-29 | Konica Minolta Holdings, Inc. | Image display apparatus and head mount display |
US8314815B2 (en) * | 2006-04-12 | 2012-11-20 | Nassir Navab | Virtual penetrating mirror device for visualizing of virtual objects within an augmented reality environment |
US8520309B2 (en) | 2008-09-04 | 2013-08-27 | Innovega Inc. | Method and apparatus to process display and non-display information |
US8294767B2 (en) * | 2009-01-30 | 2012-10-23 | Microsoft Corporation | Body scan |
US8267781B2 (en) * | 2009-01-30 | 2012-09-18 | Microsoft Corporation | Visual target tracking |
US8947455B2 (en) * | 2010-02-22 | 2015-02-03 | Nike, Inc. | Augmented reality design system |
JP2013521576A (en) * | 2010-02-28 | 2013-06-10 | オスターハウト グループ インコーポレイテッド | Local advertising content on interactive head-mounted eyepieces |
US8405680B1 (en) * | 2010-04-19 | 2013-03-26 | YDreams S.A., A Public Limited Liability Company | Various methods and apparatuses for achieving augmented reality |
EP2405402A1 (en) * | 2010-07-06 | 2012-01-11 | EADS Construcciones Aeronauticas, S.A. | Method and system for assembling components |
US8941559B2 (en) * | 2010-09-21 | 2015-01-27 | Microsoft Corporation | Opacity filter for display device |
US8860760B2 (en) * | 2010-09-25 | 2014-10-14 | Teledyne Scientific & Imaging, Llc | Augmented reality (AR) system and method for tracking parts and visually cueing a user to identify and locate parts in a scene |
US20120113141A1 (en) * | 2010-11-09 | 2012-05-10 | Cbs Interactive Inc. | Techniques to visualize products using augmented reality |
US9213405B2 (en) * | 2010-12-16 | 2015-12-15 | Microsoft Technology Licensing, Llc | Comprehension and intent-based content for augmented reality displays |
-
2011
- 2011-06-22 US US13/166,346 patent/US20120326948A1/en not_active Abandoned
-
2012
- 2012-12-03 US US13/692,861 patent/US8537075B2/en active Active
Cited By (243)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11506912B2 (en) | 2008-01-02 | 2022-11-22 | Mentor Acquisition One, Llc | Temple and ear horn assembly for headworn computer |
US9965681B2 (en) | 2008-12-16 | 2018-05-08 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US10359545B2 (en) | 2010-10-21 | 2019-07-23 | Lockheed Martin Corporation | Fresnel lens with reduced draft facet visibility |
US9632315B2 (en) | 2010-10-21 | 2017-04-25 | Lockheed Martin Corporation | Head-mounted display apparatus employing one or more fresnel lenses |
US10495790B2 (en) | 2010-10-21 | 2019-12-03 | Lockheed Martin Corporation | Head-mounted display apparatus employing one or more Fresnel lenses |
US8941603B2 (en) * | 2010-12-10 | 2015-01-27 | Sony Corporation | Touch sensitive display |
US20120188179A1 (en) * | 2010-12-10 | 2012-07-26 | Sony Ericsson Mobile Communications Ab | Touch sensitive display |
US9720228B2 (en) | 2010-12-16 | 2017-08-01 | Lockheed Martin Corporation | Collimating display with pixel lenses |
US10228561B2 (en) | 2013-06-25 | 2019-03-12 | Microsoft Technology Licensing, Llc | Eye-tracking system using a freeform prism and gaze-detection light |
US9625723B2 (en) | 2013-06-25 | 2017-04-18 | Microsoft Technology Licensing, Llc | Eye-tracking system using a freeform prism |
US10194860B2 (en) | 2013-09-11 | 2019-02-05 | Industrial Technology Research Institute | Virtual image display system |
US10073269B2 (en) | 2013-09-11 | 2018-09-11 | Industrial Technology Research Institute | Virtual image display apparatus |
US10013809B2 (en) | 2013-12-23 | 2018-07-03 | Empire Technology Development Llc | Suppression of real features in see-through display |
US9607409B2 (en) | 2013-12-23 | 2017-03-28 | Empire Technology Development Llc | Suppression of real features in see-through display |
US9696552B1 (en) * | 2014-01-10 | 2017-07-04 | Lockheed Martin Corporation | System and method for providing an augmented reality lightweight clip-on wearable device |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US11782529B2 (en) | 2014-01-17 | 2023-10-10 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11507208B2 (en) | 2014-01-17 | 2022-11-22 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US9939934B2 (en) | 2014-01-17 | 2018-04-10 | Osterhout Group, Inc. | External user interface for head worn computing |
US11231817B2 (en) | 2014-01-17 | 2022-01-25 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11169623B2 (en) | 2014-01-17 | 2021-11-09 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US10001644B2 (en) | 2014-01-21 | 2018-06-19 | Osterhout Group, Inc. | See-through computer display systems |
US9952664B2 (en) | 2014-01-21 | 2018-04-24 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9651788B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US10705339B2 (en) | 2014-01-21 | 2020-07-07 | Mentor Acquisition One, Llc | Suppression of stray light in head worn computing |
US9651789B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-Through computer display systems |
US9658458B2 (en) | 2014-01-21 | 2017-05-23 | Osterhout Group, Inc. | See-through computer display systems |
US9658457B2 (en) | 2014-01-21 | 2017-05-23 | Osterhout Group, Inc. | See-through computer display systems |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US11099380B2 (en) | 2014-01-21 | 2021-08-24 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9684171B2 (en) | 2014-01-21 | 2017-06-20 | Osterhout Group, Inc. | See-through computer display systems |
US10579140B2 (en) | 2014-01-21 | 2020-03-03 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9684165B2 (en) | 2014-01-21 | 2017-06-20 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9651784B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US11103132B2 (en) | 2014-01-21 | 2021-08-31 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9715112B2 (en) | 2014-01-21 | 2017-07-25 | Osterhout Group, Inc. | Suppression of stray light in head worn computing |
US9720227B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US9720235B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US9436006B2 (en) | 2014-01-21 | 2016-09-06 | Osterhout Group, Inc. | See-through computer display systems |
US9720234B2 (en) | 2014-01-21 | 2017-08-01 | Osterhout Group, Inc. | See-through computer display systems |
US9494800B2 (en) | 2014-01-21 | 2016-11-15 | Osterhout Group, Inc. | See-through computer display systems |
US11622426B2 (en) | 2014-01-21 | 2023-04-04 | Mentor Acquisition One, Llc | See-through computer display systems |
US11054902B2 (en) | 2014-01-21 | 2021-07-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9740012B2 (en) | 2014-01-21 | 2017-08-22 | Osterhout Group, Inc. | See-through computer display systems |
US9740280B2 (en) | 2014-01-21 | 2017-08-22 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9746676B2 (en) | 2014-01-21 | 2017-08-29 | Osterhout Group, Inc. | See-through computer display systems |
US11619820B2 (en) | 2014-01-21 | 2023-04-04 | Mentor Acquisition One, Llc | See-through computer display systems |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US9766463B2 (en) | 2014-01-21 | 2017-09-19 | Osterhout Group, Inc. | See-through computer display systems |
US9772492B2 (en) | 2014-01-21 | 2017-09-26 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11353957B2 (en) | 2014-01-21 | 2022-06-07 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US9811152B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9651783B2 (en) | 2014-01-21 | 2017-05-16 | Osterhout Group, Inc. | See-through computer display systems |
US9811159B2 (en) | 2014-01-21 | 2017-11-07 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11126003B2 (en) | 2014-01-21 | 2021-09-21 | Mentor Acquisition One, Llc | See-through computer display systems |
US9829703B2 (en) | 2014-01-21 | 2017-11-28 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9836122B2 (en) | 2014-01-21 | 2017-12-05 | Osterhout Group, Inc. | Eye glint imaging in see-through computer display systems |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US10379365B2 (en) | 2014-01-21 | 2019-08-13 | Mentor Acquisition One, Llc | See-through computer display systems |
US9523856B2 (en) | 2014-01-21 | 2016-12-20 | Osterhout Group, Inc. | See-through computer display systems |
US11719934B2 (en) | 2014-01-21 | 2023-08-08 | Mentor Acquisition One, Llc | Suppression of stray light in head worn computing |
US9529199B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | See-through computer display systems |
US9885868B2 (en) | 2014-01-21 | 2018-02-06 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9529192B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US11737666B2 (en) | 2014-01-21 | 2023-08-29 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9927612B2 (en) | 2014-01-21 | 2018-03-27 | Osterhout Group, Inc. | See-through computer display systems |
US9933622B2 (en) | 2014-01-21 | 2018-04-03 | Osterhout Group, Inc. | See-through computer display systems |
US9529195B2 (en) | 2014-01-21 | 2016-12-27 | Osterhout Group, Inc. | See-through computer display systems |
US11892644B2 (en) | 2014-01-21 | 2024-02-06 | Mentor Acquisition One, Llc | See-through computer display systems |
US9615742B2 (en) | 2014-01-21 | 2017-04-11 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US10698223B2 (en) | 2014-01-21 | 2020-06-30 | Mentor Acquisition One, Llc | See-through computer display systems |
US9958674B2 (en) | 2014-01-21 | 2018-05-01 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9594246B2 (en) | 2014-01-21 | 2017-03-14 | Osterhout Group, Inc. | See-through computer display systems |
US10866420B2 (en) | 2014-01-21 | 2020-12-15 | Mentor Acquisition One, Llc | See-through computer display systems |
US11796805B2 (en) | 2014-01-21 | 2023-10-24 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US10558050B2 (en) | 2014-01-24 | 2020-02-11 | Mentor Acquisition One, Llc | Haptic systems for head-worn computers |
US11822090B2 (en) | 2014-01-24 | 2023-11-21 | Mentor Acquisition One, Llc | Haptic systems for head-worn computers |
US9939646B2 (en) | 2014-01-24 | 2018-04-10 | Osterhout Group, Inc. | Stray light suppression for head worn computing |
US9843093B2 (en) | 2014-02-11 | 2017-12-12 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9401540B2 (en) | 2014-02-11 | 2016-07-26 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US9784973B2 (en) | 2014-02-11 | 2017-10-10 | Osterhout Group, Inc. | Micro doppler presentations in head worn computing |
US9841602B2 (en) | 2014-02-11 | 2017-12-12 | Osterhout Group, Inc. | Location indicating avatar in head worn computing |
US9928019B2 (en) | 2014-02-14 | 2018-03-27 | Osterhout Group, Inc. | Object shadowing in head worn computing |
US9547465B2 (en) | 2014-02-14 | 2017-01-17 | Osterhout Group, Inc. | Object shadowing in head worn computing |
US10191279B2 (en) | 2014-03-17 | 2019-01-29 | Osterhout Group, Inc. | Eye imaging in head worn computing |
US9423612B2 (en) | 2014-03-28 | 2016-08-23 | Osterhout Group, Inc. | Sensor dependent content position in head worn computing |
US11104272B2 (en) | 2014-03-28 | 2021-08-31 | Mentor Acquisition One, Llc | System for assisted operator safety using an HMD |
US9733478B2 (en) | 2014-04-03 | 2017-08-15 | Thomson Licensing | Ambient light management for virtual reality glasses |
US11227294B2 (en) | 2014-04-03 | 2022-01-18 | Mentor Acquisition One, Llc | Sight information collection in head worn computing |
US11880041B2 (en) | 2014-04-25 | 2024-01-23 | Mentor Acquisition One, Llc | Speaker assembly for headworn computer |
US10146772B2 (en) | 2014-04-25 | 2018-12-04 | Osterhout Group, Inc. | Language translation with head-worn computing |
US11474360B2 (en) | 2014-04-25 | 2022-10-18 | Mentor Acquisition One, Llc | Speaker assembly for headworn computer |
US9897822B2 (en) | 2014-04-25 | 2018-02-20 | Osterhout Group, Inc. | Temple and ear horn assembly for headworn computer |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US11727223B2 (en) | 2014-04-25 | 2023-08-15 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US11809022B2 (en) | 2014-04-25 | 2023-11-07 | Mentor Acquisition One, Llc | Temple and ear horn assembly for headworn computer |
US10634922B2 (en) | 2014-04-25 | 2020-04-28 | Mentor Acquisition One, Llc | Speaker assembly for headworn computer |
US10101588B2 (en) | 2014-04-25 | 2018-10-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US9672210B2 (en) | 2014-04-25 | 2017-06-06 | Osterhout Group, Inc. | Language translation with head-worn computing |
US10732434B2 (en) | 2014-04-25 | 2020-08-04 | Mentor Acquisition One, Llc | Temple and ear horn assembly for headworn computer |
US10466492B2 (en) | 2014-04-25 | 2019-11-05 | Mentor Acquisition One, Llc | Ear horn assembly for headworn computer |
US9746686B2 (en) | 2014-05-19 | 2017-08-29 | Osterhout Group, Inc. | Content position calibration in head worn computing |
US10877270B2 (en) | 2014-06-05 | 2020-12-29 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US11402639B2 (en) | 2014-06-05 | 2022-08-02 | Mentor Acquisition One, Llc | Optical configurations for head-worn see-through displays |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US11790617B2 (en) | 2014-06-09 | 2023-10-17 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11360318B2 (en) | 2014-06-09 | 2022-06-14 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10976559B2 (en) | 2014-06-09 | 2021-04-13 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9720241B2 (en) | 2014-06-09 | 2017-08-01 | Osterhout Group, Inc. | Content presentation in head worn computing |
US11887265B2 (en) | 2014-06-09 | 2024-01-30 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11022810B2 (en) | 2014-06-09 | 2021-06-01 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11663794B2 (en) | 2014-06-09 | 2023-05-30 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US10139635B2 (en) | 2014-06-09 | 2018-11-27 | Osterhout Group, Inc. | Content presentation in head worn computing |
US9575321B2 (en) | 2014-06-09 | 2017-02-21 | Osterhout Group, Inc. | Content presentation in head worn computing |
US11327323B2 (en) | 2014-06-09 | 2022-05-10 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US9810906B2 (en) | 2014-06-17 | 2017-11-07 | Osterhout Group, Inc. | External user interface for head worn computing |
US11294180B2 (en) | 2014-06-17 | 2022-04-05 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US10698212B2 (en) | 2014-06-17 | 2020-06-30 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11054645B2 (en) | 2014-06-17 | 2021-07-06 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US11789267B2 (en) | 2014-06-17 | 2023-10-17 | Mentor Acquisition One, Llc | External user interface for head worn computing |
US20160004085A1 (en) * | 2014-07-02 | 2016-01-07 | Christian Stroetmann | Head-mounted display device with air conditioning device and control approaches |
US11103122B2 (en) | 2014-07-15 | 2021-08-31 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11786105B2 (en) | 2014-07-15 | 2023-10-17 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11269182B2 (en) | 2014-07-15 | 2022-03-08 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11360314B2 (en) | 2014-08-12 | 2022-06-14 | Mentor Acquisition One, Llc | Measuring content brightness in head worn computing |
US11630315B2 (en) | 2014-08-12 | 2023-04-18 | Mentor Acquisition One, Llc | Measuring content brightness in head worn computing |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US10908422B2 (en) | 2014-08-12 | 2021-02-02 | Mentor Acquisition One, Llc | Measuring content brightness in head worn computing |
US10520996B2 (en) | 2014-09-18 | 2019-12-31 | Mentor Acquisition One, Llc | Thermal management for head-worn computer |
US11474575B2 (en) | 2014-09-18 | 2022-10-18 | Mentor Acquisition One, Llc | Thermal management for head-worn computer |
US10963025B2 (en) | 2014-09-18 | 2021-03-30 | Mentor Acquisition One, Llc | Thermal management for head-worn computer |
US9671613B2 (en) | 2014-09-26 | 2017-06-06 | Osterhout Group, Inc. | See-through computer display systems |
US10684476B2 (en) | 2014-10-17 | 2020-06-16 | Lockheed Martin Corporation | Head-wearable ultra-wide field of view display device |
US10388199B2 (en) | 2014-10-23 | 2019-08-20 | Signify Holding B.V. | Illumination perception augmentation method, computer program products, head-mountable computing device and lighting system that adjusts a light output of a light source based on a desired light condition |
US10036889B2 (en) | 2014-12-03 | 2018-07-31 | Osterhout Group, Inc. | Head worn computer display systems |
US10018837B2 (en) | 2014-12-03 | 2018-07-10 | Osterhout Group, Inc. | Head worn computer display systems |
US11262846B2 (en) | 2014-12-03 | 2022-03-01 | Mentor Acquisition One, Llc | See-through computer display systems |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US11809628B2 (en) | 2014-12-03 | 2023-11-07 | Mentor Acquisition One, Llc | See-through computer display systems |
US9684172B2 (en) | 2014-12-03 | 2017-06-20 | Osterhout Group, Inc. | Head worn computer display systems |
US10197801B2 (en) | 2014-12-03 | 2019-02-05 | Osterhout Group, Inc. | Head worn computer display systems |
USD792400S1 (en) | 2014-12-31 | 2017-07-18 | Osterhout Group, Inc. | Computer glasses |
USD794637S1 (en) | 2015-01-05 | 2017-08-15 | Osterhout Group, Inc. | Air mouse |
US10062182B2 (en) | 2015-02-17 | 2018-08-28 | Osterhout Group, Inc. | See-through computer display systems |
US11468639B2 (en) * | 2015-02-20 | 2022-10-11 | Microsoft Technology Licensing, Llc | Selective occlusion system for augmented reality devices |
US20160247319A1 (en) * | 2015-02-20 | 2016-08-25 | Andreas G. Nowatzyk | Selective occlusion system for augmented reality devices |
US11347960B2 (en) | 2015-02-26 | 2022-05-31 | Magic Leap, Inc. | Apparatus for a near-eye display |
US10878235B2 (en) | 2015-02-26 | 2020-12-29 | Magic Leap, Inc. | Apparatus for a near-eye display |
US11756335B2 (en) | 2015-02-26 | 2023-09-12 | Magic Leap, Inc. | Apparatus for a near-eye display |
US9939650B2 (en) | 2015-03-02 | 2018-04-10 | Lockheed Martin Corporation | Wearable display system |
US10482666B2 (en) | 2015-05-04 | 2019-11-19 | Beijing Zhigu Rui Tuo Tech Co., Ltd. | Display control methods and apparatuses |
US11128984B1 (en) | 2015-06-26 | 2021-09-21 | Lucasfilm Eniertainment Company Ltd. | Content presentation and layering across multiple devices |
US10078917B1 (en) | 2015-06-26 | 2018-09-18 | Lucasfilm Entertainment Company Ltd. | Augmented reality simulation |
US10484824B1 (en) * | 2015-06-26 | 2019-11-19 | Lucasfilm Entertainment Company Ltd. | Content presentation and layering across multiple devices |
US20170011557A1 (en) * | 2015-07-06 | 2017-01-12 | Samsung Electronics Co., Ltd | Method for providing augmented reality and virtual reality and electronic device using the same |
US10754156B2 (en) | 2015-10-20 | 2020-08-25 | Lockheed Martin Corporation | Multiple-eye, single-display, ultrawide-field-of-view optical see-through augmented reality system |
WO2017070226A1 (en) * | 2015-10-20 | 2017-04-27 | Lockheed Martin Corporation | Multiple-eye, single-display, ultrawide-field-of-view optical see-through augmented reality system |
US9995936B1 (en) | 2016-04-29 | 2018-06-12 | Lockheed Martin Corporation | Augmented reality systems having a virtual image overlaying an infrared portion of a live scene |
US11320656B2 (en) | 2016-05-09 | 2022-05-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10824253B2 (en) | 2016-05-09 | 2020-11-03 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US11226691B2 (en) | 2016-05-09 | 2022-01-18 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US11500212B2 (en) | 2016-05-09 | 2022-11-15 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10684478B2 (en) | 2016-05-09 | 2020-06-16 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US11022808B2 (en) | 2016-06-01 | 2021-06-01 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11460708B2 (en) | 2016-06-01 | 2022-10-04 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11586048B2 (en) | 2016-06-01 | 2023-02-21 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US10466491B2 (en) | 2016-06-01 | 2019-11-05 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US11754845B2 (en) | 2016-06-01 | 2023-09-12 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
USD945521S1 (en) | 2016-06-21 | 2022-03-08 | Symbol Technologies, Llc | Heads-up display mount |
US20170363874A1 (en) * | 2016-06-21 | 2017-12-21 | Symbol Technologies, Llc | Modular heads up display assemblies |
CN107643597A (en) * | 2016-07-22 | 2018-01-30 | 蔡清来 | Integration reality border wisdom glasses eyeshield photochopper |
CN111308714A (en) * | 2016-07-22 | 2020-06-19 | 蔡清来 | Eye-protecting and shading device for intelligent glasses integrating real environments |
US10757495B2 (en) | 2016-08-22 | 2020-08-25 | Mentor Acquisition One, Llc | Speaker systems for head-worn computer systems |
US11825257B2 (en) | 2016-08-22 | 2023-11-21 | Mentor Acquisition One, Llc | Speaker systems for head-worn computer systems |
US11350196B2 (en) | 2016-08-22 | 2022-05-31 | Mentor Acquisition One, Llc | Speaker systems for head-worn computer systems |
US10690936B2 (en) | 2016-08-29 | 2020-06-23 | Mentor Acquisition One, Llc | Adjustable nose bridge assembly for headworn computer |
US11409128B2 (en) | 2016-08-29 | 2022-08-09 | Mentor Acquisition One, Llc | Adjustable nose bridge assembly for headworn computer |
US11415856B2 (en) | 2016-09-08 | 2022-08-16 | Mentor Acquisition One, Llc | Electrochromic systems for head-worn computer systems |
US11768417B2 (en) | 2016-09-08 | 2023-09-26 | Mentor Acquisition One, Llc | Electrochromic systems for head-worn computer systems |
US10768500B2 (en) | 2016-09-08 | 2020-09-08 | Mentor Acquisition One, Llc | Electrochromic systems for head-worn computer systems |
US10146067B2 (en) * | 2016-09-09 | 2018-12-04 | Ching-Lai Tsai | Eye-protective shade for augmented reality smart glasses |
US10788686B2 (en) * | 2016-09-09 | 2020-09-29 | Ching-Lai Tsai | Eye-protective shade for augmented reality smart glasses |
US20190056604A1 (en) * | 2016-09-09 | 2019-02-21 | Ching-Lai Tsai | Eye-protective shade for augmented reality smart glasses |
USD840395S1 (en) | 2016-10-17 | 2019-02-12 | Osterhout Group, Inc. | Head-worn computer |
US10474230B2 (en) * | 2016-12-15 | 2019-11-12 | Tectus Corporation | Brightness control for an augmented reality eye-mounted display |
US10831269B2 (en) | 2016-12-15 | 2020-11-10 | Tectus Corporation | Brightness control for an augmented reality eye-mounted display |
US11790554B2 (en) | 2016-12-29 | 2023-10-17 | Magic Leap, Inc. | Systems and methods for augmented reality |
US11210808B2 (en) | 2016-12-29 | 2021-12-28 | Magic Leap, Inc. | Systems and methods for augmented reality |
US11874468B2 (en) | 2016-12-30 | 2024-01-16 | Magic Leap, Inc. | Polychromatic light out-coupling apparatus, near-eye displays comprising the same, and method of out-coupling polychromatic light |
US11199713B2 (en) | 2016-12-30 | 2021-12-14 | Magic Leap, Inc. | Polychromatic light out-coupling apparatus, near-eye displays comprising the same, and method of out-coupling polychromatic light |
USD947186S1 (en) | 2017-01-04 | 2022-03-29 | Mentor Acquisition One, Llc | Computer glasses |
USD864959S1 (en) | 2017-01-04 | 2019-10-29 | Mentor Acquisition One, Llc | Computer glasses |
USD918905S1 (en) | 2017-01-04 | 2021-05-11 | Mentor Acquisition One, Llc | Computer glasses |
US10877274B1 (en) * | 2017-01-27 | 2020-12-29 | Facebook Technologies, Llc | Composite optical element for eye tracking having beam splitter formed by coupling of junction surfaces |
JP2020515914A (en) * | 2017-04-03 | 2020-05-28 | ミラ ラボ インコーポレイテッド | Headset system and optical element |
US10816803B2 (en) * | 2017-04-03 | 2020-10-27 | Mira Labs, Inc. | Reflective lens headset |
US20180284454A1 (en) * | 2017-04-03 | 2018-10-04 | Mira Labs, Inc. | Reflective Lens Headset |
EP3413116A1 (en) * | 2017-06-07 | 2018-12-12 | Ching-Lai Tsai | Eye-protective shade for augmented reality smart glasses |
US10859834B2 (en) | 2017-07-03 | 2020-12-08 | Holovisions | Space-efficient optical structures for wide field-of-view augmented reality (AR) eyewear |
US10338400B2 (en) | 2017-07-03 | 2019-07-02 | Holovisions LLC | Augmented reality eyewear with VAPE or wear technology |
US11567324B2 (en) | 2017-07-26 | 2023-01-31 | Magic Leap, Inc. | Exit pupil expander |
US11927759B2 (en) | 2017-07-26 | 2024-03-12 | Magic Leap, Inc. | Exit pupil expander |
US11138301B1 (en) * | 2017-11-20 | 2021-10-05 | Snap Inc. | Eye scanner for user identification and security in an eyewear device |
US11280937B2 (en) | 2017-12-10 | 2022-03-22 | Magic Leap, Inc. | Anti-reflective coatings on optical waveguides |
US11762222B2 (en) | 2017-12-20 | 2023-09-19 | Magic Leap, Inc. | Insert for augmented reality viewing device |
US20190187490A1 (en) * | 2017-12-20 | 2019-06-20 | Magic Leap, Inc. | Insert for augmented reality viewing device |
US11187923B2 (en) * | 2017-12-20 | 2021-11-30 | Magic Leap, Inc. | Insert for augmented reality viewing device |
EP3729172A4 (en) * | 2017-12-20 | 2021-02-24 | Magic Leap, Inc. | Insert for augmented reality viewing device |
US11776509B2 (en) | 2018-03-15 | 2023-10-03 | Magic Leap, Inc. | Image correction due to deformation of components of a viewing device |
US11908434B2 (en) | 2018-03-15 | 2024-02-20 | Magic Leap, Inc. | Image correction due to deformation of components of a viewing device |
US11189252B2 (en) | 2018-03-15 | 2021-11-30 | Magic Leap, Inc. | Image correction due to deformation of components of a viewing device |
US11204491B2 (en) | 2018-05-30 | 2021-12-21 | Magic Leap, Inc. | Compact variable focus configurations |
US11885871B2 (en) | 2018-05-31 | 2024-01-30 | Magic Leap, Inc. | Radar head pose localization |
US11200870B2 (en) | 2018-06-05 | 2021-12-14 | Magic Leap, Inc. | Homography transformation matrices based temperature calibration of a viewing system |
US11092812B2 (en) | 2018-06-08 | 2021-08-17 | Magic Leap, Inc. | Augmented reality viewer with automated surface selection placement and content orientation placement |
US20190385372A1 (en) * | 2018-06-15 | 2019-12-19 | Microsoft Technology Licensing, Llc | Positioning a virtual reality passthrough region at a known distance |
US11579441B2 (en) | 2018-07-02 | 2023-02-14 | Magic Leap, Inc. | Pixel intensity modulation using modifying gain values |
US11856479B2 (en) | 2018-07-03 | 2023-12-26 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality along a route with markers |
US11510027B2 (en) | 2018-07-03 | 2022-11-22 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality |
US20210231951A1 (en) * | 2018-07-23 | 2021-07-29 | Magic Leap, Inc. | Systems and methods for external light management |
US11598651B2 (en) | 2018-07-24 | 2023-03-07 | Magic Leap, Inc. | Temperature dependent calibration of movement detection devices |
US11624929B2 (en) | 2018-07-24 | 2023-04-11 | Magic Leap, Inc. | Viewing device with dust seal integration |
US11112862B2 (en) | 2018-08-02 | 2021-09-07 | Magic Leap, Inc. | Viewing system with interpupillary distance compensation based on head motion |
US11630507B2 (en) | 2018-08-02 | 2023-04-18 | Magic Leap, Inc. | Viewing system with interpupillary distance compensation based on head motion |
US11609645B2 (en) | 2018-08-03 | 2023-03-21 | Magic Leap, Inc. | Unfused pose-based drift correction of a fused pose of a totem in a user interaction system |
US11216086B2 (en) | 2018-08-03 | 2022-01-04 | Magic Leap, Inc. | Unfused pose-based drift correction of a fused pose of a totem in a user interaction system |
US11521296B2 (en) | 2018-11-16 | 2022-12-06 | Magic Leap, Inc. | Image size triggered clarification to maintain image sharpness |
US10914949B2 (en) | 2018-11-16 | 2021-02-09 | Magic Leap, Inc. | Image size triggered clarification to maintain image sharpness |
US11425189B2 (en) | 2019-02-06 | 2022-08-23 | Magic Leap, Inc. | Target intent-based clock speed determination and adjustment to limit total heat generated by multiple processors |
US11762623B2 (en) | 2019-03-12 | 2023-09-19 | Magic Leap, Inc. | Registration of local content between first and second augmented reality viewers |
US11445232B2 (en) | 2019-05-01 | 2022-09-13 | Magic Leap, Inc. | Content provisioning system and method |
US11514673B2 (en) | 2019-07-26 | 2022-11-29 | Magic Leap, Inc. | Systems and methods for augmented reality |
US11737832B2 (en) | 2019-11-15 | 2023-08-29 | Magic Leap, Inc. | Viewing system for use in a surgical environment |
US11628038B2 (en) * | 2020-02-21 | 2023-04-18 | Raytrx, Llc | Multi-option all-digital 3D surgery visualization system and control |
US20210315662A1 (en) * | 2020-02-21 | 2021-10-14 | Raytrx, Llc | Multi-option all-digital 3d surgery visualization system and control |
WO2021168449A1 (en) * | 2020-02-21 | 2021-08-26 | Raytrx, Llc | All-digital multi-option 3d surgery visualization system and control |
US11579449B2 (en) | 2020-05-29 | 2023-02-14 | Microsoft Technology Licensing, Llc | Systems and methods for providing mixed-reality experiences under low light conditions |
WO2021242338A1 (en) * | 2020-05-29 | 2021-12-02 | Microsoft Technology Licensing, Llc | Systems and methods for providing mixed-reality experiences under low light conditions |
US11774770B2 (en) * | 2020-06-03 | 2023-10-03 | Universal City Studios Llc | Interface device with three-dimensional (3-D) viewing functionality |
US20210382315A1 (en) * | 2020-06-03 | 2021-12-09 | Universal City Studios Llc | Interface device with three-dimensional (3-d) viewing functionality |
Also Published As
Publication number | Publication date |
---|---|
US20130162505A1 (en) | 2013-06-27 |
US8537075B2 (en) | 2013-09-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8537075B2 (en) | Environmental-light filter for see-through head-mounted display device | |
US10573086B2 (en) | Opacity filter for display device | |
CN110073272B (en) | Free-form surface prism group and near-to-eye display device using same | |
US9696552B1 (en) | System and method for providing an augmented reality lightweight clip-on wearable device | |
CN107111136B (en) | Binocular device including monocular display device | |
JP2016536635A (en) | System and method for reconfigurable projection augmented reality / virtual reality appliance | |
CN204595328U (en) | Head-mounted display apparatus | |
WO2016101861A1 (en) | Head-worn display device | |
US10706600B1 (en) | Head-mounted display devices with transparent display panels for color deficient user | |
US20180031836A1 (en) | Smart glasses having interfering light filtering | |
JP6304589B2 (en) | Retina projection system and eyeglass-type display | |
JP2007334185A (en) | Eyeglass type video display device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CROCCO, ROBERT;SUGDEN, BEN;STONE PEREZ, KATHRYN;SIGNING DATES FROM 20110613 TO 20110621;REEL/FRAME:026484/0799 |
|
STCB | Information on status: application discontinuation |
Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034766/0509 Effective date: 20141014 |