US20120327116A1 - Total field of view classification for head-mounted display - Google Patents

Total field of view classification for head-mounted display Download PDF

Info

Publication number
US20120327116A1
US20120327116A1 US13/167,113 US201113167113A US2012327116A1 US 20120327116 A1 US20120327116 A1 US 20120327116A1 US 201113167113 A US201113167113 A US 201113167113A US 2012327116 A1 US2012327116 A1 US 2012327116A1
Authority
US
United States
Prior art keywords
hmd
regions
hmd wearer
wearer
tfov
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/167,113
Inventor
James Chia-Ming Liu
Stephen Gilchrist Latta
Anton Andrews
Benjamin Issac Vaught
Christopher Michael Novak
Sheridan Leigh Small
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Microsoft Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Corp filed Critical Microsoft Corp
Priority to US13/167,113 priority Critical patent/US20120327116A1/en
Assigned to MICROSOFT CORPORATION reassignment MICROSOFT CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LATTA, STEPHEN GILCHRIST, ANDREWS, ANTON, NOVAK, CHRISTOPHER MICHAEL, SMALL, SHERIDAN LEIGH, VAUGHT, BENJAMIN ISAAC, LIU, JAMES CHIA-MING
Priority to TW101115609A priority patent/TW201303640A/en
Priority to PCT/US2012/043178 priority patent/WO2012177657A2/en
Priority to JP2014517099A priority patent/JP2014526157A/en
Priority to KR1020137034185A priority patent/KR20140034252A/en
Priority to EP12803004.6A priority patent/EP2724191A4/en
Priority to CN201280030619.6A priority patent/CN103635849A/en
Priority to US13/692,885 priority patent/US9041623B2/en
Publication of US20120327116A1 publication Critical patent/US20120327116A1/en
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Priority to US14/706,377 priority patent/US9245501B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/02Viewing or reading apparatus
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • G09G5/38Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory with means for controlling the display position
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/0093Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for monitoring data relating to the user, e.g. head-tracking, eye-tracking
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B27/0172Head mounted characterised by optical features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/36Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of a graphic pattern, e.g. using an all-points-addressable [APA] memory
    • G09G5/37Details of the operation on graphic patterns
    • G09G5/377Details of the operation on graphic patterns for mixing or overlaying two or more graphic patterns
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0179Display position adjusting means not related to the information to be displayed
    • G02B2027/0187Display position adjusting means not related to the information to be displayed slaved to motion of at least a part of the body of the user, e.g. head, eye

Definitions

  • HMDs See-through head-mounted displays
  • the HMD augments the HMD wearer's view of the real world with virtual imagery to provide an augment reality view.
  • those virtual images can easily become distracting and obtrusive, drawing attention away from the real-world and occluding the HMD wearer's normal (non-augmented) view. This may present a challenge particularly in applications of “always-on” wearable display devices, especially when considering mobile, outdoor, and in-activity scenarios.
  • Embodiments of the present invention relate to locating virtual images in a head-mounted display (HMD) to provide an augment reality view to an HMD wearer.
  • Sensor data may be collected from on-board sensors provided on an HMD. Additionally, other data may be collected from external sources. Based on the collected sensor data and other data, the position and rotation of the HMD wearer's head relative to the HMD wearer's body and surrounding environment may be determined. After resolving the HMD wearer's head position, the HMD wearer's total field of view (TFOV) may be classified into regions. Virtual images may then be located in the classified TFOV regions to locate the virtual images relative to the HMD wearer's body and surrounding environment.
  • TFOV total field of view
  • FIG. 1 is a block diagram of an exemplary computing environment suitable for use in implementing embodiments of the present invention
  • FIG. 2 is a block diagram of an exemplary system for classifying regions of a TFOV and locating virtual objects in the regions in accordance with an embodiment of the present invention
  • FIGS. 3A and 3B provide a profile and front views of an HMD wearer's TFOV classified with primary, secondary, and tertiary regions in accordance with an embodiment of the present invention
  • FIG. 4 illustrates fixed regions for a HMD wearer's TFOV in accordance with an embodiment of the present invention
  • FIG. 5 illustrates dynamic regions for an HMD wearer's TFOV in accordance with an embodiment of the present invention
  • FIG. 6 illustrates an HMD wearer's view with virtual objects located in primary, secondary, and tertiary regions in accordance with an embodiment of the present invention
  • FIG. 7 is a flow diagram showing a method for classifying regions of an HMD wearer's TFOV in accordance with an embodiment of the present invention.
  • FIG. 8 is a flow diagram showing a method for displaying virtual objects using an HMD in accordance with an embodiment of the present invention.
  • Embodiments of the present invention are directed to the automatic and continuous placement and replacement of virtual images in an HMD's display relative to the HMD wearer's total field of view (TFOV) to provide the HMD wearer with an augmented reality view.
  • TFOV total field of view
  • TFOV accounts for complete range of rotation and translation of a person's head. This is in contrast to a person's field of view (FOV), which accounts for the extent of what the person sees at a given moment.
  • FOV field of view
  • sensor data is collected from any number of on-board sensors provided on an HMD. Additionally, other data may be collected from sources external to the HMD. The sensor data and other data may be used to track and interpret the HMD wearer's physical head as it relates the HMD wearer's physical body and to the physical environment surrounding the HMD wearer.
  • the HMD wearer's TFOV may be classified into various regions. These regions may be defined relative to the HMD wearer's body and surrounding environment. In some embodiments, the regions may include a primary region and one or more non-primary regions. Virtual images may be located within the TFOV according to the classified regions.
  • embodiments of the present invention provide a system that is aware of an HMD wearer's head relative to the HMD wearer's body and surrounding environment and may use such knowledge to locate virtual images relative to the HMD wearer's body and surrounding environment.
  • the virtual objects may be generally located away from the primary region and placed in unobtrusive regions such that the virtual images are presented in non-dangerous ways by reducing visual occlusion of the primary region.
  • the virtual objects may also be located away from the primary region in order to provide context (e.g., classify the information as less relevant, less important, etc.).
  • an embodiment of the present invention is directed to one or more computer storage media storing computer-useable instructions that, when used by one or more computing devices, cause the one or more computing device to perform a method.
  • the method includes receiving sensor data from one or more HMD on-board sensors.
  • the method also includes using the sensor data to determine an HMD wearer's head position and rotation relative to the HMD wearer's body and an environment surrounding the HMD wearer.
  • the method further includes classifying two or more regions within the HMD wearer's TFOV based on one or more pre-determined rules and the HMD wearer's head position and rotation relative to the HMD wearer's body and an environment surrounding the HMD wearer.
  • the method still further includes locating virtual images to be displayed by the HMD based on classifying the two or more regions within the HMD wearer's TFOV.
  • an aspect of the invention is directed to an HMD.
  • the HMD includes one or more on-board sensors.
  • the HMD also includes one or more processors configured to: determine a position and rotation of an HMD wearer's head relative to the HMD wearer's body based on sensor data from the one or more on-board sensors, classify two or more regions of the HMD wearer's TFOV based on the position and rotation of the HMD wearer's head relative to the HMD wearer's body, and locate one or more virtual objects in the two or more regions.
  • the HMD further includes one or more display components configured to display at least one of the one or more virtual objects to provide the HMD wearer an augmented view.
  • a further embodiment is directed to a method for classifying regions of an HMD wearer's TFOV.
  • the method include continuously: receiving sensor data from one or more sensors on board the HMD; receiving other data from one or more sources external to the HMD; and classifying regions of the HMD wearer's TFOV relative to the HMD wearer's body and surrounding environment based on the sensor data and other data.
  • FIG. 1 an exemplary operating environment for implementing embodiments of the present invention is shown and designated generally as computing device 100 .
  • Computing device 100 is but one example of a suitable computing environment and is not intended to suggest any limitation as to the scope of use or functionality of the invention. Neither should the computing device 100 be interpreted as having any dependency or requirement relating to any one or combination of components illustrated.
  • the invention may be described in the general context of computer code or machine-useable instructions, including computer-executable instructions such as program modules, being executed by a computer or other machine, such as a personal data assistant or other handheld device.
  • program modules including routines, programs, objects, components, data structures, etc., refer to code that perform particular tasks or implement particular abstract data types.
  • the invention may be practiced in a variety of system configurations, including hand-held devices, consumer electronics, general-purpose computers, more specialty computing devices, etc.
  • the invention may also be practiced in distributed computing environments where tasks are performed by remote-processing devices that are linked through a communications network.
  • computing device 100 includes a bus 110 that directly or indirectly couples the following devices: memory 112 , one or more processors 114 , one or more presentation components 116 , input/output (I/O) ports 118 , input/output components 120 , and an illustrative power supply 122 .
  • Bus 110 represents what may be one or more busses (such as an address bus, data bus, or combination thereof).
  • FIG. 1 is merely illustrative of an exemplary computing device that can be used in connection with one or more embodiments of the present invention. Distinction is not made between such categories as “workstation,” “server,” “laptop,” “hand-held device,” etc., as all are contemplated within the scope of FIG. 1 and reference to “computing device.”
  • Computer-readable media can be any available media that can be accessed by computing device 100 and includes both volatile and nonvolatile media, removable and non-removable media.
  • Computer-readable media may comprise computer storage media and communication media.
  • Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data.
  • Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 100 .
  • Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
  • modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
  • communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of any of the above should also be included within the scope of computer-readable media.
  • Memory 112 includes computer-storage media in the form of volatile and/or nonvolatile memory.
  • the memory may be removable, non-removable, or a combination thereof.
  • Exemplary hardware devices include solid-state memory, hard drives, optical-disc drives, etc.
  • Computing device 100 includes one or more processors that read data from various entities such as memory 112 or I/O components 120 .
  • Presentation component(s) 116 present data indications to a user or other device.
  • Exemplary presentation components include a display device, speaker, printing component, vibrating component, etc.
  • I/O ports 118 allow computing device 100 to be logically coupled to other devices including I/O components 120 , some of which may be built in.
  • I/O components 120 include a microphone, joystick, game pad, satellite dish, scanner, printer, wireless device, etc.
  • FIG. 2 a block diagram is provided that illustrates a system 200 for classifying an HMD wearer's TFOV and locating virtual objects in classified regions in accordance with an embodiment of the present invention.
  • this and other arrangements described herein are set forth only as examples. Other arrangements and elements (e.g., machines, interfaces, functions, orders, and groupings of functions, etc.) can be used in addition to or instead of those shown, and some elements may be omitted altogether. Further, many of the elements described herein are functional entities that may be implemented as discrete or distributed components or in conjunction with other components, and in any suitable combination and location.
  • Various functions described herein as being performed by one or more entities may be carried out by hardware, firmware, and/or software. For instance, various functions may be carried out by a processor executing instructions stored in memory.
  • the system 200 includes an HMD 202 .
  • the HMD 202 may include components such as those described above with reference to FIG. 1 for the general computing device 100 , including, for instance, memory, processors, computer-readable media, input/output components, and a power supply.
  • the HMD 202 may be provided in any of a number of different form factors, including, for instance, glasses, goggles, or a helmet.
  • the HMD 202 may generally provide an augmented view of the real world to a wearer by augmenting the wearer's view of the real world with computer-generated virtual images.
  • the HMD 202 may include a display component 204 that displays computer-generated virtual images while still allowing the wearer to view the real world.
  • this may include heads-up display (HUD) technology, which may employ, for instance, any type of projection or microdisplay technology to provide the virtual images.
  • HUD heads-up display
  • Other technologies may also be employed, such as retinal displays in which images are projected directly onto the wearer's retina while the wearer is viewing the real world.
  • the HMD 202 may also include a wireless communication component 206 that provides wireless communication capabilities to allow the HMD 202 to communicate with companion devices (e.g., smartphones), server devices, or other network components.
  • companion devices e.g., smartphones
  • server devices e.g., smartphones
  • some embodiments of the present invention may be implemented by having a cloud-based service assist the HMD 202 in classifying regions of the HMD wearer's TFOV and locating virtual objects in those regions.
  • the HMD 202 may be configured to communicate directly with server devices; while in other embodiments, the HMD 202 may communicate with server devices via a companion device local to the HMD wearer (e.g., a smartphone or other device).
  • the HMD 202 may also include a number of on-board sensors 208 that provide positional and other sensor data. Any of a variety of different types of on-board sensors 208 may be included on the HMD 202 in accordance with various embodiments of the present invention. Generally, any sensor that allows the system to determine the HMD wearer's head position and rotation relative to the HMD wearer's body and surrounding environment, eye position, or other useful situational information may be employed.
  • the on-board sensors 208 may include GPS sensors, inertial measurement unit (IMU) sensors, depth sensors, cameras, eye tracking sensors, microphones, biometric sensors, and other types of sensors.
  • IMU inertial measurement unit
  • An IMU on the HMD 202 may measure inertial acceleration and may incorporate the functionality of accelerometers, gyroscopes, magnetometer, and other devices. In one embodiment, an IMU measurement of inertial acceleration has six degrees of freedom. Data from an IMU may provide relative movement and rotation information that may be employed with other data to infer absolute location in space. The additional data may include information from GPS sensors, which may be employed to provide location information on a macro level. In some embodiments, the HMD 202 may employ assisted GPS (A-GPS).
  • A-GPS assisted GPS
  • Depth sensors and cameras may be employed on the HMD 202 to collect data regarding the environment surrounding the HMD wearer.
  • Depth sensors may generally include sensors that determine the distance to objects.
  • a depth sensor may include an infrared sensor that captures IR light emitted from a transmission source and reflected from an object.
  • Distance data may be determined using time-of-flight, triangulation, or other known principles.
  • Cameras may be IR or visible spectrum, black and white or red-green-blue (RGB) cameras. In some embodiments, the parallax between images from two different cameras can be used to measure depth much as a person's two eyes provide depth perception.
  • the HMD 202 may also include one or more eye-tracking sensors, which typically track the movement of the pupil or other portions of the eye or the area around the eye to determine the direction of a user's gaze. This may be accomplished, for example, using IR or RGB cameras aimed at the HMD wearer's eyes.
  • Eye-tracking sensors typically track the movement of the pupil or other portions of the eye or the area around the eye to determine the direction of a user's gaze. This may be accomplished, for example, using IR or RGB cameras aimed at the HMD wearer's eyes.
  • Microphones may be provided to collect audio information, including the HMD wearer's voice and surrounding environmental audio.
  • Biometric sensors may be used to collect biometric information for the HMD wearer.
  • Sensor data from the on-board sensors 208 may be provided to a virtual imaging system 210 for processing.
  • the virtual imaging system 210 may be provided on the HMD 202 .
  • the virtual imaging system 210 may be provided by a device external to the HMD 202 , such as a server device or other network component.
  • the virtual imaging system 210 may be distributed across multiple devices (e.g., the HMD 202 and an external device). Any and all such variations are contemplated to be within the scope of embodiments of the present invention.
  • the virtual imaging system 210 includes a head position component 212 that is operable to receive sensor data from the on-board sensors 208 and determine the HMD wearer's head position and rotation relative to HMD wearer's surrounding environment and the HMD wearer's body.
  • the head position component 212 may employ techniques such as simultaneous location and mapping (SLAM) using, for instance, camera data and/or depth sensor data to provide a real-time position of the HMD wearer's head relative to a mapped surrounding environment.
  • IMU data may also provide relative rotation and position information even when the camera or depth sensors are unavailable.
  • Regions of the HMD wearer's body e.g., hand, arm, torso, legs, feet, etc. may also be identified using sensor data.
  • camera data and/or depth sensor data may be obtained when the HMD wearer is looking at him or herself.
  • the HMD wearer's head position and rotation relative to the HMD wearer's body may be inferred to a useful degree of accuracy. This may include information such as whether the HMD wearer is standing, sitting, looking straight ahead relative to torso, to name a few.
  • the head position component 212 may also receive data from sensors and other sources 220 external to the HMD 212 .
  • the other sources 220 may include external cameras, external depth sensors (e.g., a KINECT sensor), other HMDs, mobile devices, and historical sensor data stored remotely from the HMD.
  • any external source of information that allows the system to determine the HMD wearer's head position and rotation relative to the HMD wearer's body and surrounding environment, eye position, or other useful situational information may be employed.
  • Data from the various on-board sensors 208 and external sources 220 may generally provide redundancies and refinement. However, it should be understood that not all sensors are necessary. Any combination of sensors indicated herein, as well as other sensors and sources of information, may be employed within the scope of embodiments of the present invention.
  • the TFOV classification component 214 classifies regions of the HMD wearer's TFOV.
  • the TFOV classification component 214 identifies regions of the HMD wearer's TFOV according to various rules.
  • the rules may be user-defined or system-defined. Additionally, the rules may be immutable, dynamic based on different inputs regarding the HMD wearer's current situation and environment, or alterable by the HMD wearer to provide the best experience for the HMD wearer based on what the HMD wearer is currently doing.
  • the regions may include a primary region and any number of additional non-primary regions.
  • the primary region may correspond with a primary area within the HMD wearer's FOV. Generally, this corresponds with an area that should not be occluded with virtual objects.
  • the non-primary regions correspond with areas that may be occluded with virtual objects.
  • FIGS. 3A and 3B illustrate a profile view and front view, respectively, of a TFOV 302 for an HMD wearer 304 .
  • the TFOV 302 has been classified into a primary region 306 , a secondary region 308 , and a tertiary region 310 .
  • the classification shown in FIGS. 3A and 3B are provided as examples only, and the TFOV may generally be classified into any number of regions.
  • the regions may be symmetrical or non-symmetrical. Additionally, the regions may be divided vertically, horizontally, or any permutation thereof.
  • the regions may be classified relative to the HMD wearer's body or body parts. For instance, the regions may comprise an upper left of body region, upper right of body region, lower left region of body, and lower right region of body.
  • FIGS. 4 and 5 The classification of regions of the TFOV may be fixed or may dynamically change.
  • FIGS. 4 and 5 examples of fixed regions versus dynamic regions are provided in FIGS. 4 and 5 .
  • FIG. 4 illustrates an example of fixed regions.
  • FIG. 5 illustrates an example of dynamic regions.
  • the regions move with the HMD wearer's head movements.
  • TFOV regions may dynamically change by automatically adjusting the size (i.e., expanding or shrinking) of regions, shifting regions within the TFOV, or completely re-classifying new regions for the TFOV.
  • the regions may automatically change based on rules triggered by events, environmental conditions, body placement and movement, and additional sensor information. For instance, if the HMD wearer is sitting, the primary region may be relatively small since the HMD wearer is stationary and it's safer to occlude more of the HMD wearer's FOV. However, if the HMD wearer stands and begins to walk, the primary region may expand such that less of the HMD wearer's FOV is occluded with virtual images.
  • the primary region may expand even further.
  • re-classifying regions suppose the HMD wearer's eyes are gazing and hands are moving in a secondary region of the TFOV. This may be indicative that the HMD wearer is performing some task in that space. As a result of these inputs, the TFOV may be re-classified such that region is now the primary region.
  • Other environmental conditions may also affect the classification of regions. For instance, biometric inputs may indicate that the HMD wearer is nervous or frightened. In response, the primary region may expand.
  • combinations of fixed regions and dynamically changing regions may be employed over time by employing rules that control whether fixed regions or dynamic regions are employed at any given time.
  • some circumstances may dictate the use of fixed regions, while other circumstances may dictate the use of dynamic regions.
  • the regions may initially be fixed for a given time period.
  • rules may then dictate dynamically changing the regions.
  • the regions may remain fixed while the HMD wearer remains sitting.
  • rules may be triggered to cause the regions to dynamically change as the user moves around in space. Accordingly, the region classifications may be environmentally-adaptive.
  • the virtual object placement component 216 may locate virtual objects in the various regions according to the classifications.
  • the virtual objects may be placed in regions according to defined rules.
  • the rules may be user-defined or system-defined. Additionally, the rules may be immutable, dynamic based on different inputs regarding the HMD wearer's current situation and environment, or alterable by the HMD wearer to provide the best experience for the HMD wearer based on what the HMD wearer is currently doing. In some embodiments, this may include placing most virtual objects into non-primary regions of the TFOV to prevent occluding the primary region. In some instances, only notifications, less intrusive virtual objects, or virtual objects deemed to be more important may be located in the primary region.
  • FIG. 6 provides a view provided to an HMD wearer in accordance with an embodiment of the present invention.
  • the HMD wearer is currently viewing another person 602 .
  • the HMD wearer's FOV covers a primary region 604 , secondary region 606 , and tertiary region 608 of the TFOV.
  • dashed lines and region labels are provided in FIG. 6 for discussion purposes, and those would not be displayed to the HMD wearer. Instead, only the virtual objects 610 would be displayed. As shown in FIG.
  • a permission set may be employed to govern placement of content within the TFOV. This may include universal rules that span across applications or circumstances that dictate to place or not place specific content in specific regions. There may be also be changing TFOV permissions based on status (e.g., rules for placement of virtual objects when the user is driving may be different from rules for placement of virtual objects when the user is in a living room). There may also be permission sets that change based on the classification of the application being run by the system. The interplay between these global/state/application-based permissions drives the content and the location of the content displayed within the TFOV.
  • virtual objects may be located in the regions in order to locate those virtual objects relative to the HMD wearer's body. For instance, a region may be classified around the HMD wearer's foot as a foot region, and a particular virtual object may be located in that region in order to located the virtual object near the HMD wearer's foot. As another example, a region surrounding the HMD wearer's torso may be classified as a torso region, and a virtual desktop may be located in that region to contain information the HMD wearer may want to access. As such, the HMD wearer may view the information by looking down at the virtual desktop surrounding the HMD wearer's torso.
  • the virtual objects may include static objects that are not interactive. For instance, virtual objects may simply display information or graphics for the HMD wearer to view. In other instances, the virtual objects may be user interface (UI) objects with which the HMD wearer may interact, using gestures, voice commands, and/or other forms of input.
  • UI user interface
  • the virtual image display component 218 may be operable to determine the HMD wearer's current FOV within the TFOV and to cause virtual objects to be displayed in the appropriate locations within the HMD wearer's FOV. For instance, if a virtual object is located in a TFOV region near the HMD wearer's feet, the virtual object would not be displayed when the HMD wearer is looking up.
  • the virtual image display component 218 would determine that the HMD wearer's FOV is over the foot region and would cause the virtual object to be displayed at the appropriate location. It should be noted that in some embodiments, the extent of the HMD wearer's FOV may simply be considered to correspond with the extent of the display area of the display component 204 of the HMD 202 .
  • the HMD's FOV may extend beyond the HMD wearer's FOV.
  • the depth sensors, cameras, and other on-board HMD sensors, as well as external sensors may be able to capture data outside of the HMD wearer's FOV.
  • the HMD's FOV may be defined by the extent to which the various sensors cover, which may extend well beyond the HMD wearer's FOV.
  • an interactive virtual object is located adjacent in a region adjacent to the HMD wearer's right hip.
  • the HMD wearer may interact with the virtual object even when the HMD wearer is looking elsewhere and the virtual object is not being displayed to the HMD wearer, as long as the area adjacent to the HMD wearer's right hip at which the virtual object is located is within the HMD's FOV such that the HMD may capture the HMD wearer's interaction with the virtual object.
  • a flow diagram is provided that illustrates a method 700 for classifying regions of an HMD wearer's TFOV in accordance with an embodiment of the present invention.
  • the method 700 may be performed, for instance, by an HMD, a device external to the HMD, or a combination thereof.
  • data from on-board sensors provided on an HMD is received.
  • the on-board sensors may include, without limitation, GPS sensors, inertial measurement unit (IMU) sensors, depth sensors, cameras, eye tracking sensors, microphones, biometric inputs, and other sensors.
  • data may be received from sources external to the HMD, as shown at block 704 .
  • this may include external cameras, external depth sensors, other HMDs, mobile devices, and historical sensor data stored remotely from the HMD.
  • the HMD wearer's head position and rotation are determined relative to the HMD wearer's body and surrounding environment, as shown at block 706 .
  • the HMD wearer's TFOV is classified into two or more regions, as shown at block 708 .
  • the TFOV is classified according to rules that may be user-defined or system-defined and immutable or dynamic.
  • Virtual objects are located in the regions of the TFOV, as shown at block 710 .
  • a number of rules may also be defined for placing the various virtual objects into the regions. These rules may also be user-defined or system-defined and immutable or dynamic.
  • the system may take into account the applications that are being accessed when determining where to locate the various virtual objects. As represented by the return to blocks 702 and 704 , the process of classifying regions of the TFOV may be repeated. In some embodiments, one or more rules may trigger reclassifying regions of the TFOV, while in other embodiment, the TFOV may be continuously reclassified.
  • FIG. 8 a flow diagram is provided that illustrates a method 800 for displaying virtual objects using an HMD in accordance with an embodiment of the present invention.
  • the HMD wearer's FOV is initially determined.
  • the extent of the HMD wearer's FOV may simply correspond with the extent the display area provided by a display component of the HMD.
  • the HMD wearer's FOV is compared with regions of a TFOV that have been determined, for instance, using the method 700 described above with reference to FIG. 7 . The comparison identifies the placement of the HMD wearer's FOV relative to the regions of the TFOV.
  • the virtual objects to display are determined at block 806 based on the location of the HMD wearer's FOV relative to the TFOV regions and the location of virtual objects in the TFOV regions. In some embodiments, a set of permissions may be applied at this point to govern usage of virtual objects based on, for instance, the current user state or the specific application(s) being engaged.
  • the virtual objects that are identified are then displayed using a display component of the HMD at the appropriate location, as shown at block 808 .
  • embodiments of the present invention provide for classifying regions of an HMD wearer's TFOV based on sensor inputs and rules and locating virtual objects in the classified regions.
  • the present invention has been described in relation to particular embodiments, which are intended in all respects to be illustrative rather than restrictive. Alternative embodiments will become apparent to those of ordinary skill in the art to which the present invention pertains without departing from its scope.

Abstract

Virtual images are located for display in a head-mounted display (HMD) to provide an augment reality view to an HMD wearer. Sensor data may be collected from on-board sensors provided on an HMD. Additionally, other day may be collected from external sources. Based on the collected sensor data and other data, the position and rotation of the HMD wearer's head relative to the HMD wearer's body and surrounding environment may be determined. After resolving the HMD wearer's head position, the HMD wearer's total field of view (TFOV) may be classified into regions. Virtual images may then be located in the classified TFOV regions to locate the virtual images relative to the HMD wearer's body and surrounding environment.

Description

    BACKGROUND
  • See-through head-mounted displays (HMDs) provide the ability to augment what the wearer sees with virtual objects. In other words, the HMD augments the HMD wearer's view of the real world with virtual imagery to provide an augment reality view. However, those virtual images can easily become distracting and obtrusive, drawing attention away from the real-world and occluding the HMD wearer's normal (non-augmented) view. This may present a challenge particularly in applications of “always-on” wearable display devices, especially when considering mobile, outdoor, and in-activity scenarios.
  • SUMMARY
  • This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
  • Embodiments of the present invention relate to locating virtual images in a head-mounted display (HMD) to provide an augment reality view to an HMD wearer. Sensor data may be collected from on-board sensors provided on an HMD. Additionally, other data may be collected from external sources. Based on the collected sensor data and other data, the position and rotation of the HMD wearer's head relative to the HMD wearer's body and surrounding environment may be determined. After resolving the HMD wearer's head position, the HMD wearer's total field of view (TFOV) may be classified into regions. Virtual images may then be located in the classified TFOV regions to locate the virtual images relative to the HMD wearer's body and surrounding environment.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention is described in detail below with reference to the attached drawing figures, wherein:
  • FIG. 1 is a block diagram of an exemplary computing environment suitable for use in implementing embodiments of the present invention;
  • FIG. 2 is a block diagram of an exemplary system for classifying regions of a TFOV and locating virtual objects in the regions in accordance with an embodiment of the present invention;
  • FIGS. 3A and 3B provide a profile and front views of an HMD wearer's TFOV classified with primary, secondary, and tertiary regions in accordance with an embodiment of the present invention;
  • FIG. 4 illustrates fixed regions for a HMD wearer's TFOV in accordance with an embodiment of the present invention;
  • FIG. 5 illustrates dynamic regions for an HMD wearer's TFOV in accordance with an embodiment of the present invention;
  • FIG. 6 illustrates an HMD wearer's view with virtual objects located in primary, secondary, and tertiary regions in accordance with an embodiment of the present invention;
  • FIG. 7 is a flow diagram showing a method for classifying regions of an HMD wearer's TFOV in accordance with an embodiment of the present invention; and
  • FIG. 8 is a flow diagram showing a method for displaying virtual objects using an HMD in accordance with an embodiment of the present invention.
  • DETAILED DESCRIPTION
  • The subject matter of the present invention is described with specificity herein to meet statutory requirements. However, the description itself is not intended to limit the scope of this patent. Rather, the inventors have contemplated that the claimed subject matter might also be embodied in other ways, to include different steps or combinations of steps similar to the ones described in this document, in conjunction with other present or future technologies. Moreover, although the terms “step” and/or “block” may be used herein to connote different elements of methods employed, the terms should not be interpreted as implying any particular order among or between various steps herein disclosed unless and except when the order of individual steps is explicitly described.
  • Embodiments of the present invention are directed to the automatic and continuous placement and replacement of virtual images in an HMD's display relative to the HMD wearer's total field of view (TFOV) to provide the HMD wearer with an augmented reality view. As used herein, “TFOV” accounts for complete range of rotation and translation of a person's head. This is in contrast to a person's field of view (FOV), which accounts for the extent of what the person sees at a given moment.
  • In accordance with embodiments of the present invention, sensor data is collected from any number of on-board sensors provided on an HMD. Additionally, other data may be collected from sources external to the HMD. The sensor data and other data may be used to track and interpret the HMD wearer's physical head as it relates the HMD wearer's physical body and to the physical environment surrounding the HMD wearer. By tracking the HMD wearer's physical head relative to the HMD wearer's body and surrounding environment, the HMD wearer's TFOV may be classified into various regions. These regions may be defined relative to the HMD wearer's body and surrounding environment. In some embodiments, the regions may include a primary region and one or more non-primary regions. Virtual images may be located within the TFOV according to the classified regions. Accordingly, embodiments of the present invention provide a system that is aware of an HMD wearer's head relative to the HMD wearer's body and surrounding environment and may use such knowledge to locate virtual images relative to the HMD wearer's body and surrounding environment. In some embodiments, the virtual objects may be generally located away from the primary region and placed in unobtrusive regions such that the virtual images are presented in non-dangerous ways by reducing visual occlusion of the primary region. The virtual objects may also be located away from the primary region in order to provide context (e.g., classify the information as less relevant, less important, etc.).
  • Accordingly, in one aspect, an embodiment of the present invention is directed to one or more computer storage media storing computer-useable instructions that, when used by one or more computing devices, cause the one or more computing device to perform a method. The method includes receiving sensor data from one or more HMD on-board sensors. The method also includes using the sensor data to determine an HMD wearer's head position and rotation relative to the HMD wearer's body and an environment surrounding the HMD wearer. The method further includes classifying two or more regions within the HMD wearer's TFOV based on one or more pre-determined rules and the HMD wearer's head position and rotation relative to the HMD wearer's body and an environment surrounding the HMD wearer. The method still further includes locating virtual images to be displayed by the HMD based on classifying the two or more regions within the HMD wearer's TFOV.
  • In another embodiment, an aspect of the invention is directed to an HMD. The HMD includes one or more on-board sensors. The HMD also includes one or more processors configured to: determine a position and rotation of an HMD wearer's head relative to the HMD wearer's body based on sensor data from the one or more on-board sensors, classify two or more regions of the HMD wearer's TFOV based on the position and rotation of the HMD wearer's head relative to the HMD wearer's body, and locate one or more virtual objects in the two or more regions. The HMD further includes one or more display components configured to display at least one of the one or more virtual objects to provide the HMD wearer an augmented view.
  • A further embodiment is directed to a method for classifying regions of an HMD wearer's TFOV. The method include continuously: receiving sensor data from one or more sensors on board the HMD; receiving other data from one or more sources external to the HMD; and classifying regions of the HMD wearer's TFOV relative to the HMD wearer's body and surrounding environment based on the sensor data and other data.
  • Having briefly described an overview of embodiments of the present invention, an exemplary operating environment in which embodiments of the present invention may be implemented is described below in order to provide a general context for various aspects of the present invention. Referring initially to FIG. 1 in particular, an exemplary operating environment for implementing embodiments of the present invention is shown and designated generally as computing device 100. Computing device 100 is but one example of a suitable computing environment and is not intended to suggest any limitation as to the scope of use or functionality of the invention. Neither should the computing device 100 be interpreted as having any dependency or requirement relating to any one or combination of components illustrated.
  • The invention may be described in the general context of computer code or machine-useable instructions, including computer-executable instructions such as program modules, being executed by a computer or other machine, such as a personal data assistant or other handheld device. Generally, program modules including routines, programs, objects, components, data structures, etc., refer to code that perform particular tasks or implement particular abstract data types. The invention may be practiced in a variety of system configurations, including hand-held devices, consumer electronics, general-purpose computers, more specialty computing devices, etc. The invention may also be practiced in distributed computing environments where tasks are performed by remote-processing devices that are linked through a communications network.
  • With reference to FIG. 1, computing device 100 includes a bus 110 that directly or indirectly couples the following devices: memory 112, one or more processors 114, one or more presentation components 116, input/output (I/O) ports 118, input/output components 120, and an illustrative power supply 122. Bus 110 represents what may be one or more busses (such as an address bus, data bus, or combination thereof). Although the various blocks of FIG. 1 are shown with lines for the sake of clarity, in reality, delineating various components is not so clear, and metaphorically, the lines would more accurately be grey and fuzzy. For example, one may consider a presentation component such as a display device to be an I/O component. Also, processors have memory. The inventors recognize that such is the nature of the art, and reiterate that the diagram of FIG. 1 is merely illustrative of an exemplary computing device that can be used in connection with one or more embodiments of the present invention. Distinction is not made between such categories as “workstation,” “server,” “laptop,” “hand-held device,” etc., as all are contemplated within the scope of FIG. 1 and reference to “computing device.”
  • Computing device 100 typically includes a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by computing device 100 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer-readable media may comprise computer storage media and communication media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 100. Communication media typically embodies computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. Combinations of any of the above should also be included within the scope of computer-readable media.
  • Memory 112 includes computer-storage media in the form of volatile and/or nonvolatile memory. The memory may be removable, non-removable, or a combination thereof. Exemplary hardware devices include solid-state memory, hard drives, optical-disc drives, etc. Computing device 100 includes one or more processors that read data from various entities such as memory 112 or I/O components 120. Presentation component(s) 116 present data indications to a user or other device. Exemplary presentation components include a display device, speaker, printing component, vibrating component, etc.
  • I/O ports 118 allow computing device 100 to be logically coupled to other devices including I/O components 120, some of which may be built in. Illustrative components include a microphone, joystick, game pad, satellite dish, scanner, printer, wireless device, etc.
  • Referring now to FIG. 2, a block diagram is provided that illustrates a system 200 for classifying an HMD wearer's TFOV and locating virtual objects in classified regions in accordance with an embodiment of the present invention. It should be understood that this and other arrangements described herein are set forth only as examples. Other arrangements and elements (e.g., machines, interfaces, functions, orders, and groupings of functions, etc.) can be used in addition to or instead of those shown, and some elements may be omitted altogether. Further, many of the elements described herein are functional entities that may be implemented as discrete or distributed components or in conjunction with other components, and in any suitable combination and location. Various functions described herein as being performed by one or more entities may be carried out by hardware, firmware, and/or software. For instance, various functions may be carried out by a processor executing instructions stored in memory.
  • As shown in FIG. 2, the system 200 includes an HMD 202. Although not shown, the HMD 202 may include components such as those described above with reference to FIG. 1 for the general computing device 100, including, for instance, memory, processors, computer-readable media, input/output components, and a power supply. The HMD 202 may be provided in any of a number of different form factors, including, for instance, glasses, goggles, or a helmet.
  • The HMD 202 may generally provide an augmented view of the real world to a wearer by augmenting the wearer's view of the real world with computer-generated virtual images. To provide the wearer with an augmented view, the HMD 202 may include a display component 204 that displays computer-generated virtual images while still allowing the wearer to view the real world. In some embodiments, this may include heads-up display (HUD) technology, which may employ, for instance, any type of projection or microdisplay technology to provide the virtual images. Other technologies may also be employed, such as retinal displays in which images are projected directly onto the wearer's retina while the wearer is viewing the real world.
  • The HMD 202 may also include a wireless communication component 206 that provides wireless communication capabilities to allow the HMD 202 to communicate with companion devices (e.g., smartphones), server devices, or other network components. For instance, some embodiments of the present invention may be implemented by having a cloud-based service assist the HMD 202 in classifying regions of the HMD wearer's TFOV and locating virtual objects in those regions. In some embodiments, the HMD 202 may be configured to communicate directly with server devices; while in other embodiments, the HMD 202 may communicate with server devices via a companion device local to the HMD wearer (e.g., a smartphone or other device).
  • The HMD 202 may also include a number of on-board sensors 208 that provide positional and other sensor data. Any of a variety of different types of on-board sensors 208 may be included on the HMD 202 in accordance with various embodiments of the present invention. Generally, any sensor that allows the system to determine the HMD wearer's head position and rotation relative to the HMD wearer's body and surrounding environment, eye position, or other useful situational information may be employed. By way of example only and not limitation, the on-board sensors 208 may include GPS sensors, inertial measurement unit (IMU) sensors, depth sensors, cameras, eye tracking sensors, microphones, biometric sensors, and other types of sensors.
  • An IMU on the HMD 202 may measure inertial acceleration and may incorporate the functionality of accelerometers, gyroscopes, magnetometer, and other devices. In one embodiment, an IMU measurement of inertial acceleration has six degrees of freedom. Data from an IMU may provide relative movement and rotation information that may be employed with other data to infer absolute location in space. The additional data may include information from GPS sensors, which may be employed to provide location information on a macro level. In some embodiments, the HMD 202 may employ assisted GPS (A-GPS).
  • Depth sensors and cameras may be employed on the HMD 202 to collect data regarding the environment surrounding the HMD wearer. Depth sensors may generally include sensors that determine the distance to objects. In some implementations, a depth sensor may include an infrared sensor that captures IR light emitted from a transmission source and reflected from an object. Distance data may be determined using time-of-flight, triangulation, or other known principles. Cameras may be IR or visible spectrum, black and white or red-green-blue (RGB) cameras. In some embodiments, the parallax between images from two different cameras can be used to measure depth much as a person's two eyes provide depth perception.
  • The HMD 202 may also include one or more eye-tracking sensors, which typically track the movement of the pupil or other portions of the eye or the area around the eye to determine the direction of a user's gaze. This may be accomplished, for example, using IR or RGB cameras aimed at the HMD wearer's eyes. Microphones may be provided to collect audio information, including the HMD wearer's voice and surrounding environmental audio. Biometric sensors may be used to collect biometric information for the HMD wearer.
  • Sensor data from the on-board sensors 208 may be provided to a virtual imaging system 210 for processing. In some embodiments of the present invention, the virtual imaging system 210 may be provided on the HMD 202. In other embodiments, the virtual imaging system 210 may be provided by a device external to the HMD 202, such as a server device or other network component. In still further embodiments, the virtual imaging system 210 may be distributed across multiple devices (e.g., the HMD 202 and an external device). Any and all such variations are contemplated to be within the scope of embodiments of the present invention.
  • The virtual imaging system 210 includes a head position component 212 that is operable to receive sensor data from the on-board sensors 208 and determine the HMD wearer's head position and rotation relative to HMD wearer's surrounding environment and the HMD wearer's body. By way of example only and not limitation, the head position component 212 may employ techniques such as simultaneous location and mapping (SLAM) using, for instance, camera data and/or depth sensor data to provide a real-time position of the HMD wearer's head relative to a mapped surrounding environment. IMU data may also provide relative rotation and position information even when the camera or depth sensors are unavailable. Regions of the HMD wearer's body (e.g., hand, arm, torso, legs, feet, etc.) may also be identified using sensor data. For instance, camera data and/or depth sensor data may be obtained when the HMD wearer is looking at him or herself. As such, the HMD wearer's head position and rotation relative to the HMD wearer's body may be inferred to a useful degree of accuracy. This may include information such as whether the HMD wearer is standing, sitting, looking straight ahead relative to torso, to name a few.
  • The head position component 212 may also receive data from sensors and other sources 220 external to the HMD 212. By way of example only and not limitation, the other sources 220 may include external cameras, external depth sensors (e.g., a KINECT sensor), other HMDs, mobile devices, and historical sensor data stored remotely from the HMD. Generally, any external source of information that allows the system to determine the HMD wearer's head position and rotation relative to the HMD wearer's body and surrounding environment, eye position, or other useful situational information may be employed.
  • Data from the various on-board sensors 208 and external sources 220 may generally provide redundancies and refinement. However, it should be understood that not all sensors are necessary. Any combination of sensors indicated herein, as well as other sensors and sources of information, may be employed within the scope of embodiments of the present invention.
  • After the head position component 212 resolves the HMD wearer's head position and rotation relative to the HMD wearer's body and surrounding environment, the TFOV classification component 214 classifies regions of the HMD wearer's TFOV. In accordance with embodiments of the present invention, the TFOV classification component 214 identifies regions of the HMD wearer's TFOV according to various rules. The rules may be user-defined or system-defined. Additionally, the rules may be immutable, dynamic based on different inputs regarding the HMD wearer's current situation and environment, or alterable by the HMD wearer to provide the best experience for the HMD wearer based on what the HMD wearer is currently doing.
  • In some embodiments of the present invention, the regions may include a primary region and any number of additional non-primary regions. The primary region may correspond with a primary area within the HMD wearer's FOV. Generally, this corresponds with an area that should not be occluded with virtual objects. The non-primary regions correspond with areas that may be occluded with virtual objects.
  • By way of example, FIGS. 3A and 3B illustrate a profile view and front view, respectively, of a TFOV 302 for an HMD wearer 304. As shown in FIGS. 3A and 3B, the TFOV 302 has been classified into a primary region 306, a secondary region 308, and a tertiary region 310. It should be noted that the classification shown in FIGS. 3A and 3B are provided as examples only, and the TFOV may generally be classified into any number of regions. The regions may be symmetrical or non-symmetrical. Additionally, the regions may be divided vertically, horizontally, or any permutation thereof. In some embodiments, the regions may be classified relative to the HMD wearer's body or body parts. For instance, the regions may comprise an upper left of body region, upper right of body region, lower left region of body, and lower right region of body.
  • The classification of regions of the TFOV may be fixed or may dynamically change. By way of illustration, examples of fixed regions versus dynamic regions are provided in FIGS. 4 and 5. Initially, FIG. 4 illustrates an example of fixed regions. As shown in FIG. 4, as the HMD wearer moves his/her head looking up and down, the regions remain fixed. In contrast, FIG. 5 illustrates an example of dynamic regions. As shown in FIG. 5, as the HMD wearer moves his/her head up and down, the regions move with the HMD wearer's head movements.
  • TFOV regions may dynamically change by automatically adjusting the size (i.e., expanding or shrinking) of regions, shifting regions within the TFOV, or completely re-classifying new regions for the TFOV. The regions may automatically change based on rules triggered by events, environmental conditions, body placement and movement, and additional sensor information. For instance, if the HMD wearer is sitting, the primary region may be relatively small since the HMD wearer is stationary and it's safer to occlude more of the HMD wearer's FOV. However, if the HMD wearer stands and begins to walk, the primary region may expand such that less of the HMD wearer's FOV is occluded with virtual images. If the HMD wearer begins running or begins to drive a car, the primary region may expand even further. As an example of re-classifying regions, suppose the HMD wearer's eyes are gazing and hands are moving in a secondary region of the TFOV. This may be indicative that the HMD wearer is performing some task in that space. As a result of these inputs, the TFOV may be re-classified such that region is now the primary region. Other environmental conditions may also affect the classification of regions. For instance, biometric inputs may indicate that the HMD wearer is nervous or frightened. In response, the primary region may expand.
  • In accordance with some embodiments of the present invention, combinations of fixed regions and dynamically changing regions may be employed over time by employing rules that control whether fixed regions or dynamic regions are employed at any given time. In particular, some circumstances may dictate the use of fixed regions, while other circumstances may dictate the use of dynamic regions. For instance, the regions may initially be fixed for a given time period. When certain inputs are received, rules may then dictate dynamically changing the regions. By way of specific example, if the HMD wearer is sitting, the regions may remain fixed while the HMD wearer remains sitting. However, once the HMD wearer gets up, rules may be triggered to cause the regions to dynamically change as the user moves around in space. Accordingly, the region classifications may be environmentally-adaptive.
  • After classifying regions of the TFOV, the virtual object placement component 216 may locate virtual objects in the various regions according to the classifications. Generally, the virtual objects may be placed in regions according to defined rules. The rules may be user-defined or system-defined. Additionally, the rules may be immutable, dynamic based on different inputs regarding the HMD wearer's current situation and environment, or alterable by the HMD wearer to provide the best experience for the HMD wearer based on what the HMD wearer is currently doing. In some embodiments, this may include placing most virtual objects into non-primary regions of the TFOV to prevent occluding the primary region. In some instances, only notifications, less intrusive virtual objects, or virtual objects deemed to be more important may be located in the primary region.
  • By way of illustration, FIG. 6 provides a view provided to an HMD wearer in accordance with an embodiment of the present invention. As shown in FIG. 6, the HMD wearer is currently viewing another person 602. The HMD wearer's FOV covers a primary region 604, secondary region 606, and tertiary region 608 of the TFOV. It should be noted that the dashed lines and region labels are provided in FIG. 6 for discussion purposes, and those would not be displayed to the HMD wearer. Instead, only the virtual objects 610 would be displayed. As shown in FIG. 6, only smaller notification virtual objects 610 are displayed within the primary region 604, while larger, more intrusive virtual objects 610 are displayed within the secondary region 606 and tertiary region 608. As a result, the primary region 604 remains generally unobstructed by virtual objects.
  • In various embodiments of the present invention, a permission set may be employed to govern placement of content within the TFOV. This may include universal rules that span across applications or circumstances that dictate to place or not place specific content in specific regions. There may be also be changing TFOV permissions based on status (e.g., rules for placement of virtual objects when the user is driving may be different from rules for placement of virtual objects when the user is in a living room). There may also be permission sets that change based on the classification of the application being run by the system. The interplay between these global/state/application-based permissions drives the content and the location of the content displayed within the TFOV.
  • In embodiments in which regions are classified relative to the HMD wearer's body, virtual objects may be located in the regions in order to locate those virtual objects relative to the HMD wearer's body. For instance, a region may be classified around the HMD wearer's foot as a foot region, and a particular virtual object may be located in that region in order to located the virtual object near the HMD wearer's foot. As another example, a region surrounding the HMD wearer's torso may be classified as a torso region, and a virtual desktop may be located in that region to contain information the HMD wearer may want to access. As such, the HMD wearer may view the information by looking down at the virtual desktop surrounding the HMD wearer's torso.
  • In some instances, the virtual objects may include static objects that are not interactive. For instance, virtual objects may simply display information or graphics for the HMD wearer to view. In other instances, the virtual objects may be user interface (UI) objects with which the HMD wearer may interact, using gestures, voice commands, and/or other forms of input.
  • Because the TFOV extends beyond the HMD wearer's FOV, only a portion of all virtual objects within the TFOV may be viewed by the HMD wearer at any give time. Accordingly, the virtual image display component 218 may be operable to determine the HMD wearer's current FOV within the TFOV and to cause virtual objects to be displayed in the appropriate locations within the HMD wearer's FOV. For instance, if a virtual object is located in a TFOV region near the HMD wearer's feet, the virtual object would not be displayed when the HMD wearer is looking up. However, if the HMD wearer were to look down, the virtual image display component 218 would determine that the HMD wearer's FOV is over the foot region and would cause the virtual object to be displayed at the appropriate location. It should be noted that in some embodiments, the extent of the HMD wearer's FOV may simply be considered to correspond with the extent of the display area of the display component 204 of the HMD 202.
  • In some embodiments, the HMD's FOV may extend beyond the HMD wearer's FOV. For instance, the depth sensors, cameras, and other on-board HMD sensors, as well as external sensors may be able to capture data outside of the HMD wearer's FOV. As such, the HMD's FOV may be defined by the extent to which the various sensors cover, which may extend well beyond the HMD wearer's FOV. Some embodiments may take advantage of the HMD's FOV by allowing an HMD wearer to interact with a virtual object that is not currently displayed to the HMD wearer because it is outside the HMD wearer's FOV, as long as the virtual object is located within the HMD's FOV. For instance, suppose an interactive virtual object is located adjacent in a region adjacent to the HMD wearer's right hip. The HMD wearer may interact with the virtual object even when the HMD wearer is looking elsewhere and the virtual object is not being displayed to the HMD wearer, as long as the area adjacent to the HMD wearer's right hip at which the virtual object is located is within the HMD's FOV such that the HMD may capture the HMD wearer's interaction with the virtual object.
  • Referring now to FIG. 7, a flow diagram is provided that illustrates a method 700 for classifying regions of an HMD wearer's TFOV in accordance with an embodiment of the present invention. The method 700 may be performed, for instance, by an HMD, a device external to the HMD, or a combination thereof. As shown at block 702, data from on-board sensors provided on an HMD is received. As discussed previously, the on-board sensors may include, without limitation, GPS sensors, inertial measurement unit (IMU) sensors, depth sensors, cameras, eye tracking sensors, microphones, biometric inputs, and other sensors. Additionally, data may be received from sources external to the HMD, as shown at block 704. By way of example only and not limitation, this may include external cameras, external depth sensors, other HMDs, mobile devices, and historical sensor data stored remotely from the HMD.
  • Based on the sensor data from the on-board sensors and other data from the external sources, the HMD wearer's head position and rotation are determined relative to the HMD wearer's body and surrounding environment, as shown at block 706. Using that information, the HMD wearer's TFOV is classified into two or more regions, as shown at block 708. In accordance with embodiments of the present invention, the TFOV is classified according to rules that may be user-defined or system-defined and immutable or dynamic. Virtual objects are located in the regions of the TFOV, as shown at block 710. A number of rules may also be defined for placing the various virtual objects into the regions. These rules may also be user-defined or system-defined and immutable or dynamic. In some embodiments, the system may take into account the applications that are being accessed when determining where to locate the various virtual objects. As represented by the return to blocks 702 and 704, the process of classifying regions of the TFOV may be repeated. In some embodiments, one or more rules may trigger reclassifying regions of the TFOV, while in other embodiment, the TFOV may be continuously reclassified.
  • Turning to FIG. 8, a flow diagram is provided that illustrates a method 800 for displaying virtual objects using an HMD in accordance with an embodiment of the present invention. As shown at block 802, the HMD wearer's FOV is initially determined. In some embodiments, the extent of the HMD wearer's FOV may simply correspond with the extent the display area provided by a display component of the HMD. As shown at block 804, the HMD wearer's FOV is compared with regions of a TFOV that have been determined, for instance, using the method 700 described above with reference to FIG. 7. The comparison identifies the placement of the HMD wearer's FOV relative to the regions of the TFOV.
  • The virtual objects to display are determined at block 806 based on the location of the HMD wearer's FOV relative to the TFOV regions and the location of virtual objects in the TFOV regions. In some embodiments, a set of permissions may be applied at this point to govern usage of virtual objects based on, for instance, the current user state or the specific application(s) being engaged. The virtual objects that are identified are then displayed using a display component of the HMD at the appropriate location, as shown at block 808.
  • As can be understood, embodiments of the present invention provide for classifying regions of an HMD wearer's TFOV based on sensor inputs and rules and locating virtual objects in the classified regions. The present invention has been described in relation to particular embodiments, which are intended in all respects to be illustrative rather than restrictive. Alternative embodiments will become apparent to those of ordinary skill in the art to which the present invention pertains without departing from its scope.
  • From the foregoing, it will be seen that this invention is one well adapted to attain all the ends and objects set forth above, together with other advantages which are obvious and inherent to the system and method. It will be understood that certain features and subcombinations are of utility and may be employed without reference to other features and subcombinations. This is contemplated by and is within the scope of the claims.

Claims (20)

1. One or more computer storage media storing computer-useable instructions that, when used by one or more computing devices, cause the one or more computing device to perform a method, the method comprising:
receiving sensor data from one or more head-mounted display (HMD) on-board sensors;
using the sensor data to determine an HMD wearer's head position and rotation relative to the HMD wearer's body and an environment surrounding the HMD wearer;
classifying two or more regions within the HMD wearer's total field of view (TFOV) based on one or more pre-determined rules and the HMD wearer's head position and rotation relative to the HMD wearer's body and an environment surrounding the HMD wearer; and
locating virtual images to be displayed by the HMD based on classifying the two or more regions within the HMD wearer's TFOV.
2. The one or more computer storage media of claim 1, wherein the one or more HMD on-board sensors comprise one or more selected from the following: a GPS sensor, an inertial measurement unit sensor, a depth sensor, a camera, an eye tracking sensor, a microphone, and a biometric sensor.
3. The one or more computer storage media of claim 1, wherein the method further comprises:
receiving other data from one or more sources external to the HMD; and
using the other data in conjunction with the sensor data to determine the HMD wearer's head position and rotation relative to the HMD wearer's body and the environment surrounding the HMD wearer.
4. The one or more computer storage media of claim 3, wherein the one or more sources external to the HMD comprise an external camera, an external depth sensor, another HMD, a mobile device, and historical sensor data stored remotely from the HMD.
5. The one or more computer storage media of claim 1, wherein the two or more regions are fixed within the HMD wearer's TFOV.
6. The one or more computer storage media of claim 1, wherein the two or more regions dynamically change within the HMD wearer's TFOV.
7. The one or more computer storage media of claim 1, wherein the two or more regions include a primary region and one or more non-primary regions.
8. The one or more computer storage media of claim 7, locating the virtual images comprises locating the virtual images to reduce occluding the primary region.
9. The one or more computer storage media of claim 1, wherein classifying the two or more regions within the HMD wearer's TFOV comprises classifying at least one of the regions relative to the HMD wearer's body.
10. The one or more computer storage media of claim 9, wherein locating the virtual reality images to be displayed by the HMD comprises locating at least one of the virtual reality images relative to HMD wearer's body based on classifying the at least one of the regions relative to the HMD wearer's body.
11. The one or more computer storage media of claim 1, wherein the method further comprises:
determining a field of view (FOV) for the HMD wearer;
comparing the FOV with the two or more regions of the HMD wearer's TFOV;
determining a subset of virtual objects to display based on a location of the FOV relative to the two or more regions of the HMD wearer's TFOV; and
displaying the subset of virtual objects via a display component of the HMD.
12. The one or more computer storage media of claim 1, wherein the virtual images comprise at least one static virtual image and at least one interactive virtual image.
13. The one or more computer storage media of claim 1, wherein the method further comprises identifying user interaction with an interactive virtual image located in a first region from the two or more regions, the first region being outside the HMD wearer's current field of view, and performing an action in response to the user interaction with the interactive virtual image.
14. A head-mounted display (HMD) comprising:
one or more on-board sensors;
one or more processors configured to: determine a position and rotation of an HMD wearer's head relative to the HMD wearer's body based on sensor data from the one or more on-board sensors, classify two or more regions of the HMD wearer's TFOV based on the position and rotation of the HMD wearer's head relative to the HMD wearer's body, and locate one or more virtual objects in the two or more regions; and
one or more display components configured to display at least one of the one or more virtual objects to provide the HMD wearer an augmented view.
15. The HMD of claim 14, wherein the HMD further comprises a wireless communication component.
16. The HMD of claim 14, wherein the wireless communication component is configured to receive other data from one or more sources external to the HMD, and wherein the one or more processors are further configured to determine the position and rotation of an HMD wearer's head relative to the HMD wearer's body based on the other data from the one or more sources external to the HMD.
17. A method for classifying regions of an head-mounted display (HMD) wearer's total field of view (TFOV), the method comprising:
continuously:
receiving sensor data from one or more sensors on board the HMD;
receiving other data from one or more sources external to the HMD; and
classifying regions of the HMD wearer's TFOV relative to the HMD wearer's body and surrounding environment based on the sensor data and other data.
18. The method of claim 17, wherein the regions include a primary region and one or more non-primary regions.
19. The method of claim 18, wherein the method further comprises locating virtual images in the regions based on classifying the regions.
20. The method of claim 19, wherein locating the virtual images comprises locating the virtual images in the regions to reduce occluding the primary region.
US13/167,113 2011-06-23 2011-06-23 Total field of view classification for head-mounted display Abandoned US20120327116A1 (en)

Priority Applications (9)

Application Number Priority Date Filing Date Title
US13/167,113 US20120327116A1 (en) 2011-06-23 2011-06-23 Total field of view classification for head-mounted display
TW101115609A TW201303640A (en) 2011-06-23 2012-05-02 Total field of view classification for head-mounted display
CN201280030619.6A CN103635849A (en) 2011-06-23 2012-06-19 Total field of view classification for head-mounted display
KR1020137034185A KR20140034252A (en) 2011-06-23 2012-06-19 Total field of view classification for head-mounted display
JP2014517099A JP2014526157A (en) 2011-06-23 2012-06-19 Classification of the total field of view of the head mounted display
PCT/US2012/043178 WO2012177657A2 (en) 2011-06-23 2012-06-19 Total field of view classification for head-mounted display
EP12803004.6A EP2724191A4 (en) 2011-06-23 2012-06-19 Total field of view classification for head-mounted display
US13/692,885 US9041623B2 (en) 2011-06-23 2012-12-03 Total field of view classification for head-mounted display
US14/706,377 US9245501B2 (en) 2011-06-23 2015-05-07 Total field of view classification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/167,113 US20120327116A1 (en) 2011-06-23 2011-06-23 Total field of view classification for head-mounted display

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US13/692,885 Continuation US9041623B2 (en) 2011-06-23 2012-12-03 Total field of view classification for head-mounted display

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US13/692,885 Continuation US9041623B2 (en) 2011-06-23 2012-12-03 Total field of view classification for head-mounted display
US14/706,377 Continuation US9245501B2 (en) 2011-06-23 2015-05-07 Total field of view classification

Publications (1)

Publication Number Publication Date
US20120327116A1 true US20120327116A1 (en) 2012-12-27

Family

ID=47361426

Family Applications (3)

Application Number Title Priority Date Filing Date
US13/167,113 Abandoned US20120327116A1 (en) 2011-06-23 2011-06-23 Total field of view classification for head-mounted display
US13/692,885 Active 2032-01-21 US9041623B2 (en) 2011-06-23 2012-12-03 Total field of view classification for head-mounted display
US14/706,377 Active US9245501B2 (en) 2011-06-23 2015-05-07 Total field of view classification

Family Applications After (2)

Application Number Title Priority Date Filing Date
US13/692,885 Active 2032-01-21 US9041623B2 (en) 2011-06-23 2012-12-03 Total field of view classification for head-mounted display
US14/706,377 Active US9245501B2 (en) 2011-06-23 2015-05-07 Total field of view classification

Country Status (7)

Country Link
US (3) US20120327116A1 (en)
EP (1) EP2724191A4 (en)
JP (1) JP2014526157A (en)
KR (1) KR20140034252A (en)
CN (1) CN103635849A (en)
TW (1) TW201303640A (en)
WO (1) WO2012177657A2 (en)

Cited By (124)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130194164A1 (en) * 2012-01-27 2013-08-01 Ben Sugden Executable virtual objects associated with real objects
US20140118631A1 (en) * 2012-10-29 2014-05-01 Lg Electronics Inc. Head mounted display and method of outputting audio signal using the same
US20140132629A1 (en) * 2012-11-13 2014-05-15 Qualcomm Incorporated Modifying virtual object display properties
WO2014116826A1 (en) * 2013-01-24 2014-07-31 The Trustees Of Columbia University In The City Of New York Mobile, neurally-assisted personal assistant
US20140232620A1 (en) * 2011-10-25 2014-08-21 Olympus Corporation Head mounted display apparatus, information terminal, and methods and information storage devices for controlling head mounted display apparatus and information terminal
US8866849B1 (en) * 2013-08-28 2014-10-21 Lg Electronics Inc. Portable device supporting videotelephony of a head mounted display and method of controlling therefor
WO2014190099A1 (en) * 2013-05-22 2014-11-27 Microsoft Corporation Body-locked placement of augmented reality objects
US8970495B1 (en) * 2012-03-09 2015-03-03 Google Inc. Image stabilization for color-sequential displays
US20150143459A1 (en) * 2013-11-15 2015-05-21 Microsoft Corporation Protecting privacy in web-based immersive augmented reality
WO2015084323A1 (en) 2013-12-03 2015-06-11 Nokia Corporation Display of information on a head mounted display
WO2015108887A1 (en) * 2014-01-17 2015-07-23 Sony Computer Entertainment America Llc Using a second screen as a private tracking heads-up display
US20150213649A1 (en) * 2012-07-27 2015-07-30 Nec Solutions Innovators, Ltd. Three-dimensional environment sharing system and three-dimensional environment sharing method
WO2015123073A1 (en) * 2014-02-12 2015-08-20 Microsoft Technology Licensing, Llc Motion modeling in visual tracking
US20150235452A1 (en) * 2013-03-15 2015-08-20 Magic Leap, Inc. Blanking techniques in augmented or virtual reality systems
US9122054B2 (en) 2014-01-24 2015-09-01 Osterhout Group, Inc. Stray light suppression for head worn computing
US9158116B1 (en) 2014-04-25 2015-10-13 Osterhout Group, Inc. Temple and ear horn assembly for headworn computer
USD743963S1 (en) 2014-12-22 2015-11-24 Osterhout Group, Inc. Air mouse
US9229234B2 (en) 2014-02-11 2016-01-05 Osterhout Group, Inc. Micro doppler presentations in head worn computing
CN105229584A (en) * 2013-05-29 2016-01-06 三菱电机株式会社 Information display device
WO2016014872A1 (en) * 2014-07-25 2016-01-28 Microsoft Technology Licensing, Llc Smart placement of virtual objects to stay in the field of view of a head mounted display
US20160026242A1 (en) 2014-07-25 2016-01-28 Aaron Burns Gaze-based object placement within a virtual reality environment
US9256987B2 (en) 2013-06-24 2016-02-09 Microsoft Technology Licensing, Llc Tracking head movement when wearing mobile device
US9286728B2 (en) 2014-02-11 2016-03-15 Osterhout Group, Inc. Spatial location presentation in head worn computing
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
US9298007B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9299194B2 (en) 2014-02-14 2016-03-29 Osterhout Group, Inc. Secure sharing in head worn computing
US9298001B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Optical configurations for head worn computing
USD753114S1 (en) 2015-01-05 2016-04-05 Osterhout Group, Inc. Air mouse
US9310610B2 (en) 2014-01-21 2016-04-12 Osterhout Group, Inc. See-through computer display systems
US9316833B2 (en) 2014-01-21 2016-04-19 Osterhout Group, Inc. Optical configurations for head worn computing
US9329387B2 (en) 2014-01-21 2016-05-03 Osterhout Group, Inc. See-through computer display systems
US9366868B2 (en) 2014-09-26 2016-06-14 Osterhout Group, Inc. See-through computer display systems
US9366867B2 (en) 2014-07-08 2016-06-14 Osterhout Group, Inc. Optical systems for see-through displays
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9423612B2 (en) 2014-03-28 2016-08-23 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US9423842B2 (en) 2014-09-18 2016-08-23 Osterhout Group, Inc. Thermal management for head-worn computer
US9448409B2 (en) 2014-11-26 2016-09-20 Osterhout Group, Inc. See-through computer display systems
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
WO2016187348A1 (en) * 2015-05-18 2016-11-24 Brian Mullins Biometric authentication in a head mounted device
US9523856B2 (en) 2014-01-21 2016-12-20 Osterhout Group, Inc. See-through computer display systems
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9532715B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9535249B2 (en) * 2012-11-19 2017-01-03 Seiko Epson Corporation Virtual image display apparatus
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US9606992B2 (en) * 2011-09-30 2017-03-28 Microsoft Technology Licensing, Llc Personal audio/visual apparatus providing resource management
US9630098B2 (en) 2013-06-09 2017-04-25 Sony Interactive Entertainment Inc. Head mounted display
US9645397B2 (en) 2014-07-25 2017-05-09 Microsoft Technology Licensing, Llc Use of surface reconstruction data to identify real world floor
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
US9720234B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US9811153B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9826299B1 (en) 2016-08-22 2017-11-21 Osterhout Group, Inc. Speaker systems for head-worn computer systems
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US9846308B2 (en) 2014-01-24 2017-12-19 Osterhout Group, Inc. Haptic systems for head-worn computers
US9852545B2 (en) 2014-02-11 2017-12-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9858720B2 (en) 2014-07-25 2018-01-02 Microsoft Technology Licensing, Llc Three-dimensional mixed-reality viewport
US9865089B2 (en) 2014-07-25 2018-01-09 Microsoft Technology Licensing, Llc Virtual reality environment with real world objects
US9880441B1 (en) 2016-09-08 2018-01-30 Osterhout Group, Inc. Electrochromic systems for head-worn computer systems
US9910284B1 (en) 2016-09-08 2018-03-06 Osterhout Group, Inc. Optical systems for head-worn computers
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US9977241B2 (en) 2015-03-17 2018-05-22 Seiko Epson Corporation Head-mounted display device, control method for head-mounted display device, and computer program
EP3330839A1 (en) * 2016-12-05 2018-06-06 THOMSON Licensing Method and device for adapting an immersive content to the field of view of a user
US10062182B2 (en) 2015-02-17 2018-08-28 Osterhout Group, Inc. See-through computer display systems
US10068374B2 (en) 2013-03-11 2018-09-04 Magic Leap, Inc. Systems and methods for a plurality of users to interact with an augmented or virtual reality systems
US20180253856A1 (en) * 2017-03-01 2018-09-06 Microsoft Technology Licensing, Llc Multi-Spectrum Illumination-and-Sensor Module for Head Tracking, Gesture Recognition and Spatial Mapping
US20180264363A1 (en) * 2017-03-17 2018-09-20 Roblox Corporation Avatar Teleport Controller
US10139966B2 (en) 2015-07-22 2018-11-27 Osterhout Group, Inc. External user interface for head worn computing
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
EP3413165A4 (en) * 2016-02-29 2019-02-06 Huawei Technologies Co., Ltd. Wearable system gesture control method and wearable system
USD840395S1 (en) 2016-10-17 2019-02-12 Osterhout Group, Inc. Head-worn computer
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US10311638B2 (en) 2014-07-25 2019-06-04 Microsoft Technology Licensing, Llc Anti-trip when immersed in a virtual reality environment
US10422995B2 (en) 2017-07-24 2019-09-24 Mentor Acquisition One, Llc See-through computer display systems with stray light management
US10451875B2 (en) 2014-07-25 2019-10-22 Microsoft Technology Licensing, Llc Smart transparency for virtual objects
USD864959S1 (en) 2017-01-04 2019-10-29 Mentor Acquisition One, Llc Computer glasses
US10466491B2 (en) 2016-06-01 2019-11-05 Mentor Acquisition One, Llc Modular systems for head-worn computers
US10466492B2 (en) 2014-04-25 2019-11-05 Mentor Acquisition One, Llc Ear horn assembly for headworn computer
EP3590028A4 (en) * 2017-06-01 2020-01-22 Samsung Electronics Co., Ltd. Systems and methods for window control in virtual reality environment
US10551920B2 (en) 2014-03-21 2020-02-04 Samsung Electronics Co., Ltd. Wearable device and method of operating the same
US20200043233A1 (en) * 2018-08-03 2020-02-06 Igt Providing interactive virtual elements within a mixed reality scene
US10578869B2 (en) 2017-07-24 2020-03-03 Mentor Acquisition One, Llc See-through computer display systems with adjustable zoom cameras
US10591728B2 (en) 2016-03-02 2020-03-17 Mentor Acquisition One, Llc Optical systems for head-worn computers
EP3594734A4 (en) * 2017-03-24 2020-04-01 Samsung Electronics Co., Ltd. Electronic device for playing content and computer-readable recording medium
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US10667981B2 (en) 2016-02-29 2020-06-02 Mentor Acquisition One, Llc Reading assistance system for visually impaired
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US10684478B2 (en) 2016-05-09 2020-06-16 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10690936B2 (en) 2016-08-29 2020-06-23 Mentor Acquisition One, Llc Adjustable nose bridge assembly for headworn computer
US10824253B2 (en) 2016-05-09 2020-11-03 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10850116B2 (en) 2016-12-30 2020-12-01 Mentor Acquisition One, Llc Head-worn therapy device
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US10878775B2 (en) 2015-02-17 2020-12-29 Mentor Acquisition One, Llc See-through computer display systems
US10936053B2 (en) * 2017-03-29 2021-03-02 Boe Technology Group Co., Ltd. Interaction system of three-dimensional space and method for operating same
US10969584B2 (en) 2017-08-04 2021-04-06 Mentor Acquisition One, Llc Image expansion optic for head-worn computer
US11009942B2 (en) * 2017-06-28 2021-05-18 Shandong University Multi-human tracking system and method with single kinect for supporting mobile virtual reality application
US11104272B2 (en) 2014-03-28 2021-08-31 Mentor Acquisition One, Llc System for assisted operator safety using an HMD
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US11119581B2 (en) 2017-06-15 2021-09-14 Microsoft Technology Licensing, Llc Displacement oriented interaction in computer-mediated reality
US11170565B2 (en) 2018-08-31 2021-11-09 Magic Leap, Inc. Spatially-resolved dynamic dimming for augmented reality device
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US11269182B2 (en) 2014-07-15 2022-03-08 Mentor Acquisition One, Llc Content presentation in head worn computing
US11409105B2 (en) 2017-07-24 2022-08-09 Mentor Acquisition One, Llc See-through computer display systems
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
CN116740318A (en) * 2023-08-14 2023-09-12 深圳风向标教育资源股份有限公司 Method, device and storage medium for placing automobile parts
US11808940B2 (en) 2019-10-10 2023-11-07 Medithinq Co., Ltd. Eye tracking system for smart glasses and method therefor
US11851177B2 (en) 2014-05-06 2023-12-26 Mentor Acquisition One, Llc Unmanned aerial vehicle launch system
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US11960095B2 (en) 2023-04-19 2024-04-16 Mentor Acquisition One, Llc See-through computer display systems

Families Citing this family (63)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9727132B2 (en) * 2011-07-01 2017-08-08 Microsoft Technology Licensing, Llc Multi-visor: managing applications in augmented reality environments
JP6287849B2 (en) * 2013-02-22 2018-03-07 ソニー株式会社 Head mounted display, image display device, and image display method
US10195058B2 (en) 2013-05-13 2019-02-05 The Johns Hopkins University Hybrid augmented reality multimodal operation neural integration environment
US9063330B2 (en) * 2013-05-30 2015-06-23 Oculus Vr, Llc Perception based predictive tracking for head mounted displays
US10137361B2 (en) 2013-06-07 2018-11-27 Sony Interactive Entertainment America Llc Systems and methods for using reduced hops to generate an augmented virtual reality scene within a head mounted system
US10905943B2 (en) * 2013-06-07 2021-02-02 Sony Interactive Entertainment LLC Systems and methods for reducing hops associated with a head mounted system
CN103353663B (en) 2013-06-28 2016-08-10 北京智谷睿拓技术服务有限公司 Imaging adjusting apparatus and method
CN103353667B (en) 2013-06-28 2015-10-21 北京智谷睿拓技术服务有限公司 Imaging adjustment Apparatus and method for
CN103353677B (en) 2013-06-28 2015-03-11 北京智谷睿拓技术服务有限公司 Imaging device and method thereof
JP2015015563A (en) * 2013-07-04 2015-01-22 セイコーエプソン株式会社 Image display device
CN103424891B (en) 2013-07-31 2014-12-17 北京智谷睿拓技术服务有限公司 Imaging device and method
CN103431840B (en) 2013-07-31 2016-01-20 北京智谷睿拓技术服务有限公司 Eye optical parameter detecting system and method
CN103431980A (en) 2013-08-22 2013-12-11 北京智谷睿拓技术服务有限公司 Eyesight protection imaging system and method
CN103439801B (en) 2013-08-22 2016-10-26 北京智谷睿拓技术服务有限公司 Sight protectio imaging device and method
CN103605208B (en) 2013-08-30 2016-09-28 北京智谷睿拓技术服务有限公司 content projection system and method
CN103500331B (en) 2013-08-30 2017-11-10 北京智谷睿拓技术服务有限公司 Based reminding method and device
US9448689B2 (en) * 2013-08-30 2016-09-20 Paypal, Inc. Wearable user device enhanced display system
CN103558909B (en) * 2013-10-10 2017-03-29 北京智谷睿拓技术服务有限公司 Interaction projection display packing and interaction projection display system
US9996221B2 (en) * 2013-12-01 2018-06-12 Upskill, Inc. Systems and methods for look-initiated communication
WO2015099215A1 (en) * 2013-12-24 2015-07-02 엘지전자 주식회사 Head-mounted display apparatus and method for operating same
TWI486631B (en) * 2014-01-24 2015-06-01 Quanta Comp Inc Head mounted display and control method thereof
JP6376807B2 (en) * 2014-04-02 2018-08-22 キヤノン株式会社 Display device, display control method, and program
US9959591B2 (en) * 2014-07-31 2018-05-01 Seiko Epson Corporation Display apparatus, method for controlling display apparatus, and program
KR102243656B1 (en) * 2014-09-26 2021-04-23 엘지전자 주식회사 Mobile device, head mounted display and system
KR102358548B1 (en) * 2014-10-15 2022-02-04 삼성전자주식회사 Method and appratus for processing screen using device
CN112530025A (en) * 2014-12-18 2021-03-19 脸谱科技有限责任公司 System, apparatus and method for providing a user interface for a virtual reality environment
KR102144515B1 (en) 2015-01-07 2020-08-14 삼성전자주식회사 Master device, slave device and control method thereof
US9911232B2 (en) 2015-02-27 2018-03-06 Microsoft Technology Licensing, Llc Molding and anchoring physically constrained virtual environments to real-world environments
WO2016141373A1 (en) 2015-03-05 2016-09-09 Magic Leap, Inc. Systems and methods for augmented reality
US10838207B2 (en) 2015-03-05 2020-11-17 Magic Leap, Inc. Systems and methods for augmented reality
US9836117B2 (en) 2015-05-28 2017-12-05 Microsoft Technology Licensing, Llc Autonomous drones for tactile feedback in immersive virtual reality
US9898864B2 (en) 2015-05-28 2018-02-20 Microsoft Technology Licensing, Llc Shared tactile interaction and user safety in shared space multi-person immersive virtual reality
KR102410449B1 (en) * 2015-06-30 2022-06-16 매직 립, 인코포레이티드 Techniques for more efficient display of text in virtual imaging systems
KR102402048B1 (en) 2015-10-14 2022-05-26 삼성전자주식회사 Electronic apparatus and the controlling method thereof
US9864194B2 (en) 2015-11-24 2018-01-09 Honeywell International Inc. Systems and methods for displaying FOV boundaries on HUDs
AU2016365422A1 (en) 2015-12-04 2018-06-28 Magic Leap, Inc. Relocalization systems and methods
US10147235B2 (en) 2015-12-10 2018-12-04 Microsoft Technology Licensing, Llc AR display with adjustable stereo overlap zone
US10134188B2 (en) * 2015-12-21 2018-11-20 Intel Corporation Body-centric mobile point-of-view augmented and virtual reality
KR102610120B1 (en) 2016-01-20 2023-12-06 삼성전자주식회사 Head mounted display and control method thereof
KR102524641B1 (en) 2016-01-22 2023-04-21 삼성전자주식회사 Head mounted display device and method for controlling the same
US10518172B2 (en) * 2016-03-07 2019-12-31 Htc Corporation Accessory management of virtual reality system
US10133345B2 (en) * 2016-03-22 2018-11-20 Microsoft Technology Licensing, Llc Virtual-reality navigation
CN105912123A (en) 2016-04-15 2016-08-31 北京小鸟看看科技有限公司 Interface layout method and device under three-dimension immersion environment
TWI628634B (en) * 2016-05-25 2018-07-01 國立中央大學 Interactive teaching systems and methods thereof
CN109314800B (en) * 2016-06-13 2022-02-01 索尼互动娱乐有限责任公司 Method and system for directing user attention to location-based game play companion application
CA3032567A1 (en) * 2016-08-02 2018-02-08 Magic Leap, Inc. Fixed-distance virtual and augmented reality systems and methods
US10864423B2 (en) * 2016-11-10 2020-12-15 National Taiwan University Augmented learning system for tai-chi chuan with head-mounted display
US10812936B2 (en) 2017-01-23 2020-10-20 Magic Leap, Inc. Localization determination for mixed reality systems
US10769752B2 (en) 2017-03-17 2020-09-08 Magic Leap, Inc. Mixed reality system with virtual content warping and method of generating virtual content using same
US10861237B2 (en) 2017-03-17 2020-12-08 Magic Leap, Inc. Mixed reality system with multi-source virtual content compositing and method of generating virtual content using same
AU2018234921B2 (en) 2017-03-17 2021-10-07 Magic Leap, Inc. Mixed reality system with color virtual content warping and method of generating virtual content using same
CN106997242B (en) * 2017-03-28 2020-10-30 联想(北京)有限公司 Interface management method and head-mounted display device
US11158126B1 (en) 2017-06-30 2021-10-26 Apple Inc. Redirected walking in virtual reality environments
TWI658291B (en) * 2018-01-25 2019-05-01 宏碁股份有限公司 Head-mounted display and operation method thereof
CN108198249B (en) * 2018-03-07 2024-02-02 上海交通大学医学院附属第九人民医院 Visual field defect simulation device and method based on virtual reality technology
US10572002B2 (en) * 2018-03-13 2020-02-25 Facebook Technologies, Llc Distributed artificial reality system with contextualized hand tracking
WO2020023383A1 (en) 2018-07-23 2020-01-30 Magic Leap, Inc. Mixed reality system with virtual content warping and method of generating virtual content using same
US11347303B2 (en) 2018-11-30 2022-05-31 Sony Interactive Entertainment Inc. Systems and methods for determining movement of a controller with respect to an HMD
US10802281B2 (en) 2018-12-20 2020-10-13 General Electric Company Periodic lenses systems for augmented reality
KR102174794B1 (en) * 2019-01-31 2020-11-05 주식회사 알파서클 Method and device for controlling transit time of virtual reality video
US20210192799A1 (en) * 2019-12-19 2021-06-24 Facebook Technologies, Llc Passthrough window object locator in an artificial reality system
KR20220039113A (en) * 2020-09-21 2022-03-29 삼성전자주식회사 Method and apparatus for transmitting video content using edge computing service
TW202301083A (en) 2021-06-28 2023-01-01 見臻科技股份有限公司 Optical system providing accurate eye-tracking and related method

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1995005620A1 (en) * 1993-08-12 1995-02-23 Seiko Epson Corporation Head mount type image display device and information processor equipped with the device
JP3701053B2 (en) * 1995-07-13 2005-09-28 コニカミノルタホールディングス株式会社 Video display device
JPH1184307A (en) * 1997-09-01 1999-03-26 M R Syst Kenkyusho:Kk Head-mounted optical device
AUPQ896000A0 (en) * 2000-07-24 2000-08-17 Seeing Machines Pty Ltd Facial image processing system
US20020044152A1 (en) 2000-10-16 2002-04-18 Abbott Kenneth H. Dynamic integration of computer generated and real world images
DE10131720B4 (en) * 2001-06-30 2017-02-23 Robert Bosch Gmbh Head-Up Display System and Procedures
JP2004234253A (en) * 2003-01-29 2004-08-19 Canon Inc Method for presenting composite sense of reality
US7561966B2 (en) * 2003-12-17 2009-07-14 Denso Corporation Vehicle information display system
CN1922651A (en) * 2004-06-10 2007-02-28 松下电器产业株式会社 Wearable type information presentation device
SE527257C2 (en) * 2004-06-21 2006-01-31 Totalfoersvarets Forskningsins Device and method for presenting an external image
US7369317B2 (en) 2005-03-07 2008-05-06 Himax Technologies, Inc. Head-mounted display utilizing an LCOS panel with a color filter attached thereon
HU0500357D0 (en) * 2005-04-04 2005-05-30 Innoracio Fejlesztoe Es Kutata Dynamic display and method for enhancing the effective resolution of displays
JP2008217119A (en) * 2007-02-28 2008-09-18 Canon Inc System, image processor and image processing method
JP5067850B2 (en) * 2007-08-02 2012-11-07 キヤノン株式会社 System, head-mounted display device, and control method thereof
IL196078A (en) * 2007-12-20 2014-09-30 Raytheon Co Imaging system
US9600067B2 (en) 2008-10-27 2017-03-21 Sri International System and method for generating a mixed reality environment
US20100182340A1 (en) * 2009-01-19 2010-07-22 Bachelder Edward N Systems and methods for combining virtual and real-time physical environments

Cited By (331)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11506912B2 (en) 2008-01-02 2022-11-22 Mentor Acquisition One, Llc Temple and ear horn assembly for headworn computer
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US9606992B2 (en) * 2011-09-30 2017-03-28 Microsoft Technology Licensing, Llc Personal audio/visual apparatus providing resource management
US9746671B2 (en) * 2011-10-25 2017-08-29 Olympus Corporation Head mounted display apparatus, information terminal, and methods and information storage devices for controlling head mounted display apparatus and information terminal
US20140232620A1 (en) * 2011-10-25 2014-08-21 Olympus Corporation Head mounted display apparatus, information terminal, and methods and information storage devices for controlling head mounted display apparatus and information terminal
US9594537B2 (en) 2012-01-27 2017-03-14 Microsoft Technology Licensing, Llc Executable virtual objects associated with real objects
US9836889B2 (en) 2012-01-27 2017-12-05 Microsoft Technology Licensing, Llc Executable virtual objects associated with real objects
US8963805B2 (en) * 2012-01-27 2015-02-24 Microsoft Corporation Executable virtual objects associated with real objects
US9201243B2 (en) 2012-01-27 2015-12-01 Microsoft Technology Licensing, Llc Executable virtual objects associated with real objects
US20130194164A1 (en) * 2012-01-27 2013-08-01 Ben Sugden Executable virtual objects associated with real objects
US8970495B1 (en) * 2012-03-09 2015-03-03 Google Inc. Image stabilization for color-sequential displays
US20150213649A1 (en) * 2012-07-27 2015-07-30 Nec Solutions Innovators, Ltd. Three-dimensional environment sharing system and three-dimensional environment sharing method
US20140118631A1 (en) * 2012-10-29 2014-05-01 Lg Electronics Inc. Head mounted display and method of outputting audio signal using the same
US9374549B2 (en) * 2012-10-29 2016-06-21 Lg Electronics Inc. Head mounted display and method of outputting audio signal using the same
US9727996B2 (en) 2012-11-13 2017-08-08 Qualcomm Incorporated Modifying virtual object display properties to increase power performance of augmented reality devices
US9619911B2 (en) * 2012-11-13 2017-04-11 Qualcomm Incorporated Modifying virtual object display properties
US9448404B2 (en) 2012-11-13 2016-09-20 Qualcomm Incorporated Modifying virtual object display properties to increase power performance of augmented reality devices
US20140132629A1 (en) * 2012-11-13 2014-05-15 Qualcomm Incorporated Modifying virtual object display properties
US9535249B2 (en) * 2012-11-19 2017-01-03 Seiko Epson Corporation Virtual image display apparatus
WO2014116826A1 (en) * 2013-01-24 2014-07-31 The Trustees Of Columbia University In The City Of New York Mobile, neurally-assisted personal assistant
US11663789B2 (en) 2013-03-11 2023-05-30 Magic Leap, Inc. Recognizing objects in a passable world model in augmented or virtual reality systems
US10068374B2 (en) 2013-03-11 2018-09-04 Magic Leap, Inc. Systems and methods for a plurality of users to interact with an augmented or virtual reality systems
US10126812B2 (en) 2013-03-11 2018-11-13 Magic Leap, Inc. Interacting with a network to transmit virtual image data in augmented or virtual reality systems
US10163265B2 (en) 2013-03-11 2018-12-25 Magic Leap, Inc. Selective light transmission for augmented or virtual reality
US10282907B2 (en) 2013-03-11 2019-05-07 Magic Leap, Inc Interacting with a network to transmit virtual image data in augmented or virtual reality systems
US11087555B2 (en) 2013-03-11 2021-08-10 Magic Leap, Inc. Recognizing objects in a passable world model in augmented or virtual reality systems
US10234939B2 (en) 2013-03-11 2019-03-19 Magic Leap, Inc. Systems and methods for a plurality of users to interact with each other in augmented or virtual reality systems
US10629003B2 (en) 2013-03-11 2020-04-21 Magic Leap, Inc. System and method for augmented and virtual reality
US20150235452A1 (en) * 2013-03-15 2015-08-20 Magic Leap, Inc. Blanking techniques in augmented or virtual reality systems
US10134186B2 (en) 2013-03-15 2018-11-20 Magic Leap, Inc. Predicting head movement for rendering virtual objects in augmented or virtual reality systems
US10510188B2 (en) 2013-03-15 2019-12-17 Magic Leap, Inc. Over-rendering techniques in augmented or virtual reality systems
US10453258B2 (en) 2013-03-15 2019-10-22 Magic Leap, Inc. Adjusting pixels to compensate for spacing in augmented or virtual reality systems
US10304246B2 (en) * 2013-03-15 2019-05-28 Magic Leap, Inc. Blanking techniques in augmented or virtual reality systems
US10553028B2 (en) 2013-03-15 2020-02-04 Magic Leap, Inc. Presenting virtual objects based on head movements in augmented or virtual reality systems
US11854150B2 (en) 2013-03-15 2023-12-26 Magic Leap, Inc. Frame-by-frame rendering for augmented or virtual reality systems
US11205303B2 (en) 2013-03-15 2021-12-21 Magic Leap, Inc. Frame-by-frame rendering for augmented or virtual reality systems
WO2014190099A1 (en) * 2013-05-22 2014-11-27 Microsoft Corporation Body-locked placement of augmented reality objects
US9367960B2 (en) 2013-05-22 2016-06-14 Microsoft Technology Licensing, Llc Body-locked placement of augmented reality objects
CN105229584A (en) * 2013-05-29 2016-01-06 三菱电机株式会社 Information display device
US10987574B2 (en) 2013-06-09 2021-04-27 Sony Interactive Entertainment Inc. Head mounted display
US10173129B2 (en) 2013-06-09 2019-01-08 Sony Interactive Entertainment Inc. Methods for rendering interactive content to a head mounted display
US10525335B2 (en) 2013-06-09 2020-01-07 Sony Interactive Entertainment Inc. Head mounted display
US9630098B2 (en) 2013-06-09 2017-04-25 Sony Interactive Entertainment Inc. Head mounted display
US9256987B2 (en) 2013-06-24 2016-02-09 Microsoft Technology Licensing, Llc Tracking head movement when wearing mobile device
US8866849B1 (en) * 2013-08-28 2014-10-21 Lg Electronics Inc. Portable device supporting videotelephony of a head mounted display and method of controlling therefor
US20150143459A1 (en) * 2013-11-15 2015-05-21 Microsoft Corporation Protecting privacy in web-based immersive augmented reality
US9679144B2 (en) * 2013-11-15 2017-06-13 Microsoft Technology Licensing, Llc Protecting privacy in web-based immersive augmented reality
WO2015084323A1 (en) 2013-12-03 2015-06-11 Nokia Corporation Display of information on a head mounted display
US10386921B2 (en) 2013-12-03 2019-08-20 Nokia Technologies Oy Display of information on a head mounted display
EP3078019A4 (en) * 2013-12-03 2017-06-14 Nokia Technologies Oy Display of information on a head mounted display
US10001645B2 (en) 2014-01-17 2018-06-19 Sony Interactive Entertainment America Llc Using a second screen as a private tracking heads-up display
WO2015108887A1 (en) * 2014-01-17 2015-07-23 Sony Computer Entertainment America Llc Using a second screen as a private tracking heads-up display
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US11782529B2 (en) 2014-01-17 2023-10-10 Mentor Acquisition One, Llc External user interface for head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US11507208B2 (en) 2014-01-17 2022-11-22 Mentor Acquisition One, Llc External user interface for head worn computing
US11231817B2 (en) 2014-01-17 2022-01-25 Mentor Acquisition One, Llc External user interface for head worn computing
US11169623B2 (en) 2014-01-17 2021-11-09 Mentor Acquisition One, Llc External user interface for head worn computing
US9720234B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9927612B2 (en) 2014-01-21 2018-03-27 Osterhout Group, Inc. See-through computer display systems
US9594246B2 (en) 2014-01-21 2017-03-14 Osterhout Group, Inc. See-through computer display systems
US9298001B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Optical configurations for head worn computing
US9538915B2 (en) 2014-01-21 2017-01-10 Osterhout Group, Inc. Eye imaging in head worn computing
US9615742B2 (en) 2014-01-21 2017-04-11 Osterhout Group, Inc. Eye imaging in head worn computing
US9532714B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US11947126B2 (en) 2014-01-21 2024-04-02 Mentor Acquisition One, Llc See-through computer display systems
US9651789B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-Through computer display systems
US9298002B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Optical configurations for head worn computing
US9651788B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651783B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US9658457B2 (en) 2014-01-21 2017-05-23 Osterhout Group, Inc. See-through computer display systems
US9658458B2 (en) 2014-01-21 2017-05-23 Osterhout Group, Inc. See-through computer display systems
US10481393B2 (en) 2014-01-21 2019-11-19 Mentor Acquisition One, Llc See-through computer display systems
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US9532715B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9529199B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US10579140B2 (en) 2014-01-21 2020-03-03 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9684171B2 (en) 2014-01-21 2017-06-20 Osterhout Group, Inc. See-through computer display systems
US9684165B2 (en) 2014-01-21 2017-06-20 Osterhout Group, Inc. Eye imaging in head worn computing
US9298007B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
US10698223B2 (en) 2014-01-21 2020-06-30 Mentor Acquisition One, Llc See-through computer display systems
US9529192B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. Eye imaging in head worn computing
US9720227B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9720235B2 (en) 2014-01-21 2017-08-01 Osterhout Group, Inc. See-through computer display systems
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US10705339B2 (en) 2014-01-21 2020-07-07 Mentor Acquisition One, Llc Suppression of stray light in head worn computing
US9740012B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. See-through computer display systems
US9740280B2 (en) 2014-01-21 2017-08-22 Osterhout Group, Inc. Eye imaging in head worn computing
US9523856B2 (en) 2014-01-21 2016-12-20 Osterhout Group, Inc. See-through computer display systems
US9746676B2 (en) 2014-01-21 2017-08-29 Osterhout Group, Inc. See-through computer display systems
US9310610B2 (en) 2014-01-21 2016-04-12 Osterhout Group, Inc. See-through computer display systems
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US10379365B2 (en) 2014-01-21 2019-08-13 Mentor Acquisition One, Llc See-through computer display systems
US9316833B2 (en) 2014-01-21 2016-04-19 Osterhout Group, Inc. Optical configurations for head worn computing
US9772492B2 (en) 2014-01-21 2017-09-26 Osterhout Group, Inc. Eye imaging in head worn computing
US9329387B2 (en) 2014-01-21 2016-05-03 Osterhout Group, Inc. See-through computer display systems
US11796805B2 (en) 2014-01-21 2023-10-24 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11796799B2 (en) 2014-01-21 2023-10-24 Mentor Acquisition One, Llc See-through computer display systems
US9811153B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9811159B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9811152B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US10866420B2 (en) 2014-01-21 2020-12-15 Mentor Acquisition One, Llc See-through computer display systems
US10890760B2 (en) 2014-01-21 2021-01-12 Mentor Acquisition One, Llc See-through computer display systems
US9829703B2 (en) 2014-01-21 2017-11-28 Osterhout Group, Inc. Eye imaging in head worn computing
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US10222618B2 (en) 2014-01-21 2019-03-05 Osterhout Group, Inc. Compact optics with reduced chromatic aberrations
US11002961B2 (en) 2014-01-21 2021-05-11 Mentor Acquisition One, Llc See-through computer display systems
US11054902B2 (en) 2014-01-21 2021-07-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US10191284B2 (en) 2014-01-21 2019-01-29 Osterhout Group, Inc. See-through computer display systems
US9377625B2 (en) 2014-01-21 2016-06-28 Osterhout Group, Inc. Optical configurations for head worn computing
US11099380B2 (en) 2014-01-21 2021-08-24 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11719934B2 (en) 2014-01-21 2023-08-08 Mentor Acquisition One, Llc Suppression of stray light in head worn computing
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9885868B2 (en) 2014-01-21 2018-02-06 Osterhout Group, Inc. Eye imaging in head worn computing
US10139632B2 (en) 2014-01-21 2018-11-27 Osterhout Group, Inc. See-through computer display systems
US11103132B2 (en) 2014-01-21 2021-08-31 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11650416B2 (en) 2014-01-21 2023-05-16 Mentor Acquisition One, Llc See-through computer display systems
US11126003B2 (en) 2014-01-21 2021-09-21 Mentor Acquisition One, Llc See-through computer display systems
US10073266B2 (en) 2014-01-21 2018-09-11 Osterhout Group, Inc. See-through computer display systems
US9933622B2 (en) 2014-01-21 2018-04-03 Osterhout Group, Inc. See-through computer display systems
US11353957B2 (en) 2014-01-21 2022-06-07 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9958674B2 (en) 2014-01-21 2018-05-01 Osterhout Group, Inc. Eye imaging in head worn computing
US11622426B2 (en) 2014-01-21 2023-04-04 Mentor Acquisition One, Llc See-through computer display systems
US9971156B2 (en) 2014-01-21 2018-05-15 Osterhout Group, Inc. See-through computer display systems
US11619820B2 (en) 2014-01-21 2023-04-04 Mentor Acquisition One, Llc See-through computer display systems
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US10001644B2 (en) 2014-01-21 2018-06-19 Osterhout Group, Inc. See-through computer display systems
US9436006B2 (en) 2014-01-21 2016-09-06 Osterhout Group, Inc. See-through computer display systems
US10007118B2 (en) 2014-01-21 2018-06-26 Osterhout Group, Inc. Compact optical system with improved illumination
US10012840B2 (en) 2014-01-21 2018-07-03 Osterhout Group, Inc. See-through computer display systems
US10012838B2 (en) 2014-01-21 2018-07-03 Osterhout Group, Inc. Compact optical system with improved contrast uniformity
US11822090B2 (en) 2014-01-24 2023-11-21 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US9846308B2 (en) 2014-01-24 2017-12-19 Osterhout Group, Inc. Haptic systems for head-worn computers
US10558050B2 (en) 2014-01-24 2020-02-11 Mentor Acquisition One, Llc Haptic systems for head-worn computers
US9400390B2 (en) 2014-01-24 2016-07-26 Osterhout Group, Inc. Peripheral lighting for head worn computing
US9939646B2 (en) 2014-01-24 2018-04-10 Osterhout Group, Inc. Stray light suppression for head worn computing
US10578874B2 (en) 2014-01-24 2020-03-03 Mentor Acquisition One, Llc Stray light suppression for head worn computing
US9122054B2 (en) 2014-01-24 2015-09-01 Osterhout Group, Inc. Stray light suppression for head worn computing
US11782274B2 (en) 2014-01-24 2023-10-10 Mentor Acquisition One, Llc Stray light suppression for head worn computing
US9784973B2 (en) 2014-02-11 2017-10-10 Osterhout Group, Inc. Micro doppler presentations in head worn computing
US9286728B2 (en) 2014-02-11 2016-03-15 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9229234B2 (en) 2014-02-11 2016-01-05 Osterhout Group, Inc. Micro doppler presentations in head worn computing
US9229233B2 (en) 2014-02-11 2016-01-05 Osterhout Group, Inc. Micro Doppler presentations in head worn computing
US9841602B2 (en) 2014-02-11 2017-12-12 Osterhout Group, Inc. Location indicating avatar in head worn computing
US9843093B2 (en) 2014-02-11 2017-12-12 Osterhout Group, Inc. Spatial location presentation in head worn computing
US10558420B2 (en) 2014-02-11 2020-02-11 Mentor Acquisition One, Llc Spatial location presentation in head worn computing
US11599326B2 (en) 2014-02-11 2023-03-07 Mentor Acquisition One, Llc Spatial location presentation in head worn computing
US9852545B2 (en) 2014-02-11 2017-12-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
US9767609B2 (en) 2014-02-12 2017-09-19 Microsoft Technology Licensing, Llc Motion modeling in visual tracking
WO2015123073A1 (en) * 2014-02-12 2015-08-20 Microsoft Technology Licensing, Llc Motion modeling in visual tracking
US10140079B2 (en) 2014-02-14 2018-11-27 Osterhout Group, Inc. Object shadowing in head worn computing
US9547465B2 (en) 2014-02-14 2017-01-17 Osterhout Group, Inc. Object shadowing in head worn computing
US9299194B2 (en) 2014-02-14 2016-03-29 Osterhout Group, Inc. Secure sharing in head worn computing
US9928019B2 (en) 2014-02-14 2018-03-27 Osterhout Group, Inc. Object shadowing in head worn computing
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US10551920B2 (en) 2014-03-21 2020-02-04 Samsung Electronics Co., Ltd. Wearable device and method of operating the same
US10955919B2 (en) 2014-03-21 2021-03-23 Samsung Electronics Co., Ltd. Wearable device and method of operating the same
US9423612B2 (en) 2014-03-28 2016-08-23 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US11104272B2 (en) 2014-03-28 2021-08-31 Mentor Acquisition One, Llc System for assisted operator safety using an HMD
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US11809022B2 (en) 2014-04-25 2023-11-07 Mentor Acquisition One, Llc Temple and ear horn assembly for headworn computer
US11727223B2 (en) 2014-04-25 2023-08-15 Mentor Acquisition One, Llc Language translation with head-worn computing
US10732434B2 (en) 2014-04-25 2020-08-04 Mentor Acquisition One, Llc Temple and ear horn assembly for headworn computer
US9158116B1 (en) 2014-04-25 2015-10-13 Osterhout Group, Inc. Temple and ear horn assembly for headworn computer
US11474360B2 (en) 2014-04-25 2022-10-18 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US10634922B2 (en) 2014-04-25 2020-04-28 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US10146772B2 (en) 2014-04-25 2018-12-04 Osterhout Group, Inc. Language translation with head-worn computing
US11880041B2 (en) 2014-04-25 2024-01-23 Mentor Acquisition One, Llc Speaker assembly for headworn computer
US10466492B2 (en) 2014-04-25 2019-11-05 Mentor Acquisition One, Llc Ear horn assembly for headworn computer
US9897822B2 (en) 2014-04-25 2018-02-20 Osterhout Group, Inc. Temple and ear horn assembly for headworn computer
US10101588B2 (en) 2014-04-25 2018-10-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US11851177B2 (en) 2014-05-06 2023-12-26 Mentor Acquisition One, Llc Unmanned aerial vehicle launch system
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US11402639B2 (en) 2014-06-05 2022-08-02 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US10877270B2 (en) 2014-06-05 2020-12-29 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11327323B2 (en) 2014-06-09 2022-05-10 Mentor Acquisition One, Llc Content presentation in head worn computing
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US11022810B2 (en) 2014-06-09 2021-06-01 Mentor Acquisition One, Llc Content presentation in head worn computing
US10139635B2 (en) 2014-06-09 2018-11-27 Osterhout Group, Inc. Content presentation in head worn computing
US10976559B2 (en) 2014-06-09 2021-04-13 Mentor Acquisition One, Llc Content presentation in head worn computing
US11663794B2 (en) 2014-06-09 2023-05-30 Mentor Acquisition One, Llc Content presentation in head worn computing
US11790617B2 (en) 2014-06-09 2023-10-17 Mentor Acquisition One, Llc Content presentation in head worn computing
US11887265B2 (en) 2014-06-09 2024-01-30 Mentor Acquisition One, Llc Content presentation in head worn computing
US11360318B2 (en) 2014-06-09 2022-06-14 Mentor Acquisition One, Llc Content presentation in head worn computing
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
US9720241B2 (en) 2014-06-09 2017-08-01 Osterhout Group, Inc. Content presentation in head worn computing
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US11789267B2 (en) 2014-06-17 2023-10-17 Mentor Acquisition One, Llc External user interface for head worn computing
US10698212B2 (en) 2014-06-17 2020-06-30 Mentor Acquisition One, Llc External user interface for head worn computing
US11054645B2 (en) 2014-06-17 2021-07-06 Mentor Acquisition One, Llc External user interface for head worn computing
US11294180B2 (en) 2014-06-17 2022-04-05 Mentor Acquisition One, Llc External user interface for head worn computing
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US11940629B2 (en) 2014-07-08 2024-03-26 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US9366867B2 (en) 2014-07-08 2016-06-14 Osterhout Group, Inc. Optical systems for see-through displays
US9798148B2 (en) 2014-07-08 2017-10-24 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US10564426B2 (en) 2014-07-08 2020-02-18 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11409110B2 (en) 2014-07-08 2022-08-09 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US10775630B2 (en) 2014-07-08 2020-09-15 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11786105B2 (en) 2014-07-15 2023-10-17 Mentor Acquisition One, Llc Content presentation in head worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US11269182B2 (en) 2014-07-15 2022-03-08 Mentor Acquisition One, Llc Content presentation in head worn computing
US9645397B2 (en) 2014-07-25 2017-05-09 Microsoft Technology Licensing, Llc Use of surface reconstruction data to identify real world floor
US10649212B2 (en) 2014-07-25 2020-05-12 Microsoft Technology Licensing Llc Ground plane adjustment in a virtual reality environment
US9766460B2 (en) 2014-07-25 2017-09-19 Microsoft Technology Licensing, Llc Ground plane adjustment in a virtual reality environment
US9904055B2 (en) 2014-07-25 2018-02-27 Microsoft Technology Licensing, Llc Smart placement of virtual objects to stay in the field of view of a head mounted display
US10311638B2 (en) 2014-07-25 2019-06-04 Microsoft Technology Licensing, Llc Anti-trip when immersed in a virtual reality environment
US10451875B2 (en) 2014-07-25 2019-10-22 Microsoft Technology Licensing, Llc Smart transparency for virtual objects
US9858720B2 (en) 2014-07-25 2018-01-02 Microsoft Technology Licensing, Llc Three-dimensional mixed-reality viewport
US20160026242A1 (en) 2014-07-25 2016-01-28 Aaron Burns Gaze-based object placement within a virtual reality environment
US10416760B2 (en) 2014-07-25 2019-09-17 Microsoft Technology Licensing, Llc Gaze-based object placement within a virtual reality environment
WO2016014872A1 (en) * 2014-07-25 2016-01-28 Microsoft Technology Licensing, Llc Smart placement of virtual objects to stay in the field of view of a head mounted display
US10096168B2 (en) 2014-07-25 2018-10-09 Microsoft Technology Licensing, Llc Three-dimensional mixed-reality viewport
US9865089B2 (en) 2014-07-25 2018-01-09 Microsoft Technology Licensing, Llc Virtual reality environment with real world objects
US11360314B2 (en) 2014-08-12 2022-06-14 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US10908422B2 (en) 2014-08-12 2021-02-02 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US11630315B2 (en) 2014-08-12 2023-04-18 Mentor Acquisition One, Llc Measuring content brightness in head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US9423842B2 (en) 2014-09-18 2016-08-23 Osterhout Group, Inc. Thermal management for head-worn computer
US11474575B2 (en) 2014-09-18 2022-10-18 Mentor Acquisition One, Llc Thermal management for head-worn computer
US10963025B2 (en) 2014-09-18 2021-03-30 Mentor Acquisition One, Llc Thermal management for head-worn computer
US10520996B2 (en) 2014-09-18 2019-12-31 Mentor Acquisition One, Llc Thermal management for head-worn computer
US9366868B2 (en) 2014-09-26 2016-06-14 Osterhout Group, Inc. See-through computer display systems
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US10078224B2 (en) 2014-09-26 2018-09-18 Osterhout Group, Inc. See-through computer display systems
US9448409B2 (en) 2014-11-26 2016-09-20 Osterhout Group, Inc. See-through computer display systems
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US11262846B2 (en) 2014-12-03 2022-03-01 Mentor Acquisition One, Llc See-through computer display systems
US10036889B2 (en) 2014-12-03 2018-07-31 Osterhout Group, Inc. Head worn computer display systems
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
US10018837B2 (en) 2014-12-03 2018-07-10 Osterhout Group, Inc. Head worn computer display systems
US11809628B2 (en) 2014-12-03 2023-11-07 Mentor Acquisition One, Llc See-through computer display systems
US10197801B2 (en) 2014-12-03 2019-02-05 Osterhout Group, Inc. Head worn computer display systems
USD743963S1 (en) 2014-12-22 2015-11-24 Osterhout Group, Inc. Air mouse
USD792400S1 (en) 2014-12-31 2017-07-18 Osterhout Group, Inc. Computer glasses
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
USD753114S1 (en) 2015-01-05 2016-04-05 Osterhout Group, Inc. Air mouse
USD794637S1 (en) 2015-01-05 2017-08-15 Osterhout Group, Inc. Air mouse
US10878775B2 (en) 2015-02-17 2020-12-29 Mentor Acquisition One, Llc See-through computer display systems
US11721303B2 (en) 2015-02-17 2023-08-08 Mentor Acquisition One, Llc See-through computer display systems
US10062182B2 (en) 2015-02-17 2018-08-28 Osterhout Group, Inc. See-through computer display systems
US9977241B2 (en) 2015-03-17 2018-05-22 Seiko Epson Corporation Head-mounted display device, control method for head-mounted display device, and computer program
US10175484B2 (en) 2015-03-17 2019-01-08 Seiko Epson Corporation Head-mounted display device, control method for head-mounted display device, and computer program
WO2016187348A1 (en) * 2015-05-18 2016-11-24 Brian Mullins Biometric authentication in a head mounted device
US11209939B2 (en) 2015-07-22 2021-12-28 Mentor Acquisition One, Llc External user interface for head worn computing
US11816296B2 (en) 2015-07-22 2023-11-14 Mentor Acquisition One, Llc External user interface for head worn computing
US10139966B2 (en) 2015-07-22 2018-11-27 Osterhout Group, Inc. External user interface for head worn computing
US10948994B2 (en) 2016-02-29 2021-03-16 Huawei Technologies Co., Ltd. Gesture control method for wearable system and wearable system
US10849817B2 (en) 2016-02-29 2020-12-01 Mentor Acquisition One, Llc Providing enhanced images for navigation
US11654074B2 (en) 2016-02-29 2023-05-23 Mentor Acquisition One, Llc Providing enhanced images for navigation
US11298288B2 (en) 2016-02-29 2022-04-12 Mentor Acquisition One, Llc Providing enhanced images for navigation
US10667981B2 (en) 2016-02-29 2020-06-02 Mentor Acquisition One, Llc Reading assistance system for visually impaired
EP3413165A4 (en) * 2016-02-29 2019-02-06 Huawei Technologies Co., Ltd. Wearable system gesture control method and wearable system
US10591728B2 (en) 2016-03-02 2020-03-17 Mentor Acquisition One, Llc Optical systems for head-worn computers
US11592669B2 (en) 2016-03-02 2023-02-28 Mentor Acquisition One, Llc Optical systems for head-worn computers
US11156834B2 (en) 2016-03-02 2021-10-26 Mentor Acquisition One, Llc Optical systems for head-worn computers
US11500212B2 (en) 2016-05-09 2022-11-15 Mentor Acquisition One, Llc User interface systems for head-worn computers
US11226691B2 (en) 2016-05-09 2022-01-18 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10824253B2 (en) 2016-05-09 2020-11-03 Mentor Acquisition One, Llc User interface systems for head-worn computers
US11320656B2 (en) 2016-05-09 2022-05-03 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10684478B2 (en) 2016-05-09 2020-06-16 Mentor Acquisition One, Llc User interface systems for head-worn computers
US10466491B2 (en) 2016-06-01 2019-11-05 Mentor Acquisition One, Llc Modular systems for head-worn computers
US11754845B2 (en) 2016-06-01 2023-09-12 Mentor Acquisition One, Llc Modular systems for head-worn computers
US11022808B2 (en) 2016-06-01 2021-06-01 Mentor Acquisition One, Llc Modular systems for head-worn computers
US11586048B2 (en) 2016-06-01 2023-02-21 Mentor Acquisition One, Llc Modular systems for head-worn computers
US11460708B2 (en) 2016-06-01 2022-10-04 Mentor Acquisition One, Llc Modular systems for head-worn computers
US11825257B2 (en) 2016-08-22 2023-11-21 Mentor Acquisition One, Llc Speaker systems for head-worn computer systems
US10757495B2 (en) 2016-08-22 2020-08-25 Mentor Acquisition One, Llc Speaker systems for head-worn computer systems
US9826299B1 (en) 2016-08-22 2017-11-21 Osterhout Group, Inc. Speaker systems for head-worn computer systems
US11350196B2 (en) 2016-08-22 2022-05-31 Mentor Acquisition One, Llc Speaker systems for head-worn computer systems
US10690936B2 (en) 2016-08-29 2020-06-23 Mentor Acquisition One, Llc Adjustable nose bridge assembly for headworn computer
US11409128B2 (en) 2016-08-29 2022-08-09 Mentor Acquisition One, Llc Adjustable nose bridge assembly for headworn computer
US11415856B2 (en) 2016-09-08 2022-08-16 Mentor Acquisition One, Llc Electrochromic systems for head-worn computer systems
US10768500B2 (en) 2016-09-08 2020-09-08 Mentor Acquisition One, Llc Electrochromic systems for head-worn computer systems
US9880441B1 (en) 2016-09-08 2018-01-30 Osterhout Group, Inc. Electrochromic systems for head-worn computer systems
US11366320B2 (en) 2016-09-08 2022-06-21 Mentor Acquisition One, Llc Optical systems for head-worn computers
US11768417B2 (en) 2016-09-08 2023-09-26 Mentor Acquisition One, Llc Electrochromic systems for head-worn computer systems
US10534180B2 (en) 2016-09-08 2020-01-14 Mentor Acquisition One, Llc Optical systems for head-worn computers
US11604358B2 (en) 2016-09-08 2023-03-14 Mentor Acquisition One, Llc Optical systems for head-worn computers
US9910284B1 (en) 2016-09-08 2018-03-06 Osterhout Group, Inc. Optical systems for head-worn computers
USD840395S1 (en) 2016-10-17 2019-02-12 Osterhout Group, Inc. Head-worn computer
EP3330839A1 (en) * 2016-12-05 2018-06-06 THOMSON Licensing Method and device for adapting an immersive content to the field of view of a user
US10850116B2 (en) 2016-12-30 2020-12-01 Mentor Acquisition One, Llc Head-worn therapy device
US11771915B2 (en) 2016-12-30 2023-10-03 Mentor Acquisition One, Llc Head-worn therapy device
USD918905S1 (en) 2017-01-04 2021-05-11 Mentor Acquisition One, Llc Computer glasses
USD947186S1 (en) 2017-01-04 2022-03-29 Mentor Acquisition One, Llc Computer glasses
USD864959S1 (en) 2017-01-04 2019-10-29 Mentor Acquisition One, Llc Computer glasses
US20180253856A1 (en) * 2017-03-01 2018-09-06 Microsoft Technology Licensing, Llc Multi-Spectrum Illumination-and-Sensor Module for Head Tracking, Gesture Recognition and Spatial Mapping
US10628950B2 (en) * 2017-03-01 2020-04-21 Microsoft Technology Licensing, Llc Multi-spectrum illumination-and-sensor module for head tracking, gesture recognition and spatial mapping
US10252164B2 (en) * 2017-03-17 2019-04-09 Roblox Corporation Avatar teleport controller
US20180264363A1 (en) * 2017-03-17 2018-09-20 Roblox Corporation Avatar Teleport Controller
US10159900B2 (en) * 2017-03-17 2018-12-25 Roblox Corporation Avatar teleport controller
EP3594734A4 (en) * 2017-03-24 2020-04-01 Samsung Electronics Co., Ltd. Electronic device for playing content and computer-readable recording medium
US11194390B2 (en) 2017-03-24 2021-12-07 Samsung Electronics Co., Ltd. Electronic device for playing content and computer-readable recording medium
US10936053B2 (en) * 2017-03-29 2021-03-02 Boe Technology Group Co., Ltd. Interaction system of three-dimensional space and method for operating same
US10747386B2 (en) 2017-06-01 2020-08-18 Samsung Electronics Co., Ltd. Systems and methods for window control in virtual reality environment
EP3590028A4 (en) * 2017-06-01 2020-01-22 Samsung Electronics Co., Ltd. Systems and methods for window control in virtual reality environment
US11366561B2 (en) 2017-06-01 2022-06-21 Samsung Electronics Co., Ltd. Systems and methods for window control in virtual reality environment
US11119581B2 (en) 2017-06-15 2021-09-14 Microsoft Technology Licensing, Llc Displacement oriented interaction in computer-mediated reality
US11009942B2 (en) * 2017-06-28 2021-05-18 Shandong University Multi-human tracking system and method with single kinect for supporting mobile virtual reality application
US10422995B2 (en) 2017-07-24 2019-09-24 Mentor Acquisition One, Llc See-through computer display systems with stray light management
US11567328B2 (en) 2017-07-24 2023-01-31 Mentor Acquisition One, Llc See-through computer display systems with adjustable zoom cameras
US11789269B2 (en) 2017-07-24 2023-10-17 Mentor Acquisition One, Llc See-through computer display systems
US11226489B2 (en) 2017-07-24 2022-01-18 Mentor Acquisition One, Llc See-through computer display systems with stray light management
US11668939B2 (en) 2017-07-24 2023-06-06 Mentor Acquisition One, Llc See-through computer display systems with stray light management
US11409105B2 (en) 2017-07-24 2022-08-09 Mentor Acquisition One, Llc See-through computer display systems
US10578869B2 (en) 2017-07-24 2020-03-03 Mentor Acquisition One, Llc See-through computer display systems with adjustable zoom cameras
US11550157B2 (en) 2017-07-24 2023-01-10 Mentor Acquisition One, Llc See-through computer display systems
US11042035B2 (en) 2017-07-24 2021-06-22 Mentor Acquisition One, Llc See-through computer display systems with adjustable zoom cameras
US10969584B2 (en) 2017-08-04 2021-04-06 Mentor Acquisition One, Llc Image expansion optic for head-worn computer
US11500207B2 (en) 2017-08-04 2022-11-15 Mentor Acquisition One, Llc Image expansion optic for head-worn computer
US11947120B2 (en) 2017-08-04 2024-04-02 Mentor Acquisition One, Llc Image expansion optic for head-worn computer
US11195334B2 (en) * 2018-08-03 2021-12-07 Igt Providing interactive virtual elements within a mixed reality scene
US20200043233A1 (en) * 2018-08-03 2020-02-06 Igt Providing interactive virtual elements within a mixed reality scene
US11170565B2 (en) 2018-08-31 2021-11-09 Magic Leap, Inc. Spatially-resolved dynamic dimming for augmented reality device
US11676333B2 (en) 2018-08-31 2023-06-13 Magic Leap, Inc. Spatially-resolved dynamic dimming for augmented reality device
US11461961B2 (en) 2018-08-31 2022-10-04 Magic Leap, Inc. Spatially-resolved dynamic dimming for augmented reality device
US11808940B2 (en) 2019-10-10 2023-11-07 Medithinq Co., Ltd. Eye tracking system for smart glasses and method therefor
US11960089B2 (en) 2022-06-27 2024-04-16 Mentor Acquisition One, Llc Optical configurations for head-worn see-through displays
US11960095B2 (en) 2023-04-19 2024-04-16 Mentor Acquisition One, Llc See-through computer display systems
CN116740318A (en) * 2023-08-14 2023-09-12 深圳风向标教育资源股份有限公司 Method, device and storage medium for placing automobile parts

Also Published As

Publication number Publication date
EP2724191A2 (en) 2014-04-30
JP2014526157A (en) 2014-10-02
US20150235632A1 (en) 2015-08-20
TW201303640A (en) 2013-01-16
US20130093789A1 (en) 2013-04-18
CN103635849A (en) 2014-03-12
WO2012177657A2 (en) 2012-12-27
US9041623B2 (en) 2015-05-26
WO2012177657A3 (en) 2013-05-02
KR20140034252A (en) 2014-03-19
EP2724191A4 (en) 2015-03-25
US9245501B2 (en) 2016-01-26

Similar Documents

Publication Publication Date Title
US9245501B2 (en) Total field of view classification
US9898075B2 (en) Visual stabilization system for head-mounted displays
CN107209386B (en) Augmented reality view object follower
US9135508B2 (en) Enhanced user eye gaze estimation
US20200004401A1 (en) Gesture-based content sharing in artifical reality environments
US9165381B2 (en) Augmented books in a mixed reality environment
US10831268B1 (en) Systems and methods for using eye tracking to improve user interactions with objects in artificial reality
US11579689B2 (en) Reducing head mounted display power consumption and heat generation through predictive rendering of content
US11017257B2 (en) Information processing device, information processing method, and program
US10789779B2 (en) Location-based holographic experience
US20130241805A1 (en) Using Convergence Angle to Select Among Different UI Elements
US20190212828A1 (en) Object enhancement in artificial reality via a near eye display interface
US11022794B2 (en) Visual indicators of user attention in AR/VR environment
Lander et al. hEYEbrid: A hybrid approach for mobile calibration-free gaze estimation
US11321926B2 (en) Method and device for content placement
US20230316671A1 (en) Attention-based content visualization for an extended reality environment
US20240029329A1 (en) Mitigation of Animation Disruption in Artificial Reality
WO2024064909A2 (en) Methods, systems, and computer program products for alignment of a wearable device
WO2023192254A1 (en) Attention-based content visualization for an extended reality environment

Legal Events

Date Code Title Description
AS Assignment

Owner name: MICROSOFT CORPORATION, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIU, JAMES CHIA-MING;LATTA, STEPHEN GILCHRIST;ANDREWS, ANTON;AND OTHERS;SIGNING DATES FROM 20110614 TO 20110622;REEL/FRAME:026489/0078

STCB Information on status: application discontinuation

Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION

AS Assignment

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034766/0509

Effective date: 20141014