US20060244677A1 - Industrial augmented reality - Google Patents
Industrial augmented reality Download PDFInfo
- Publication number
- US20060244677A1 US20060244677A1 US10/046,468 US4646801A US2006244677A1 US 20060244677 A1 US20060244677 A1 US 20060244677A1 US 4646801 A US4646801 A US 4646801A US 2006244677 A1 US2006244677 A1 US 2006244677A1
- Authority
- US
- United States
- Prior art keywords
- operator
- view
- visual markers
- field
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/01—Head-up displays
- G02B27/017—Head mounted
Definitions
- This invention relates to information processing systems and methods in an industrial environment. More particularly, the invention relates to systems and methods for displaying information for use by human operators in a mobile working environment.
- hand-held or wearable computers allow the worker to move freely about the workspace and retrieve information from computer networks accessible at their fingertips, while they maintain in proximity to the object of the task.
- Examples of these include retail operations where sales assistants or inventory control clerks carry hand-held computers with bar code scanners that can identify products by scanning the bar code and then displaying information associated with that product.
- Another example includes car rental return agents who key information into a wearable computer in the parking lot of the rental agency when the car is returned, and then print out a receipt from a wearable printer to be able to give to a customer standing in the parking lot by the rental car.
- Hand-held computers require the worker to devote one or both hands to the task of inputting information into the computer.
- Wearable computers generally require the use of at least one hand for inputting data into a computer, as well as requiring the worker to focus his gaze and attention to the hand-held computer rather than on the object or task before him. While these wearable computer solutions are an advance over stationary kiosks and strategically located catalogs, wall charts or instruction manual books accessible to the worker, there is still much room for improvement to free up the hands and attention of the worker, to thereby increase the workers' productivity.
- the present invention addresses the aforementioned needs by providing a method and system for viewing data comprising detecting visual markers within view of a wearable camera worn by a human worker, and displaying data associated with at least one of those visual markers on a wearable display worn by the operator.
- Another aspect of the invention provides a method and system for coordinating movement of human workers comprising detecting of one or more visual markers within view of a camera worn by the worker, and determining the location of the worker from the known location of the visual marker within view of the camera.
- Many physical tasks in an industrial work environment can be made more efficient by providing workers with more data about their tasks and their environment. These tasks may include such things as order picking for factory floor activities.
- This invention makes such tasks more efficient by equipping workers with a small wearable display and a small wearable camera.
- both the wearable display and wearable camera are mounted on a pair eyeglasses or goggles.
- the camera and display are connected to a small wearable computer.
- low cost visual markers may be strategically placed to identify locations or objects in proximity to the markers. When the visual marker is within the field of view of the camera, the wearable computer system can “see” and identify the visual marker. Once identified, the wearable display can be used to give the worker instructions or information regarding the identified object or location.
- FIG. 1 depicts a typical system in accordance with one embodiment of the invention.
- a pair of eyeglasses 10 includes a wearable see-through display 12 with a wearable camera 14 both of which are attached to the eyeglasses 10 .
- a headband 16 assists with keeping the glasses on the worker's head.
- the wearable camera 14 is positioned to detect visual markers 18 within the field of view of the worker.
- the camera and display are preferably connected through cables 20 and 22 to a wearable computer 24 .
- the wearable computer 24 includes a belt 26 for strapping it to the waist or an appendage (e.g., arm or leg) of a worker.
- the wearable computer 24 includes a wireless transmitter and antenna 28 for wireless connectivity to a computer network 30 in the industrial environment.
- the databases may include, but are not limited to, a visual marker library 36 , visual marker location maps 38 , production and operation reference manuals 40 , and employee preference and profile library 40 .
- the wearable display 12 may be a see-through or see-around micro-optical display, such as eyeglass displays commercially available form Microoptical Corp., Boston, Mass. USA.
- the small camera 14 may be attached to or integrated with the eyeglass displays. Numerous such cameras are commercially available.
- the wearable computer 24 may be one such as commercially available from Xybernaut Corp., Fairfax, Va., USA, with a wireless communication transceiver, for example using an 802.11(b) protocol.
- the wearable computer 24 or the processor 34 preferably includes a visual marker or pattern recognition system such as, but not limited to, the ARToolkit developed by and available from the University of Washington.
- the invention also provides various methods for applying augmented reality in an industrial work environment.
- One embodiment of the invention includes a method for viewing data, which includes directing a wearable camera worn by a human operator towards one or more labeled objects within a field of view of the operator, detecting one or more visual markers within a field of view of the camera with at least one of the visual markers associated with and proximate to each of the labeled objects, selecting data from a memory storage with the data being associated with one of the objects associated with one of the visual markers, and then displaying the data on a wearable display worn by the operator.
- the data is selected by identifying which of the visual markers is located within a predetermined zone of the field of view of the camera for a predetermined amount of time of between about 0.02 seconds and 0.5 seconds.
- the zone is preferably the central 50% of the field of view of the camera.
- the data is displayed on a see-through display such that the data are superimposed on a real image seen in the field of view of the operator.
- the display is a heads-up see-through display attached to a pair of eyeglasses or goggles worn by the operator on which additional data associated with the object is displayed in response to a request sent by the operator.
- the request may be an electric signal generated by a voice command from the operator or by a physical touch by the operator, or the request may be a focused gaze upon the object for an extended period of time.
- the “query” gaze may be between 1 and 2 second of steady gaze for fast-paced activity, such as working in a warehouse, to 5 seconds or longer for slow-paced activity, such as working on a small parts assembly bench, where a worker's gaze does not change much.
- the method preferably includes processing the image of the detected visual marker to determine a unique identifier associated with the marker, and transmitting the unique identifier to a processor for selecting the data appropriate for the workers present task based on contextual clues observed either through the visual markers presently in view or recently viewed.
- the image processing and pattern recognition is conducted by the wasable computer. This processing may also be done by a workstation computer on the network, with the wearable computer merely transmitting the images viewed through the camera.
- visual markers placed on stationary objects such as walls or stationary equipment allow the computer system to determine where the worker is physically located whenever such marker is “seen” by the system. This allows for efficient tracking of workers or operators and efficient routing of tasks.
- This invention thereby, provides for the coordinating of movement of workers in a work environment.
- containers and walls can be marked with visual markers.
- the system can track where they are and where they are going.
- the system can identify the nearest worker and provide them with instructions on their wearable display to go to the point of need.
- the system can identify that object as the correct object of interest, and provide additional instructions for the worker to follow with regard to that object.
- a second embodiment of this invention includes a method for coordinating the movement of human workers in a working environment comprising directing a wearable camera worn by a human operator towards a one or more labeled objects within a field of view of the operator, and detecting one or more visual markers within a field of view of the camera with at least one of the visual markers associated with and proximate to each of the labeled objects. Further, the method includes processing the image of the visual markers and determining the unique identifier associated with each marker, obtaining the physical location of the markers maintained in a memory storage; and determining the location of the operator from the locations of the markers within the field of view of the camera.
- the various embodiments of this invention are not limited to a factory floor environment, but can be applied to a wide variety of industrial work places. Similar scenarios can be described in situations such as hospitals, factories, oil fields, chemical plants or anywhere that are physical tasks that require detailed knowledge to make the task safer and/or more efficient.
- This invention allows the operator or worker to interact with the work environment in a very natural manner.
- the wearable camera is focused within the same field of view as the operator when focusing straight ahead in a work activity. In this way the system can see the objects that the operator is focusing on.
- the system is thus programmed to respond to objects that are within the field of view of that camera for at least a predetermined amount of time. If a person stares at an object, that usually means they are trying to obtain more information about the object. Naturally, in response, the system within identify the object of regard, retrieve information about that object from a database, and display that information on the wearable display of the operator. The longer that the operator gazes upon the object the system may optionally be programmed to provide additional information regarding the object. Alternatively, the system may be responsive to voice command of the operator or mechanical inputs from the touch of the operator to provide additional information upon request of the operator.
- the system can also be programmed to be responsive to the personal preferences and other personal information maintained in a profile associated with the operator. For example, a worker that has special skills and responsibilities as a mechanical maintenance worker would receive information relating to the repair of an object within his view. On the other hand, a person whose job responsibilities involve the operation or manipulation of objects, such as an assembly line operator, will receive information on their display associated with the operation of an object within their field of view. Likewise, a firefighter staring at the same object may see information about safety precaution or about the volatility of the contents of the object.
- Visual based markers are preferred for the system because they are extremely low cost.
- the visual markers can be bar codes, dot matrix codes, or other recognizable patterns, such as letters and numbers. While it is possible to use other type of non-visual markers such as active or passive radio frequency tags, such tags can be expensive. Also such RF tags usually only give general proximity information. Visual markers can be located actually on objects. For many applications this type of tracking and interaction can be more powerful and/or cost effective.
- visual pattern recognition systems can be incorporated into this invention.
- visual quality control inspection systems rely upon pattern recognitions of common elements by comparison with template pattern of such elements.
- pick and place machines can detect objects out of location by comparison of objects with a database of commonly seen objects.
Abstract
Description
- This invention relates to information processing systems and methods in an industrial environment. More particularly, the invention relates to systems and methods for displaying information for use by human operators in a mobile working environment.
- An increasing number of workers today in mobile working environments are assisted by hand-held or wearable computer systems. Rather than using computer kiosks or workstations at locations throughout the work environment, the hand-held or wearable computers allow the worker to move freely about the workspace and retrieve information from computer networks accessible at their fingertips, while they maintain in proximity to the object of the task. Examples of these include retail operations where sales assistants or inventory control clerks carry hand-held computers with bar code scanners that can identify products by scanning the bar code and then displaying information associated with that product. Another example includes car rental return agents who key information into a wearable computer in the parking lot of the rental agency when the car is returned, and then print out a receipt from a wearable printer to be able to give to a customer standing in the parking lot by the rental car.
- While these systems are useful, they have limited capabilities. Hand-held computers require the worker to devote one or both hands to the task of inputting information into the computer. Wearable computers generally require the use of at least one hand for inputting data into a computer, as well as requiring the worker to focus his gaze and attention to the hand-held computer rather than on the object or task before him. While these wearable computer solutions are an advance over stationary kiosks and strategically located catalogs, wall charts or instruction manual books accessible to the worker, there is still much room for improvement to free up the hands and attention of the worker, to thereby increase the workers' productivity.
- The present invention addresses the aforementioned needs by providing a method and system for viewing data comprising detecting visual markers within view of a wearable camera worn by a human worker, and displaying data associated with at least one of those visual markers on a wearable display worn by the operator.
- Another aspect of the invention provides a method and system for coordinating movement of human workers comprising detecting of one or more visual markers within view of a camera worn by the worker, and determining the location of the worker from the known location of the visual marker within view of the camera.
-
FIG. 1 is a system in accordance with one embodiment of the invention. - Many physical tasks in an industrial work environment can be made more efficient by providing workers with more data about their tasks and their environment. These tasks may include such things as order picking for factory floor activities. This invention makes such tasks more efficient by equipping workers with a small wearable display and a small wearable camera. Preferably both the wearable display and wearable camera are mounted on a pair eyeglasses or goggles. Preferably the camera and display are connected to a small wearable computer. Throughout the work environment, low cost visual markers may be strategically placed to identify locations or objects in proximity to the markers. When the visual marker is within the field of view of the camera, the wearable computer system can “see” and identify the visual marker. Once identified, the wearable display can be used to give the worker instructions or information regarding the identified object or location.
-
FIG. 1 depicts a typical system in accordance with one embodiment of the invention. A pair ofeyeglasses 10 includes a wearable see-throughdisplay 12 with awearable camera 14 both of which are attached to theeyeglasses 10. Aheadband 16 assists with keeping the glasses on the worker's head. Thewearable camera 14 is positioned to detectvisual markers 18 within the field of view of the worker. - The camera and display are preferably connected through
cables wearable computer 24. However, technological developments may permit the connections to be wireless, obviating the need for cables. Thewearable computer 24 includes abelt 26 for strapping it to the waist or an appendage (e.g., arm or leg) of a worker. Preferably, thewearable computer 24 includes a wireless transmitter andantenna 28 for wireless connectivity to acomputer network 30 in the industrial environment. - Also connected with the
network 30 are anetwork server 32, aprocessor 34, and several databases. The databases may include, but are not limited to, avisual marker library 36, visualmarker location maps 38, production andoperation reference manuals 40, and employee preference andprofile library 40. - A typical system as described above may include several components that are presently commercially available. The
wearable display 12 may be a see-through or see-around micro-optical display, such as eyeglass displays commercially available form Microoptical Corp., Boston, Mass. USA. Thesmall camera 14 may be attached to or integrated with the eyeglass displays. Numerous such cameras are commercially available. Thewearable computer 24 may be one such as commercially available from Xybernaut Corp., Fairfax, Va., USA, with a wireless communication transceiver, for example using an 802.11(b) protocol. Thewearable computer 24 or theprocessor 34 preferably includes a visual marker or pattern recognition system such as, but not limited to, the ARToolkit developed by and available from the University of Washington. - The system also includes a
computer network 30 wirelessly accessible to the wearable computer. The computer network includes aprocessor 34 in communication with a database located on memory storage devices in communication with aserver 32. Thedatabases - The invention also provides various methods for applying augmented reality in an industrial work environment. One embodiment of the invention includes a method for viewing data, which includes directing a wearable camera worn by a human operator towards one or more labeled objects within a field of view of the operator, detecting one or more visual markers within a field of view of the camera with at least one of the visual markers associated with and proximate to each of the labeled objects, selecting data from a memory storage with the data being associated with one of the objects associated with one of the visual markers, and then displaying the data on a wearable display worn by the operator.
- Preferably, the data is selected by identifying which of the visual markers is located within a predetermined zone of the field of view of the camera for a predetermined amount of time of between about 0.02 seconds and 0.5 seconds. The zone is preferably the central 50% of the field of view of the camera.
- Preferably, the data is displayed on a see-through display such that the data are superimposed on a real image seen in the field of view of the operator. More preferably, the display is a heads-up see-through display attached to a pair of eyeglasses or goggles worn by the operator on which additional data associated with the object is displayed in response to a request sent by the operator. The request may be an electric signal generated by a voice command from the operator or by a physical touch by the operator, or the request may be a focused gaze upon the object for an extended period of time. Depending on the job activity, the “query” gaze may be between 1 and 2 second of steady gaze for fast-paced activity, such as working in a warehouse, to 5 seconds or longer for slow-paced activity, such as working on a small parts assembly bench, where a worker's gaze does not change much.
- Additionally, the method preferably includes processing the image of the detected visual marker to determine a unique identifier associated with the marker, and transmitting the unique identifier to a processor for selecting the data appropriate for the workers present task based on contextual clues observed either through the visual markers presently in view or recently viewed. Preferably, the image processing and pattern recognition is conducted by the werable computer. This processing may also be done by a workstation computer on the network, with the wearable computer merely transmitting the images viewed through the camera.
- In addition, visual markers placed on stationary objects such as walls or stationary equipment allow the computer system to determine where the worker is physically located whenever such marker is “seen” by the system. This allows for efficient tracking of workers or operators and efficient routing of tasks.
- This invention, thereby, provides for the coordinating of movement of workers in a work environment. For example, in a warehouse environment, containers and walls can be marked with visual markers. As workers walk around the warehouse, the system can track where they are and where they are going. When a worker is needed in a specific area, the system can identify the nearest worker and provide them with instructions on their wearable display to go to the point of need. When the worker arrives and looks at a specific box or object, the system can identify that object as the correct object of interest, and provide additional instructions for the worker to follow with regard to that object.
- Accordingly, a second embodiment of this invention includes a method for coordinating the movement of human workers in a working environment comprising directing a wearable camera worn by a human operator towards a one or more labeled objects within a field of view of the operator, and detecting one or more visual markers within a field of view of the camera with at least one of the visual markers associated with and proximate to each of the labeled objects. Further, the method includes processing the image of the visual markers and determining the unique identifier associated with each marker, obtaining the physical location of the markers maintained in a memory storage; and determining the location of the operator from the locations of the markers within the field of view of the camera.
- In addition, the method may preferably include tracking the movement of the operator by monitoring the change of the location of the markers within the field of view of the camera. The movement of the worker is preferably coordinated by displaying data on a see-through display such that the data are superimposed on a real image seen in the field of view of the operator with the data including instructions directing the operator to a new location.
- The various embodiments of this invention are not limited to a factory floor environment, but can be applied to a wide variety of industrial work places. Similar scenarios can be described in situations such as hospitals, factories, oil fields, chemical plants or anywhere that are physical tasks that require detailed knowledge to make the task safer and/or more efficient.
- This invention allows the operator or worker to interact with the work environment in a very natural manner. Preferably the wearable camera is focused within the same field of view as the operator when focusing straight ahead in a work activity. In this way the system can see the objects that the operator is focusing on. The system is thus programmed to respond to objects that are within the field of view of that camera for at least a predetermined amount of time. If a person stares at an object, that usually means they are trying to obtain more information about the object. Naturally, in response, the system within identify the object of regard, retrieve information about that object from a database, and display that information on the wearable display of the operator. The longer that the operator gazes upon the object the system may optionally be programmed to provide additional information regarding the object. Alternatively, the system may be responsive to voice command of the operator or mechanical inputs from the touch of the operator to provide additional information upon request of the operator.
- The system can also be programmed to be responsive to the personal preferences and other personal information maintained in a profile associated with the operator. For example, a worker that has special skills and responsibilities as a mechanical maintenance worker would receive information relating to the repair of an object within his view. On the other hand, a person whose job responsibilities involve the operation or manipulation of objects, such as an assembly line operator, will receive information on their display associated with the operation of an object within their field of view. Likewise, a firefighter staring at the same object may see information about safety precaution or about the volatility of the contents of the object.
- Visual based markers are preferred for the system because they are extremely low cost. The visual markers can be bar codes, dot matrix codes, or other recognizable patterns, such as letters and numbers. While it is possible to use other type of non-visual markers such as active or passive radio frequency tags, such tags can be expensive. Also such RF tags usually only give general proximity information. Visual markers can be located actually on objects. For many applications this type of tracking and interaction can be more powerful and/or cost effective.
- Depending on the work environment and commonality of objects seen therein, more complex visual pattern recognition systems can be incorporated into this invention. For example, visual quality control inspection systems rely upon pattern recognitions of common elements by comparison with template pattern of such elements. Also, pick and place machines can detect objects out of location by comparison of objects with a database of commonly seen objects.
- While this invention has been shown and described in connection with the preferred embodiments, it is apparent that certain changes and modifications, in addition to those mentioned above may be made from the basic features of this invention. In addition, there are many different types of computer software and hardware that may be utilized in practicing the invention, and the invention is not limited to the examples described above. Accordingly, it is the intention of the applicants to protect all variations and modifications within the valid scope of the present invention. It is intended that the invention be defined by the following claims, including all equivalents.
Claims (30)
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/046,468 US7126558B1 (en) | 2001-10-19 | 2001-10-19 | Industrial augmented reality |
ES02773811.1T ES2525679T3 (en) | 2001-10-19 | 2002-10-18 | Augmented Industrial Reality |
PCT/US2002/033466 WO2003034397A1 (en) | 2001-10-19 | 2002-10-18 | Industrial augmented reality |
US10/492,437 US7372451B2 (en) | 2001-10-19 | 2002-10-18 | Industrial augmented reality |
EP02773811.1A EP1436800B1 (en) | 2001-10-19 | 2002-10-18 | Industrial augmented reality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/046,468 US7126558B1 (en) | 2001-10-19 | 2001-10-19 | Industrial augmented reality |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10492437 Continuation-In-Part | 2002-10-18 |
Publications (2)
Publication Number | Publication Date |
---|---|
US7126558B1 US7126558B1 (en) | 2006-10-24 |
US20060244677A1 true US20060244677A1 (en) | 2006-11-02 |
Family
ID=21943625
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/046,468 Expired - Lifetime US7126558B1 (en) | 2001-10-19 | 2001-10-19 | Industrial augmented reality |
US10/492,437 Expired - Lifetime US7372451B2 (en) | 2001-10-19 | 2002-10-18 | Industrial augmented reality |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/492,437 Expired - Lifetime US7372451B2 (en) | 2001-10-19 | 2002-10-18 | Industrial augmented reality |
Country Status (4)
Country | Link |
---|---|
US (2) | US7126558B1 (en) |
EP (1) | EP1436800B1 (en) |
ES (1) | ES2525679T3 (en) |
WO (1) | WO2003034397A1 (en) |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050174470A1 (en) * | 2004-02-06 | 2005-08-11 | Olympus Corporation | Head-mounted camera |
US20060271292A1 (en) * | 2003-12-02 | 2006-11-30 | Buck Michael A | Mobile Geographical Information Display System |
US20080191846A1 (en) * | 2007-02-12 | 2008-08-14 | Wayne Chang | Methods and apparatus to visualize locations of radio frequency identification (rfid) tagged items |
US20080266323A1 (en) * | 2007-04-25 | 2008-10-30 | Board Of Trustees Of Michigan State University | Augmented reality user interaction system |
US20080306756A1 (en) * | 2007-06-08 | 2008-12-11 | Sorensen Associates Inc | Shopper view tracking and analysis system and method |
US20090154293A1 (en) * | 2007-12-18 | 2009-06-18 | Anandraj Sengupta | System and method for augmented reality inspection and data visualization |
US20090300535A1 (en) * | 2003-12-31 | 2009-12-03 | Charlotte Skourup | Virtual control panel |
US20120078686A1 (en) * | 2010-09-27 | 2012-03-29 | Bashani Gilad G | Parking status system |
US20120154195A1 (en) * | 2010-12-17 | 2012-06-21 | Sony Ericsson Mobile Communications Ab | System and method for remote controlled device selection |
WO2013085927A1 (en) * | 2011-12-07 | 2013-06-13 | Sheridan Martin Small | Updating printed content with personalized virtual data |
AU2013200187B2 (en) * | 2011-07-22 | 2013-06-20 | Microsoft Technology Licensing, Llc | Automatic text scrolling on a head-mounted display |
US20140111346A1 (en) * | 2009-06-29 | 2014-04-24 | Commscope, Inc. Of North Carolina | Methods of Automatically Recording Patching Changes at Passive Patch Panels and Network Equipment |
US20140291396A1 (en) * | 2013-04-01 | 2014-10-02 | Ncr Corporation | Headheld Scanner and POS Display with Mobile Phone |
US9165381B2 (en) | 2012-05-31 | 2015-10-20 | Microsoft Technology Licensing, Llc | Augmented books in a mixed reality environment |
US9183807B2 (en) | 2011-12-07 | 2015-11-10 | Microsoft Technology Licensing, Llc | Displaying virtual data as printed content |
US9182815B2 (en) | 2011-12-07 | 2015-11-10 | Microsoft Technology Licensing, Llc | Making static printed content dynamic with virtual data |
US20160132532A1 (en) * | 2014-11-11 | 2016-05-12 | Commscope, Inc. Of North Carolina | Multi-tier intelligent infrastructure management systems for communications systems and related equipment and methods |
DE102017219067A1 (en) * | 2017-10-25 | 2019-04-25 | Bayerische Motoren Werke Aktiengesellschaft | DEVICE AND METHOD FOR THE VISUAL SUPPORT OF A USER IN A WORKING ENVIRONMENT |
GB2549874B (en) * | 2015-02-13 | 2022-03-23 | Halliburton Energy Services Inc | Distributing information using role-specific augmented reality devices |
Families Citing this family (150)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003308117A (en) * | 2002-04-17 | 2003-10-31 | Hitachi Zosen Corp | System for maintenance |
US7050078B2 (en) * | 2002-12-19 | 2006-05-23 | Accenture Global Services Gmbh | Arbitrary object tracking augmented reality applications |
US20050063563A1 (en) * | 2003-09-23 | 2005-03-24 | Soliman Samir S. | System and method for geolocation using imaging techniques |
JP2005115868A (en) * | 2003-10-10 | 2005-04-28 | Sony Corp | Private information storage device and method, and private information management device and method |
ITTO20031055A1 (en) * | 2003-12-30 | 2005-06-30 | Fiat Ricerche | SYSTEM FOR REMOTE ASSISTANCE OF AN OPERATOR IN A WORK PHASE. |
WO2005107274A1 (en) * | 2004-04-28 | 2005-11-10 | Augmented Media Aps | A system for generating virtual three dimensional projections |
JP4532982B2 (en) * | 2004-05-14 | 2010-08-25 | キヤノン株式会社 | Arrangement information estimation method and information processing apparatus |
US7508840B2 (en) * | 2004-05-28 | 2009-03-24 | Bae Systems Information And Electronic Systems Integration Inc. | Mobile temporary incident area network for local communications interoperability |
FR2875989B1 (en) * | 2004-09-24 | 2007-02-23 | Renault Sas | VIDEO DEVICE FOR INCREASING REALITY INCREASED AND METHOD FOR COMPARING TWO ENVIRONMENTS |
CA2624946A1 (en) * | 2004-11-10 | 2006-05-18 | Bae Systems Information And Electronic Systems Integration Inc. | Wearable portable device for establishing communications interoperability at an incident site |
US7845560B2 (en) * | 2004-12-14 | 2010-12-07 | Sky-Trax Incorporated | Method and apparatus for determining position and rotational orientation of an object |
US20060181482A1 (en) * | 2005-02-03 | 2006-08-17 | Iaquinto John M | Apparatus for providing visual data during an operation |
ES2325374T3 (en) * | 2005-05-03 | 2009-09-02 | Seac02 S.R.L. | INCREASED REALITY SYSTEM WITH IDENTIFICATION OF REAL MARKER OBJECT. |
DE102005046762A1 (en) * | 2005-09-29 | 2007-04-05 | Siemens Ag | System and method for displaying user information, in particular augmented reality information, with the aid of tracking information stored in RFID data memories |
US8381982B2 (en) * | 2005-12-03 | 2013-02-26 | Sky-Trax, Inc. | Method and apparatus for managing and controlling manned and automated utility vehicles |
US9323055B2 (en) * | 2006-05-26 | 2016-04-26 | Exelis, Inc. | System and method to display maintenance and operational instructions of an apparatus using augmented reality |
EP1865455A1 (en) * | 2006-06-07 | 2007-12-12 | Seac02 S.r.l. | A virtual advertising system |
EP1887526A1 (en) * | 2006-08-11 | 2008-02-13 | Seac02 S.r.l. | A digitally-augmented reality video system |
US20080064326A1 (en) * | 2006-08-24 | 2008-03-13 | Stephen Joseph Foster | Systems and Methods for Casting Captions Associated With A Media Stream To A User |
US20080120561A1 (en) * | 2006-11-21 | 2008-05-22 | Eric Charles Woods | Network connected media platform |
US9217868B2 (en) | 2007-01-12 | 2015-12-22 | Kopin Corporation | Monocular display device |
CN101632033B (en) | 2007-01-12 | 2013-07-31 | 寇平公司 | Helmet type monocular display device |
US20080218331A1 (en) * | 2007-03-08 | 2008-09-11 | Itt Manufacturing Enterprises, Inc. | Augmented reality-based system and method to show the location of personnel and sensors inside occluded structures and provide increased situation awareness |
US7898504B2 (en) * | 2007-04-06 | 2011-03-01 | Sony Corporation | Personal theater display |
US8210435B2 (en) * | 2008-01-14 | 2012-07-03 | Sky-Trax, Inc. | Optical position marker apparatus |
US8565913B2 (en) * | 2008-02-01 | 2013-10-22 | Sky-Trax, Inc. | Apparatus and method for asset tracking |
FR2928805B1 (en) * | 2008-03-14 | 2012-06-01 | Alcatel Lucent | METHOD FOR IMPLEMENTING VIDEO ENRICHED ON MOBILE TERMINALS |
AT10520U3 (en) * | 2008-09-05 | 2013-10-15 | Knapp Systemintegration Gmbh | DEVICE AND METHOD FOR THE VISUAL SUPPORT OF PICKING PROCESSES |
US20100315329A1 (en) * | 2009-06-12 | 2010-12-16 | Southwest Research Institute | Wearable workspace |
US20100328471A1 (en) * | 2009-06-24 | 2010-12-30 | Justin Boland | Wearable Multi-Channel Camera |
JP5282709B2 (en) * | 2009-09-25 | 2013-09-04 | ブラザー工業株式会社 | Imaging data utilization system including head mounted display |
US20110079639A1 (en) * | 2009-10-06 | 2011-04-07 | Samsung Electronics Co. Ltd. | Geotagging using barcodes |
US9129295B2 (en) | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
US9366862B2 (en) | 2010-02-28 | 2016-06-14 | Microsoft Technology Licensing, Llc | System and method for delivering content to a group of see-through near eye display eyepieces |
US9134534B2 (en) | 2010-02-28 | 2015-09-15 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including a modular image source |
US9128281B2 (en) | 2010-09-14 | 2015-09-08 | Microsoft Technology Licensing, Llc | Eyepiece with uniformly illuminated reflective display |
US8472120B2 (en) | 2010-02-28 | 2013-06-25 | Osterhout Group, Inc. | See-through near-eye display glasses with a small scale image source |
US10180572B2 (en) | 2010-02-28 | 2019-01-15 | Microsoft Technology Licensing, Llc | AR glasses with event and user action control of external applications |
US8477425B2 (en) | 2010-02-28 | 2013-07-02 | Osterhout Group, Inc. | See-through near-eye display glasses including a partially reflective, partially transmitting optical element |
US9097891B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses including an auto-brightness control for the display brightness based on the brightness in the environment |
US9285589B2 (en) | 2010-02-28 | 2016-03-15 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered control of AR eyepiece applications |
US9341843B2 (en) | 2010-02-28 | 2016-05-17 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a small scale image source |
US20120249797A1 (en) | 2010-02-28 | 2012-10-04 | Osterhout Group, Inc. | Head-worn adaptive display |
EP2539759A1 (en) | 2010-02-28 | 2013-01-02 | Osterhout Group, Inc. | Local advertising content on an interactive head-mounted eyepiece |
US9223134B2 (en) | 2010-02-28 | 2015-12-29 | Microsoft Technology Licensing, Llc | Optical imperfections in a light transmissive illumination system for see-through near-eye display glasses |
US9091851B2 (en) | 2010-02-28 | 2015-07-28 | Microsoft Technology Licensing, Llc | Light control in head mounted displays |
US9182596B2 (en) | 2010-02-28 | 2015-11-10 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with the optical assembly including absorptive polarizers or anti-reflective coatings to reduce stray light |
US9097890B2 (en) | 2010-02-28 | 2015-08-04 | Microsoft Technology Licensing, Llc | Grating in a light transmissive illumination system for see-through near-eye display glasses |
US8488246B2 (en) | 2010-02-28 | 2013-07-16 | Osterhout Group, Inc. | See-through near-eye display glasses including a curved polarizing film in the image source, a partially reflective, partially transmitting optical element and an optically flat film |
US8482859B2 (en) | 2010-02-28 | 2013-07-09 | Osterhout Group, Inc. | See-through near-eye display glasses wherein image light is transmitted to and reflected from an optically flat film |
US9759917B2 (en) | 2010-02-28 | 2017-09-12 | Microsoft Technology Licensing, Llc | AR glasses with event and sensor triggered AR eyepiece interface to external devices |
US9229227B2 (en) | 2010-02-28 | 2016-01-05 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a light transmissive wedge shaped illumination system |
US8467133B2 (en) | 2010-02-28 | 2013-06-18 | Osterhout Group, Inc. | See-through display with an optical assembly including a wedge-shaped illumination system |
US20150309316A1 (en) | 2011-04-06 | 2015-10-29 | Microsoft Technology Licensing, Llc | Ar glasses with predictive control of external device based on event input |
US8610771B2 (en) | 2010-03-08 | 2013-12-17 | Empire Technology Development Llc | Broadband passive tracking for augmented reality |
JP4913913B2 (en) * | 2010-04-28 | 2012-04-11 | 新日鉄ソリューションズ株式会社 | Information processing system, information processing method, and program |
KR101909742B1 (en) | 2010-06-15 | 2018-10-18 | 티켓마스터 엘엘씨 | Methods and systems for computer aided event and venue setup and modeling and interactive maps |
US9781170B2 (en) | 2010-06-15 | 2017-10-03 | Live Nation Entertainment, Inc. | Establishing communication links using routing protocols |
US10096161B2 (en) | 2010-06-15 | 2018-10-09 | Live Nation Entertainment, Inc. | Generating augmented reality images using sensor and location data |
US9605857B2 (en) | 2010-08-11 | 2017-03-28 | Evergreen Telemetry Llc | Wireless sensors system and method of using same |
US8860760B2 (en) | 2010-09-25 | 2014-10-14 | Teledyne Scientific & Imaging, Llc | Augmented reality (AR) system and method for tracking parts and visually cueing a user to identify and locate parts in a scene |
US8561897B2 (en) | 2010-11-18 | 2013-10-22 | Sky-Trax, Inc. | Load tracking utilizing load identifying indicia and spatial discrimination |
US9277248B1 (en) | 2011-01-26 | 2016-03-01 | Amdocs Software Systems Limited | System, method, and computer program for receiving device instructions from one user to be overlaid on an image or video of the device for another user |
US8188880B1 (en) | 2011-03-14 | 2012-05-29 | Google Inc. | Methods and devices for augmenting a field of view |
US8601380B2 (en) * | 2011-03-16 | 2013-12-03 | Nokia Corporation | Method and apparatus for displaying interactive preview information in a location-based user interface |
US8738754B2 (en) | 2011-04-07 | 2014-05-27 | International Business Machines Corporation | Systems and methods for managing computing systems utilizing augmented reality |
US8913086B2 (en) * | 2011-04-07 | 2014-12-16 | International Business Machines Corporation | Systems and methods for managing errors utilizing augmented reality |
US9477348B2 (en) | 2011-04-15 | 2016-10-25 | Hewlett-Packard Development Company, L.P. | Focus-based touch and hover detection |
US9965564B2 (en) * | 2011-07-26 | 2018-05-08 | Schneider Electric It Corporation | Apparatus and method of displaying hardware status using augmented reality |
US8963956B2 (en) | 2011-08-19 | 2015-02-24 | Microsoft Technology Licensing, Llc | Location based skins for mixed reality displays |
KR20220032059A (en) | 2011-09-19 | 2022-03-15 | 아이사이트 모빌 테크놀로지 엘티디 | Touch free interface for augmented reality systems |
US8941560B2 (en) | 2011-09-21 | 2015-01-27 | Google Inc. | Wearable computer with superimposed controls and instructions for external device |
US9255813B2 (en) | 2011-10-14 | 2016-02-09 | Microsoft Technology Licensing, Llc | User controlled real object disappearance in a mixed reality display |
US8761811B2 (en) * | 2012-04-27 | 2014-06-24 | Oracle International Corporation | Augmented reality for maintenance management, asset management, or real estate management |
US9519640B2 (en) * | 2012-05-04 | 2016-12-13 | Microsoft Technology Licensing, Llc | Intelligent translations in personal see through display |
US9583032B2 (en) * | 2012-06-05 | 2017-02-28 | Microsoft Technology Licensing, Llc | Navigating content using a physical object |
AT513130B1 (en) | 2012-08-24 | 2014-02-15 | Tgw Logistics Group Gmbh | Method and system for assisting a worker in a facility for manipulating goods |
JP6061586B2 (en) * | 2012-09-26 | 2017-01-18 | 株式会社アマダホールディングス | Support system and method |
US9019174B2 (en) | 2012-10-31 | 2015-04-28 | Microsoft Technology Licensing, Llc | Wearable emotion detection and feedback system |
US20140152530A1 (en) * | 2012-12-03 | 2014-06-05 | Honeywell International Inc. | Multimedia near to eye display system |
MX2012015026A (en) * | 2012-12-18 | 2014-06-20 | Ruben Librado Rodriguez Chavana | System for the interaction of customized graphic articles in real time by detecting a printed reference marker. |
EP2936445A2 (en) | 2012-12-20 | 2015-10-28 | Accenture Global Services Limited | Context based augmented reality |
US20140204121A1 (en) * | 2012-12-27 | 2014-07-24 | Schlumberger Technology Corporation | Augmented reality for oilfield |
US9812046B2 (en) | 2013-01-10 | 2017-11-07 | Microsoft Technology Licensing, Llc | Mixed reality display accommodation |
US9449340B2 (en) * | 2013-01-30 | 2016-09-20 | Wal-Mart Stores, Inc. | Method and system for managing an electronic shopping list with gestures |
US9082149B2 (en) * | 2013-02-19 | 2015-07-14 | Wal-Mart Stores, Inc. | System and method for providing sales assistance to a consumer wearing an augmented reality device in a physical store |
US9959190B2 (en) | 2013-03-12 | 2018-05-01 | International Business Machines Corporation | On-site visualization of component status |
JP6160290B2 (en) | 2013-06-21 | 2017-07-12 | 富士通株式会社 | Information processing apparatus, determination method, and determination program |
KR102165444B1 (en) * | 2013-08-28 | 2020-10-14 | 엘지전자 주식회사 | Apparatus and Method for Portable Device displaying Augmented Reality image |
US9404904B2 (en) * | 2013-11-05 | 2016-08-02 | The Boeing Company | Methods and systems for non-destructive inspection |
AU2013273829A1 (en) | 2013-12-23 | 2015-07-09 | Canon Kabushiki Kaisha | Time constrained augmented reality |
JP6393986B2 (en) * | 2013-12-26 | 2018-09-26 | セイコーエプソン株式会社 | Head-mounted display device, image display system, and method for controlling head-mounted display device |
US9523853B1 (en) * | 2014-02-20 | 2016-12-20 | Google Inc. | Providing focus assistance to users of a head mounted display |
US9626709B2 (en) | 2014-04-16 | 2017-04-18 | At&T Intellectual Property I, L.P. | In-store field-of-view merchandising and analytics |
WO2015160515A1 (en) | 2014-04-16 | 2015-10-22 | Exxonmobil Upstream Research Company | Methods and systems for providing procedures in real-time |
TWI645230B (en) | 2014-08-03 | 2018-12-21 | 帕戈技術股份有限公司 | Wearable camera systems and apparatus and method for attaching camera systems or other electronic devices to wearable articles |
US9635222B2 (en) | 2014-08-03 | 2017-04-25 | PogoTec, Inc. | Wearable camera systems and apparatus for aligning an eyewear camera |
US10134049B2 (en) | 2014-11-20 | 2018-11-20 | At&T Intellectual Property I, L.P. | Customer service based upon in-store field-of-view and analytics |
CN107003135B (en) * | 2014-12-09 | 2020-04-21 | 旭化成株式会社 | Position and orientation detection device and position and orientation detection program |
MX2017008519A (en) | 2014-12-23 | 2018-06-27 | Pogotec Inc | Wireless camera system and methods. |
US9869996B2 (en) * | 2015-01-08 | 2018-01-16 | The Boeing Company | System and method for using an internet of things network for managing factory production |
GB2535727A (en) * | 2015-02-25 | 2016-08-31 | Bae Systems Plc | Interactive information system |
US11809383B2 (en) * | 2015-03-26 | 2023-11-07 | Invisible Holdings, Llc | Social identity of objects |
US9390346B1 (en) * | 2015-06-10 | 2016-07-12 | Robert Edwin Sheets, Jr. | Object location using a mobile device |
AU2016274951A1 (en) | 2015-06-10 | 2018-01-04 | PogoTec, Inc. | Eyewear with magnetic track for electronic wearable device |
US10481417B2 (en) | 2015-06-10 | 2019-11-19 | PogoTec, Inc. | Magnetic attachment mechanism for electronic wearable device |
US20180211445A1 (en) * | 2015-07-17 | 2018-07-26 | Sharp Kabushiki Kaisha | Information processing device, terminal, and remote communication system |
US11172273B2 (en) | 2015-08-10 | 2021-11-09 | Delta Energy & Communications, Inc. | Transformer monitor, communications and data collection device |
US10055869B2 (en) | 2015-08-11 | 2018-08-21 | Delta Energy & Communications, Inc. | Enhanced reality system for visualizing, evaluating, diagnosing, optimizing and servicing smart grids and incorporated components |
WO2017041093A1 (en) | 2015-09-03 | 2017-03-09 | Delta Energy & Communications, Inc. | System and method for determination and remediation of energy diversion in a smart grid network |
EP3139247A1 (en) * | 2015-09-03 | 2017-03-08 | Siemens Aktiengesellschaft | Method of and system for performing buyoff operations in a computer-managed production facility |
EP3757916A1 (en) | 2015-09-09 | 2020-12-30 | Dematic Corp. | Heads up display for material handling systems |
JP2017062748A (en) * | 2015-09-25 | 2017-03-30 | 富士通株式会社 | Alignment apparatus, alignment method, and alignment program |
MX2018004053A (en) | 2015-10-02 | 2018-12-17 | Delta Energy & Communications Inc | Supplemental and alternative digital data delivery and receipt mesh network realized through the placement of enhanced transformer mounted monitoring devices. |
US9961572B2 (en) | 2015-10-22 | 2018-05-01 | Delta Energy & Communications, Inc. | Augmentation, expansion and self-healing of a geographically distributed mesh network using unmanned aerial vehicle (UAV) technology |
WO2017070646A1 (en) | 2015-10-22 | 2017-04-27 | Delta Energy & Communications, Inc. | Data transfer facilitation across a distributed mesh network using light and optical based technology |
WO2017075405A1 (en) | 2015-10-29 | 2017-05-04 | PogoTec, Inc. | Hearing aid adapted for wireless power reception |
US10791020B2 (en) | 2016-02-24 | 2020-09-29 | Delta Energy & Communications, Inc. | Distributed 802.11S mesh network using transformer module hardware for the capture and transmission of data |
US9965841B2 (en) | 2016-02-29 | 2018-05-08 | Schneider Electric USA, Inc. | Monitoring system based on image analysis of photos |
US11558538B2 (en) | 2016-03-18 | 2023-01-17 | Opkix, Inc. | Portable camera system |
US9952432B2 (en) | 2016-04-08 | 2018-04-24 | Symbol Technologies, Llc | Arrangement for, and method of, calibrating a wearable apparatus to electro-optically read targets |
CN105929948B (en) * | 2016-04-14 | 2018-12-04 | 佛山市威格特电气设备有限公司 | Based on augmented reality self learning type intelligent helmet and its operation method |
US10210661B2 (en) | 2016-04-25 | 2019-02-19 | Microsoft Technology Licensing, Llc | Location-based holographic experience |
US10142410B2 (en) | 2016-04-29 | 2018-11-27 | Raytheon Company | Multi-mode remote collaboration |
BE1024192B1 (en) * | 2016-05-09 | 2017-12-12 | Safran Aero Boosters S.A. | METHOD AND TURBOMACHINE TEST BENCH WITH INCREASED REALITY |
WO2017195286A1 (en) * | 2016-05-11 | 2017-11-16 | 三菱電機株式会社 | Air conditioning visualization system |
US10652633B2 (en) | 2016-08-15 | 2020-05-12 | Delta Energy & Communications, Inc. | Integrated solutions of Internet of Things and smart grid network pertaining to communication, data and asset serialization, and data modeling algorithms |
US10762468B2 (en) * | 2016-09-22 | 2020-09-01 | Amazon Technologies, Inc. | Adaptive process for guiding human-performed inventory tasks |
JP6658430B2 (en) * | 2016-09-28 | 2020-03-04 | 株式会社デンソー | Inspection device |
CN107944781A (en) | 2016-10-12 | 2018-04-20 | 菜鸟智能物流控股有限公司 | Method and device for providing prompt information of stored object |
US10088902B2 (en) * | 2016-11-01 | 2018-10-02 | Oculus Vr, Llc | Fiducial rings in virtual reality |
EP3539285A4 (en) | 2016-11-08 | 2020-09-02 | Pogotec, Inc. | A smart case for electronic wearable device |
US20180211447A1 (en) * | 2017-01-24 | 2018-07-26 | Lonza Limited | Methods and Systems for Using a Virtual or Augmented Reality Display to Perform Industrial Maintenance |
DE102017108622A1 (en) * | 2017-04-23 | 2018-10-25 | Goodly Innovations GmbH | SYSTEM FOR SUPPORTING TEAMWORK BY AUGMENTED REALITY |
US20180349837A1 (en) * | 2017-05-19 | 2018-12-06 | Hcl Technologies Limited | System and method for inventory management within a warehouse |
US10685324B2 (en) * | 2017-05-19 | 2020-06-16 | Hcl Technologies Limited | Method and system for optimizing storage and retrieval of a stock keeping unit (SKU) |
US10748443B2 (en) | 2017-06-08 | 2020-08-18 | Honeywell International Inc. | Apparatus and method for visual-assisted training, collaboration, and monitoring in augmented/virtual reality in industrial automation systems and other systems |
JP6585665B2 (en) * | 2017-06-29 | 2019-10-02 | ファナック株式会社 | Virtual object display system |
DE102017213638A1 (en) * | 2017-08-07 | 2019-02-07 | Siemens Aktiengesellschaft | marker |
US10403046B2 (en) | 2017-10-20 | 2019-09-03 | Raytheon Company | Field of view (FOV) and key code limited augmented reality to enforce data capture and transmission compliance |
US11222081B2 (en) | 2017-11-27 | 2022-01-11 | Evoqua Water Technologies Llc | Off-line electronic documentation solutions |
ES2699999A1 (en) * | 2018-04-25 | 2019-02-13 | Mendez Francisco Jose Cuadrado | PROCEDURE FOR THE CREATION AND HANDLING OF MUSIC AND SOUND FROM THE INTERACTION WITH TANGIBLE OBJECTS (TUI) AND A SYSTEM OF INCREASED REALITY, WITH SPECIAL APPLICABILITY TO THE SCOPE OF TEACHING. (Machine-translation by Google Translate, not legally binding) |
US11145123B1 (en) | 2018-04-27 | 2021-10-12 | Splunk Inc. | Generating extended reality overlays in an industrial environment |
US11847773B1 (en) | 2018-04-27 | 2023-12-19 | Splunk Inc. | Geofence-based object identification in an extended reality environment |
JP6781201B2 (en) | 2018-06-05 | 2020-11-04 | ファナック株式会社 | Virtual object display system |
US11300857B2 (en) | 2018-11-13 | 2022-04-12 | Opkix, Inc. | Wearable mounts for portable camera |
US11696011B2 (en) | 2021-10-21 | 2023-07-04 | Raytheon Company | Predictive field-of-view (FOV) and cueing to enforce data capture and transmission compliance in real and near real time video |
US11792499B2 (en) | 2021-10-21 | 2023-10-17 | Raytheon Company | Time-delay to enforce data capture and transmission compliance in real and near real time video |
US11700448B1 (en) | 2022-04-29 | 2023-07-11 | Raytheon Company | Computer/human generation, validation and use of a ground truth map to enforce data capture and transmission compliance in real and near real time video of a local scene |
CN116320364B (en) * | 2023-05-25 | 2023-08-01 | 四川中绳矩阵技术发展有限公司 | Virtual reality shooting method and display method based on multi-layer display |
Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5696521A (en) * | 1994-06-22 | 1997-12-09 | Astounding Technologies (M) Sdn. Bhd. | Video headset |
US5731766A (en) * | 1995-08-18 | 1998-03-24 | Agency Of Industrial Science And Technology | Route guide system and method |
US5886822A (en) * | 1996-10-08 | 1999-03-23 | The Microoptical Corporation | Image combining system for eyeglasses and face masks |
US6046712A (en) * | 1996-07-23 | 2000-04-04 | Telxon Corporation | Head mounted communication system for providing interactive visual communications with a remote system |
US6091546A (en) * | 1997-10-30 | 2000-07-18 | The Microoptical Corporation | Eyeglass interface system |
US6094625A (en) * | 1997-07-03 | 2000-07-25 | Trimble Navigation Limited | Augmented vision for survey work and machine control |
US6111597A (en) * | 1996-12-28 | 2000-08-29 | Olympus Optical Co., Ltd. | Stereo image forming apparatus |
US6172657B1 (en) * | 1996-02-26 | 2001-01-09 | Seiko Epson Corporation | Body mount-type information display apparatus and display method using the same |
US6301050B1 (en) * | 1999-10-13 | 2001-10-09 | Optics Wireless Led, Inc. | Image enhancement system for scaled viewing at night or under other vision impaired conditions |
US6356437B1 (en) * | 1999-03-29 | 2002-03-12 | Siemens Dematic Postal Automation, L.P. | System, apparatus and method for providing a portable customizable maintenance support instruction system |
US20020046368A1 (en) * | 1999-03-02 | 2002-04-18 | Wolfgang Friedrich | System for, and method of, situation-relevant asistance to interaction with the aid of augmented-reality technologies |
US6396961B1 (en) * | 1997-11-12 | 2002-05-28 | Sarnoff Corporation | Method and apparatus for fixating a camera on a target point using image alignment |
US6522312B2 (en) * | 1997-09-01 | 2003-02-18 | Canon Kabushiki Kaisha | Apparatus for presenting mixed reality shared among operators |
US6803887B1 (en) * | 1999-07-22 | 2004-10-12 | Swisscom Mobile Ag | Method and corresponding devices for delivering useful data concerning observed objects |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2673004A1 (en) | 1991-02-15 | 1992-08-21 | X Ial | ROAD INFORMATION VISUALIZATION SET FOR THE DRIVING STATION OF A RAILWAY VEHICLE. |
US6078928A (en) * | 1997-12-12 | 2000-06-20 | Missouri Botanical Garden | Site-specific interest profiling system |
US6123259A (en) * | 1998-04-30 | 2000-09-26 | Fujitsu Limited | Electronic shopping system including customer relocation recognition |
US6826554B2 (en) * | 1998-06-29 | 2004-11-30 | Fujitsu Limited | System and method for adaptively configuring a shopping display in response to a recognized customer profile |
GB0007474D0 (en) * | 2000-03-29 | 2000-05-17 | Hewlett Packard Co | Location-Dependent User Interface |
US6396691B1 (en) * | 2000-04-17 | 2002-05-28 | Circa Telecom, Usa, Inc. | Thermal cover for T1/HDSL repeater case |
-
2001
- 2001-10-19 US US10/046,468 patent/US7126558B1/en not_active Expired - Lifetime
-
2002
- 2002-10-18 EP EP02773811.1A patent/EP1436800B1/en not_active Expired - Lifetime
- 2002-10-18 ES ES02773811.1T patent/ES2525679T3/en not_active Expired - Lifetime
- 2002-10-18 US US10/492,437 patent/US7372451B2/en not_active Expired - Lifetime
- 2002-10-18 WO PCT/US2002/033466 patent/WO2003034397A1/en not_active Application Discontinuation
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5696521A (en) * | 1994-06-22 | 1997-12-09 | Astounding Technologies (M) Sdn. Bhd. | Video headset |
US5731766A (en) * | 1995-08-18 | 1998-03-24 | Agency Of Industrial Science And Technology | Route guide system and method |
US6172657B1 (en) * | 1996-02-26 | 2001-01-09 | Seiko Epson Corporation | Body mount-type information display apparatus and display method using the same |
US6046712A (en) * | 1996-07-23 | 2000-04-04 | Telxon Corporation | Head mounted communication system for providing interactive visual communications with a remote system |
US5886822A (en) * | 1996-10-08 | 1999-03-23 | The Microoptical Corporation | Image combining system for eyeglasses and face masks |
US6111597A (en) * | 1996-12-28 | 2000-08-29 | Olympus Optical Co., Ltd. | Stereo image forming apparatus |
US6094625A (en) * | 1997-07-03 | 2000-07-25 | Trimble Navigation Limited | Augmented vision for survey work and machine control |
US6522312B2 (en) * | 1997-09-01 | 2003-02-18 | Canon Kabushiki Kaisha | Apparatus for presenting mixed reality shared among operators |
US6091546A (en) * | 1997-10-30 | 2000-07-18 | The Microoptical Corporation | Eyeglass interface system |
US6396961B1 (en) * | 1997-11-12 | 2002-05-28 | Sarnoff Corporation | Method and apparatus for fixating a camera on a target point using image alignment |
US20020046368A1 (en) * | 1999-03-02 | 2002-04-18 | Wolfgang Friedrich | System for, and method of, situation-relevant asistance to interaction with the aid of augmented-reality technologies |
US20020067372A1 (en) * | 1999-03-02 | 2002-06-06 | Wolfgang Friedrich | Utilizing augmented reality-based technologies to provide situation-related assistance to a skilled operator from remote experts |
US6356437B1 (en) * | 1999-03-29 | 2002-03-12 | Siemens Dematic Postal Automation, L.P. | System, apparatus and method for providing a portable customizable maintenance support instruction system |
US6803887B1 (en) * | 1999-07-22 | 2004-10-12 | Swisscom Mobile Ag | Method and corresponding devices for delivering useful data concerning observed objects |
US6301050B1 (en) * | 1999-10-13 | 2001-10-09 | Optics Wireless Led, Inc. | Image enhancement system for scaled viewing at night or under other vision impaired conditions |
Cited By (43)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060271292A1 (en) * | 2003-12-02 | 2006-11-30 | Buck Michael A | Mobile Geographical Information Display System |
US8225226B2 (en) * | 2003-12-31 | 2012-07-17 | Abb Research Ltd. | Virtual control panel |
US20090300535A1 (en) * | 2003-12-31 | 2009-12-03 | Charlotte Skourup | Virtual control panel |
US7573525B2 (en) * | 2004-02-06 | 2009-08-11 | Olympus Corporation | Camera and photographing method for setting focal distance of photographing optical system so as to correspond to information that indicates photographic range inputted via an operation section |
US20050174470A1 (en) * | 2004-02-06 | 2005-08-11 | Olympus Corporation | Head-mounted camera |
US8564441B2 (en) | 2007-02-12 | 2013-10-22 | At&T Intellectual Property I, L.P. | Methods and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US8970379B2 (en) | 2007-02-12 | 2015-03-03 | At&T Intellectual Property I, L.P. | Method and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US9898770B2 (en) | 2007-02-12 | 2018-02-20 | At&T Intellectual Property I, L.P. | Method and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US10592951B2 (en) | 2007-02-12 | 2020-03-17 | At&T Intellectual Property I, L.P. | Method and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US7639138B2 (en) | 2007-02-12 | 2009-12-29 | At&T Intellectual Property I, L.P. | Methods and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US7986239B2 (en) | 2007-02-12 | 2011-07-26 | At&T Intellectual Property I, L.P. | Methods and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US9411996B2 (en) | 2007-02-12 | 2016-08-09 | At&T Intellectual Property I, L.P. | Method and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US20080191846A1 (en) * | 2007-02-12 | 2008-08-14 | Wayne Chang | Methods and apparatus to visualize locations of radio frequency identification (rfid) tagged items |
WO2008100668A1 (en) * | 2007-02-12 | 2008-08-21 | At & T Knowledge Ventures, L.P. | Methods and apparatus to visualize locations of radio frequency identification (rfid) tagged items |
US8237564B2 (en) | 2007-02-12 | 2012-08-07 | At&T Intellectual Property I, L.P. | Methods and apparatus to visualize locations of radio frequency identification (RFID) tagged items |
US20080266323A1 (en) * | 2007-04-25 | 2008-10-30 | Board Of Trustees Of Michigan State University | Augmented reality user interaction system |
US9076149B2 (en) * | 2007-06-08 | 2015-07-07 | Shopper Scientist Llc | Shopper view tracking and analysis system and method |
US20080306756A1 (en) * | 2007-06-08 | 2008-12-11 | Sorensen Associates Inc | Shopper view tracking and analysis system and method |
US20090154293A1 (en) * | 2007-12-18 | 2009-06-18 | Anandraj Sengupta | System and method for augmented reality inspection and data visualization |
US8485038B2 (en) | 2007-12-18 | 2013-07-16 | General Electric Company | System and method for augmented reality inspection and data visualization |
US20140111346A1 (en) * | 2009-06-29 | 2014-04-24 | Commscope, Inc. Of North Carolina | Methods of Automatically Recording Patching Changes at Passive Patch Panels and Network Equipment |
US9123217B2 (en) * | 2009-06-29 | 2015-09-01 | Commscope, Inc. Of North Carolina | Methods of automatically recording patching changes at passive patch panels and network equipment |
US9811494B2 (en) | 2009-06-29 | 2017-11-07 | Commscope Inc. Of North Carolina | Methods of automatically recording patching changes at passive patch panels and network equipment |
US10372651B2 (en) * | 2009-06-29 | 2019-08-06 | Commscope, Inc. Of North Carolina | Methods of automatically recording patching changes at passive patch panels and network equipment |
US9338525B2 (en) | 2009-06-29 | 2016-05-10 | Commscope, Inc. Of North Carolina | Methods of automatically recording patching changes at passive patch panels and network equipment |
US20120078686A1 (en) * | 2010-09-27 | 2012-03-29 | Bashani Gilad G | Parking status system |
US20120154195A1 (en) * | 2010-12-17 | 2012-06-21 | Sony Ericsson Mobile Communications Ab | System and method for remote controlled device selection |
US8963694B2 (en) * | 2010-12-17 | 2015-02-24 | Sony Corporation | System and method for remote controlled device selection based on device position data and orientation data of a user |
AU2013200187A9 (en) * | 2011-07-22 | 2013-07-25 | Microsoft Technology Licensing, Llc | Automatic text scrolling on a head-mounted display |
AU2013200187B9 (en) * | 2011-07-22 | 2013-07-25 | Microsoft Technology Licensing, Llc | Automatic text scrolling on a head-mounted display |
AU2013200187B2 (en) * | 2011-07-22 | 2013-06-20 | Microsoft Technology Licensing, Llc | Automatic text scrolling on a head-mounted display |
WO2013085927A1 (en) * | 2011-12-07 | 2013-06-13 | Sheridan Martin Small | Updating printed content with personalized virtual data |
US9229231B2 (en) | 2011-12-07 | 2016-01-05 | Microsoft Technology Licensing, Llc | Updating printed content with personalized virtual data |
US9182815B2 (en) | 2011-12-07 | 2015-11-10 | Microsoft Technology Licensing, Llc | Making static printed content dynamic with virtual data |
US9183807B2 (en) | 2011-12-07 | 2015-11-10 | Microsoft Technology Licensing, Llc | Displaying virtual data as printed content |
US9165381B2 (en) | 2012-05-31 | 2015-10-20 | Microsoft Technology Licensing, Llc | Augmented books in a mixed reality environment |
US9361501B2 (en) * | 2013-04-01 | 2016-06-07 | Ncr Corporation | Headheld scanner and POS display with mobile phone |
US20140291396A1 (en) * | 2013-04-01 | 2014-10-02 | Ncr Corporation | Headheld Scanner and POS Display with Mobile Phone |
US20160132532A1 (en) * | 2014-11-11 | 2016-05-12 | Commscope, Inc. Of North Carolina | Multi-tier intelligent infrastructure management systems for communications systems and related equipment and methods |
US9870773B2 (en) * | 2014-11-11 | 2018-01-16 | Commscope, Inc. Of North Carolina | Multi-tier intelligent infrastructure management systems for communications systems and related equipment and methods |
US10262656B2 (en) | 2014-11-11 | 2019-04-16 | Commscope Inc. Of North Carolina | Multi-tier intelligent infrastructure management systems for communications systems and related equipment and methods |
GB2549874B (en) * | 2015-02-13 | 2022-03-23 | Halliburton Energy Services Inc | Distributing information using role-specific augmented reality devices |
DE102017219067A1 (en) * | 2017-10-25 | 2019-04-25 | Bayerische Motoren Werke Aktiengesellschaft | DEVICE AND METHOD FOR THE VISUAL SUPPORT OF A USER IN A WORKING ENVIRONMENT |
Also Published As
Publication number | Publication date |
---|---|
US7126558B1 (en) | 2006-10-24 |
EP1436800B1 (en) | 2014-09-17 |
EP1436800A1 (en) | 2004-07-14 |
ES2525679T3 (en) | 2014-12-29 |
WO2003034397A1 (en) | 2003-04-24 |
EP1436800A4 (en) | 2005-12-21 |
US20040183751A1 (en) | 2004-09-23 |
US7372451B2 (en) | 2008-05-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7126558B1 (en) | Industrial augmented reality | |
KR102451653B1 (en) | System and method for dynamic inventory management | |
US9299013B1 (en) | Visual task feedback for workstations in materials handling facilities | |
US10282696B1 (en) | Augmented reality enhanced interaction system | |
US20200143315A1 (en) | Inventory Management System With Position Determination of Storage Goods and Associated Storage Regions | |
US11315073B1 (en) | Event aspect determination | |
US20040046711A1 (en) | User-controlled linkage of information within an augmented reality system | |
WO2019006116A1 (en) | Methods and systems for automatically mapping a retail location | |
US11000953B2 (en) | Robot gamification for improvement of operator performance | |
US20160045276A1 (en) | Systems and methods for identifying instruments | |
JP7268155B2 (en) | Customer assistance robot picking | |
WO2016051185A1 (en) | System and method for surveying display units in a retail store | |
Etonam et al. | Augmented reality (AR) application in manufacturing encompassing quality control and maintenance | |
CA2962142A1 (en) | System and method for monitoring display unit compliance | |
EP3929855A1 (en) | Store system, status determination method, and program | |
CN114897455A (en) | Unmanned warehouse management system and method based on display enhancement technology | |
GB2530770A (en) | System and method for monitoring display unit compliance | |
Singh et al. | AIoT-based e-commerce | |
EP3929854A1 (en) | Commodity management assisting system, commodity management assisting method, and program | |
Colabella et al. | Applications of Augmented Reality to Reduce Risk in Supply Chain Management | |
JP2021087029A (en) | Position detection system, position detection device, and position detection method | |
EP4086831A1 (en) | Ultra-wideband, real time locating system | |
KR20220110416A (en) | Work information management system for ar glasses that can be used in an offline environment using nfc tags | |
US20230169684A1 (en) | System and Method Using a System | |
WO2020195763A1 (en) | Information processing device, information processing method, and recording medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ACCENTURE GLOBAL SERVICES GMBH, SWITZERLAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DEMPSKI, KELLY L.;REEL/FRAME:012733/0997 Effective date: 20020301 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
CC | Certificate of correction | ||
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: ACCENTURE GLOBAL SERVICES LIMITED, IRELAND Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ACCENTURE GLOBAL SERVICES GMBH;REEL/FRAME:025700/0287 Effective date: 20100901 |
|
FPAY | Fee payment |
Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553) Year of fee payment: 12 |