Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20050204438 A1
Publication typeApplication
Application numberUS 10/962,829
Publication date15 Sep 2005
Filing date11 Oct 2004
Priority date26 Feb 2004
Also published asUS20100115418, US20110301759
Publication number10962829, 962829, US 2005/0204438 A1, US 2005/204438 A1, US 20050204438 A1, US 20050204438A1, US 2005204438 A1, US 2005204438A1, US-A1-20050204438, US-A1-2005204438, US2005/0204438A1, US2005/204438A1, US20050204438 A1, US20050204438A1, US2005204438 A1, US2005204438A1
InventorsYulun Wang, Charles Jordan, Jonathan Southard, Marco Pinter
Original AssigneeYulun Wang, Jordan Charles S., Jonathan Southard, Marco Pinter
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Graphical interface for a remote presence system
US 20050204438 A1
Abstract
A robot system that includes a robot and a remote station. The remote station may be a personal computer coupled to the robot through a broadband network. A user at the remote station may receive both video and audio from a camera and microphone of the robot, respectively. The remote station may include a display user interface that has a variety of viewable fields and selectable buttons.
Images(7)
Previous page
Next page
Claims(100)
1. A robot system, comprising:
a mobile robot that has a camera that captures an image; and, a remote station that is coupled to said robot, said remote station includes a display user interface that displays the image and provides a first graphical input that can be selected to vary the image.
2. The system of claim 1, wherein said mobile robot can be controlled from said remote station.
3. The system of claim 1, wherein said first graphical input can be selected to view a still picture image.
4. The system of claim 1, wherein said first graphical input is a slide bar that can be selected to view a plurality of still picture images.
5. The system of claim 1, wherein a still picture image provided by said mobile robot camera can be stored at said remote station by selecting said first graphical input.
6. The system of claim 1, wherein selecting said first graphical input can initiate a storage of a video segment of said image.
7. The system of claim 1, wherein said mobile robot includes a battery and said display user interface depicts how much energy is left in said battery.
8. The system of claim 1, further comprising a base station that is coupled to said remote station and wirelessly coupled to said mobile robot, said display user interface depicts a signal strength of a signal transmitted between said base station and said mobile robot.
9. The system of claim 1, further comprising a second graphical input that can be selected to vary a characteristic of the image.
10. The system of claim 1, wherein mobile robot includes a microphone and said remote station includes a speaker, said display user interface including a second graphical input that can be selected to vary a characteristic of sound provided by said microphone.
11. A method for operating robot system, comprising:
moving a mobile robot that has a camera that captures an image;
transmitting the image to a remote station; and,
displaying a display user interface at the remote station, the display user interface displays an image and provides a first graphical input that can be selected to vary the image.
12. The method of claim 11, wherein movement of the mobile robot is controlled by the remote station.
13. The method of claim 11, further comprising selecting the first graphical input to display a still picture image.
14. The method of claim 11, wherein the first graphical input is a slide bar, and further comprising selecting said slide bar to view a plurality of still picture images.
15. The method of claim 11, further comprising selecting the first graphical input to store a still picture image provided by the mobile robot camera.
16. The method of claim 11, further comprising selecting said first graphical input to initiate a storage of a video segment of said image.
17. The method of claim 11, wherein the mobile robot includes a battery and the display user interface depicts how much energy is left in the battery.
18. The method of claim 11, further comprising transmitting a signal between the mobile robot and a base station that is coupled to the remote station and wirelessly coupled to said mobile robot, and displaying a signal strength of a signal on the display user interface.
19. The method of claim 11, further comprising selecting a second graphical input displayed by the display user interface to vary a characteristic of the image.
20. The method of claim 11, wherein the mobile robot includes a microphone and the remote station includes a speaker, and further comprising selecting a second graphical input to vary a characteristic of sound provided by said microphone.
21. A robot system, comprising:
a mobile robot;
a base station that transmits a wireless signal to said mobile robot; and,
a remote station that is coupled to said mobile robot through said base station, said remote station includes a display user interface that displays a signal strength of the wireless signal.
22. The system of claim 21, wherein said display user interface displays an image and a first graphical input that can be selected to vary the image.
23. The system of claim 22, wherein the image is provided by a camera of said mobile robot.
24. The system of claim 22, wherein said first graphical input can be selected to view a still picture image.
25. The system of claim 22, wherein said first graphical input is a slide bar that can be selected to view a plurality of still picture images.
26. The system of claim 22, wherein a still picture image provided by said mobile robot camera can be stored at said remote station by selecting said first graphical input.
27. The system of claim 21, wherein selecting said first graphical input can initiate a storage of a video segment of said image.
28. The system of claim 21, wherein said mobile robot includes a battery and said display user interface depicts how much energy is left in said battery.
29. The system of claim 22, wherein said display user interface includes a second graphical input that can be selected to vary a characteristic of the image.
30. The system of claim 21, wherein mobile robot includes a microphone and said remote station includes a speaker, said display user interface includes a graphical input that can be selected to vary a characteristic of sound provided by said microphone.
31. A method for operating a robot system, comprising:
transmitting a wireless signal from a base station to a mobile robot;
moving the mobile robot; and,
displaying a display user interface at a remote station that is coupled to the base station, the display user interface displays a signal strength of the wireless signal.
32. The method of claim 31, wherein the display user interface displays an image and a first graphical input that can be selected to vary the image.
33. The method of claim 32, further comprising selecting the first graphical input to display a still picture image.
34. The method of claim 32, wherein the first graphical input is a slide bar and further comprising selecting the slide bar to view a plurality of still picture images.
35. The method of claim 32, further comprising selecting the first graphical input to store a still picture image provided by the mobile robot camera.
36. The method of claim 31, further comprising selecting said first graphical input to initiate a storage of a video segment of the image.
37. The method of claim 31, wherein movement of the mobile robot is controlled by the remote station.
38. The method of claim 31, wherein the mobile robot includes a battery and the display user interface depicts an amount of energy within the battery.
39. The method of claim 32, further comprising selecting a second graphical input displayed by the graphical user interface to vary a characteristic of the image.
40. The method of claim 31, wherein the mobile robot includes a microphone and the remote station includes a speaker, and further comprising selecting a graphical input to vary a characteristic of sound provided by said microphone.
41. A robot system, comprising:
a mobile robot that contains a battery; and,
a remote station that is coupled to said mobile robot, said remote station includes a display user interface that displays an amount of energy within said battery.
42. The system of claim 41, wherein said display user interface displays an image and a first graphical input that can be selected to vary the image.
43. The system of claim 42, wherein said image is provided by a camera of said mobile robot.
44. The system of claim 42, wherein said first graphical input can be selected to view a still picture image.
45. The system of claim 42, wherein said first graphical input is a slide bar that can be selected to view a plurality of still picture images.
46. The system of claim 42, wherein a still picture image provided by said mobile robot camera can be stored at said remote station by selecting said first graphical input.
47. The system of claim 41, wherein selecting said first graphical input can initiate a storage of a video segment of said image.
48. The system of claim 42, wherein said display user interface includes a second graphical input that can be selected to vary a characteristic of the image.
49. The system of claim 41, wherein mobile robot includes a microphone and said remote station includes a speaker, said display user interface including a graphical input that can be selected to vary a characteristic of sound provided by said microphone.
50. A method for operating a robot system, comprising:
moving a mobile robot that has a battery; and,
presenting a display user interface at a remote station, the display user interface displays an amount of energy within the battery.
51. The method of claim 50, wherein the display user interface displays an image and a first graphical input selected to vary the image.
52. The method of claim 51, further comprising selecting the first graphical input to display a still picture image.
53. The method of claim 51, wherein the first graphical input is a slide bar and further comprising selecting the slide bar to view a plurality of still picture images.
54. The method of claim 50, further comprising selecting the first graphical input to store a still picture image provided by the mobile robot camera.
55. The method of claim 51, further comprising selecting said first graphical input to initiate a storage of a video segment of said image.
56. The method of claim 51, wherein movement of the mobile robot is controlled by the remote station.
57. The method of claim 51, further comprising selecting a graphical input displayed by the display user interface to vary a characteristic of the image.
58. The method of claim 51, wherein the mobile robot includes a microphone and the remote station includes a speaker, and further comprising selecting a second graphical input to vary a characteristic of sound provided by said microphone.
59. A robot system, comprising:
a mobile robot with a sensor that provides an indication of when said robot is in proximity to an object; and,
a remote station that is coupled to said mobile robot, said remote station includes a display user interface that displays a graphical representation of when said sensor provides said indication.
60. The system of claim 59, wherein said display user interface displays an image provided by said mobile robot.
61. The system of claim 59, wherein said display user interface displays a graphical feature that indicates a direction of movement of said mobile robot.
62. The system of claim 59, wherein said display user interface displays a graphical input that can be selected to transmit an image to said mobile robot.
63. The system of claim 60, wherein a user can annotate the image.
64. The system of claim 59, wherein said mobile robot includes a screen that displays an image and a user can annotate the image.
65. The system of claim 59, wherein said display user interface includes a graphical input that can be selected to select a mobile robot from a plurality of mobile robots.
66. A method for operating a robot system, comprising:
moving a mobile robot with a sensor that provides an indication of when the mobile robot is in proximity to an object; and,
presenting a display user interface at a remote station, the display user interface displays a graphical representation of when the sensor provides the indication.
67. The method of claim 66, wherein the display user interface displays an image.
68. The method of claim 66, further comprising displaying a graphical feature that indicates a direction of movement of the mobile robot.
69. The method of claim 66, further comprising selecting a graphical input to transmit an image to the mobile robot.
70. The method of claim 67, further comprising annotating the image.
71. The method of claim 66, further comprising selecting a graphical input to select a mobile robot from a plurality of mobile robots.
72. A robot system, comprising:
a mobile robot; and,
a remote station that is coupled to said mobile robot, said remote station includes a display user interface that displays a graphical feature that indicates a direction of movement of said mobile robot.
73. The system of claim 72, wherein said display user interface displays an image provided by said mobile robot.
74. The system of claim 72, wherein said display user interface displays a graphical input that can be selected to transmit an image to said mobile robot.
75. The system of claim 73, wherein a user can annotate the image.
76. The system of claim 72, wherein said mobile robot includes a screen that displays an image and a user can annotate the image.
77. The system of claim 72, wherein said display user interface includes a graphical input that can be selected to select a mobile robot from a plurality of mobile robots.
78. A method for operating a robot system, comprising:
moving a mobile robot; and,
presenting a display user interface at a remote station, the display user interface displays a graphical feature that indicates a direction of movement of the mobile robot.
79. The method of claim 78, wherein the display user interface displays an image.
80. The method of claim 78, further comprising selecting a graphical input to transmit an image to the mobile robot.
81. The method of claim 79, further comprising annotating the image.
82. The method of claim 78, further comprising selecting a graphical input to select a mobile robot from a plurality of mobile robots.
83. A robot system, comprising:
a mobile robot that has a screen; and,
a remote station that is coupled to said mobile robot, said remote station includes a display user interface that displays a graphical input that can be selected to transmit an image to said mobile robot.
84. The system of claim 83, wherein said display user interface displays an image provided by said mobile robot.
85. The system of claim 84, wherein a user can annotate the image.
86. The system of claim 83, wherein said robot screen displays the image and a user can annotate the image.
87. The system of claim 82, wherein said display user interface includes a graphical input that can be selected to select a mobile robot from a plurality of mobile robots.
88. A method for operating a robot system, comprising:
moving a mobile robot that has a screen; and,
presenting a display user interface at a remote station, the display user interface displays a graphical input that can be selected to transmit an image to the mobile robot.
89. The method of claim 88, wherein the display user interface displays an image.
90. The method of claim 88, further comprising annotating the image.
91. The method of claim 88, wherein the image is displayed by a robot screen and a user can annotate the image.
92. The method of claim 88, further comprising selecting a graphical input to select a robot from a plurality of robots.
93. A robot system, comprising:
a plurality of mobile robots; and,
a remote station that is coupled to at least one of said mobile robots, said remote station includes a display user interface that displays a graphical input that can be selected to select one of said mobile robots.
94. The system of claim 93, wherein said display user interface displays an image provided by said mobile robot.
95. The system of claim 94, wherein a user can annotate the image.
96. The system of claim 93, wherein said mobile robot includes a screen that displays an image and a user can annotate the image.
97. A method for operating a robot system, comprising:
providing a plurality of mobile robots; and,
presenting a display user interface at a remote station, the display user interface displays a graphical input that can be selected to select one of the mobile robots.
98. The method of claim 97, wherein the display user interface displays an image.
99. The method of claim 97, further comprising annotating the image.
100. The method of claim 97, wherein the image is displayed by a robot screen and a user can annotate the image.
Description
REFERENCE TO CROSS-RELATED APPLICATIONS

This application claims priority to Application No. 60/548,561 filed on Feb. 26, 2004.

BACKGROUND OF THE INVENTION

1. Field of the Invention

The subject matter disclosed generally relates to the field of mobile two-way teleconferencing.

2. Background Information

There is a growing need to provide remote health care to patients that have a variety of ailments ranging from Alzheimers to stress disorders. To minimize costs it is desirable to-provide home care for such patients. Home care typically requires a periodic visit by a health care provider such as a nurse or some type of assistant. Due to financial and/or staffing issues the health care provider may not be there when the patient needs some type of assistance. Additionally, existing staff must be continuously trained, which can create a burden on training personnel. It would be desirable to provide a system that would allow a health care provider to remotely care for a patient without being physically present.

Robots have been used in a variety of applications ranging from remote control of hazardous material to assisting in the performance of surgery. For example, U.S. Pat. No. 5,762,458 issued to Wang et al. discloses a system that allows a surgeon to perform minimally invasive medical procedures through the use of robotically controlled instruments. One of the robotic arms in the Wang system moves an endoscope that has a camera. The camera allows a surgeon to view a surgical area of a patient.

Tele-robots such as hazardous waste handlers and bomb detectors may contain a camera that allows the operator to view the remote site. Canadian Pat. No. 2289697 issued to Treviranus, et al. discloses a teleconferencing platform that has both a camera and a monitor. The platform includes mechanisms to both pivot and raise the camera and monitor. The Treviranus patent also discloses embodiments with a mobile platform, and different mechanisms to move the camera and the monitor.

There has been marketed a mobile robot introduced by InTouch-Health, Inc., the assignee of this application, under the trademarks COMPANION and RP-6. The InTouch robot is controlled by a user at a remote station. The remote station may be a personal computer with a joystick that allows the user to remotely control the movement of the robot. Both the robot and remote station have cameras, monitors, speakers and microphones to allow for two-way video/audio communication.

U.S. patent application Pub. No. US 2001/0054071 filed in the name of Loeb, discloses a video-conferencing system that includes a number of graphical user interfaces (“GUIs”) that can be used to establish a video-conference. One of the GUIs has an icon that can be selected to make a call. The Loeb application discloses stationary video-conferencing equipment such as a television. There is no discussion in Loeb about the use of robotics.

BRIEF SUMMARY OF THE INVENTION

A robot system that includes a remote station and a robot. The remote station includes a display user interface that can be used to operate the system.

BRIEF DESCRIPTION OF THE DRAWINGS

FIG. 1 is an illustration of a robotic system;

FIG. 2 is a schematic of an electrical system of a robot;

FIG. 3 is a further schematic of the electrical system of the robot;

FIG. 4 is a display user interface of a remote station;

FIG. 5 is a display user interface showing an electronic medical record;

FIG. 6 is a display user interface showing an image and an electronic medical record being simultaneously displayed.

the robot 12 may have a direct connection to the network thru for example a satellite.

The remote control station 16 may include a computer 22 that has a monitor 24, a camera 26, a microphone 28 and a speaker 30. The computer 22 may also contain an input device 32 such as a joystick or a mouse. The control station 16 is typically located in a place that is remote from the robot 12. Although only one remote control station 16 is shown, the system 10 may include a plurality of remote stations. In general any number of robots 12 may be controlled by any number of remote stations 16 or other robots 12. For example, one remote station 16 may be coupled to a plurality of robots 12, or one robot 12 may be coupled to a plurality of remote stations 16, or a plurality of robots 12.

Each robot 12 includes a movement platform 34 that is attached to a robot housing 36. Also attached to the robot housing 36 are a camera 38, a monitor 40, a microphone(s) 42 and a speaker(s) 44. The microphone 42 and speaker 30 may create a stereophonic sound. The robot 12 may also have an antenna 46 that is wirelessly coupled to an antenna 48 of the base station 14. The system 10 allows a user at the remote control station 16 to move the robot 12 through operation of the input device 32. The robot camera 38 is coupled to the remote monitor 24 so that a user at the remote station 16 can view a patient. Likewise, the robot monitor 40 is coupled to the remote camera 26 so that the patient may view the user. The microphones 28 and 42, and speakers 30 and 44, allow for audible communication between the patient and the user.

The remote station computer 22 may operate Microsoft OS software and WINDOWS XP or other operating systems such as LINUX. The remote computer 22 may also operate a video driver, a camera driver, an audio driver and a joystick driver. The video images may be transmitted and received with compression software such as MPEG CODEC.

The robot 12 may be coupled to one or more medical monitoring devices 50. The medical monitoring device 50 can take medical data from a patient. By of example, the medical monitoring device 50 may be a stethoscope, a pulse oximeter and/or an EKG monitor. The medical monitoring device 50 may contain a wireless transmitter 52 that transmits the patient data to the robot 12. The wirelessly transmitted data may be received by antennae 46, or a separate antennae (not shown). The robot 12 can then transmit the data to the remote station 16.

The wireless transmission from the medical monitoring device 50 may be in accord with various wireless standards such as IEEE. The standard used to transmit data from the medical monitoring device 50 should not interfere with the wireless communication between the robot 12 and the base station 14. Although wireless transmission is shown and described, it is to be understood that the medical monitoring device 50 can be coupled to the robot 12 by wires (not shown).

The remote station 16 may be coupled to a server 54 through the network 18. The server 54 may contain electronic medical records of a patient. By way of example, the electronic medical records may include written records of treatment, patient history, medication information, x-rays, EKGs, laboratory results, physician notes, etc. The medical records can be retrieved from the server 54 and displayed by the monitor 24 of the remote station. In lieu of, or in addition to, the medical records can be stored in the mobile robot 12. The remote station 16 may allow the physician to modify the records and then store the modified records back in the server 54 and/or robot 12.

FIGS. 2 and 3 show an embodiment of a robot 12. Each robot 12 may include a high level control system 60 and a low level control system 62. The high level control system 60 may include a processor 64 that is connected to a bus 66. The bus is coupled to the camera 38 by an input/output (I/O) port 68, and to the monitor 40 by a serial output port 70 and a VGA driver 72. The monitor 40 may include a touchscreen function that allows the patient to enter input by touching the monitor screen.

The speaker 44 is coupled to the bus 66 by a digital to analog converter 74. The microphone 42 is coupled to the bus 66 by an analog to digital converter 76. The high level controller 60 may also contain random access memory (RAM) device 78, a non-volatile RAM device 80 and a mass storage device 82 that are all coupled to the bus 72. The mass storage device 82 may contain medical files of the patient that can be accessed by the user at the remote control station 16. For example, the mass storage device 82 may contain a picture of the patient. The user, particularly a health care provider, can recall the old picture and make a side by side comparison on the monitor 24 with a present video image of the patient provided by the camera 38. The robot antennae 46 may be coupled to a wireless transceiver 84. By way of example, the transceiver 84 may transmit and receive information in accordance with IEEE 802.11b. The transceiver 84 may also process signals from the medical monitoring device in accordance with IEEE also known as Bluetooth. The robot may have a separate antennae to receive the wireless signals from the medical monitoring device.

The controller 64 may operate with a LINUX OS operating system. The controller 64 may also operate MS WINDOWS along with video, camera and audio drivers for communication with the remote control station 16. Video information may be transceived using MPEG CODEC compression techniques. The software may allow the user to send e-mail to the patient and vice versa, or allow the patient to access the Internet. In general the high level controller 60 operates to control communication between the robot 12 and the remote control station 16.

The high level controller 60 may be linked to the low level controller 62 by serial ports 86 and 88. The low level controller 62 includes a processor 90 that is coupled to a RAM device 92 and non-volatile RAM device 94 by a bus 96. Each robot 12 contains a plurality of motors 98 and motor encoders 100. The motors 98 can activate the movement platform and move other parts of the robot such as the monitor and camera. The encoders 100 provide feedback information regarding the output of the motors 98. The motors 98 can be coupled to the bus 96 by a digital to analog converter 102 and a driver amplifier 104. The encoders 100 can be coupled to the bus 96 by a decoder 106. Each robot 12 also has a number of proximity sensors 108 (see also FIG. 1). The position sensors 108 can be coupled to the bus 96 by a signal conditioning circuit 110 and an analog to digital converter 112.

The low level controller 62 runs software routines that mechanically actuate the robot 12. For example, the low level controller 62 provides instructions to actuate the movement platform to move the robot 12. The low level controller 62 may receive movement instructions from the high level controller 60. The movement instructions may be received as movement commands from the remote control station or another robot. Although two controllers are shown, it is to be understood that each robot 12 may have one controller, or more than two controllers, controlling the high and low level functions.

The various electrical devices of each robot 12 may be powered by a battery(ies) 114. The battery 114 may be recharged by a battery recharger station 116. The low level controller 62 may include a battery control circuit 118 that senses the power level of the battery 114. The low level controller 62 can sense when the power falls below a threshold and then send a message to the high level controller 60.

The system may be the same or similar to a robotic system provided by the assignee InTouch-Health, Inc. of Santa Barbara, Calif. under the name RP-6, which is hereby incorporated by reference. The system may also be the same or similar to the system disclosed in application No. 10/206,457 published on Jan. 29, 2004, which is hereby incorporated by reference.

FIG. 4 shows a display user interface (“DUI”) 120 that can be displayed at the remote station 16 and/or the robot 12. The DUI 120 may include a robot view field 122 that displays a video image captured by the camera of the robot. The DUI 120 may also include a station view field 124 that displays a video image provided by the camera of the remote station 16. The DUI 120 may be part of an application program stored and operated by the computer 22 of the remote station 16.

The DUI 120 may include a graphic button 126 that can be selected to display an electronic medical record as shown in FIG. 5. The button 126 can be toggled to sequentially view the video image and the electronic medical record. Alternatively, the view field 122 may be split to simultaneously display both the video image and the electronic medical record as shown in FIG. 6. The viewing field may allow the physician to modify the medical record by adding, changing or deleting all or part of the record. The remote clinician can also add to the medical record still images or video captured by the camera of the robot.

The DUI 120 may have a monitor data field 128 that can display the data generated by the medical monitoring device(s) and transmitted to the remote station. The data can be added to the electronic medical record, either automatically or through user input. For example, the data can be added to a record by “dragging” a monitor data field 128 into the viewing field 122.

The DUI 120 may include alert input icons 130 and 132. Alert icon 130 can be selected by the user at the remote station to generate an alert indicator such as a sound from the speaker of the robot. Selection of the icon generates an alert input to the robot. The robot generates a sound through its speaker in response to the alert input. By way of example, the sound may simulate the noise of a horn. Consequently, the icon may have the appearance of a horn. The remote station user may select the horn shaped icon 130 while remotely moving the robot to alert persons to the presence of the moving robot.

Alert icon 132 can be selected to request access to the video images from the robot. The default state of the robot may be to not send video information to the remote station. Selecting the alert icon 132 sends an alert input such as an access request to the robot. The robot then generates an alert indicator. The alert indicator can be a sound generated by the robot speaker, and/or a visual prompt on the robot monitor. By way of example, the visual prompt may be a “flashing” graphical icon. The sound may simulate the knocking of a door. Consequently, the alert icon 128 may have the appearance of a door knocker.

In response to the alert indicator the user may provide a user input such as the depression of a button on the robot, or the selection of a graphical image on the robot monitor, to allow access to the robot camera. The robot may also have a voice recognition system that allows the user to grant access with a voice command. The user input causes the robot to begin transmitting video images from the robot camera to the remote station that requested access to the robot. A voice communication may be established before the cycle of the alert input and response, to allow the user at the remote station to talk to the caller recipient at the robot.

The DUI 120 may include a graphical “battery meter” 134 that indicates the amount of energy left in the robot battery. A graphical “signal strength meter” 136 may indicate the strength of the wireless signal transmitted between the robot and the base station (see FIG. 1).

The DUI 120 may include a location display 138 that provides the location of the robot. The CHANGE button 140 can be selected to change the default robot in a new session. The CHANGE button 140 can be used to select and control a different robot in a system that has multiple robots. The user can initiate and terminate a session by selecting box 142. The box 142 changes from CONNECT to DISCONNECT when the user selects the box to initiate a session. System settings and support can be selected through buttons 144 and 146.

Both the robot view field 122 and the station view field 124 may have associated graphics to vary the video and audio displays. Each field may have an associated graphical audio slide bar 148 to vary the audio level of the microphone and another slide bar 152 to vary the volume of the speakers.

The DUI 120 may have slide bars 150, 154 and 156 to vary the zoom, focus and brightness of the cameras, respectively. A still picture may be taken at either the robot or remote station by selecting one of the graphical camera icons 158. The still picture may be the image presented at the corresponding field 122 or 124 at the time the camera icon 158 is selected. Capturing and playing back video can be taken through graphical icons 160. A return to real time video can be resumed, after the taking of a still picture, captured video, or reviewing a slide show, by selecting a graphical LIVE button 162.

A still picture can be loaded from disk for viewing through selection of icon 164. Stored still images can be reviewed by selecting buttons 166. The number of the image displayed relative to the total number of images is shown by graphical boxes 168. The user can rapidly move through the still images in a slide show fashion or move through a captured video clip by moving the slide bar 170. A captured video image can be paused through the selection of circle 174. Play can be resumed through the same button 174. Video or still images may be dismissed from the active list through button 172. Video or still images may be transferred to the robot by selecting icon 176. For example, a doctor at the remote station may transfer an x-ray to the screen of the robot.

A graphical depiction of the base of the robot can be shown in sensor field 178. The sensor may have various sensors that sense contact with another object. The sensor field 178 can provide a visual display of the sensors that detect the object. By way of example, the field may have one or more graphical dots 180 that display where on the robot the sensors detected an object. This provides the user with a sense of the robot environment that is outside the view of the robot camera.

The graphical depiction of the robot base may contain a graphical vector overlay 182 that indicates the direction of robot movement. The direction of movement may be different than the direction the camera is facing. The vector can provide a visual aid when driving the robot.

The system may provide the ability to annotate 184 the image displayed in field 122 and/or 124. For example, a doctor at the remote station may annotate some portion of the image captured by the robot camera. The annotated image may be stored by the system. The system may also allow for annotation of images sent to the robot through icon 176. For example, a doctor may send an x-ray to the robot which is displayed by the robot screen. The doctor can annotate the x-ray to point out a portion of the x-ray to personnel located at the robot site. This can assist in allowing the doctor to instruct personnel at the robot site.

The display user interface may include graphical inputs 186 that allow the operator to turn the views of the remote station and remote cameras on and off.

While certain exemplary embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that this invention not be limited to the specific constructions and arrangements shown and described, since various other modifications may occur to those ordinarily skilled in the art.

Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US4572594 *8 Feb 198425 Feb 1986Schwartz C BruceArthroscopy support stand
US4638445 *8 Jun 198420 Jan 1987Mattaboni Paul JAutonomous mobile robot
US4652204 *2 Aug 198524 Mar 1987Arnett Edward MApparatus for handling hazardous materials
US4733737 *29 Aug 198529 Mar 1988Reza FalamakDrivable steerable platform for industrial, domestic, entertainment and like uses
US4797557 *14 May 198710 Jan 1989Aktiebolaget ElectroluxPosition sensing system for a moving object wherein a lens focuses light onto a radiation sensitive matrix
US4803625 *30 Jun 19867 Feb 1989Buddy Systems, Inc.Personal health monitor
US5084828 *29 Sep 198928 Jan 1992Healthtech Services Corp.Interactive medication delivery system
US5182641 *17 Jun 199126 Jan 1993The United States Of America As Represented By The Administrator Of The National Aeronautics And Space AdministrationComposite video and graphics display for camera viewing systems in robotics and teleoperation
US5186270 *24 Oct 199116 Feb 1993Massachusetts Institute Of TechnologyOmnidirectional vehicle
US5193143 *7 Nov 19899 Mar 1993Honeywell Inc.Problem state monitoring
US5486853 *13 Dec 199423 Jan 1996Picturetel CorporationElectrical cable interface for electronic camera
US5544649 *15 Mar 199513 Aug 1996Cardiomedix, Inc.Ambulatory patient health monitoring techniques utilizing interactive visual communication
US5553609 *9 Feb 199510 Sep 1996Visiting Nurse Service, Inc.Intelligent remote visual monitoring system for home health care service
US5594859 *21 Feb 199514 Jan 1997Digital Equipment CorporationGraphical user interface for video teleconferencing
US5600573 *2 Dec 19944 Feb 1997Discovery Communications, Inc.Operations center with video storage for a television program packaging and delivery system
US5657246 *7 Mar 199512 Aug 1997Vtel CorporationMethod and apparatus for a video conference user interface
US5758079 *7 Jun 199626 May 1998Vicor, Inc.Call control in video conferencing allowing acceptance and identification of participants in a new incoming call during an active teleconference
US5857534 *5 Jun 199712 Jan 1999Kansas State University Research FoundationRobotic inspection apparatus and method
US5867653 *18 Apr 19962 Feb 1999International Business Machines CorporationMethod and apparatus for multi-cast based video conferencing
US5871451 *30 Jun 199716 Feb 1999Siemens Medical Systems, Inc.Apparatus and method for providing dual output signals in a telemetry transmitter
US5876325 *30 Sep 19972 Mar 1999Olympus Optical Co., Ltd.Surgical manipulation system
US5983263 *2 Jan 19989 Nov 1999Intel CorporationMethod and apparatus for transmitting images during a multimedia teleconference
US6036812 *8 Dec 199714 Mar 2000Automated Prescription Systems, Inc.Pill dispensing system
US6170929 *2 Dec 19989 Jan 2001Ronald H. WilsonAutomated medication-dispensing cart
US6175779 *29 Sep 199816 Jan 2001J. Todd BarrettComputerized unit dose medication dispensing cart
US6201984 *26 Jan 199513 Mar 2001International Business Machines CorporationSystem and method for augmentation of endoscopic surgery
US6346950 *20 May 199912 Feb 2002Compaq Computer CorporationSystem and method for display images using anamorphic video
US6346962 *27 Feb 199812 Feb 2002International Business Machines CorporationControl of video conferencing system with pointing device
US6507773 *14 Jun 200114 Jan 2003Sharper Image CorporationMulti-functional robot with remote and video system
US6522906 *14 Dec 199918 Feb 2003Intuitive Surgical, Inc.Devices and methods for presenting and regulating auxiliary information on an image display of a telesurgical system to assist an operator in performing a surgical procedure
US6523629 *21 Dec 200025 Feb 2003Sandia CorporationTandem mobile robot system
US6526332 *10 Jun 200225 Feb 2003Sony CorporationRobot control system and method for introducing robot control software
US6529765 *24 May 19994 Mar 2003Neutar L.L.C.Instrumented and actuated guidance fixture for sterotactic surgery
US6529802 *23 Jun 19994 Mar 2003Sony CorporationRobot and information processing system
US6532404 *1 Mar 200211 Mar 2003Colens AndreMobile robots and their control system
US6535182 *7 Dec 199818 Mar 2003Koninklijke Philips Electronics N.V.Head-mounted projection display system
US6535793 *1 May 200118 Mar 2003Irobot CorporationMethod and system for remote control of mobile robot
US6684129 *11 Apr 200227 Jan 2004Intuitive Surgical, Inc.Master having redundant degrees of freedom
US6691000 *25 Jul 200210 Feb 2004Communications Research Laboratory, Independent Administrative InstitutionRobot-arm telemanipulating system presenting auditory information
US6839612 *7 Dec 20014 Jan 2005Institute Surgical, Inc.Microwrist system for surgical procedures
US6840904 *11 Oct 200111 Jan 2005Jason GoldbergMedical monitoring device and system
US6845297 *9 Jan 200318 Jan 2005Irobot CorporationMethod and system for remote control of mobile robot
US6852107 *16 Jan 20028 Feb 2005Computer Motion, Inc.Minimally invasive surgical training using robotics and tele-collaboration
US6853878 *8 Feb 20018 Feb 2005Kabushiki Kaisha Yaskawa DenkiRobot controller
US6853880 *22 Aug 20028 Feb 2005Honda Giken Kogyo Kabushiki KaishaAutonomous action robot
US6995664 *20 Jun 20017 Feb 2006Jeffrey DarlingRemote supervision system and method
US7156809 *17 Jun 20052 Jan 2007Q-Tec Systems LlcMethod and apparatus for health and disease management combining patient data monitoring with wireless internet connectivity
US7158859 *14 May 20032 Jan 2007Intouch Technologies, Inc.5 degrees of freedom mobile robot
US7158860 *12 Sep 20032 Jan 2007Intouch Technologies, Inc.Healthcare tele-robotic system which allows parallel remote station observation
US7158861 *18 Sep 20032 Jan 2007Intouch Technologies, Inc.Tele-robotic system used to provide remote consultation services
US7161322 *18 Nov 20039 Jan 2007Intouch Technologies, Inc.Robot with a manipulator arm
US7162338 *17 Dec 20039 Jan 2007Evolution Robotics, Inc.Systems and methods for computing a relative pose for global localization in a visual simultaneous localization and mapping system
US7164969 *2 Jan 200416 Jan 2007Intouch Technologies, Inc.Apparatus and method for patient rounding with a remote controlled robot
US7171286 *12 Sep 200330 Jan 2007Intouch Technologies, Inc.Healthcare tele-robotic system with a robot that also functions as a remote station
US7174238 *2 Sep 20036 Feb 2007Stephen Eliot ZweigMobile robotic system with web server and digital radio links
US7184559 *23 Feb 200127 Feb 2007Hewlett-Packard Development Company, L.P.System and method for audio telepresence
US7321807 *22 Nov 200622 Jan 2008Abb Inc.Robotic wash cell using recycled pure water
US7391432 *5 Feb 200224 Jun 2008Fujifilm CorporationVideoconference system
US7421470 *26 Nov 20032 Sep 2008Avistar Communications CorporationMethod for real-time communication between plural users
US7433921 *26 Nov 20037 Oct 2008Avistar Communications CorporationSystem for real-time communication between plural users
US7643051 *9 Sep 20055 Jan 2010Roy Benjamin SandbergMobile video teleconferencing system and control method
US7647320 *31 May 200212 Jan 2010Peoplechart CorporationPatient directed system and method for managing medical information
US7956894 *19 Feb 20037 Jun 2011William Rex AkersApparatus and method for computerized multi-media medical and pharmaceutical data organization and transmission
US8116910 *23 Aug 200714 Feb 2012Intouch Technologies, Inc.Telepresence robot with a printer
US20010010541 *29 Mar 20012 Aug 2001Fernandez Dennis SungaIntegrated network for monitoring remote objects
US20010051881 *22 Dec 200013 Dec 2001Aaron G. FillerSystem, method and article of manufacture for managing a medical services network
US20020010596 *13 Apr 200124 Jan 2002Matory Yvedt L.Remote patient care
US20020015296 *22 Dec 20007 Feb 2002Howell Charles A.Surgical theater system having light, monitors, and cameras
US20020027597 *5 Sep 20017 Mar 2002John SachauSystem for mobile videoconferencing
US20020033880 *27 Dec 200021 Mar 2002Dong-Myung SulMethod for performing multipoint video conference in video conferencing system
US20020044201 *18 Oct 200118 Apr 2002Intel CorporationMethod and apparatus for controlling a remote video camera in a video conferencing system
US20020109770 *5 Feb 200215 Aug 2002Masahiro TeradaVideoconference system
US20020183598 *30 May 20025 Dec 2002Nobuyuki TerauraRemote care service technique, care recipient monitoring terminal for use in the technique, and program for use in the terminal
US20030030397 *20 Sep 200113 Feb 2003Simmons John CastleNatural robot control
US20030048481 *23 Aug 200213 Mar 2003Takashi KobayashiElectronic apparatus
US20030050733 *7 Sep 200113 Mar 2003Yulun WangModularity system for computer assisted surgery
US20030060808 *22 Feb 200127 Mar 2003Wilk Peter J.Telemedical method and system
US20040010344 *9 Jul 200315 Jan 2004Kawasaki Jukogyo Kabushiki KaishaRemote control method and system for robot controller
US20040012362 *26 Mar 200322 Jan 2004Shingo TsurumiElectrical charging system, electrical charging controlling method, robot apparatus, electrical charging device, electrical charging controlling program and recording medium
US20040013295 *13 Mar 200322 Jan 2004Kohtaro SabeObstacle recognition apparatus and method, obstacle recognition program, and mobile robot apparatus
US20040017475 *19 Feb 200329 Jan 2004Akers William RexApparatus and method for computerized multi-media data organization and transmission
US20040019406 *25 Jul 200229 Jan 2004Yulun WangMedical tele-robotic system
US20040024490 *16 Apr 20035 Feb 2004Mclurkin JamesSystem amd methods for adaptive control of robotic devices
US20040041604 *25 Feb 20034 Mar 2004Kizer Jade M.Phase jumping locked loop circuit
US20040107254 *26 Nov 20033 Jun 2004Collaboration Properties, Inc.Method for real-time communication between plural users
US20040107255 *26 Nov 20033 Jun 2004Collaboration Properties, Inc.System for real-time communication between plural users
US20040150725 *27 Oct 20035 Aug 2004Canon Kabushiki KaishaVideo system for use with video telephone and video conferencing
US20050000330 *1 Jul 20036 Jan 2005Schimmels William J.Adjustable blind trimming apparatus and method of operating the same
US20050003330 *2 Jul 20036 Jan 2005Mehdi AsgarinejadInteractive virtual classroom
US20050004708 *5 May 20046 Jan 2005Goldenberg Andrew A.Mobile robot hybrid communication link
US20050007445 *11 Jul 200313 Jan 2005Foote Jonathan T.Telepresence system and method for video teleconferencing
US20050013149 *20 May 200420 Jan 2005David TrossellLibrary partitioning module operable to modify a request to move the medium
US20050021182 *6 Aug 200427 Jan 2005Yulun WangMedical tele-robotic system
US20050021183 *6 Aug 200427 Jan 2005Yulun WangMedical tele-robotic system
US20050021187 *6 Aug 200427 Jan 2005Yulun WangMedical tele-robotic system
US20050021309 *6 Apr 200427 Jan 2005Vigilos, Inc.Method and process for configuring a premises for monitoring
US20050024485 *31 Jul 20033 Feb 2005Polycom, Inc.Graphical user interface for system status alert on videoconference terminal
US20050027567 *28 Jul 20043 Feb 2005Taha Amer JamilSystem and method for health care data collection and management
US20050027794 *29 Jul 20033 Feb 2005Far Touch Inc.Remote control of a wireless device using a web browser
US20050028221 *28 Jul 20033 Feb 2005Fuji Xerox Co., Ltd.Video enabled tele-presence control host
US20050035862 *12 Apr 200417 Feb 2005Wildman Timothy D.Article locating and tracking apparatus and method
US20050038416 *23 Sep 200417 Feb 2005Computer Motion, Inc.Minimally invasive surgical training using robotics and telecollaboration
US20050038564 *11 Aug 200417 Feb 2005Burick Thomas J.Robot with removable mounting elements
US20060007943 *7 Jul 200512 Jan 2006Fellman Ronald DMethod and system for providing site independent real-time multimedia transport over packet-switched networks
US20060013263 *19 Jul 200519 Jan 2006Fellman Ronald DSystem and method for clock synchronization over packet-switched networks
US20060013469 *13 Jul 200419 Jan 2006Yulun WangMobile robot with a head-based movement mapping scheme
US20060013488 *8 Aug 200319 Jan 2006Hisashi InoueInformation-embedding apparatus, encoder, tamper-detecting apparatus, information-embedding method, tamper-detecting method, and recording medium having program recorded therein for executing both of the methods
US20060020694 *14 Jul 200526 Jan 2006Prominence Networks, Inc.Administering a communication network
US20060029065 *19 Jul 20059 Feb 2006Fellman Ronald DSystem and method for low-latency content-sensitive forward error correction
US20070021871 *27 Sep 200625 Jan 2007Yulun WangMedical tele-robotic system
US20080011904 *5 Feb 200717 Jan 2008United States Of America As Represented By The Administrator Of The NasaMethod and Associated Apparatus for Capturing, Servicing, and De-Orbiting Earth Satellites Using Robotics
US20090030552 *12 Feb 200329 Jan 2009Japan Science And Technology AgencyRobotics visual and auditory system
US20090055023 *23 Aug 200726 Feb 2009Derek WaltersTelepresence robot with a printer
US20100010672 *10 Jul 200814 Jan 2010Yulun WangDocking system for a tele-presence robot
US20100010673 *11 Jul 200814 Jan 2010Yulun WangTele-presence robot system with multi-cast features
US20100019715 *17 Apr 200828 Jan 2010David Bjorn RoeMobile tele-presence system with a microphone system
US20120023506 *20 Jul 201026 Jan 2012Apple Inc.Maintaining Data States Upon Forced Exit
AU1216200A * Title not available
JP2002321180A * Title not available
WO2000025516A1 *22 Oct 19994 May 2000Vtel CorpGraphical menu items for a user interface menu of a video teleconferencing system
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US72119805 Jul 20061 May 2007Battelle Energy Alliance, LlcRobotic follow system and method
US75840205 Jul 20061 Sep 2009Battelle Energy Alliance, LlcOccupancy change detection system and method
US75872605 Jul 20068 Sep 2009Battelle Energy Alliance, LlcAutonomous navigation system and method
US76204775 Jul 200617 Nov 2009Battelle Energy Alliance, LlcRobotic intelligence kernel
US76430519 Sep 20055 Jan 2010Roy Benjamin SandbergMobile video teleconferencing system and control method
US76686215 Jul 200623 Feb 2010The United States Of America As Represented By The United States Department Of EnergyRobotic guarded motion system and method
US7761185 *3 Oct 200620 Jul 2010Intouch Technologies, Inc.Remote presence display through remotely controlled robot
US776949222 Feb 20063 Aug 2010Intouch Technologies, Inc.Graphical interface for a remote presence system
US78016445 Jul 200621 Sep 2010Battelle Energy Alliance, LlcGeneric robot architecture
US8073564 *5 Jul 20066 Dec 2011Battelle Energy Alliance, LlcMulti-robot control interface
US8306638 *30 Nov 20056 Nov 2012The Invention Science Fund I, LlcMote presentation affecting
US8352072 *7 Aug 20088 Jan 2013Wave Group Ltd.System for extending the observation, surveillance, and navigational capabilities of a robot
US852585324 Jul 20123 Sep 2013Google Inc.Methods and systems for generating a layered display of a device
US86700174 Mar 201011 Mar 2014Intouch Technologies, Inc.Remote presence system including a cart that supports a robot face and an overhead camera
US890227825 Jul 20122 Dec 2014Intouch Technologies, Inc.Systems and methods for visualizing and managing telepresence devices in healthcare networks
US901484822 Feb 201121 Apr 2015Irobot CorporationMobile robot system
US908997216 Jan 201428 Jul 2015Intouch Technologies, Inc.Remote presence system including a cart that supports a robot face and an overhead camera
US909861114 Mar 20134 Aug 2015Intouch Technologies, Inc.Enhanced video interaction for a user interface of a telepresence network
US20070122783 *10 Oct 200631 May 2007Habashi Nader MOn-line healthcare consultation services system and method of using same
US20100010673 *11 Jul 200814 Jan 2010Yulun WangTele-presence robot system with multi-cast features
US20110035054 *7 Aug 200810 Feb 2011Wave Group Ltd.System for Extending The Observation, Surveillance, and Navigational Capabilities of a Robot
US20110246551 *20 Jun 20086 Oct 2011Space Software Italia S.P.A.Adaptive multifunction mission system
US20120209431 *16 Aug 2012Wright Timothy CRobotic based health care system
US20130218346 *21 Aug 201222 Aug 2013Timothy D. RootMethod & apparatus for remotely operating a robotic device linked to a communications network
WO2006127297A2 *11 May 200630 Nov 2006Intouch Technologies IncTelerobotic system with a dual application screen presentation
WO2009128997A1 *9 Mar 200922 Oct 2009Intouch Technologies, Inc.A robotic based health care system
Classifications
U.S. Classification455/67.7, 901/1
International ClassificationH04B17/00, B25J5/00
Cooperative ClassificationB25J9/1689, B25J5/00, B25J11/009
European ClassificationB25J11/00S2, B25J9/16T4, B25J5/00
Legal Events
DateCodeEventDescription
5 May 2005ASAssignment
Owner name: INTOUCH-HEALTH, INC., CALIFORNIA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WANG, YULAN;JORDAN, CHARLES S.;SOUTHARD, JONATHAN;AND OTHERS;REEL/FRAME:016529/0987;SIGNING DATES FROM 20050405 TO 20050414
15 Jun 2005ASAssignment
Owner name: INTOUCH TECHNOLOGIES, INC., CALIFORNIA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTOUCH HEALTH, INC.;REEL/FRAME:016686/0356
Effective date: 20050531