WO2012050897A1 - Head-mounted text display system and method for the hearing impaired - Google Patents

Head-mounted text display system and method for the hearing impaired Download PDF

Info

Publication number
WO2012050897A1
WO2012050897A1 PCT/US2011/053713 US2011053713W WO2012050897A1 WO 2012050897 A1 WO2012050897 A1 WO 2012050897A1 US 2011053713 W US2011053713 W US 2011053713W WO 2012050897 A1 WO2012050897 A1 WO 2012050897A1
Authority
WO
WIPO (PCT)
Prior art keywords
hearing impaired
text
spoken
user
head
Prior art date
Application number
PCT/US2011/053713
Other languages
French (fr)
Inventor
Mahmoud M. Ghulman
Original Assignee
32211.01 Patent Trust
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 32211.01 Patent Trust filed Critical 32211.01 Patent Trust
Publication of WO2012050897A1 publication Critical patent/WO2012050897A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B21/00Teaching, or communicating with, the blind, deaf or mute
    • G09B21/009Teaching or communicating with deaf persons
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/014Head-up displays characterised by optical features comprising information/image processing systems
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • G02B2027/0178Eyeglass type
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2380/00Specific applications
    • G09G2380/08Biomedical applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present invention relates to devices to assist the hearing impaired, and particularly to a head-mounted text display system and method for the hearing impaired that uses a speech-to-text system or speech recognition system to convert speech into a visual textual display that is displayed to the user on a head-mounted display in passages containing a selected number of words.
  • BACKGROUND ART Devices that provide visual cues to hearing impaired persons are known. Such visual devices are typically mounted upon a pair of spectacles to be worn by the hearing impaired person. These devices are typically provided for live performances and are wired into a centralized hub for delivering text or visual cues to the wearer throughout the performance. Such devices, though, typically have limited display capabilities and are not synchronized to the actual speech of the performance. Accordingly, there remains a need to provide sufficient information within a wearer's field of view, which can be synchronized with a performance or presentation.
  • heads-up displays for pilots and the like are known.
  • Such systems are bulky, complicated and expensive, and are generally limited to providing parametric information, such as speed, range, fuel, and the like.
  • Such devices fail to provide sequences of several words that can be synchronized to a performance or presentation being viewed by the wearer.
  • Other considerations such as the aesthetic undesirability of using a bulky heads-up display in a classroom, movie theater or the like, also prevents such devices from being commercially acceptable. Therefore, conventional heads-up displays fail to address the needs of hearing-impaired persons or those wishing to view a performance or presentation in a language other than that in which the presentation is being made.
  • a head-mounted text display system and method for the hearing impaired solving the aforementioned problems is desired.
  • the head-mounted text display system for the hearing impaired is a speech-to-text system in which spoken words are converted into a visual textual display and displayed to the user in passages containing a selected number of words.
  • the head-mounted text display system for the hearing impaired includes a head-mounted visual display, such as eyeglass- type dual liquid crystal displays (dual LCDs) or the like, and a controller.
  • the controller includes an audio receiver, such as a microphone or the like, for receiving spoken language and converting the spoken language into electrical signals representative of the spoken language.
  • the controller further includes a speech-to-text module for converting the electrical signals representative of the spoken language to a textual data signal representative of individual words.
  • a receiver is in communication with the head-mounted visual display, and a transmitter associated with the controller transmits the textual data signal to the receiver.
  • the textual data representative of the individual words is then displayed to the user in passages containing a selected number of individual words, e.g., a display of three words at a time.
  • the controller further includes memory containing a database of video data representative of individual words, such as graphical depictions of sign language. Following speech-to-text conversion, the controller further matches each word to a corresponding visual image in the database.
  • the textual data signal and the corresponding video data are transmitted simultaneously to the receiver, and the textual data and the corresponding video images may then be displayed simultaneously to the user.
  • Fig. 1 is an environmental, perspective view of a head-mounted text display system for the hearing impaired according to the present invention.
  • Fig. 2A is a front view of an exemplary visual display presented to the user by the head- mounted text display system for the hearing impaired of Fig. 1.
  • Fig. 2B is a front view of an exemplary subsequent visual display presented to the user by the head-mounted text display system for the hearing impaired following the display shown in Fig. 2A, Figs. 2A and 2B representing a single spoken phrase.
  • Fig. 3 is a block diagram illustrating elements of a controller of the head-mounted text display system for the hearing impaired according to the present invention.
  • Fig. 4 is a perspective view of a head-mounted display of the head-mounted text display system for the hearing impaired according to the present invention.
  • the head- mounted text display system for the hearing impaired 10 is a speech-to-text system in which spoken words are converted into a visual textual display and displayed to the user in passages containing a selected number of words.
  • the head-mounted text display system for the hearing impaired 10 includes a head-mounted visual display 12 and a controller 14.
  • the head-mounted visual display 12 is shown as an eyeglass-type dual liquid crystal display (dual LCD).
  • such a display 12 includes a pair of liquid crystal displays D, mounted in an eyeglass type frame, with each display D covering a respective one of the user's eyes.
  • Such displays are well known in the field of virtual reality displays.
  • MYVU® Shades 301 manufactured by the MicroOptical Corporation of Westwood, Massachusetts.
  • a similar display is shown in PCT patent application WO 99/23524, published on May 14, 1999 to the MicroOptical Corporation, which is hereby incorporated by reference in its entirety. It should be understood that any suitable type of visual display may be utilized.
  • the controller 14 includes an audio receiver 20, such as a microphone or the like, for receiving spoken language and converting the spoken language into electrical signals representative of the spoken language. It should be understood that any suitable type of audio receiver, microphone or sensor may be used. Further, although shown as being body- mounted in Fig. 1, it should be understood that the controller 14 may be a stand-alone unit (i.e., not carried by the user), or may be integrated into the head-mounted display 12.
  • the controller 14 further includes a speech-to-text module 44 for converting the electrical signals (produced by microphone 20) representative of the spoken language to a textual data signal representative of individual words.
  • the speech-to- text module 44 may be a stand-alone unit, or may be in the form of speech recognition software stored in computer readable memory 46 and executable by the processor 48.
  • Speech-to-text systems and modules are well known in the art, and it should be understood that any suitable type of speech-to-text system or module may be utilized. Examples of such systems are shown in U.S. Patent Nos. 5,475,798; 5,857,099; and 7,047,191, each of which is herein incorporated by reference in its entirety.
  • the controller 14 preferably includes a processor 48 in communication with computer readable memory 46.
  • the speech-to-text module 44 may be a stand-alone unit in communication with processor 48 and memory 46, or may be in the form of software stored in memory 46 and implemented by the processor 48.
  • Speech-to-text or speech recognition software is well known in the art, and any suitable such software may be utilized. An example of such software is Dragon Naturally Speaking, manufactured by Nuance® Communications, LLC of Burlington, Massachusetts.
  • controller 14 may be, or may incorporate, any suitable computer system or controller, such as that diagrammatically shown in Fig. 3.
  • Data may be entered into the controller 14 by any suitable type of user interface, along with the input signal generated by the microphone 20, and may be stored in memory 46, which may be any suitable type of computer readable and programmable memory.
  • processor 48 which may be any suitable type of computer processor, microprocessor, microcontroller, digital signal processor, or the like, and may be transmitted to the head-mounted display 12 by any suitable type of wireless transmitter 16, which is preferably a wireless transmitter.
  • the processor 48 may be associated with, or incorporated into, any suitable type of computing device, for example, a personal computer or a programmable logic controller.
  • the transmitter 16, the microphone 20, the speech-to-text module 44, the processor 48, the memory 46 and any associated computer readable recording media are in communication with one another by any suitable type of data bus, as is well known in the art.
  • Examples of computer-readable recording media include a magnetic recording apparatus, an optical disk, a magneto-optical disk, and/or a semiconductor memory (for example, RAM, ROM, etc.).
  • Examples of magnetic recording apparatus that may be used in addition to memory 46, or in place of memory 46, include a hard disk device (HDD), a flexible disk (ED), and a magnetic tape (MT).
  • Examples of the optical disk include a DVD (Digital Versatile Disc), a DVD-RAM, a CD-ROM (Compact Disc-Read Only Memory), and a CD-R (Recordable)/RW.
  • the wireless signal S containing the textual data generated by transmitter 16 is received by a receiver 18 in communication with the head-mounted visual display 12.
  • the textual data representative of the individual words is then displayed to the user in passages containing a selected number of individual words, e.g., a display of three words at a time.
  • exemplary three-word passages 30, 32, respectively are shown being displayed on a display D.
  • the words are presented to the user three words at a time, allowing the user to easily read each passage, regardless of the speed in which the original speaker speaks the spoken language or the display speed of the particular head- mounted display device.
  • the memory 46 of controller 14 includes a database of video data representative of individual words, such as graphical depictions of sign language.
  • the processor 48 of controller 14 further matches each word to a corresponding visual image in the database.
  • the textual data signal and the corresponding video data are transmitted simultaneously to the receiver 18, and the textual data and the corresponding video images may then be displayed simultaneously to the user.
  • a sign language display 40 is shown adjacent the textual displays 30, 32.
  • the graphical display 40 allows for simultaneous display of sign language with the textual display.
  • the user may selectively display only text, only the graphical display, or both simultaneously.
  • the system 10 may also provide translation capability.
  • the speech-to-text subsystem may be in

Abstract

The head-mounted text display system for the hearing impaired (10) is a speech-to-text system, in which spoken words are converted into a visual textual display and displayed to the user in passages containing a selected number of words. The system includes a head-mounted visual display (12), such as eyeglass-type dual liquid crystal displays (D) or the like, and a controller (14). The controller (14) includes an audio receiver (20), such as a microphone or the like, for receiving spoken language and converting the spoken language into electrical signals. The controller (14) further includes a speech-to-text module (44) for converting the electrical signals representative of the spoken language to a textual data signal (S) representative of individual words. A transmitter (16) associated with the controller (14) transmits the textual data signal (S) to a receiver (18) associated with the head-mounted display (12).

Description

HEAD-MOUNTED TEXT DISPLAY SYSTEM AND METHOD
FOR THE HEARING IMPAIRED TECHNICAL FIELD
The present invention relates to devices to assist the hearing impaired, and particularly to a head-mounted text display system and method for the hearing impaired that uses a speech-to-text system or speech recognition system to convert speech into a visual textual display that is displayed to the user on a head-mounted display in passages containing a selected number of words.
BACKGROUND ART Devices that provide visual cues to hearing impaired persons are known. Such visual devices are typically mounted upon a pair of spectacles to be worn by the hearing impaired person. These devices are typically provided for live performances and are wired into a centralized hub for delivering text or visual cues to the wearer throughout the performance. Such devices, though, typically have limited display capabilities and are not synchronized to the actual speech of the performance. Accordingly, there remains a need to provide sufficient information within a wearer's field of view, which can be synchronized with a performance or presentation.
Additionally, heads-up displays for pilots and the like are known. However, such systems are bulky, complicated and expensive, and are generally limited to providing parametric information, such as speed, range, fuel, and the like. Such devices fail to provide sequences of several words that can be synchronized to a performance or presentation being viewed by the wearer. Other considerations, such as the aesthetic undesirability of using a bulky heads-up display in a classroom, movie theater or the like, also prevents such devices from being commercially acceptable. Therefore, conventional heads-up displays fail to address the needs of hearing-impaired persons or those wishing to view a performance or presentation in a language other than that in which the presentation is being made. Thus, a head-mounted text display system and method for the hearing impaired solving the aforementioned problems is desired.
DISCLOSURE OF INVENTION The head-mounted text display system for the hearing impaired is a speech-to-text system in which spoken words are converted into a visual textual display and displayed to the user in passages containing a selected number of words. The head-mounted text display system for the hearing impaired includes a head-mounted visual display, such as eyeglass- type dual liquid crystal displays (dual LCDs) or the like, and a controller. The controller includes an audio receiver, such as a microphone or the like, for receiving spoken language and converting the spoken language into electrical signals representative of the spoken language.
The controller further includes a speech-to-text module for converting the electrical signals representative of the spoken language to a textual data signal representative of individual words. A receiver is in communication with the head-mounted visual display, and a transmitter associated with the controller transmits the textual data signal to the receiver. The textual data representative of the individual words is then displayed to the user in passages containing a selected number of individual words, e.g., a display of three words at a time.
Preferably, the controller further includes memory containing a database of video data representative of individual words, such as graphical depictions of sign language. Following speech-to-text conversion, the controller further matches each word to a corresponding visual image in the database. The textual data signal and the corresponding video data are transmitted simultaneously to the receiver, and the textual data and the corresponding video images may then be displayed simultaneously to the user.
These and other features of the present invention will become readily apparent upon further review of the following specification and drawings.
BRIEF DESCRIPTION OF DRAWINGS
Fig. 1 is an environmental, perspective view of a head-mounted text display system for the hearing impaired according to the present invention.
Fig. 2A is a front view of an exemplary visual display presented to the user by the head- mounted text display system for the hearing impaired of Fig. 1.
Fig. 2B is a front view of an exemplary subsequent visual display presented to the user by the head-mounted text display system for the hearing impaired following the display shown in Fig. 2A, Figs. 2A and 2B representing a single spoken phrase.
Fig. 3 is a block diagram illustrating elements of a controller of the head-mounted text display system for the hearing impaired according to the present invention. Fig. 4 is a perspective view of a head-mounted display of the head-mounted text display system for the hearing impaired according to the present invention.
Similar reference characters denote corresponding features consistently throughout the attached drawings.
BEST MODES FOR CARRYING OUT THE INVENTION
The head- mounted text display system for the hearing impaired 10 is a speech-to-text system in which spoken words are converted into a visual textual display and displayed to the user in passages containing a selected number of words. As shown in Fig. 1, the head- mounted text display system for the hearing impaired 10 includes a head-mounted visual display 12 and a controller 14. In Fig. 1, the head-mounted visual display 12 is shown as an eyeglass-type dual liquid crystal display (dual LCD). As best shown in Fig. 4, such a display 12 includes a pair of liquid crystal displays D, mounted in an eyeglass type frame, with each display D covering a respective one of the user's eyes. Such displays are well known in the field of virtual reality displays. One such display is the MYVU® Shades 301, manufactured by the MicroOptical Corporation of Westwood, Massachusetts. A similar display is shown in PCT patent application WO 99/23524, published on May 14, 1999 to the MicroOptical Corporation, which is hereby incorporated by reference in its entirety. It should be understood that any suitable type of visual display may be utilized.
The controller 14 includes an audio receiver 20, such as a microphone or the like, for receiving spoken language and converting the spoken language into electrical signals representative of the spoken language. It should be understood that any suitable type of audio receiver, microphone or sensor may be used. Further, although shown as being body- mounted in Fig. 1, it should be understood that the controller 14 may be a stand-alone unit (i.e., not carried by the user), or may be integrated into the head-mounted display 12.
As best shown in Fig. 3, the controller 14 further includes a speech-to-text module 44 for converting the electrical signals (produced by microphone 20) representative of the spoken language to a textual data signal representative of individual words. The speech-to- text module 44 may be a stand-alone unit, or may be in the form of speech recognition software stored in computer readable memory 46 and executable by the processor 48.
Speech-to-text systems and modules are well known in the art, and it should be understood that any suitable type of speech-to-text system or module may be utilized. Examples of such systems are shown in U.S. Patent Nos. 5,475,798; 5,857,099; and 7,047,191, each of which is herein incorporated by reference in its entirety. The controller 14 preferably includes a processor 48 in communication with computer readable memory 46. As noted above, the speech-to-text module 44 may be a stand-alone unit in communication with processor 48 and memory 46, or may be in the form of software stored in memory 46 and implemented by the processor 48. Speech-to-text or speech recognition software is well known in the art, and any suitable such software may be utilized. An example of such software is Dragon Naturally Speaking, manufactured by Nuance® Communications, LLC of Burlington, Massachusetts.
It should be understood that the controller 14 may be, or may incorporate, any suitable computer system or controller, such as that diagrammatically shown in Fig. 3. Data may be entered into the controller 14 by any suitable type of user interface, along with the input signal generated by the microphone 20, and may be stored in memory 46, which may be any suitable type of computer readable and programmable memory. Calculations and processing are performed by a processor 48, which may be any suitable type of computer processor, microprocessor, microcontroller, digital signal processor, or the like, and may be transmitted to the head-mounted display 12 by any suitable type of wireless transmitter 16, which is preferably a wireless transmitter.
The processor 48 may be associated with, or incorporated into, any suitable type of computing device, for example, a personal computer or a programmable logic controller. The transmitter 16, the microphone 20, the speech-to-text module 44, the processor 48, the memory 46 and any associated computer readable recording media are in communication with one another by any suitable type of data bus, as is well known in the art.
Examples of computer-readable recording media include a magnetic recording apparatus, an optical disk, a magneto-optical disk, and/or a semiconductor memory (for example, RAM, ROM, etc.). Examples of magnetic recording apparatus that may be used in addition to memory 46, or in place of memory 46, include a hard disk device (HDD), a flexible disk (ED), and a magnetic tape (MT). Examples of the optical disk include a DVD (Digital Versatile Disc), a DVD-RAM, a CD-ROM (Compact Disc-Read Only Memory), and a CD-R (Recordable)/RW.
The wireless signal S containing the textual data generated by transmitter 16 is received by a receiver 18 in communication with the head-mounted visual display 12. The textual data representative of the individual words is then displayed to the user in passages containing a selected number of individual words, e.g., a display of three words at a time. In Figs. 2A and 2B, exemplary three-word passages 30, 32, respectively, are shown being displayed on a display D. As shown, the words are presented to the user three words at a time, allowing the user to easily read each passage, regardless of the speed in which the original speaker speaks the spoken language or the display speed of the particular head- mounted display device.
Preferably, the memory 46 of controller 14 includes a database of video data representative of individual words, such as graphical depictions of sign language. Following speech-to-text conversion, the processor 48 of controller 14 further matches each word to a corresponding visual image in the database. The textual data signal and the corresponding video data are transmitted simultaneously to the receiver 18, and the textual data and the corresponding video images may then be displayed simultaneously to the user. In Figs. 2 A and 2B, a sign language display 40 is shown adjacent the textual displays 30, 32. The graphical display 40 allows for simultaneous display of sign language with the textual display. The user may selectively display only text, only the graphical display, or both simultaneously. In addition to providing the option of the graphical display, the system 10 may also provide translation capability. The speech-to-text subsystem may be in
communication with one or more databases containing language translation, allowing the user to select a particular language to be displayed to the user, independent of the language of the speaker. Such speech-to-text translation systems and software are well known in the art. An example of such a system is shown in U.S. Patent No. 7,747,434, which is herein
incorporated by reference in its entirety.
It is to be understood that the present invention is not limited to the embodiments described above, but encompasses any and all embodiments within the scope of the following claims.

Claims

CLAIMS I claim:
1. A method of visually displaying spoken text for the hearing impaired, comprising the steps of:
receiving spoken language;
converting the spoken language to textual data representative of individual words; transmitting the textual data to a receiver in communication with a visual display; and displaying the textual data to the user, wherein the textual data is displayed to the user in passages containing a selected number of individual words.
2. The method of visually displaying spoken text for the hearing impaired as recited in claim 1, further comprising the step of mounting the visual display and the receiver on the user's head.
3. The method of visually displaying spoken text for the hearing impaired as recited in claim 2, further comprising the step of covering at least one of the user's eyes with the visual display.
4. The method of visually displaying spoken text for the hearing impaired as recited in claim 3, further comprising the steps of:
converting the spoken language to video data representative of the individual words; transmitting the video data to the receiver; and
displaying the video data simultaneously with the display of the textual data, wherein the video data corresponds to the textual data being displayed to the user.
5. The method of visually displaying spoken text for the hearing impaired as recited in claim 4, wherein the step of converting the spoken language to the video data
representative of the individual words comprises converting the spoken language to a graphical representation of sign language.
6. The method of visually displaying spoken text for the hearing impaired as recited in claim 5, wherein the steps of transmitting the textual and video data to the receiver comprise wirelessly transmitting the textual and video data.
7. The method of visually displaying spoken text for the hearing impaired as recited in claim 1, wherein the step of displaying the textual data to the user comprises displaying the textual data in passages containing three words at a time.
8. A method of visually displaying spoken text for the hearing impaired, comprising the steps of:
receiving spoken language;
converting the spoken language to textual data representative of individual words; converting the spoken language to video data representative of the individual words; transmitting the textual data and the video data to a receiver in communication with a visual display; and
simultaneously displaying the textual data and the video data to the user, wherein the textual data is displayed to the user in passages containing a selected number of individual words, the video data corresponding to the textual data being displayed to the user.
9. The method of visually displaying spoken text for the hearing impaired as recited in claim 8, further comprising the step of mounting the visual display and the receiver on the user's head.
10. The method of visually displaying spoken text for the hearing impaired as recited in claim 9, further comprising the step of covering at least one of the user's eyes with the visual display.
11. The method of visually displaying spoken text for the hearing impaired as recited in claim 10, wherein the step of converting the spoken language to the video data representative of the individual words comprises converting the spoken language to a graphical representation of sign language.
12. The method of visually displaying spoken text for the hearing impaired as recited in claim 11, further comprising the step of translating the spoken language into a selected second language, the textual data being displayed to the user in the second language.
13. The method of visually displaying spoken text for the hearing impaired as recited in claim 12, wherein the step of simultaneously displaying the textual data and the video data to the user comprises displaying the textual data in passages containing three words at a time.
14. A head-mounted text display system for the hearing impaired, comprising:
a head-mounted visual display;
an audio receiver having a transducer for receiving spoken language and converting the spoken language into electrical signals representative of the spoken language;
means for converting the electrical signals representative of the spoken language to a textual data signal representative of individual words;
a receiver in communication with the head-mounted visual display;
a transmitter for transmitting the textual data signal to the receiver; and means for displaying the textual data representative of the individual words to the user in passages containing a selected number of individual words.
15. The head-mounted text display system for the hearing impaired as recited in claim 14, further comprising:
means for converting the spoken language to video data representative of the individual words, the video data being transmitted to the receiver with the textual data signal; and
means for displaying the video data simultaneously with the display of the textual data, wherein the video data corresponds to the textual data being displayed to the user.
16. The head-mounted text display system for the hearing impaired as recited in claim 15, wherein the video data comprises a graphical representation of sign language.
17. The head-mounted text display system for the hearing impaired as recited in claim 15, wherein the transmitter is a wireless transmitter.
18. The head-mounted text display system for the hearing impaired as recited in claim 17, wherein the receiver is a wireless receiver.
19. The head-mounted text display system for the hearing impaired as recited in claim 18, wherein the textual data is displayed to the user in passages containing three words at a time.
20. The head-mounted text display system for the hearing impaired as recited in claim 19, further comprising means for translating the spoken language into a selected second language, the textual data being displayed to the user in the second language.
PCT/US2011/053713 2010-09-28 2011-09-28 Head-mounted text display system and method for the hearing impaired WO2012050897A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/892,711 US20120078628A1 (en) 2010-09-28 2010-09-28 Head-mounted text display system and method for the hearing impaired
US12/892,711 2010-09-28

Publications (1)

Publication Number Publication Date
WO2012050897A1 true WO2012050897A1 (en) 2012-04-19

Family

ID=45871525

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/053713 WO2012050897A1 (en) 2010-09-28 2011-09-28 Head-mounted text display system and method for the hearing impaired

Country Status (2)

Country Link
US (1) US20120078628A1 (en)
WO (1) WO2012050897A1 (en)

Families Citing this family (78)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9229233B2 (en) 2014-02-11 2016-01-05 Osterhout Group, Inc. Micro Doppler presentations in head worn computing
US9400390B2 (en) 2014-01-24 2016-07-26 Osterhout Group, Inc. Peripheral lighting for head worn computing
US9298007B2 (en) 2014-01-21 2016-03-29 Osterhout Group, Inc. Eye imaging in head worn computing
US20150205111A1 (en) 2014-01-21 2015-07-23 Osterhout Group, Inc. Optical configurations for head worn computing
US9952664B2 (en) 2014-01-21 2018-04-24 Osterhout Group, Inc. Eye imaging in head worn computing
US9965681B2 (en) 2008-12-16 2018-05-08 Osterhout Group, Inc. Eye imaging in head worn computing
US9715112B2 (en) 2014-01-21 2017-07-25 Osterhout Group, Inc. Suppression of stray light in head worn computing
JP5229209B2 (en) * 2009-12-28 2013-07-03 ブラザー工業株式会社 Head mounted display
JP5699649B2 (en) * 2011-02-04 2015-04-15 セイコーエプソン株式会社 Virtual image display device
AU2011204946C1 (en) * 2011-07-22 2012-07-26 Microsoft Technology Licensing, Llc Automatic text scrolling on a head-mounted display
AT519733B1 (en) * 2012-06-06 2019-08-15 Agfa Nv Radiation curable inkjet inks and industrial inkjet printing processes
US9966075B2 (en) 2012-09-18 2018-05-08 Qualcomm Incorporated Leveraging head mounted displays to enable person-to-person interactions
US10043535B2 (en) * 2013-01-15 2018-08-07 Staton Techiya, Llc Method and device for spectral expansion for an audio signal
US9536453B2 (en) 2013-05-03 2017-01-03 Brigham Young University Computer-implemented communication assistant for the hearing-impaired
US9424843B2 (en) * 2013-09-24 2016-08-23 Starkey Laboratories, Inc. Methods and apparatus for signal sharing to improve speech understanding
US9848260B2 (en) * 2013-09-24 2017-12-19 Nuance Communications, Inc. Wearable communication enhancement device
CN103646587B (en) * 2013-12-05 2017-02-22 北京京东方光电科技有限公司 deaf-mute people
US10191279B2 (en) 2014-03-17 2019-01-29 Osterhout Group, Inc. Eye imaging in head worn computing
US9529195B2 (en) 2014-01-21 2016-12-27 Osterhout Group, Inc. See-through computer display systems
US9299194B2 (en) 2014-02-14 2016-03-29 Osterhout Group, Inc. Secure sharing in head worn computing
US11227294B2 (en) 2014-04-03 2022-01-18 Mentor Acquisition One, Llc Sight information collection in head worn computing
US9829707B2 (en) 2014-08-12 2017-11-28 Osterhout Group, Inc. Measuring content brightness in head worn computing
US20150277118A1 (en) 2014-03-28 2015-10-01 Osterhout Group, Inc. Sensor dependent content position in head worn computing
US10254856B2 (en) 2014-01-17 2019-04-09 Osterhout Group, Inc. External user interface for head worn computing
US10684687B2 (en) 2014-12-03 2020-06-16 Mentor Acquisition One, Llc See-through computer display systems
US9575321B2 (en) 2014-06-09 2017-02-21 Osterhout Group, Inc. Content presentation in head worn computing
US9594246B2 (en) 2014-01-21 2017-03-14 Osterhout Group, Inc. See-through computer display systems
US9810906B2 (en) 2014-06-17 2017-11-07 Osterhout Group, Inc. External user interface for head worn computing
US11103122B2 (en) 2014-07-15 2021-08-31 Mentor Acquisition One, Llc Content presentation in head worn computing
US10649220B2 (en) 2014-06-09 2020-05-12 Mentor Acquisition One, Llc Content presentation in head worn computing
US9380374B2 (en) 2014-01-17 2016-06-28 Okappi, Inc. Hearing assistance systems configured to detect and provide protection to the user from harmful conditions
US9841599B2 (en) 2014-06-05 2017-12-12 Osterhout Group, Inc. Optical configurations for head-worn see-through displays
US9746686B2 (en) 2014-05-19 2017-08-29 Osterhout Group, Inc. Content position calibration in head worn computing
US20160019715A1 (en) 2014-07-15 2016-01-21 Osterhout Group, Inc. Content presentation in head worn computing
US9939934B2 (en) 2014-01-17 2018-04-10 Osterhout Group, Inc. External user interface for head worn computing
US9671613B2 (en) 2014-09-26 2017-06-06 Osterhout Group, Inc. See-through computer display systems
US9448409B2 (en) 2014-11-26 2016-09-20 Osterhout Group, Inc. See-through computer display systems
US9532714B2 (en) 2014-01-21 2017-01-03 Osterhout Group, Inc. Eye imaging in head worn computing
US9494800B2 (en) 2014-01-21 2016-11-15 Osterhout Group, Inc. See-through computer display systems
US9811159B2 (en) 2014-01-21 2017-11-07 Osterhout Group, Inc. Eye imaging in head worn computing
US9651784B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US11892644B2 (en) 2014-01-21 2024-02-06 Mentor Acquisition One, Llc See-through computer display systems
US9651788B2 (en) 2014-01-21 2017-05-16 Osterhout Group, Inc. See-through computer display systems
US20150205135A1 (en) 2014-01-21 2015-07-23 Osterhout Group, Inc. See-through computer display systems
US9766463B2 (en) 2014-01-21 2017-09-19 Osterhout Group, Inc. See-through computer display systems
US9753288B2 (en) 2014-01-21 2017-09-05 Osterhout Group, Inc. See-through computer display systems
US9836122B2 (en) 2014-01-21 2017-12-05 Osterhout Group, Inc. Eye glint imaging in see-through computer display systems
US11669163B2 (en) 2014-01-21 2023-06-06 Mentor Acquisition One, Llc Eye glint imaging in see-through computer display systems
US11737666B2 (en) 2014-01-21 2023-08-29 Mentor Acquisition One, Llc Eye imaging in head worn computing
US11487110B2 (en) 2014-01-21 2022-11-01 Mentor Acquisition One, Llc Eye imaging in head worn computing
US9846308B2 (en) 2014-01-24 2017-12-19 Osterhout Group, Inc. Haptic systems for head-worn computers
US20150241963A1 (en) 2014-02-11 2015-08-27 Osterhout Group, Inc. Eye imaging in head worn computing
US9401540B2 (en) 2014-02-11 2016-07-26 Osterhout Group, Inc. Spatial location presentation in head worn computing
WO2015143114A1 (en) * 2014-03-21 2015-09-24 Thomson Licensing Sign language translation apparatus with smart glasses as display featuring a camera and optionally a microphone
US20160187651A1 (en) 2014-03-28 2016-06-30 Osterhout Group, Inc. Safety for a vehicle operator with an hmd
US9672210B2 (en) 2014-04-25 2017-06-06 Osterhout Group, Inc. Language translation with head-worn computing
US9651787B2 (en) 2014-04-25 2017-05-16 Osterhout Group, Inc. Speaker assembly for headworn computer
US9423842B2 (en) 2014-09-18 2016-08-23 Osterhout Group, Inc. Thermal management for head-worn computer
US10853589B2 (en) 2014-04-25 2020-12-01 Mentor Acquisition One, Llc Language translation with head-worn computing
US10663740B2 (en) 2014-06-09 2020-05-26 Mentor Acquisition One, Llc Content presentation in head worn computing
JP2016033757A (en) * 2014-07-31 2016-03-10 セイコーエプソン株式会社 Display device, method for controlling display device, and program
EP3220372B1 (en) * 2014-11-12 2019-10-16 Fujitsu Limited Wearable device, display control method, and display control program
US9684172B2 (en) 2014-12-03 2017-06-20 Osterhout Group, Inc. Head worn computer display systems
USD743963S1 (en) 2014-12-22 2015-11-24 Osterhout Group, Inc. Air mouse
USD751552S1 (en) 2014-12-31 2016-03-15 Osterhout Group, Inc. Computer glasses
USD753114S1 (en) 2015-01-05 2016-04-05 Osterhout Group, Inc. Air mouse
US20160239985A1 (en) 2015-02-17 2016-08-18 Osterhout Group, Inc. See-through computer display systems
US20150319546A1 (en) * 2015-04-14 2015-11-05 Okappi, Inc. Hearing Assistance System
CH711334A2 (en) * 2015-07-15 2017-01-31 Cosson Patrick A method and apparatus for helping to understand an auditory sensory message by transforming it into a visual message.
KR102450803B1 (en) * 2016-02-11 2022-10-05 한국전자통신연구원 Duplex sign language translation apparatus and the apparatus for performing the duplex sign language translation method
JP6255524B2 (en) * 2016-06-09 2017-12-27 株式会社Qdレーザ Image projection system, image projection apparatus, image projection method, image projection program, and server apparatus
CN106125922B (en) * 2016-06-22 2023-11-07 齐齐哈尔大学 Dumb speech and spoken speech image information communication system
US10690936B2 (en) 2016-08-29 2020-06-23 Mentor Acquisition One, Llc Adjustable nose bridge assembly for headworn computer
USD864959S1 (en) 2017-01-04 2019-10-29 Mentor Acquisition One, Llc Computer glasses
US11069368B2 (en) * 2018-12-18 2021-07-20 Colquitt Partners, Ltd. Glasses with closed captioning, voice recognition, volume of speech detection, and translation capabilities
US11264035B2 (en) 2019-01-05 2022-03-01 Starkey Laboratories, Inc. Audio signal processing for automatic transcription using ear-wearable device
US11264029B2 (en) 2019-01-05 2022-03-01 Starkey Laboratories, Inc. Local artificial intelligence assistant system with ear-wearable device
WO2020250110A1 (en) * 2019-06-08 2020-12-17 Pankaj Raut A system and a method for generating a 3d visualization using mixed reality-based hmd

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5647834A (en) * 1995-06-30 1997-07-15 Ron; Samuel Speech-based biofeedback method and system
US20020087322A1 (en) * 2000-11-15 2002-07-04 Fletcher Samuel G. Method for utilizing oral movement and related events
US20080288022A1 (en) * 2003-12-22 2008-11-20 Cochlear Limited Hearing System Prostheses
US20090259277A1 (en) * 2008-02-26 2009-10-15 Universidad Autonoma Metropolitana Systems and Methods for Detecting and Using an Electrical Cochlear Response ("ECR") in Analyzing Operation of a Cochlear Stimulation System

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH07168851A (en) * 1993-12-16 1995-07-04 Canon Inc Method and device for image display
US6330540B1 (en) * 1999-05-27 2001-12-11 Louis Dischler Hand-held computer device having mirror with negative curvature and voice recognition
US7221405B2 (en) * 2001-01-31 2007-05-22 International Business Machines Corporation Universal closed caption portable receiver
US7076429B2 (en) * 2001-04-27 2006-07-11 International Business Machines Corporation Method and apparatus for presenting images representative of an utterance with corresponding decoded speech
US7746986B2 (en) * 2006-06-15 2010-06-29 Verizon Data Services Llc Methods and systems for a sign language graphical interpreter

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5647834A (en) * 1995-06-30 1997-07-15 Ron; Samuel Speech-based biofeedback method and system
US20020087322A1 (en) * 2000-11-15 2002-07-04 Fletcher Samuel G. Method for utilizing oral movement and related events
US20080288022A1 (en) * 2003-12-22 2008-11-20 Cochlear Limited Hearing System Prostheses
US20090259277A1 (en) * 2008-02-26 2009-10-15 Universidad Autonoma Metropolitana Systems and Methods for Detecting and Using an Electrical Cochlear Response ("ECR") in Analyzing Operation of a Cochlear Stimulation System

Also Published As

Publication number Publication date
US20120078628A1 (en) 2012-03-29

Similar Documents

Publication Publication Date Title
US20120078628A1 (en) Head-mounted text display system and method for the hearing impaired
US10019993B2 (en) Multi-level voice menu
Peng et al. Speechbubbles: Enhancing captioning experiences for deaf and hard-of-hearing people in group conversations
KR102002979B1 (en) Leveraging head mounted displays to enable person-to-person interactions
US20190146753A1 (en) Automatic Speech Recognition (ASR) Feedback For Head Mounted Displays (HMD)
US9519640B2 (en) Intelligent translations in personal see through display
US11068668B2 (en) Natural language translation in augmented reality(AR)
US8515728B2 (en) Language translation of visual and audio input
US20170303052A1 (en) Wearable auditory feedback device
US20170188173A1 (en) Method and apparatus for presenting to a user of a wearable apparatus additional information related to an audio scene
US20140236594A1 (en) Assistive device for converting an audio signal into a visual representation
WO2004049312A1 (en) Method and apparatus for providing an animated display with translated speech
CN114760555A (en) User configurable voice commands
CN203858414U (en) Head-worn type voice recognition projection device and system
JP2002153684A (en) Head mounted display for watching public subtitles and closed caption text in movie theater
WO2019237427A1 (en) Method, apparatus and system for assisting hearing-impaired people, and augmented reality glasses
US20120088211A1 (en) Method And System For Acquisition Of Literacy
US20170186431A1 (en) Speech to Text Prosthetic Hearing Aid
US20060183088A1 (en) Audio-visual language teaching material and audio-visual languages teaching method
JP2015041101A (en) Foreign language learning system using smart spectacles and its method
WO2019237429A1 (en) Method, apparatus and system for assisting communication, and augmented reality glasses
KR20230079846A (en) Augmented reality smart glass and method for controlling the output of smart glasses
CA2214243C (en) Communication device and method for deaf and mute persons
Nuorivaara Finnish Voice Command in Head Mounted Display Devices
Massaro et al. Optimizing visual feature perception for an automatic wearable speech supplement in face-to-face communication and classroom situations

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11833050

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11833050

Country of ref document: EP

Kind code of ref document: A1