US20110111805A1 - Synthesized audio message over communication links - Google Patents

Synthesized audio message over communication links Download PDF

Info

Publication number
US20110111805A1
US20110111805A1 US12/613,636 US61363609A US2011111805A1 US 20110111805 A1 US20110111805 A1 US 20110111805A1 US 61363609 A US61363609 A US 61363609A US 2011111805 A1 US2011111805 A1 US 2011111805A1
Authority
US
United States
Prior art keywords
end user
text
communication device
speech
far
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/613,636
Inventor
Baptiste P. Paquier
Aram M. Lindahl
Phillip G. Tamchina
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Apple Inc
Original Assignee
Apple Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Apple Inc filed Critical Apple Inc
Priority to US12/613,636 priority Critical patent/US20110111805A1/en
Assigned to APPLE INC. reassignment APPLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TAMCHINA, PHILLIP G., LINDAHL, ARAM M., PAQUIER, BAPTISTE P.
Publication of US20110111805A1 publication Critical patent/US20110111805A1/en
Priority to US14/849,433 priority patent/US9666180B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72436User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for text messaging, e.g. SMS or e-mail
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L21/0216Noise filtering characterised by the method used for estimating noise
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M3/00Automatic or semi-automatic exchanges
    • H04M3/42Systems providing special services or facilities to subscribers
    • H04M3/42382Text-based messaging services in telephone networks such as PSTN/ISDN, e.g. User-to-User Signalling or Short Message Service for fixed networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L2021/02087Noise filtering the noise being separate speech, e.g. cocktail party
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2201/00Electronic components, circuits, software, systems or apparatus used in telephone systems
    • H04M2201/40Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M2203/00Aspects of automatic or semi-automatic exchanges
    • H04M2203/45Aspects of automatic or semi-automatic exchanges related to voicemail messaging
    • H04M2203/4536Voicemail combined with text-based messaging

Definitions

  • the near-end communication device e.g., the telephone module 238 of FIG. 2
  • the near-end communication device may be configured or programmed by the user to support one or more of the above-described features.

Abstract

A communication device establishes an audio connection with a far-end user via a communication network. The communication device receives text input from a near-end user, and converts the text input into speech signals. The speech signals are transmitted to the far-end user using the established audio connection while muting audio input to its microphone. Other embodiments are also described and claimed.

Description

    FIELD
  • An embodiment of the invention relates to audio synthesis by a communication device. Other embodiments are also described.
  • BACKGROUND
  • A user of a communication device (e.g., a telephone) may sometimes have to make or answer a phone call in a noisy environment. Noise can interfere with a phone conversation to a degree that the conversation is no longer intelligible to either conversing party. A user in the noisy environment may try to scream into the phone over the noise, but the screaming and the noise may render the voice signal unintelligible at the other end.
  • For example, a user may be talking on the phone in a busy restaurant. The user may not be able to shout loud enough into the phone to cover the noise in the restaurant. The user may not even be able to hear when the other end is talking. The noise may render the conversation unintelligible and may lead to a termination of the telephone conversation.
  • In another scenario, it may be inconvenient for a user to talk on a phone. For example, the user may be in a meeting and does not want to draw attention to himself by speaking into the phone. The user may try to whisper into the phone, but the whispering may render the conversation unintelligible. The user may choose to send a text message to the other party, but the other party may be on a landline where texting is unavailable, or may not have a texting plan.
  • It can be frustrating to conduct a telephone conversation when the environment is noisy or the circumstance is inappropriate for a user to speak.
  • SUMMARY
  • An embodiment of the invention is directed to a communication device, which establishes an audio connection with a far-end user via a communication network. The communication device receives text input from a near-end user, and converts the text input into speech signals. The speech signals are transmitted to the far-end user using the established audio connection while muting audio input to its audio receiving component.
  • In one embodiment, the communication device detects the noise level at the near end. When the noise level is above a threshold, the communication device can automatically activate or prompt the near-end user to activate text-to-speech conversion at any point of a communication such as a phone call. Alternatively, the communication device may playback a pre-recorded message to inform the far-end user of the near-end user's inability to speak due to the excessive noise at the near end.
  • In another embodiment, the near-end user can activate text-to-speech conversion whenever necessary regardless of the detected noise level. The near-end user can enter a text message, which is converted into speech signals for transmission via the established audio connection to the far-end user.
  • In yet another embodiment, the communication device can also perform speech-to-text conversion to convert the far-end user's speech into text for display on the communication device. This feature can be used when the far-end communication device cannot, or is not enabled to, send or receive text messages. The speech-to-text conversion and the text-to-speech conversion can be activated at the same time, or can be activated independent of each other. The far-end communication device communicates with the near-end communication device in audio signals, regardless of whether the speech-to-text conversion or the text-to-speech conversion is activated.
  • The communication device may be configured or programmed by its user, to support one or more of the above-described features.
  • The above summary does not include an exhaustive list of all aspects of embodiments of the present invention. It is contemplated that embodiments of the invention includes all systems and methods that can be practiced from all suitable combinations of the various aspects summarized above, as well as those disclosed in the Detailed Description below and particularly pointed out in the claims filed with the application. Such combinations have particular advantages not specifically recited in the above summary.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The embodiments of the invention are illustrated by way of example and not by way of limitation in the figures of the accompanying drawings in which like references indicate similar elements. It should be noted that references to “an” or “one” embodiment of the invention in this disclosure are not necessarily to the same embodiment, and they mean at least one.
  • FIG. 1 is a diagram of a communication device operating to answer or receive a phone call.
  • FIG. 2 is a block diagram illustrating an embodiment of the components of the communication device.
  • FIG. 3 is an example of a timeline of events that occur while processing a phone call for a near-end user in a noisy environment.
  • FIG. 4 shows an example of a communication device that provides an option for a near-end user to activate text-to-speech conversion.
  • FIG. 5 shows an example of a communication device that is activated to convert a near-end user's text input to speech.
  • FIG. 6 shows an example of a communication device that is activated to convert a near-end user's text input to speech and a far-end user's speech into text.
  • FIG. 7 is a flow diagram illustrating an embodiment of a method for generating and transmitting synthesized speech signals.
  • DETAILED DESCRIPTION
  • FIG. 1 is a diagram illustrating a communication environment in which a near-end communication device (e.g., a near-end phone 100) is engaged in, or about to be engaged in, a communication (e.g., phone call) with a far-end communication device (e.g., a far-end phone 98) via a communication network (e.g., wireless network 120). The term “communication device” herein broadly refers to various real-time communication devices, e.g., landline telephone system (POTS) end stations, voice-over-IP end stations, cellular handsets, smart phones, computing devices, etc.
  • It should be understood that for ease of description in the present disclosure that the communication device is embodied as and described in terms of a phone. However, it should further be understood that this description should in no way be construed that the communication device is limited to this embodiment and instead can be embodied as a variety of communication devices as described above.
  • In one embodiment, the near-end communication device is a portable communication device that provides real-time mobile telephonic connections. For example, the near-end communication device can be a mobile phone or a mobile multi-functional device that can send and receive voice signals in a cellular communication network. In an alternative embodiment, the near-end communication device can be a computing device having communication capabilities. For example, the computing device may be a personal computer, workstation, personal digital assistant (PDA), etc.
  • The near-end communication device communicates with the far-end communication device over a communication network, for example, a wireless network 120, POTS 130, and a VoIP network 140. Communications between the near-end communication device and the wireless network 120 may be in accordance with known cellular telephone communication network protocols including, for example, global system for mobile communications (GSM), enhanced data rate for GSM evolution (EDGE), and worldwide interoperability for microwave access (WiMAX). The near-end communication device may also have a subscriber identity module (SIM) card, which is a detachable smart card that contains the subscription information of its user (“near-end user”), and may also contain a contacts list of the user. The near-end user may own the near-end communication device or may otherwise be its primary user. The near-end communication device may be assigned a unique address by a wireless or wireline telephony network operator, or an Internet Service Provider (ISP). For example, the unique address may be a domestic or international telephone number, an Internet Protocol (IP) address, or other unique identifiers. In other embodiments, the communication network may be embodied as a wired network, a wireless network, or a combination therein.
  • In one embodiment, the near-end communication device includes a microphone, a processor and a communication network interface. The processor is coupled to the microphone to detect a noise level received by the microphone, and to determine whether the noise level exceeds a threshold. The communication network interface transmits an audio signal to the far-end user using an audio connection that has been established between the near-end user and the far-end user. The audio signal conveys a message of the near-end user that is generated internally by the near-end communication device.
  • In an embodiment shown in FIG. 1, the near-end communication device is shown as the phone 100. Although the near-end phone 100 is shown and described, it is understood that the near-end phone 100 can be any of the communication devices mentioned above. The exterior of the near-end phone 100 is made of a housing 149 within which are integrated several components including a display screen 112, a receiver 111 (e.g., an earpiece speaker for generating sound) and one or more audio receiving components, such as a microphone 113 (e.g., a mouthpiece for picking up a near-end user's voice). Although one microphone 113 is shown and described, it is understood that the near-end phone 100 can include multiple microphones. Therefore, the term “microphone 113” described herein can be understood to represent one or more audio receiving components, such as one or more microphones. The near-end phone 100 may also implement noise suppression and/or acoustic echo cancellation (AEC) techniques to improve the sound quality received into the microphone 113. In one embodiment, the microphone 113 can be used to monitor the noise level in the environment surrounding the near-end phone 100. In an alternative embodiment, a separate microphone can be used to monitor the environmental noise. A noise meter 152 may be shown on the display screen 112 to indicate the detected noise level. The noise meter 152 may be shown when a phone call is made or received, when the noise level reaches the vicinity of a pre-determined threshold, or as long as the near-end phone 100 is powered on. The noise level may be indicated by the noise meter 152 by colors, numeral values, height or length of a bar indicator, etc.
  • Audio input to the microphone 113 can be analyzed by the near-end phone 100 to detect whether the noise level in the input exceeds a pre-determined threshold. For example, if the noise level exceeds the threshold, the near-end phone 100 can notify the user; by flashing the display screen 112 or displaying a popup on the display screen 112. The near-end phone 100 can provide an option to the near-end user to switch to text input. Alternatively, the near-end phone 100 can automatically switch the input mode from speech to text for the near-end user. In one embodiment, the near-end phone 100 mutes the microphone 113 and any other sound pick-up mechanisms (e.g., if there are multiple microphones) on the phone 100 after switched into text input mode. In one embodiment, muting the microphone 113 means that the audio input picked up by the microphone 113 is not transmitted to the far-end communication device. The muted microphone 113 may continue to detect the noise level in the surrounding environment.
  • The near-end phone 100 can convert the near-end user's text input into a speech signal, and transmits the speech signal to the far-end communication device. For example, the speech signal is transmitted to the far-end phone 98 via the communication network (e.g., the wireless network 120, the POTS 130, or the VoIP network 140). The text-to-speech conversion can be activated before the near-end user starts to speak to the near-end phone 100, or in the middle of a phone conversation. As the transmitted speech signal is synthesized internally by the near-end phone 100, the signal quality can be far better than the original speech of the near-end user in the noisy environment.
  • In an alternative embodiment, the near-end phone 100 may playback a pre-recorded audio message, automatically or under the near-end user's direction, when a high noise level is detected. With the use of either the text input or the pre-recorded message, the near-end user's message can be conveyed to a far-end user as noise-free audio signals via an audio connection in the communication network.
  • Turning to the far-end communication device, an embodiment of which is shown in FIG. 1 as the far-end phone 98, this device may receive or initiate a phone call with the near-end user. For example, the far-end phone 98 may be a mobile device or a land-based device that is coupled to a telephony network or other communication networks through wires or cables. The far-end phone 98 may be identified with a unique address, such as a telephone number within the public switched telephone network. The far-end phone 98 may also have an Internet protocol (IP) address if it performs calls through a voice over IP (VOIP) protocol. The far-end phone 98 may be a cellular handset, a telephone service (POTS), analog telephone, a VOIP telephone station, or a desktop or notebook computer running telephony or other communication software. The far-end phone 98 may or may not have the capabilities to receive text messages.
  • The far-end phone 98 is associated with a far-end user. The far-end user can be a person, or an automated entity with synthesized speech capability (e.g., an interactive voice response system).
  • FIG. 2 is a block diagram illustrating an embodiment of the near-end phone 100. The near-end phone 100 includes a communication network interface 235 for receiving and transmitting communication signals, e.g., audio, video and/or data signals. The near-end phone 100 also includes the receiver 111 for generating audio signals in response to an incoming call signal and the microphone 113 for picking up the speech signal of the near-end user. The near-end phone 100 also includes a user interface 230. The user interface 230 includes the display screen 112 and touch sensors 213 for sensing user's touch and motion. The near-end phone 100 may include a physical keyboard 214 for receiving keystroke input from the user, or a virtual keyboard displayed by display screen 12 for accepting user input via the touch sensors 213. The touch sensors 213 may be based on resistive sensing, capacitive sensing, optical sensing, force sensing, surface acoustic wave sensing, and/or other sensing techniques or combinations of sensing techniques. The coordinates of the touch sensors 213 that respond to the user's touch and motion (e.g. gestures) represent a specific user input. The touch sensors 213 may be embedded in the display screen 112, or may be embedded in a touch-sensing panel separate from the display screen 112. In other embodiments, the near-end phone 100 may include other types of sensors for accepting user inputs other than touch input including a motion sensor such as an accelerometer. For example a user may provide an input by shaking the near-end phone 100 or moving the phone in a particular manner.
  • In one embodiment, the user input interface 230 may include one or more buttons (referred to as “T->S selector 255”) for invoking a text-to-speech feature. The T->S selector 255 may be a physical button or a virtual button. The physical button may be a dedicated “text-to-speech” button, or one or more buttons identified by the text shown on the display screen 112 (e.g., “press #*1 to activate text-to-speech conversion”). In an embodiment where the T->S selector 255 is a virtual button, the virtual button may be embedded in the display screen 112 that includes the touch sensors 213. The display screen 112 may show a graphical “text-to-speech” virtual button that can be pressed by the user to invoke text-to-speech conversion. In an alternative embodiment, the T->S selector 255 may be a virtual button implemented on a touch-sensing panel separate from the display screen 112. A user can use the touch-sensing panel to direct a cursor on the display screen 112 to select a graphical “text-to-speech” button shown on the display screen 112. In alternative embodiments, text-to-speech conversion may be activated by a combination of one or more physical buttons and virtual buttons. When a near-end user selects the T->S selector 255, a text-to-speech converter 208 of the near-end phone 100 is activated. The text-to-speech converter 208 can be used to convert the near-end user's text input into speech for transmission to a far-end user.
  • In one embodiment, the user input interface 230 also includes one or more buttons (referred to as “S->T selector 256”) for invoking speech-to-text conversion. The S->T selector 256 can be implemented by physical or virtual button mechanisms, similar to the implementation of the T->S selector 255. When a near-end user selects the S->T selector 256, a speech-to-text converter 206 of the near-end phone 100 is activated. The speech-to-text converter 206 can be used to convert a far-end user's speech into text for displaying on the display screen 112.
  • The text-to-speech converter 208 may be implemented with hardware circuitry, firmware, software or any combination thereof. If implemented in software and/or firmware, the software and firmware may be stored within memory 240 in the near-end phone 100 and executed by a processor 220. Alternatively, the firmware/software may not be stored within the memory 240 in the near-end phone 100. The text-to-speech converter 208 converts the text input into a speech signal based on one or more speech synthesis techniques. The synthesized speech can be created by concatenating pieces of recorded speech stored in the memory 240. The text-to-speech converter 208 may be activated and deactivated by input to the user interface 230.
  • The speech-to-text converter 206 may be implemented with hardware circuitry, firmware, software or any combination thereof. If implemented in software and/or firmware, the software and firmware may be stored within the memory 240 in the near-end phone 100 and executed by the processor 220. Alternatively, the firmware/software may not be stored within memory in the near-end phone 100. The speech-to-text converter 206 identifies the words in an input speech signal based on one or more speech recognition techniques, and causes the display screen 112 to show the recognized words in text. The speech-to-text converter 206 may be activated and deactivated by input to the user interface 230.
  • Still referring to FIG. 2, in one embodiment, the near-end phone 100 also includes a telephone module 238 which is responsible for coordinating various tasks involved in a phone call. The telephone module 238 may be implemented with hardware circuitry, firmware, software or any combination thereof. If implemented in software and/or firmware, the software and firmware may be stored within memory 240 in the near-end phone 100 and executed by the processor 220. Although one processor 220 is shown, it is understood that any number of processors or data processing elements may be included in the near-end phone 100. The telephone module 238 coordinates tasks such as receiving an incoming call signal, placing an outgoing call, activating speech-to-text conversion, activating text-to-speech conversion, playing back pre-recorded messages 207 in the memory 240 and directing a call to a voicemail system.
  • In one embodiment, the telephone module 238 includes a signal analyzer 239 to analyze the noise received at the near-end phone 100. The signal analyzer 239 compares the noise level with a pre-determined threshold, which may be configurable by the near-end user, to determine whether to alert the near-end user of the high noise level, whether to trigger speech synthesis, and/or whether to playback a pre-recorded message.
  • The term “noise” herein refers to broadband environmental noise signals surrounding the near-end phone 100. The signal analyzer 239 receives the audio signals picked up by the microphone 113, and determines whether the noise level in the audio signal exceeds a threshold. The term “noise level” or “audio noise level” herein refers to an absolute level of the noise, a relative level of the noise with respect to the desired signal (e.g., the near-end user's speech), or a combination of the absolute level and the relative level of the noise. The noise level can be determined based on one or more factors, including but not limited to, signal-to-noise (SNR) ratio, noise volume (e.g., 70-80 db), combined signal volume and the noise volume, etc. The SNR can be used to measure whether the speech signal strength of the near-end user is relatively low in comparison to the noise signal strength. The combined signal and noise volume can be used to determine whether the microphone 113 has reached or come close to a saturation point that distorts the received audio signal. If the near-end user has not started to speak on the near-end phone 100, the noise level may be determined purely based on the detected noise volume. If the near-end user has already started to speak on the near-end phone 100, the noise level may be determined based on a combination of the near-end user's speech signals and the environmental noise.
  • In response to the detection of a particular or relative noise level, the near-end phone 100 may provide the near-end user an option to activate text-to-speech conversion. The user interface 230 may display a virtual button implementing the T->S selector 225 on the display screen 112, or may display a message indicating the physical buttons for activating the text-to-speech conversion. Alternatively, the activation of the text-to-speech conversion may be automatic upon detection of a particular or relative noise level at the near end. When the noise level exceeds a particular or relative threshold, the near-end phone 100 may automatically mute the microphone 113 and prompt the near-end user to enter text or to select a text-message stored in the memory 240. In one embodiment, all of the signals picked up by the microphone 113 are bypassed without being transmitted to the far-end user.
  • The text-to-speech conversion can occur anytime after an audio connection in the communication network (e.g., the wireless network 120, the POTS 130, or the VOIP network 140) is established between the near-end user and the far-end user. The conversion causes no interruption to the established audio connection. For example, the near-end user may have placed or picked up a phone call but have not begun to speak with the far-end user. Alternatively, the near-end user may be in the middle of a phone conversation with the far-end user when he activates the text-to-speech conversion.
  • In yet another alternative embodiment, upon being alerted of the particular or relative noise level, the near-end user may select one of the pre-recorded messages 207 in the memory 240. The pre-recorded message 207 may inform the far-end user that the text-to-speech feature is to be activated, or may ask the far-end user to leave a message or to call back at a later time.
  • Additional circuitry, including a combination of hardware circuitry and software, can be included to obtain the needed functionality described herein. These are not described in detail as they would be readily apparent to those of ordinary skill in the art of mobile phone circuits and software.
  • An example of a timeline for operating the near-end phone 100 in a noisy environment is shown in FIG. 3. Referring to FIG. 3, operation may begin when the near-end phone 100 receives a request to establish an audio connection with the far-end phone 98 via a communication network (301). The request may originate from the near-end user (e.g., when the near-end user dials the telephone number of the far-end phone 98), or from the far-end user (e.g., when the far-end user dials the telephone number of the near-end phone 100). In response to the request, the near-end phone 100 begins the process for establishing an audio connection, and, in the meantime, may be monitoring the noise level in the environment where the near-end phone 100 is located (302). The near-end phone 100 may continuously monitor the noise level after the near-end phone 100 is powered on, or may start monitoring the noise level when triggered by an event, such as when receiving the aforementioned request to establish an audio connection, when the cover of the near-end phone 100 is flipped or slid open, or when a button of the near-end phone 100 is pressed. It should be noted that noise level may be actively monitored or passively detected.
  • After the audio connection is established, the near-end phone 100 continues to monitor the noise level received by the near-end phone 100. If the near-end phone 100 detects that the noise level at the near end is above a pre-determined threshold, the near-end phone 100 alerts the near-end user of the high noise level (303). The near-end phone 100 may automatically switch to text input mode or present the user with an option to switch to text input. When the input mode is switched to text, the text-to-speech conversion is activated and the microphone 113 is muted (304). At this time, the user can begin to enter a text message (305), which is converted by the text-to-speech converter 208 into a speech signal. The near-end phone 100 may playback a voice message before the transmission of the converted speech to indicate to the far-end user that speech synthesis at the near end has been activated. The far-end user may continue to speak into the far-end phone 98 without being affected by the speech synthesis performed at the near end.
  • In one scenario, the noise level may drop at the near end before the phone call ends. The near-end phone 100 may alert the near-end user upon detection of reduced noise level (306). The near-end phone 100 may automatically switch back to speech input mode or present the user with an option to switch back to speech input. When the input mode is switched back to speech, the text-to-speech conversion is deactivated and the microphone 113 is turned back on (307). The near-end user and the far-end user can continue the phone conversation (308) until one or both of the parties end the call (309).
  • The text-to-speech conversion may also be used in a non-noisy environment. In one embodiment, the near-end phone 100 allows the near-end user to activate text-to-speech conversion when the detected noise level is still below the threshold. This feature can be useful when the near-end user is in an environment where talking on the phone is inappropriate (e.g., in a meeting). Referring to FIG. 3, if the detected noise level is below the threshold at time 302, the near-end user may switch to text input after an audio connection is established with the far-end user. The near-end user may use an option provided by the near-end phone to activate text-to-speech conversion. If, during the phone call, the near-end user wishes to switch back to speech input, the near-end user can deactivate the text-to-speech conversion, which turns the microphone 113 back on. Operation of the near-end phone 100 in a non-noisy environment can be the same as the operation described above in 301, 304, 305 and 307-309 of FIG. 3.
  • In one scenario, the near-end user may be wearing a headset such that the voice of the far-end user can be heard clearly in the presence of the external noise. Thus, the near-end user can hear the far-end user's speech and enter text messages in response. In some other scenarios, the near-end user may not have a headset when the noise is loud. Alternatively, in a non-noisy environment (e.g., in a meeting), it may be inappropriate for the near-end user to hold the phone to the ear or place an earphone into the ear. In these scenarios, the near-end user may activate speech-to-text conversion at the near-end phone 100, which converts the far-end user's speech into text and displays the text on the display screen 112. With the activation of both speech-to-text and text-to-speech conversions, the near-end user can enter and receive text messages, and the far-end user can receive and answer the near-end user in regular speech.
  • The activation of the speech-to-text conversion may occur before the start of a phone conversation or during the phone conversation. The speech-to-text conversion may be activated by the near-end user or automatically when the detected noise level is above a threshold. The speech-to-text conversion may be activated independently of the activation of text-to-speech conversion, or may be activated at the same time with the text-to-speech conversion. The activation of the speech-to-text conversion does not interrupt the established audio connection between the near-end and far-end users.
  • The activation of the speech-to-text conversion may be controlled by the near-end user or may be automatic. In one embodiment, when a relative or particular noise level is detected at the near end, the near-end phone 100 may automatically convert the far-end user's speech to text. Additionally, the near-end phone 100 may also automatically deactivate the speech-to-text conversion of the far-end user's speech, or provide the near-end user an option to stop converting the far-end user's speech into text.
  • FIG. 4 shows an example of the near-end phone 100 that provides a number of options for the near-end user, including an option to activate text-to-speech conversion. Although the near-end phone 100 is shown in FIG. 4 as a mobile phone, it is understood that other communication devices can also be used.
  • Referring to FIG. 4, in response to the detection of the relative or particular noise level at the near end, the near-end phone 100 displays a number of options for the user to choose. The options may include: text-to-speech 401, two-way text 402, play (pre-recorded) message 403, and voicemail 404. The user may select one of these options using a physical button or a virtual button. In one embodiment, the near-end phone 100 also displays the noise meter 152 on the display screen 112 to provide a visual indication of the environmental noise level at the near-end.
  • The display screen 112 also shows “INCOMING CALL . . . ” or “OUTGOING CALL . . . ” depending on the direction of the call, as well as the number assigned to the far-end user. In one embodiment where the near-end phone 100 stores a list of contacts including their nicknames and numbers, the display screen 112 may show the nickname of the far-end user in addition to or instead of the assigned number.
  • Referring to FIG. 5, if the near-end user selects the text-to-speech option 401, the display screen 112 may show “TEXT TO SPEECH” to indicate that the text-to-speech conversion has been activated. The near-end user may use a physical keyboard or a virtual keyboard (such as a virtual keyboard 512 presented on the display screen 112 in which touch sensors are embedded) to input text messages. The display screen 112 also shows an outgoing message area 511 that displays the text entered by the near-end user. As the near-end user inputs the text, the text-to-speech converter 208 (FIG. 2) automatically converts the text into speech. The near-end phone 100 transmits the converted speech signal to the far-end user, utilizing the audio connection that has already been established between the near-end user and the far-end user.
  • In some scenarios, the near-end user may wish to activate text-to-speech conversion even though the external noise level has not exceeded a threshold. In one embodiment, the near-end user may be provided with the text-to-speech option 401 after an audio connection is established with a far-end user regardless of the noise level at the near end. Alternatively or additionally, the near-end user may press one or more physical buttons on the near-end phone 100 to activate the text-to-speech conversion, or to cause the display of the text-to-speech option 401 on the display screen 112.
  • After activation of the text-to-speech conversion, if the noise level drops below a threshold, the near-end phone 100 may present a resume option 510 for the near-end user to resume speaking to near-end phone 100. Once the resume option 510 is selected, the text-to-speech conversion is deactivated and the microphone 113 is turned back on. In an alternative embodiment, the resume option 510 may be provided regardless of the detected noise level.
  • If the near-end user wishes to receive text messages instead of speech signals from the far-end user, the near-end user may activate speech-to-text conversion after an audio connection is established with the far-end user. The example of FIG. 4 shows an embodiment in which the speech-to-text conversion and the text-to-speech conversion can be activated at the same time by selecting the two-way text 402 option. In an alternative embodiment, the speech-to-text conversion may be activated independently of the text-to-speech conversion.
  • Referring to FIG. 6, if the user selects the two-way text 402 option, the display screen 112 may show “TWO-WAY TEXT” to indicate that both of the text-to-speech and speech-to-text conversions have been activated. The near-end user may use a physical keyboard or a virtual keyboard (such as a virtual keyboard 614 presented on the display screen 112 in which touch sensors are embedded) to input text messages. The display screen 112 shows an incoming message area 612 for displaying the text converted from the far-end user, and an outgoing message area 613 for displaying the text entered by the near-end user. The established audio connection carries two-way voice signals between the near-end and the far-end users. The conversions from text to speech and from speech to text are performed by the near-end phone 100. The far-end user can speak to the far-end phone 98 in the same way as in a normal telephone conversation that does not involve text messages.
  • In a scenario where the near-end user wishes to activate speech-to-text conversion even though the external noise level has not exceeded a threshold, the near-end user may be provided with the two-way text option 402 after an audio connection is established with a far-end user. Alternatively or additionally, the near-end user may be provided with a separate speech-to-text option (not shown), or press one or more physical buttons on the near-end phone 100 to activate the speech-to-text conversion.
  • After activation of the two-way text-to-speech and speech-to-text conversions, if the noise level drops below a threshold, the near-end phone 100 may present a resume option 610 for the near-end user to resume a two-way speech conversation. Once the resume option 610 is selected, the two-way conversions are deactivated and the microphone 113 is turned back on. In an alternative embodiment, the resume option 610 may be provided regardless of the detected noise level.
  • Referring again to FIG. 4, if the near-end user selects the play message option 403, a pre-recorded default message can be played to the far-end user. For example, the default message may indicate that it is too noisy for the near-end user to speak on the phone and the near-end user will call back later. The near-end user can record any suitable message as the default message when setting up the near-end phone 100. Alternatively, the near-end phone 100 may provide the near-end user with a number of additional options, each associated with a different pre-recorded message. The near-end user may select a pre-recorded message that best suits the current situation. After the message is played to the far-end user, the near-end phone 100 may terminate the call, or hold the call to receive further input from the near-end user.
  • If the near-end user selects the voicemail option 404, the far-end user will be re-directed to a voicemail system to leave a voice message. The near-end user will be able to retrieve the message from the voicemail system at a later time. The message left by the far-end user may be stored in the memory 240 of the near-end phone 100, or it may be stored in a remote voicemail server.
  • Regardless of which option is selected in FIGS. 4-6, the established audio connection between the near-end and far-end users is uninterrupted until one or both users terminate the phone call.
  • FIG. 7 is a flow diagram illustrating an embodiment of a method 700 for generating synthesized speech signals at a near-end communication device (e.g., the near-end phone 100), and transmitting the synthesized speech signals to a far-end communication device (e.g., the far-end phone 98). The method 700 may be performed by a communication device that may comprise hardware (e.g., circuitry, dedicated logic, programmable logic, microcode, etc.), firmware, software (e.g., instructions run on a processing device to perform hardware simulation), or a combination thereof.
  • The method 700 begins when a near-end communication device receives a request for an audio connection (710). In one embodiment, the near-end communication device detects the noise level in the surrounding environment as exceeding a threshold, and causes the activation of or the recommendation to active the text-to-speech conversion (720). In alternative embodiment, the near-end communication device may receive a near-end user's input to activate the text-to-speech conversion even if the detected noise level is below the threshold. After the activation of the text-to-speech conversion, the near-end communication device receives text input from the near-end user (730). The near-end communication device then converts the text input into speech signals using speech synthesis techniques (740). The near-end communication device transmits the speech signal to the far-end communication device via the established audio connection, while muting the audio input to the near-end communication device (750).
  • In general, the near-end communication device (e.g., the telephone module 238 of FIG. 2) may be configured or programmed by the user to support one or more of the above-described features.
  • To conclude, various ways of using speech synthesis in a communication device (e.g., a portable communication device, mobile phone etc.) have been described. These techniques allow a person to use his phone when the environment is noisy or the circumstance is inappropriate for him to speak. As explained above, an embodiment of the invention may be a machine-readable storage medium (such as the memory 240 of FIG. 2) having stored thereon instructions which program a processor to perform some of the operations described above. In other embodiments, some of these operations might be performed by specific hardware components that contain hardwired logic. Those operations might alternatively be performed by any combination of programmed data processing components and custom hardware components.
  • The invention is not limited to the specific embodiments described above. Accordingly, other embodiments are within the scope of the claims.

Claims (20)

1. A method of a communication device, the method comprising:
establishing an audio connection between a near-end user of the communication device and a far-end user via a communication network;
receiving text input from the near-end user and converting the text input into speech signals; and
transmitting the speech signals to the far-end user using the audio connection while muting audio input to the communication device.
2. The method of claim 1, further comprising:
detecting a noise level in the audio input to the communication device; and
in response to the noise level exceeding a threshold, causing one of activation of and a recommendation to activate text-to-speech conversion.
3. The method of claim 2, wherein said detecting a noise level comprises:
detecting the noise level based on a combination of signal to noise ratio, noise volume and combined signal and noise volume.
4. The method of claim 2, further comprising:
detecting the noise level as below the threshold after the text-to-speech conversion is activated; and
in response to the noise level below the threshold, deactivating the text-to-speech conversion.
5. The method of claim 1, further comprising:
detecting a noise level in the audio input to the communication device; and
in response to the noise level exceeding a threshold, causing one of activation of and a recommendation to activate speech-to-text conversion to convert voice of the far-end user into text.
6. The method of claim 1, further comprising:
activating speech-to-text conversion after establishment of the audio connection between the near-end user and the far-end user;
converting voice of the far-end user transmitted over the audio connection into text; and
displaying the text on a display of the communication device.
7. The method of claim 1, wherein the communication device is a mobile phone.
8. A communication device, comprising:
a microphone;
a processor coupled to the microphone to detect a noise level and to determine whether the noise level exceeds a threshold; and
a communication network interface to transmit an audio signal to a far-end user using an audio connection that has been established between a near-end user and the far-end user via a communication network, wherein the processor is to cause audio input from the microphone to be muted when the noise level is above the threshold and while the audio signal conveys a message of the near-end user that is generated internally by the communication device.
9. The communication device of claim 8, further comprising:
a text-to-speech converter to convert text input of the near-end user into the message that is conveyed by the audio signal.
10. The communication device of claim 8, further comprising:
a speech-to-text converter to convert voice of the far-end user transmitted over the audio connection into text; and
a display screen to show the text to the near-end user.
11. The communication device of claim 8, further comprising:
a user interface to enable the near-end user of the communication device to activate text-to-speech conversion of a text message one of selected and written by the near-end user for conveyance to the far-end user over the audio connection.
12. The communication device of claim 8, further comprising:
a user interface to enable the near-end user of the communication device to playback a pre-recorded message stored in memory for conveyance to the far-end user over the audio connection.
13. The communication device of claim 8, wherein the display screen is enabled to show a noise meter that indicates to the near-end user the noise level in the environment.
14. The communication device of claim 8, wherein communication device is portable.
15. The communication device of claim 8, wherein communication device is a mobile phone.
16. An article of manufacture comprising:
machine-readable storage medium having stored therein instructions that when executed cause a communication device to:
establish an audio connection between a near-end user of the communication device and a far-end user via a communication network,
indicate audio noise level of an environment surrounding the communication device,
in response to a detected noise level, generate an audio signal that conveys a message of the near-end user, and
transmit the audio signal to the far-end user using the audio connection while muting audio input to the communication device.
17. The article of manufacture of claim 16, wherein the instructions further cause the communication device to:
automatically activate text-to-speech conversion when the detected noise level exceeds a threshold; and
prompt the near-end user to enter text input.
18. The article of manufacture of claim 16, wherein the instructions further program the communication device to:
enable the near-end user to activate text-to-speech conversion independently of the detected noise level.
19. The article of manufacture of claim 16, wherein the instructions further program the communication device to:
activate speech-to-text conversion to convert voice of the far-end user into text for display on the communication device.
20. The article of manufacture of claim 16, wherein the instructions further program the communication device to:
playback a pre-recorded message for conveyance to the far-end user over the audio connection.
US12/613,636 2009-11-06 2009-11-06 Synthesized audio message over communication links Abandoned US20110111805A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/613,636 US20110111805A1 (en) 2009-11-06 2009-11-06 Synthesized audio message over communication links
US14/849,433 US9666180B2 (en) 2009-11-06 2015-09-09 Synthesized audio message over communication links

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US12/613,636 US20110111805A1 (en) 2009-11-06 2009-11-06 Synthesized audio message over communication links

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/849,433 Continuation US9666180B2 (en) 2009-11-06 2015-09-09 Synthesized audio message over communication links

Publications (1)

Publication Number Publication Date
US20110111805A1 true US20110111805A1 (en) 2011-05-12

Family

ID=43974545

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/613,636 Abandoned US20110111805A1 (en) 2009-11-06 2009-11-06 Synthesized audio message over communication links
US14/849,433 Active US9666180B2 (en) 2009-11-06 2015-09-09 Synthesized audio message over communication links

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/849,433 Active US9666180B2 (en) 2009-11-06 2015-09-09 Synthesized audio message over communication links

Country Status (1)

Country Link
US (2) US20110111805A1 (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110287806A1 (en) * 2010-05-18 2011-11-24 Preetha Prasanna Vasudevan Motion-based tune composition on a mobile device
CN102347913A (en) * 2011-07-08 2012-02-08 个信互动(北京)网络科技有限公司 Method for realizing voice and text content mixed message
US8170537B1 (en) * 2009-12-15 2012-05-01 Google Inc. Playing local device information over a telephone connection
US8265938B1 (en) 2011-05-24 2012-09-11 Verna Ip Holdings, Llc Voice alert methods, systems and processor-readable media
US8438023B1 (en) * 2011-09-30 2013-05-07 Google Inc. Warning a user when voice input to a device is likely to fail because of background or other noise
US20130142056A1 (en) * 2011-12-06 2013-06-06 Qualcomm Incorporated Recommending information associated with a user equipment or a communication group in a communications system
CN103167066A (en) * 2011-12-16 2013-06-19 富泰华工业(深圳)有限公司 Cellphone and noise detection method thereof
US8566100B2 (en) 2011-06-21 2013-10-22 Verna Ip Holdings, Llc Automated method and system for obtaining user-selected real-time information on a mobile communication device
US20130278824A1 (en) * 2012-04-24 2013-10-24 Mobitv, Inc. Closed captioning management system
CN103596155A (en) * 2012-08-16 2014-02-19 三星电子株式会社 Method for providing voice call using text data and electronic device thereof
US8970400B2 (en) 2011-05-24 2015-03-03 Verna Ip Holdings, Llc Unmanned vehicle civil communications systems and methods
WO2015017029A3 (en) * 2013-08-02 2015-07-23 Whatsapp Inc. Voice communications with real-time status notifications
US20150294639A1 (en) * 2014-04-14 2015-10-15 Sony Corporation Method and system for switching a mode of an electronic device
WO2015175119A1 (en) * 2014-05-12 2015-11-19 Intel Corporation Far-end context dependent pre-processing
WO2015191990A1 (en) * 2014-06-14 2015-12-17 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
US9226121B2 (en) 2013-08-02 2015-12-29 Whatsapp Inc. Voice communications with real-time status notifications
US9311911B2 (en) * 2014-07-30 2016-04-12 Google Technology Holdings Llc. Method and apparatus for live call text-to-speech
US9378737B2 (en) 2012-11-05 2016-06-28 Mitsubishi Electric Corporation Voice recognition device
US9380146B1 (en) * 2012-03-14 2016-06-28 Amazon Technologies, Inc. System and method to facilitate conversion between voice calls and text calls
US9666180B2 (en) 2009-11-06 2017-05-30 Apple Inc. Synthesized audio message over communication links
KR20190026704A (en) * 2019-02-27 2019-03-13 삼성전자주식회사 Method for providing voice communication using character data and an electronic device thereof
US20190116408A1 (en) * 2017-10-18 2019-04-18 Alibaba Group Holding Limited Guard apparatus and interaction method for sound pickup device
US10389885B2 (en) * 2017-02-01 2019-08-20 Cisco Technology, Inc. Full-duplex adaptive echo cancellation in a conference endpoint
US10431201B1 (en) 2018-03-20 2019-10-01 International Business Machines Corporation Analyzing messages with typographic errors due to phonemic spellings using text-to-speech and speech-to-text algorithms
US10666800B1 (en) * 2014-03-26 2020-05-26 Open Invention Network Llc IVR engagements and upfront background noise
US20200168203A1 (en) * 2018-11-26 2020-05-28 International Business Machines Corporation Sharing confidential information with privacy using a mobile phone
US10769923B2 (en) 2011-05-24 2020-09-08 Verna Ip Holdings, Llc Digitized voice alerts
US10805669B2 (en) * 2018-01-02 2020-10-13 Boe Technology Group Co., Ltd. Method, system and apparatus for automatically starting recording
US11011165B2 (en) * 2017-08-30 2021-05-18 Kyocera Document Solutions Inc. Voice input device, non-transitory computer readable medium storing voice input program, and voice input system
US11170757B2 (en) * 2016-09-30 2021-11-09 T-Mobile Usa, Inc. Systems and methods for improved call handling
US20220164161A1 (en) * 2019-04-26 2022-05-26 Sony Interactive Entertainment Inc. Information processing system, information processing apparatus, control method for information processing apparatus, and program
US11568867B2 (en) 2013-06-27 2023-01-31 Amazon Technologies, Inc. Detecting self-generated wake expressions

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3217638B1 (en) * 2016-03-11 2019-11-20 Sony Mobile Communications Inc Transferring information from a sender to a recipient during a telephone call under noisy environment
CN107342088B (en) * 2017-06-19 2021-05-18 联想(北京)有限公司 Method, device and equipment for converting voice information
KR102512614B1 (en) * 2018-12-12 2023-03-23 삼성전자주식회사 Electronic device audio enhancement and method thereof
US11115075B2 (en) * 2019-07-30 2021-09-07 Ppip Llc Safe case with security choke point control
US11595462B2 (en) * 2019-09-09 2023-02-28 Motorola Mobility Llc In-call feedback to far end device of near end device constraints
US11475873B2 (en) * 2020-07-01 2022-10-18 International Business Machines Corporation Translating inputs for voice communication

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5860064A (en) * 1993-05-13 1999-01-12 Apple Computer, Inc. Method and apparatus for automatic generation of vocal emotion in a synthetic text-to-speech system
US20020025832A1 (en) * 2000-02-18 2002-02-28 Durian Michael B. Controlling data transmission involving a wireless telephone
US20030149565A1 (en) * 2001-03-09 2003-08-07 Chang Steve S. System, method and computer program product for spelling fallback during large-scale speech recognition
US6868385B1 (en) * 1999-10-05 2005-03-15 Yomobile, Inc. Method and apparatus for the provision of information signals based upon speech recognition
US20060095848A1 (en) * 2004-11-04 2006-05-04 Apple Computer, Inc. Audio user interface for computing devices
US20060149558A1 (en) * 2001-07-17 2006-07-06 Jonathan Kahn Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile
US20080043996A1 (en) * 2006-08-07 2008-02-21 Dolph Blaine H Systems And Arrangements For Controlling Audio Levels Based On User Selectable Parameters
US7356470B2 (en) * 2000-11-10 2008-04-08 Adam Roth Text-to-speech and image generation of multimedia attachments to e-mail
US7546083B2 (en) * 2006-01-24 2009-06-09 Apple Inc. Multimedia data transfer for a personal communication device
US7890128B1 (en) * 2003-12-08 2011-02-15 Ipventure, Inc. Adaptable communication techniques for electronic devices

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100831024B1 (en) * 2006-12-20 2008-05-20 삼성전자주식회사 Method for displaying function of equipped touch screen and mobile terminal
US20110111805A1 (en) 2009-11-06 2011-05-12 Apple Inc. Synthesized audio message over communication links

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5860064A (en) * 1993-05-13 1999-01-12 Apple Computer, Inc. Method and apparatus for automatic generation of vocal emotion in a synthetic text-to-speech system
US6868385B1 (en) * 1999-10-05 2005-03-15 Yomobile, Inc. Method and apparatus for the provision of information signals based upon speech recognition
US20020025832A1 (en) * 2000-02-18 2002-02-28 Durian Michael B. Controlling data transmission involving a wireless telephone
US7356470B2 (en) * 2000-11-10 2008-04-08 Adam Roth Text-to-speech and image generation of multimedia attachments to e-mail
US20030149565A1 (en) * 2001-03-09 2003-08-07 Chang Steve S. System, method and computer program product for spelling fallback during large-scale speech recognition
US20060149558A1 (en) * 2001-07-17 2006-07-06 Jonathan Kahn Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile
US7890128B1 (en) * 2003-12-08 2011-02-15 Ipventure, Inc. Adaptable communication techniques for electronic devices
US20060095848A1 (en) * 2004-11-04 2006-05-04 Apple Computer, Inc. Audio user interface for computing devices
US7546083B2 (en) * 2006-01-24 2009-06-09 Apple Inc. Multimedia data transfer for a personal communication device
US20080043996A1 (en) * 2006-08-07 2008-02-21 Dolph Blaine H Systems And Arrangements For Controlling Audio Levels Based On User Selectable Parameters

Cited By (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9666180B2 (en) 2009-11-06 2017-05-30 Apple Inc. Synthesized audio message over communication links
US8335496B1 (en) * 2009-12-15 2012-12-18 Google Inc. Playing local device information over a telephone connection
US8170537B1 (en) * 2009-12-15 2012-05-01 Google Inc. Playing local device information over a telephone connection
US8583093B1 (en) * 2009-12-15 2013-11-12 Google Inc. Playing local device information over a telephone connection
US9531854B1 (en) * 2009-12-15 2016-12-27 Google Inc. Playing local device information over a telephone connection
US20110287806A1 (en) * 2010-05-18 2011-11-24 Preetha Prasanna Vasudevan Motion-based tune composition on a mobile device
US8265938B1 (en) 2011-05-24 2012-09-11 Verna Ip Holdings, Llc Voice alert methods, systems and processor-readable media
US10769923B2 (en) 2011-05-24 2020-09-08 Verna Ip Holdings, Llc Digitized voice alerts
US10282960B2 (en) 2011-05-24 2019-05-07 Verna Ip Holdings, Llc Digitized voice alerts
US11403932B2 (en) 2011-05-24 2022-08-02 Verna Ip Holdings, Llc Digitized voice alerts
US8970400B2 (en) 2011-05-24 2015-03-03 Verna Ip Holdings, Llc Unmanned vehicle civil communications systems and methods
US9883001B2 (en) 2011-05-24 2018-01-30 Verna Ip Holdings, Llc Digitized voice alerts
US9305542B2 (en) 2011-06-21 2016-04-05 Verna Ip Holdings, Llc Mobile communication device including text-to-speech module, a touch sensitive screen, and customizable tiles displayed thereon
US8566100B2 (en) 2011-06-21 2013-10-22 Verna Ip Holdings, Llc Automated method and system for obtaining user-selected real-time information on a mobile communication device
CN102347913A (en) * 2011-07-08 2012-02-08 个信互动(北京)网络科技有限公司 Method for realizing voice and text content mixed message
US8438023B1 (en) * 2011-09-30 2013-05-07 Google Inc. Warning a user when voice input to a device is likely to fail because of background or other noise
US9497599B2 (en) 2011-12-06 2016-11-15 Qualcomm Incorporated Recommending information associated with a user equipment or a communication group in a communications system
US20130142056A1 (en) * 2011-12-06 2013-06-06 Qualcomm Incorporated Recommending information associated with a user equipment or a communication group in a communications system
US9025474B2 (en) * 2011-12-06 2015-05-05 Qualcomm Incorporated Recommending information associated with a user equipment or a communication group in a communications system
US9203943B2 (en) * 2011-12-16 2015-12-01 Fu Tai Hua Industry (Shenzhen) Co., Ltd. Apparatus and method for detecting time division duplex noise of communication device
US20130155886A1 (en) * 2011-12-16 2013-06-20 Yang Xin Apparatus and method for detecting time division duplex noise of communication device
TWI566567B (en) * 2011-12-16 2017-01-11 鴻海精密工業股份有限公司 Mobile phone and noise detection method thereof
CN103167066A (en) * 2011-12-16 2013-06-19 富泰华工业(深圳)有限公司 Cellphone and noise detection method thereof
US10115390B2 (en) 2012-03-14 2018-10-30 Amazon Technologies, Inc. System and method to facilitate conversion between voice calls and text communications
US9380146B1 (en) * 2012-03-14 2016-06-28 Amazon Technologies, Inc. System and method to facilitate conversion between voice calls and text calls
US10523896B2 (en) * 2012-04-24 2019-12-31 Mobitv, Inc. Closed captioning management system
US11196960B2 (en) 2012-04-24 2021-12-07 Tivo Corporation Closed captioning management system
US10122961B2 (en) 2012-04-24 2018-11-06 Mobitv, Inc. Closed captioning management system
US11736659B2 (en) 2012-04-24 2023-08-22 Tivo Corporation Closed captioning management system
US20130278824A1 (en) * 2012-04-24 2013-10-24 Mobitv, Inc. Closed captioning management system
US9516371B2 (en) * 2012-04-24 2016-12-06 Mobitv, Inc. Closed captioning management system
US9078111B2 (en) 2012-08-16 2015-07-07 Samsung Electronics Co., Ltd. Method for providing voice call using text data and electronic device thereof
KR20140023080A (en) * 2012-08-16 2014-02-26 삼성전자주식회사 Method for providing voice communication using character data and an electronic device thereof
CN108430043A (en) * 2012-08-16 2018-08-21 三星电子株式会社 Method and its electronic device for providing the voice communication for using text data
EP2698787A3 (en) * 2012-08-16 2014-02-26 Samsung Electronics Co., Ltd Method for providing voice call using text data and electronic device thereof
KR101954774B1 (en) * 2012-08-16 2019-03-06 삼성전자주식회사 Method for providing voice communication using character data and an electronic device thereof
CN103596155A (en) * 2012-08-16 2014-02-19 三星电子株式会社 Method for providing voice call using text data and electronic device thereof
DE112012007103B4 (en) * 2012-11-05 2017-02-02 Mitsubishi Electric Corporation Voice recognition device
US9378737B2 (en) 2012-11-05 2016-06-28 Mitsubishi Electric Corporation Voice recognition device
US11600271B2 (en) * 2013-06-27 2023-03-07 Amazon Technologies, Inc. Detecting self-generated wake expressions
US11568867B2 (en) 2013-06-27 2023-01-31 Amazon Technologies, Inc. Detecting self-generated wake expressions
WO2015017029A3 (en) * 2013-08-02 2015-07-23 Whatsapp Inc. Voice communications with real-time status notifications
US9419935B2 (en) 2013-08-02 2016-08-16 Whatsapp Inc. Voice communications with real-time status notifications
US9226121B2 (en) 2013-08-02 2015-12-29 Whatsapp Inc. Voice communications with real-time status notifications
US10608978B2 (en) 2013-08-02 2020-03-31 Whatsapp Inc. Voice communications with real-time status notifications
US10666800B1 (en) * 2014-03-26 2020-05-26 Open Invention Network Llc IVR engagements and upfront background noise
US20150294639A1 (en) * 2014-04-14 2015-10-15 Sony Corporation Method and system for switching a mode of an electronic device
WO2015175119A1 (en) * 2014-05-12 2015-11-19 Intel Corporation Far-end context dependent pre-processing
US11228834B2 (en) 2014-06-14 2022-01-18 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
US10567875B2 (en) 2014-06-14 2020-02-18 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
US10555080B2 (en) 2014-06-14 2020-02-04 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
US10856077B2 (en) * 2014-06-14 2020-12-01 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
WO2015191990A1 (en) * 2014-06-14 2015-12-17 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
US10750282B2 (en) 2014-06-14 2020-08-18 Polycom, Inc. Acoustic perimeter for reducing noise transmitted by a communication device in an open-plan environment
US9311911B2 (en) * 2014-07-30 2016-04-12 Google Technology Holdings Llc. Method and apparatus for live call text-to-speech
US11170757B2 (en) * 2016-09-30 2021-11-09 T-Mobile Usa, Inc. Systems and methods for improved call handling
US11399100B2 (en) 2017-02-01 2022-07-26 Cisco Technology, Inc. Full-duplex adaptive echo cancellation in a conference endpoint
US10389885B2 (en) * 2017-02-01 2019-08-20 Cisco Technology, Inc. Full-duplex adaptive echo cancellation in a conference endpoint
US11011165B2 (en) * 2017-08-30 2021-05-18 Kyocera Document Solutions Inc. Voice input device, non-transitory computer readable medium storing voice input program, and voice input system
US10681442B2 (en) * 2017-10-18 2020-06-09 Alibaba Group Holding Limited Guard apparatus and interaction method for sound pickup device
US20190116408A1 (en) * 2017-10-18 2019-04-18 Alibaba Group Holding Limited Guard apparatus and interaction method for sound pickup device
US10805669B2 (en) * 2018-01-02 2020-10-13 Boe Technology Group Co., Ltd. Method, system and apparatus for automatically starting recording
US10431201B1 (en) 2018-03-20 2019-10-01 International Business Machines Corporation Analyzing messages with typographic errors due to phonemic spellings using text-to-speech and speech-to-text algorithms
US20200168203A1 (en) * 2018-11-26 2020-05-28 International Business Machines Corporation Sharing confidential information with privacy using a mobile phone
US10891939B2 (en) * 2018-11-26 2021-01-12 International Business Machines Corporation Sharing confidential information with privacy using a mobile phone
KR102075750B1 (en) * 2019-02-27 2020-03-02 삼성전자 주식회사 Method for providing voice communication using character data and an electronic device thereof
KR20190026704A (en) * 2019-02-27 2019-03-13 삼성전자주식회사 Method for providing voice communication using character data and an electronic device thereof
US20220164161A1 (en) * 2019-04-26 2022-05-26 Sony Interactive Entertainment Inc. Information processing system, information processing apparatus, control method for information processing apparatus, and program
US11880633B2 (en) * 2019-04-26 2024-01-23 Sony Interactive Entertainment Inc. Information processing system, information processing apparatus, control method for information processing apparatus, and program

Also Published As

Publication number Publication date
US20150379982A1 (en) 2015-12-31
US9666180B2 (en) 2017-05-30

Similar Documents

Publication Publication Date Title
US9666180B2 (en) Synthesized audio message over communication links
US20110111735A1 (en) Phone hold mechanism
US8451312B2 (en) Automatic video stream selection
US8660248B2 (en) Method and structure for controlling incoming communication notification
EP1732294A1 (en) Voice activated phone mute reminder method and apparatus
CA3076243C (en) Transparent fax call processing in a mobile device
JP2007520943A (en) Extended use of phones in noisy environments
US20050232404A1 (en) Method of determining a user presence state
EP3217638B1 (en) Transferring information from a sender to a recipient during a telephone call under noisy environment
KR100605651B1 (en) Selective mute method and mobile phone using the same
EP2636212B1 (en) Controlling audio signals
CA2534774C (en) System and method of safe and automatic acoustic volume adjustment for handsfree operation
US8195135B2 (en) Mobile communication device and audio processing method thereof
JP4600643B2 (en) Videophone device having character display function and voice character conversion display method in videophone device
KR101155764B1 (en) Mobile having a restrain function of high-pitch voice
JP2005244579A (en) Telephone
GB2409605A (en) Notification at calling terminal of called party answering call
JP2009232232A (en) Communication device
JP2015226198A (en) Telephone, hands-free call system, and control method for telephone
JP2008104044A (en) Telephone device
KR20070019676A (en) Enhanced usage of telephones in noisy surroundings

Legal Events

Date Code Title Description
AS Assignment

Owner name: APPLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PAQUIER, BAPTISTE P.;LINDAHL, ARAM M.;TAMCHINA, PHILLIP G.;SIGNING DATES FROM 20091102 TO 20091103;REEL/FRAME:023489/0867

STCB Information on status: application discontinuation

Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION