US20100030557A1 - Voice and text communication system, method and apparatus - Google Patents

Voice and text communication system, method and apparatus Download PDF

Info

Publication number
US20100030557A1
US20100030557A1 US11/497,011 US49701106A US2010030557A1 US 20100030557 A1 US20100030557 A1 US 20100030557A1 US 49701106 A US49701106 A US 49701106A US 2010030557 A1 US2010030557 A1 US 2010030557A1
Authority
US
United States
Prior art keywords
speech
text
operable
packets
mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/497,011
Inventor
Stephen Molloy
Khaled Helmi El-Maleh
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qualcomm Inc
Original Assignee
Qualcomm Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qualcomm Inc filed Critical Qualcomm Inc
Priority to US11/497,011 priority Critical patent/US20100030557A1/en
Assigned to QUALCOMM INCORPORATED A DELWARE CORPORATION reassignment QUALCOMM INCORPORATED A DELWARE CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MOLLOY, STEPHEN, EL-MALEH, KHALED HELMI
Priority to JP2009523030A priority patent/JP5149292B2/en
Priority to KR1020097003859A priority patent/KR101136769B1/en
Priority to EP07840622A priority patent/EP2047459A2/en
Priority to CN2007800282788A priority patent/CN101496096B/en
Priority to PCT/US2007/074902 priority patent/WO2008016949A2/en
Publication of US20100030557A1 publication Critical patent/US20100030557A1/en
Priority to US14/981,298 priority patent/US9940923B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/027Concept to speech synthesisers; Generation of natural phrases from machine-based concepts
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Processing of the speech or voice signal to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering
    • G10L2021/02082Noise filtering the noise being echo, reverberation of the speech

Definitions

  • the disclosure relates to communications and, more particularly, to a voice and text communication system, method and apparatus.
  • a cellular phone may include an audio capture device, such as a microphone and/or speech synthesizer, and an audio encoder to generate audio packets or frames.
  • the phone may use communication protocol layers and modules to transmit packets across a wireless communication channel to a network or another communication device.
  • One aspect relates to an apparatus comprising a vocoder, a speech to text conversion engine, a text to speech conversion engine, and a user interface.
  • the vocoder is operable to convert speech signals into packets and convert packets into speech signals.
  • the speech to text conversion engine is operable to convert speech to text.
  • the text to speech conversion engine is operable to convert text to speech.
  • the user interface is operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the speech to text conversion engine, a second mode enables the text to speech conversion engine, and a third mode enables the speech to text conversion engine and the text to speech conversion engine.
  • Another aspect relates to an apparatus comprising: a vocoder operable to convert speech signals into packets and convert packets into speech signals; a speech to text conversion engine operable to convert speech to text; a user interface operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the vocoder, and a second mode enables the speech to text conversion engine; and a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
  • a network apparatus comprising: a vocoder operable to convert packets into speech signals; a speech to text conversion engine operable to convert speech to text; a selection unit operable to switch between first and second modes, wherein the first mode enables the vocoder, and a second mode enables the vocoder and the speech to text conversion engine; and a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
  • Another aspect relates to a method comprising: receiving encoded speech packets; converting the received encoded speech packets into speech signals; and receiving a user selection of a mode from among a plurality of modes, wherein a first mode enables speech to text conversion, a second mode enables text to speech conversion, and a third mode enables speech to text and text to speech conversion.
  • FIG. 1 illustrates a system comprising a first communication device, a network, and a second communication device.
  • FIG. 2 illustrates a method of using the second device of FIG. 1 .
  • FIG. 3 illustrates another configuration of the first communication device of FIG. 1 .
  • FIG. 4 illustrates another configuration of the network of FIG. 1 .
  • Receiving a call on a mobile device in a meeting, airplane, train, theater, restaurant, church or other place may be disruptive to others. It may be much less disruptive if a user could select another mode on the mobile device to receive the call and/or respond to the call. In one mode, the device receives the call and converts speech/voice signals to text without requiring the caller on the other end to input text.
  • FIG. 1 illustrates a system comprising a first communication device 100 , a network 110 , and a second communication device 120 .
  • the system may include other components.
  • the system may use any type of wireless communication, such as Global System for Mobile communications (GSM), code division multiple access (CDMA), CDMA2000, CDMA2000 1x EV-DO, Wideband CDMA (WCDMA), orthogonal frequency division multiple access (OFDMA), Bluetooth, WiFi, WiMax, etc.
  • GSM Global System for Mobile communications
  • CDMA code division multiple access
  • CDMA2000 Code division multiple access
  • CDMA2000 1x EV-DO Wideband CDMA
  • OFDMA orthogonal frequency division multiple access
  • Bluetooth WiFi, WiMax, etc.
  • the first communication device 100 comprises a voice coder (vocoder) 102 and a transceiver 104 .
  • the first communication device 100 may include other components in addition to or instead of the components shown in FIG. 1 .
  • the first communication device 100 may represent or be implemented in a landline (non-wireless) phone, a wireless communication device, a personal digital assistant (PDA), a handheld device, a laptop computer, a desktop computer, a digital camera, a digital recording device, a network-enabled digital television, a mobile phone, a cellular phone, a satellite telephone, a camera phone, a terrestrial-based radiotelephone, a direct two-way communication device (sometimes referred to as a “walkie-talkie”), a camcorder, etc.
  • a landline (non-wireless) phone a wireless communication device
  • PDA personal digital assistant
  • a handheld device a laptop computer
  • a desktop computer a digital camera
  • a digital recording device a network-enabled
  • the vocoder 102 may include an encoder to encode speech signals into packets and a decoder to decode packets into speech signals.
  • the vocoder 102 may be any type of vocoder, such as an enhanced variable rate coder (EVRC), Adaptive Multi-Rate (AMR), Fourth Generation vocoder (4GV), etc.
  • EVRC enhanced variable rate coder
  • AMR Adaptive Multi-Rate
  • 4GV Fourth Generation vocoder
  • Vocoders are described in co-assigned U.S. Pat. Nos. 6,397,175, 6,434,519, 6,438,518, 6,449,592, 6,456,964, 6,477,502, 6,584,438, 6,678,649, 6,691,084, 6,804,218, 6,947,888, which are hereby incorporated by reference.
  • the transceiver 104 may wirelessly transmit and receive packets containing encoded speech.
  • the network 110 may represent one or more base stations, base station controllers (BSCs), mobile switching centers (MSCs), etc. If the first device 100 is a landline phone, then network 110 may include components in a plain old telephone service (POTS) network.
  • POTS plain old telephone service
  • the network 110 comprises a vocoder 112 and a transceiver 114 .
  • the network 110 may include other components in addition to or instead of the components shown in FIG. 1 .
  • the second communication device 120 may represent or be implemented in a wireless communication device, a personal digital assistant (PDA), a handheld device, a laptop computer, a desktop computer, a digital camera, a digital recording device, a network-enabled digital television, a mobile phone, a cellular phone, a satellite telephone, a camera phone, a terrestrial-based radiotelephone, a direct two-way communication device (sometimes referred to as a “walkie-talkie”), a camcorder, etc.
  • PDA personal digital assistant
  • handheld device a laptop computer
  • a desktop computer a digital camera
  • a digital recording device a network-enabled digital television
  • a mobile phone a cellular phone
  • satellite telephone a satellite telephone
  • camera phone a terrestrial-based radiotelephone
  • a direct two-way communication device sometimes referred to as a “walkie-talkie”
  • camcorder a camcorder
  • the second communication device 120 comprises a transceiver 124 , a speech and text unit 140 , a speaker 142 , a display 128 , a user input interface 130 , e.g., a keypad, and a microphone 146 .
  • the speech and text unit 140 comprises a vocoder 122 , a speech to text conversion engine 126 , a controller 144 , a text to speech conversion engine 132 , and a voice synthesizer 134 .
  • the speech and text unit 140 may include other components in addition to or instead of the components shown in FIG. 1 .
  • One or more of the components or functions in the speech and text unit 140 may be integrated into a single module, unit, component, or software.
  • the speech to text conversion engine 126 may be combined with the vocoder 122 .
  • the text to speech conversion engine 132 may be combined with the vocoder 122 , such that text is converted into encoded speech packets.
  • the voice synthesizer 134 may be combined with the vocoder 122 and/or the text to speech conversion engine 132 .
  • the speech to text conversion engine 126 may convert voice/speech to text.
  • the text to speech conversion engine 132 may convert text to speech.
  • the controller 144 may control operations and parameters of one or more components in the speech and text unit 140 .
  • the device 120 may provide several modes of communication for a user to receive calls and/or respond to calls, as shown in the table below and in FIG. 2 .
  • a normal mode (blocks 202 and 210 ) the user of the second device 120 receives a call from the first device 100 , listens to speech from the speaker 142 , and speaks into the microphone 146 .
  • FIG. 2 illustrates a method of using the second device 120 of FIG. 1 .
  • a user of the second device 120 can select one of the modes via the user interface 130 in block 200 .
  • the user may switch between modes in block 200 before the second device 120 receives a call from another device. For example, if the user of the second device 120 enters a meeting, airplane, train, theater, restaurant, church or other place where incoming calls may be disruptive to others, the user may switch from the normal mode to one of the other three modes.
  • the user of the second device 130 may listen to speech from the first device 100 , such as using an ear piece, headset, or headphones, but not talk. Instead, the user of the second device 130 may type on the keypad 130 or use a writing stylus to enter handwritten text on the display 128 .
  • the display 128 or the text to speech conversion engine 132 may have a module that recognizes handwritten text and characters.
  • the device 120 may (a) send the text to the first device 100 or (b) convert the text to speech with the text to speech conversion engine 132 .
  • the voice synthesizer 134 may synthesize the speech to produce personalized speech signals to substantially match the user's natural voice.
  • the voice synthesizer 134 may include a memory that stores characteristics of the user's voice, such as pitch.
  • a voice synthesizer is described in co-assigned U.S. Pat. No. 6,950,799, which is incorporated by reference.
  • Another voice synthesizer is described in co-assigned U.S. patent application Ser. No. 11/398,364, which is incorporated by reference.
  • the vocoder 122 encodes the speech into packets. There may or may not be a short delay. In one configuration, other than a short time delay, communication with the second device 120 may appear seamless to the user of the first device 100 . If the user of the second device 120 is in a meeting, the conversation may be more message-based than seamless.
  • the device 120 receives a call, and the speech to text conversion engine 126 converts speech/voice signals to text for display on the display 128 .
  • the third and fourth modes may allow the user of the first device 100 to continue talking and not require the user of the first device 100 to switch to a text input mode.
  • the speech to text conversion engine 126 may include a voice recognition module to recognize words and sounds to convert them to text.
  • the device 120 allows the user to speak into the microphone 146 , which passes speech to the vocoder 122 to encode into packets.
  • the user of the second device 130 may type on the keypad 130 or use a writing stylus to enter handwritten text on the display 128 .
  • the device 120 may (a) send the text to the first device 100 or (b) convert the text to speech with the text to speech conversion engine 132 .
  • the voice synthesizer 134 may synthesize the speech to produce personalized speech signals to substantially match the user's natural voice.
  • the vocoder 122 encodes the speech into packets.
  • the second device 120 may be configured to play a pre-recorded message by the user to inform the first device 100 that the user of the second device 120 is in a meeting and will respond using text to speech conversion.
  • the second and fourth modes may provide one or more advantages, such as transmitting speech without background noise, no need or reduced need for echo cancellation, no need or reduced need for noise suppression, faster encoding, less processing, etc.
  • FIG. 1 shows an example where changes (new functions and/or elements) may be implemented in only the second communication device 120 .
  • the second communication device 120 has a vocoder 122 , a speech-to-text engine 126 , a text-to-speech engine 132 , etc.
  • the system can support the new modes without any changes in the network 110 and conventional phones 100 (landline, mobile phones, etc.).
  • the device 120 may receive and send voice packets regardless of the mode selected by the user.
  • FIG. 3 illustrates another configuration 100 A of the first communication device 100 of FIG. 1 .
  • the first communication device 100 A comprises a speech to text conversion engine 300 , an encoder 302 , a transceiver 104 , a decoder 304 , and a user interface 330 .
  • the speech to text conversion engine 300 may convert voice/speech to text to be transmitted by the transceiver 104 to the network 110 .
  • the first communication device 100 A of FIG. 3 may allow the second device 120 to be designed without a speech to text conversion engine 126 .
  • the first communication device 100 A of FIG. 3 may save bandwidth by sending text instead of speech to the network 110 .
  • the user interface 330 may be operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the vocoder 302 , 304 , and a second mode enables the speech to text conversion engine 300 .
  • FIG. 4 illustrates another configuration 110 A of the network 110 of FIG. 1 .
  • the network 110 A comprises a voice coder/decoder 400 , a transceiver 114 and a speech to text conversion engine 402 .
  • the network 110 A may further comprise a text to speech conversion engine 404 , a voice synthesizer 402 and a controller 444 .
  • the vocoder 400 decodes speech packets to provide speech signals.
  • the speech to text conversion engine 402 may convert voice/speech to text to be transmitted by the transceiver 114 to the second device 120 .
  • the network 110 A of FIG. 4 may allow the second device 120 to be designed without a speech to text conversion engine 126 or allow the speech to text conversion engine 126 to be deactivated.
  • the network 110 A of FIG. 4 may save bandwidth by sending text instead of speech to the second device 120 .
  • the network 110 A in FIG. 4 may acquire knowledge of a configuration, situation or preference of the receiving device 120 . If the network 110 A realizes that the receiving device 120 will not benefit from receiving voice packets (e.g., sensing a user preference or place of the call, for example, an extremely noisy environment and it is difficult to listen to received speech), then the network 110 A will transform voice packets to text packets. Even if the receiving device 120 has the ability to change voice packets to text packets (using a speech-to-text engine 126 ), it can be a waste of bandwidth and device power to do this transformation (from voice to text) if the user is in a text-receiving mode (a meeting, or silent communication in general).
  • the network 110 A in FIG. 4 may be used in a system where changes (new features and/or elements) are implemented only in the network 110 A, i.e., no changes in communication devices or handsets.
  • the network 110 A may take care of changing voice packets into text and vice versa where the mobile handsets do not have speech to text conversion units; or if the mobile handsets do have speech to text conversion units, the handsets prefer not to do the conversion or cannot do the conversion due to a lack of computational resources, battery power, etc.
  • the first device 100 in FIG. 1 can send ⁇ receive voice packets (i.e., first mode), while the second device 120 sends ⁇ receives text (i.e., fourth mode).
  • the second device 120 may not have unit 140 (or just have a vocoder 122 ) or have unit 140 deactivated.
  • the network 110 A in FIG. 4 will change the first device's voice packets into text packets (using the speech-to-text engine 402 ) to send to the second device 120 and will change text packets from the second device 120 to voice packets (using the text-to-speech engine 404 ) to send to the first device 100 .
  • the second device 120 can signal (in-band for example) a desired mode to the network 110 A and thus ask the network 110 A to convert between speech and text, i.e., do the functions of unit 140 .
  • Personalized speech synthesis may be done in the network 110 A.
  • the unit 140 in FIG. 1 has a voice synthesizer 134 to change the output of the text-to-speech engine 132 to personalized speech (the user's voice).
  • the second device 120 may send stored voice packets (at the beginning of using second or fourth modes) that have the spectral parameters and pitch information of the user to the network 110 A. These few transmitted voice packets (preceding the text packets) can be used by the network 110 A to produce personalized voice packets.
  • the second device 120 transmits to the network 110 A at the beginning of using these “text modes” (second or fourth modes) user pre-stored voice packets (N packets) plus a mode of operation (1, 2, 3, or 4; request to do the conversion).
  • the second device 120 may then send text packets.
  • the network 110 A will enable the text ⁇ speech conversion after sensing (e.g., receiving a request via signaling) the capability of the receiving device 120 , which does the conversion or lets the network 110 A or receiving device 100 A does the conversion.
  • PTT push to talk
  • a PTT device allows a user to push a button on the device and talk, while the device converts speech to text and transmits text packets to a network or directly to another communication device.
  • PTT communication is “message based,” rather than continuous, such as a standard voice call. A time period over which a user holds down the PTT button on the device may nicely frame the message that is then converted to text, etc.
  • the device 120 may have a dedicated memory for storing instructions and data, as well as dedicated hardware, software, firmware, or combinations thereof. If implemented in software, the techniques may be embodied as instructions on a computer-readable medium such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), FLASH memory, or the like.
  • RAM random access memory
  • ROM read-only memory
  • NVRAM non-volatile random access memory
  • EEPROM electrically erasable programmable read-only memory
  • FLASH memory or the like.
  • the instructions cause one or more processors to perform certain aspects of the functionality described in this disclosure.
  • the techniques described in this disclosure may be implemented within a general purpose microprocessor, digital signal processor (DSP), application specific integrated circuit (ASIC), field programmable gate array (FPGA), or other equivalent logic devices.
  • DSP digital signal processor
  • ASIC application specific integrated circuit
  • FPGA field programmable gate array
  • the speech and text unit 140 and associated components and modules may be implemented as parts of an encoding process, or coding/decoding (CODEC) process, running on a digital signal processor (DSP) or other processing device.
  • components described as modules may form programmable features of such a process, or a separate process.
  • the speech and text unit 140 may have a dedicated memory for storing instructions and data, as well as dedicated hardware, software, firmware, or combinations thereof. If implemented in software, the techniques may be embodied as instructions executable by one or more processors.
  • the instructions may be stored on a computer-readable medium such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), FLASH memory, magnetic or optical data storage device, or the like.
  • the instructions cause one or more processors to perform certain aspects of the functionality described in this disclosure.

Abstract

The disclosure relates to systems, methods and apparatus to convert speech to text and vice versa. One apparatus comprises a vocoder, a speech to text conversion engine, a text to speech conversion engine, and a user interface. The vocoder is operable to convert speech signals into packets and convert packets into speech signals. The speech to text conversion engine is operable to convert speech to text. The text to speech conversion engine is operable to convert text to speech. The user interface is operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the speech to text conversion engine, a second mode enables the text to speech conversion engine, and a third mode enables the speech to text conversion engine and the text to speech conversion engine.

Description

    TECHNICAL FIELD
  • The disclosure relates to communications and, more particularly, to a voice and text communication system, method and apparatus.
  • BACKGROUND
  • A cellular phone may include an audio capture device, such as a microphone and/or speech synthesizer, and an audio encoder to generate audio packets or frames. The phone may use communication protocol layers and modules to transmit packets across a wireless communication channel to a network or another communication device.
  • SUMMARY
  • One aspect relates to an apparatus comprising a vocoder, a speech to text conversion engine, a text to speech conversion engine, and a user interface. The vocoder is operable to convert speech signals into packets and convert packets into speech signals. The speech to text conversion engine is operable to convert speech to text. The text to speech conversion engine is operable to convert text to speech. The user interface is operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the speech to text conversion engine, a second mode enables the text to speech conversion engine, and a third mode enables the speech to text conversion engine and the text to speech conversion engine.
  • Another aspect relates to an apparatus comprising: a vocoder operable to convert speech signals into packets and convert packets into speech signals; a speech to text conversion engine operable to convert speech to text; a user interface operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the vocoder, and a second mode enables the speech to text conversion engine; and a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
  • Another aspect relates to a network apparatus comprising: a vocoder operable to convert packets into speech signals; a speech to text conversion engine operable to convert speech to text; a selection unit operable to switch between first and second modes, wherein the first mode enables the vocoder, and a second mode enables the vocoder and the speech to text conversion engine; and a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
  • Another aspect relates to a method comprising: receiving encoded speech packets; converting the received encoded speech packets into speech signals; and receiving a user selection of a mode from among a plurality of modes, wherein a first mode enables speech to text conversion, a second mode enables text to speech conversion, and a third mode enables speech to text and text to speech conversion.
  • The details of one or more embodiments are set forth in the accompanying drawings and the description below.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 illustrates a system comprising a first communication device, a network, and a second communication device.
  • FIG. 2 illustrates a method of using the second device of FIG. 1.
  • FIG. 3 illustrates another configuration of the first communication device of FIG. 1.
  • FIG. 4 illustrates another configuration of the network of FIG. 1.
  • DETAILED DESCRIPTION
  • Receiving a call on a mobile device in a meeting, airplane, train, theater, restaurant, church or other place may be disruptive to others. It may be much less disruptive if a user could select another mode on the mobile device to receive the call and/or respond to the call. In one mode, the device receives the call and converts speech/voice signals to text without requiring the caller on the other end to input text.
  • FIG. 1 illustrates a system comprising a first communication device 100, a network 110, and a second communication device 120. The system may include other components. The system may use any type of wireless communication, such as Global System for Mobile communications (GSM), code division multiple access (CDMA), CDMA2000, CDMA2000 1x EV-DO, Wideband CDMA (WCDMA), orthogonal frequency division multiple access (OFDMA), Bluetooth, WiFi, WiMax, etc.
  • The first communication device 100 comprises a voice coder (vocoder) 102 and a transceiver 104. The first communication device 100 may include other components in addition to or instead of the components shown in FIG. 1. The first communication device 100 may represent or be implemented in a landline (non-wireless) phone, a wireless communication device, a personal digital assistant (PDA), a handheld device, a laptop computer, a desktop computer, a digital camera, a digital recording device, a network-enabled digital television, a mobile phone, a cellular phone, a satellite telephone, a camera phone, a terrestrial-based radiotelephone, a direct two-way communication device (sometimes referred to as a “walkie-talkie”), a camcorder, etc.
  • The vocoder 102 may include an encoder to encode speech signals into packets and a decoder to decode packets into speech signals. The vocoder 102 may be any type of vocoder, such as an enhanced variable rate coder (EVRC), Adaptive Multi-Rate (AMR), Fourth Generation vocoder (4GV), etc. Vocoders are described in co-assigned U.S. Pat. Nos. 6,397,175, 6,434,519, 6,438,518, 6,449,592, 6,456,964, 6,477,502, 6,584,438, 6,678,649, 6,691,084, 6,804,218, 6,947,888, which are hereby incorporated by reference.
  • The transceiver 104 may wirelessly transmit and receive packets containing encoded speech.
  • The network 110 may represent one or more base stations, base station controllers (BSCs), mobile switching centers (MSCs), etc. If the first device 100 is a landline phone, then network 110 may include components in a plain old telephone service (POTS) network. The network 110 comprises a vocoder 112 and a transceiver 114. The network 110 may include other components in addition to or instead of the components shown in FIG. 1.
  • The second communication device 120 may represent or be implemented in a wireless communication device, a personal digital assistant (PDA), a handheld device, a laptop computer, a desktop computer, a digital camera, a digital recording device, a network-enabled digital television, a mobile phone, a cellular phone, a satellite telephone, a camera phone, a terrestrial-based radiotelephone, a direct two-way communication device (sometimes referred to as a “walkie-talkie”), a camcorder, etc.
  • The second communication device 120 comprises a transceiver 124, a speech and text unit 140, a speaker 142, a display 128, a user input interface 130, e.g., a keypad, and a microphone 146. The speech and text unit 140 comprises a vocoder 122, a speech to text conversion engine 126, a controller 144, a text to speech conversion engine 132, and a voice synthesizer 134. The speech and text unit 140 may include other components in addition to or instead of the components shown in FIG. 1.
  • One or more of the components or functions in the speech and text unit 140 may be integrated into a single module, unit, component, or software. For example, the speech to text conversion engine 126 may be combined with the vocoder 122. The text to speech conversion engine 132 may be combined with the vocoder 122, such that text is converted into encoded speech packets. The voice synthesizer 134 may be combined with the vocoder 122 and/or the text to speech conversion engine 132.
  • The speech to text conversion engine 126 may convert voice/speech to text. The text to speech conversion engine 132 may convert text to speech. The controller 144 may control operations and parameters of one or more components in the speech and text unit 140.
  • The device 120 may provide several modes of communication for a user to receive calls and/or respond to calls, as shown in the table below and in FIG. 2.
  • Mode Listen Speak
    Normal mode Yes Yes
    Second mode Yes No - transmit text or
    synthesized speech
    Third mode No - convert incoming Yes
    speech to text
    Fourth mode No - convert incoming No - transmit text or
    speech to text synthesized speech

    In a normal mode (blocks 202 and 210), the user of the second device 120 receives a call from the first device 100, listens to speech from the speaker 142, and speaks into the microphone 146.
  • FIG. 2 illustrates a method of using the second device 120 of FIG. 1. When the second device 120 receives a call from the first device 100, a user of the second device 120 can select one of the modes via the user interface 130 in block 200. Alternatively, the user may switch between modes in block 200 before the second device 120 receives a call from another device. For example, if the user of the second device 120 enters a meeting, airplane, train, theater, restaurant, church or other place where incoming calls may be disruptive to others, the user may switch from the normal mode to one of the other three modes.
  • In a second mode (blocks 204 and 212), the user of the second device 130 may listen to speech from the first device 100, such as using an ear piece, headset, or headphones, but not talk. Instead, the user of the second device 130 may type on the keypad 130 or use a writing stylus to enter handwritten text on the display 128. The display 128 or the text to speech conversion engine 132 may have a module that recognizes handwritten text and characters. The device 120 may (a) send the text to the first device 100 or (b) convert the text to speech with the text to speech conversion engine 132.
  • The voice synthesizer 134 may synthesize the speech to produce personalized speech signals to substantially match the user's natural voice. The voice synthesizer 134 may include a memory that stores characteristics of the user's voice, such as pitch. A voice synthesizer is described in co-assigned U.S. Pat. No. 6,950,799, which is incorporated by reference. Another voice synthesizer is described in co-assigned U.S. patent application Ser. No. 11/398,364, which is incorporated by reference.
  • The vocoder 122 encodes the speech into packets. There may or may not be a short delay. In one configuration, other than a short time delay, communication with the second device 120 may appear seamless to the user of the first device 100. If the user of the second device 120 is in a meeting, the conversation may be more message-based than seamless.
  • In third and fourth modes ( blocks 206, 208, 214 and 216), the device 120 receives a call, and the speech to text conversion engine 126 converts speech/voice signals to text for display on the display 128. In one configuration, the third and fourth modes may allow the user of the first device 100 to continue talking and not require the user of the first device 100 to switch to a text input mode. The speech to text conversion engine 126 may include a voice recognition module to recognize words and sounds to convert them to text.
  • In the third mode, the device 120 allows the user to speak into the microphone 146, which passes speech to the vocoder 122 to encode into packets.
  • In the fourth mode, the user of the second device 130 may type on the keypad 130 or use a writing stylus to enter handwritten text on the display 128. The device 120 may (a) send the text to the first device 100 or (b) convert the text to speech with the text to speech conversion engine 132. The voice synthesizer 134 may synthesize the speech to produce personalized speech signals to substantially match the user's natural voice. The vocoder 122 encodes the speech into packets.
  • In the second and fourth modes, if the second device 120 is set to convert text to speech and synthesize speech, there may be a time delay between when the second device 120 accepts a call from the first device 100 and when the first device 100 receives speech packets. The second device 120 may be configured to play a pre-recorded message by the user to inform the first device 100 that the user of the second device 120 is in a meeting and will respond using text to speech conversion.
  • The second and fourth modes may provide one or more advantages, such as transmitting speech without background noise, no need or reduced need for echo cancellation, no need or reduced need for noise suppression, faster encoding, less processing, etc.
  • FIG. 1 shows an example where changes (new functions and/or elements) may be implemented in only the second communication device 120. To realize the new modes (second, third and fourth modes) of communication, the second communication device 120 has a vocoder 122, a speech-to-text engine 126, a text-to-speech engine 132, etc. With this device 120, the system can support the new modes without any changes in the network 110 and conventional phones 100 (landline, mobile phones, etc.). The device 120 may receive and send voice packets regardless of the mode selected by the user.
  • FIG. 3 illustrates another configuration 100A of the first communication device 100 of FIG. 1. In FIG. 3, the first communication device 100A comprises a speech to text conversion engine 300, an encoder 302, a transceiver 104, a decoder 304, and a user interface 330. The speech to text conversion engine 300 may convert voice/speech to text to be transmitted by the transceiver 104 to the network 110. The first communication device 100A of FIG. 3 may allow the second device 120 to be designed without a speech to text conversion engine 126. The first communication device 100A of FIG. 3 may save bandwidth by sending text instead of speech to the network 110. The user interface 330 may be operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the vocoder 302, 304, and a second mode enables the speech to text conversion engine 300.
  • FIG. 4 illustrates another configuration 110A of the network 110 of FIG. 1. In FIG. 4, the network 110A comprises a voice coder/decoder 400, a transceiver 114 and a speech to text conversion engine 402. In another configuration, the network 110A may further comprise a text to speech conversion engine 404, a voice synthesizer 402 and a controller 444. The vocoder 400 decodes speech packets to provide speech signals. The speech to text conversion engine 402 may convert voice/speech to text to be transmitted by the transceiver 114 to the second device 120. The network 110A of FIG. 4 may allow the second device 120 to be designed without a speech to text conversion engine 126 or allow the speech to text conversion engine 126 to be deactivated. The network 110A of FIG. 4 may save bandwidth by sending text instead of speech to the second device 120.
  • The network 110A in FIG. 4 may acquire knowledge of a configuration, situation or preference of the receiving device 120. If the network 110A realizes that the receiving device 120 will not benefit from receiving voice packets (e.g., sensing a user preference or place of the call, for example, an extremely noisy environment and it is difficult to listen to received speech), then the network 110A will transform voice packets to text packets. Even if the receiving device 120 has the ability to change voice packets to text packets (using a speech-to-text engine 126), it can be a waste of bandwidth and device power to do this transformation (from voice to text) if the user is in a text-receiving mode (a meeting, or silent communication in general).
  • Thus, the network 110A in FIG. 4 may be used in a system where changes (new features and/or elements) are implemented only in the network 110A, i.e., no changes in communication devices or handsets. The network 110A may take care of changing voice packets into text and vice versa where the mobile handsets do not have speech to text conversion units; or if the mobile handsets do have speech to text conversion units, the handsets prefer not to do the conversion or cannot do the conversion due to a lack of computational resources, battery power, etc.
  • For example, the first device 100 in FIG. 1 can send\receive voice packets (i.e., first mode), while the second device 120 sends\receives text (i.e., fourth mode). The second device 120 may not have unit 140 (or just have a vocoder 122) or have unit 140 deactivated. To allow the second device 120 to operate in the fourth mode, the network 110A in FIG. 4 will change the first device's voice packets into text packets (using the speech-to-text engine 402) to send to the second device 120 and will change text packets from the second device 120 to voice packets (using the text-to-speech engine 404) to send to the first device 100.
  • If the second device 120 does not have the unit 140, the second device 120 can signal (in-band for example) a desired mode to the network 110A and thus ask the network 110A to convert between speech and text, i.e., do the functions of unit 140.
  • Personalized speech synthesis may be done in the network 110A. As described above, the unit 140 in FIG. 1 has a voice synthesizer 134 to change the output of the text-to-speech engine 132 to personalized speech (the user's voice). In a system with the network 110A of FIG. 4, to produce voice packets that carry a voice signature of the user of the second device 120, the second device 120 may send stored voice packets (at the beginning of using second or fourth modes) that have the spectral parameters and pitch information of the user to the network 110A. These few transmitted voice packets (preceding the text packets) can be used by the network 110A to produce personalized voice packets.
  • An example of transmitting packets for second or fourth modes from the second device 120 to the network 110A is described. The second device 120 transmits to the network 110A at the beginning of using these “text modes” (second or fourth modes) user pre-stored voice packets (N packets) plus a mode of operation (1, 2, 3, or 4; request to do the conversion). The second device 120 may then send text packets.
  • A combination of the two configurations (FIG. 1 and FIG. 4) is also possible. When using one of these modes, the network 110A will enable the text\speech conversion after sensing (e.g., receiving a request via signaling) the capability of the receiving device 120, which does the conversion or lets the network 110A or receiving device 100A does the conversion.
  • One or more components and features described above may be implemented in a push to talk (PTT) or push to read communication device. A PTT device allows a user to push a button on the device and talk, while the device converts speech to text and transmits text packets to a network or directly to another communication device. PTT communication is “message based,” rather than continuous, such as a standard voice call. A time period over which a user holds down the PTT button on the device may nicely frame the message that is then converted to text, etc.
  • The device 120 may have a dedicated memory for storing instructions and data, as well as dedicated hardware, software, firmware, or combinations thereof. If implemented in software, the techniques may be embodied as instructions on a computer-readable medium such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), FLASH memory, or the like. The instructions cause one or more processors to perform certain aspects of the functionality described in this disclosure.
  • The techniques described in this disclosure may be implemented within a general purpose microprocessor, digital signal processor (DSP), application specific integrated circuit (ASIC), field programmable gate array (FPGA), or other equivalent logic devices. For example, the speech and text unit 140 and associated components and modules, may be implemented as parts of an encoding process, or coding/decoding (CODEC) process, running on a digital signal processor (DSP) or other processing device. Accordingly, components described as modules may form programmable features of such a process, or a separate process.
  • The speech and text unit 140 may have a dedicated memory for storing instructions and data, as well as dedicated hardware, software, firmware, or combinations thereof. If implemented in software, the techniques may be embodied as instructions executable by one or more processors. The instructions may be stored on a computer-readable medium such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), FLASH memory, magnetic or optical data storage device, or the like. The instructions cause one or more processors to perform certain aspects of the functionality described in this disclosure.
  • Various embodiments have been described. These and other embodiments are within the scope of the following claims.

Claims (21)

1. An apparatus comprising:
a vocoder operable to convert speech signals into packets and convert packets into speech signals;
a speech to text conversion engine operable to convert speech to text;
a text to speech conversion engine operable to convert text to speech; and
a user interface operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the speech to text conversion engine, a second mode enables the text to speech conversion engine, and a third mode enables the speech to text conversion engine and the text to speech conversion engine.
2. The apparatus of claim 1, further comprising a display to display text from the speech to text conversion engine.
3. The apparatus of claim 1, further comprising a keypad to receive input text from a user.
4. The apparatus of claim 1, wherein the user interface is operable to receive a user selection of a mode before the apparatus receives a call from another apparatus.
5. The apparatus of claim 1, wherein the user interface is operable to receive a user selection of a mode after the apparatus receives a call from another apparatus.
6. The apparatus of claim 1, further comprising a voice synthesizer to synthesize a user's voice.
7. The apparatus of claim 1, further comprising a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
8. An apparatus comprising:
a vocoder operable to convert speech signals into packets and convert packets into speech signals;
a speech to text conversion engine operable to convert speech to text;
a user interface operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the vocoder, and a second mode enables the speech to text conversion engine; and
a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
9. The apparatus of claim 8, further comprising a display to display text from the speech to text conversion engine.
10. The apparatus of claim 8, further comprising a keypad to receive input text from a user.
11. The apparatus of claim 8, wherein the user interface is operable to receive a user selection of a mode before the apparatus receives a call from another apparatus.
12. The apparatus of claim 8, wherein the user interface is operable to receive a user selection of a mode after the apparatus receives a call from another apparatus.
13. A network apparatus comprising:
a vocoder operable to convert packets into speech signals;
a speech to text conversion engine operable to convert speech to text;
a selection unit operable to switch between first and second modes, wherein the first mode enables the vocoder, and a second mode enables the vocoder and the speech to text conversion engine; and
a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
14. The network apparatus of claim 13, further comprising a text to speech conversion engine operable to convert text to speech, wherein the selection unit is operable to switch to a third mode where the vocoder and both conversion engines are enabled.
15. The network apparatus of claim 14, further comprising a voice synthesizer operable to synthesize a user's voice from text converted to speech.
16. The network apparatus of claim 15, wherein the voice synthesizer is operable to receive and store voice characteristics of a user's voice.
17. The network apparatus of claim 13, further comprising a controller operable to receive a request from a communication device to convert speech to text.
18. The network apparatus of claim 13, further comprising a controller operable to receive a request from a communication device to convert text to speech.
19. A method comprising:
receiving encoded speech packets;
converting the received encoded speech packets into speech signals; and
receiving a user selection of a mode from among a plurality of modes, wherein a first mode enables speech to text conversion, a second mode enables text to speech conversion, and a third mode enables speech to text and text to speech conversion.
20. The method of claim 19, further comprising receiving a user selection for a mode before receiving an incoming call.
21. The method of claim 19, further comprising receiving a user selection for a mode after receiving an incoming call.
US11/497,011 2006-07-31 2006-07-31 Voice and text communication system, method and apparatus Abandoned US20100030557A1 (en)

Priority Applications (7)

Application Number Priority Date Filing Date Title
US11/497,011 US20100030557A1 (en) 2006-07-31 2006-07-31 Voice and text communication system, method and apparatus
JP2009523030A JP5149292B2 (en) 2006-07-31 2007-07-31 Voice and text communication system, method and apparatus
KR1020097003859A KR101136769B1 (en) 2006-07-31 2007-07-31 Voice and text communication system, method and apparatus
EP07840622A EP2047459A2 (en) 2006-07-31 2007-07-31 Voice and text communication system, method and apparatus
CN2007800282788A CN101496096B (en) 2006-07-31 2007-07-31 Voice and text communication system, method and apparatus
PCT/US2007/074902 WO2008016949A2 (en) 2006-07-31 2007-07-31 Voice and text communication system, method and apparatus
US14/981,298 US9940923B2 (en) 2006-07-31 2015-12-28 Voice and text communication system, method and apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/497,011 US20100030557A1 (en) 2006-07-31 2006-07-31 Voice and text communication system, method and apparatus

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US14/981,298 Division US9940923B2 (en) 2006-07-31 2015-12-28 Voice and text communication system, method and apparatus

Publications (1)

Publication Number Publication Date
US20100030557A1 true US20100030557A1 (en) 2010-02-04

Family

ID=38871584

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/497,011 Abandoned US20100030557A1 (en) 2006-07-31 2006-07-31 Voice and text communication system, method and apparatus
US14/981,298 Active US9940923B2 (en) 2006-07-31 2015-12-28 Voice and text communication system, method and apparatus

Family Applications After (1)

Application Number Title Priority Date Filing Date
US14/981,298 Active US9940923B2 (en) 2006-07-31 2015-12-28 Voice and text communication system, method and apparatus

Country Status (6)

Country Link
US (2) US20100030557A1 (en)
EP (1) EP2047459A2 (en)
JP (1) JP5149292B2 (en)
KR (1) KR101136769B1 (en)
CN (1) CN101496096B (en)
WO (1) WO2008016949A2 (en)

Cited By (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090276214A1 (en) * 2008-04-30 2009-11-05 Motorola, Inc. Method for dual channel monitoring on a radio device
US20110015930A1 (en) * 2009-07-08 2011-01-20 Intelepeer, Inc. Unified communication system
US20110092187A1 (en) * 2009-10-15 2011-04-21 Larry Miller Configurable phone with interactive voice response engine
US20110112834A1 (en) * 2009-11-10 2011-05-12 Samsung Electronics Co., Ltd. Communication method and terminal
US7945236B1 (en) 2001-10-18 2011-05-17 Iwao Fujisaki Communication device
US7996038B1 (en) 2003-09-26 2011-08-09 Iwao Fujisaki Communication device
US8064964B1 (en) 2001-10-18 2011-11-22 Iwao Fujisaki Communication device
US8081962B1 (en) 2004-03-23 2011-12-20 Iwao Fujisaki Communication device
US8121635B1 (en) 2003-11-22 2012-02-21 Iwao Fujisaki Communication device
US8150473B1 (en) 2001-10-18 2012-04-03 Iwao Fujisaki Communication device
US8208954B1 (en) 2005-04-08 2012-06-26 Iwao Fujisaki Communication device
US8229512B1 (en) 2003-02-08 2012-07-24 Iwao Fujisaki Communication device
US8241128B1 (en) 2003-04-03 2012-08-14 Iwao Fujisaki Communication device
US8340726B1 (en) 2008-06-30 2012-12-25 Iwao Fujisaki Communication device
US20130080917A1 (en) * 2011-09-28 2013-03-28 Royce A. Levien Multi-Modality communication modification
US8423366B1 (en) * 2012-07-18 2013-04-16 Google Inc. Automatically training speech synthesizers
CN103108092A (en) * 2011-11-15 2013-05-15 希姆通信息技术(上海)有限公司 Communication terminal and information transmitting method between communication terminals
US8452307B1 (en) 2008-07-02 2013-05-28 Iwao Fujisaki Communication device
US8472935B1 (en) 2007-10-29 2013-06-25 Iwao Fujisaki Communication device
US8543157B1 (en) 2008-05-09 2013-09-24 Iwao Fujisaki Communication device which notifies its pin-point location or geographic area in accordance with user selection
US8584388B1 (en) 2008-05-09 2013-11-19 Iwao Fujisaki Firearm
US8639214B1 (en) 2007-10-26 2014-01-28 Iwao Fujisaki Communication device
US8676273B1 (en) * 2007-08-24 2014-03-18 Iwao Fujisaki Communication device
US8825090B1 (en) 2007-05-03 2014-09-02 Iwao Fujisaki Communication device
US8825026B1 (en) 2007-05-03 2014-09-02 Iwao Fujisaki Communication device
US9002937B2 (en) 2011-09-28 2015-04-07 Elwha Llc Multi-party multi-modality communication
US20150100315A1 (en) * 2013-08-29 2015-04-09 Vonage Network, Llc Methods and apparatus for conducting internet protocol telephony communications
US20150199956A1 (en) * 2014-01-14 2015-07-16 Interactive Intelligence Group, Inc. System and method for synthesis of speech from provided text
WO2015119492A1 (en) * 2014-02-05 2015-08-13 Alvarez Heinemeyer Marco System, method and computer-readable medium for a language interface
US9139089B1 (en) 2007-12-27 2015-09-22 Iwao Fujisaki Inter-vehicle middle point maintaining implementer
US20160171982A1 (en) * 2014-12-10 2016-06-16 Honeywell International Inc. High intelligibility voice announcement system
US9477943B2 (en) 2011-09-28 2016-10-25 Elwha Llc Multi-modality communication
US20160372114A1 (en) * 2015-06-18 2016-12-22 Airbus Operations Gmbh Announcement signaling on board an aircraft
US9699632B2 (en) 2011-09-28 2017-07-04 Elwha Llc Multi-modality communication with interceptive conversion
US9762524B2 (en) 2011-09-28 2017-09-12 Elwha Llc Multi-modality communication participation
US9788349B2 (en) 2011-09-28 2017-10-10 Elwha Llc Multi-modality communication auto-activation
US9906927B2 (en) 2011-09-28 2018-02-27 Elwha Llc Multi-modality communication initiation
US20180096334A1 (en) * 2016-10-03 2018-04-05 Paypal, Inc. Voice activated remittances
US9940923B2 (en) 2006-07-31 2018-04-10 Qualcomm Incorporated Voice and text communication system, method and apparatus
US10805774B2 (en) 2016-12-19 2020-10-13 Samsung Electronics Co., Ltd. Methods and apparatus for managing control data
US11449868B2 (en) 2016-10-03 2022-09-20 Paypal, Inc. Voice activated remittances

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101237489A (en) * 2008-03-05 2008-08-06 北京邮电大学 Processing method and device based on voice communication content
CN102687583B (en) 2009-12-31 2016-02-24 诺基亚技术有限公司 For performing the method and apparatus of multiple communication form in a session
CN102984672A (en) * 2011-09-07 2013-03-20 比亚迪股份有限公司 Mobile terminal and communication method thereof
US10129340B2 (en) 2012-08-07 2018-11-13 Nokia Technologies Oy Method and apparatus for performing multiple forms of communications in one session
KR101398806B1 (en) * 2012-11-21 2014-05-27 오준수 Peer to peer communication method between mobile terminal equipments and recording medium thereof
CN103973542B (en) * 2013-02-01 2017-06-13 腾讯科技(深圳)有限公司 A kind of voice information processing method and device
KR102225401B1 (en) * 2014-05-23 2021-03-09 삼성전자주식회사 System and method for providing voice-message call service
EP3496377B1 (en) 2014-05-23 2020-09-30 Samsung Electronics Co., Ltd. System and method of providing voice-message call service
CN104123932B (en) * 2014-07-29 2017-11-07 科大讯飞股份有限公司 A kind of speech conversion system and method
CN105530351B (en) * 2014-10-23 2019-09-10 中兴通讯股份有限公司 Call handling method and device
US9491270B1 (en) * 2015-11-13 2016-11-08 Motorola Solutions, Inc. Method and apparatus for muting an audio output interface of a portable communications device
CN109600307A (en) * 2019-01-29 2019-04-09 北京百度网讯科技有限公司 Instant communication method, terminal, equipment, computer-readable medium
US11468896B2 (en) * 2019-06-12 2022-10-11 Nvoq Incorporated Systems, methods, and apparatus for real-time dictation and transcription with multiple remote endpoints
CN112967719A (en) * 2019-12-12 2021-06-15 上海棋语智能科技有限公司 Computer terminal access equipment of standard radio station hand microphone

Citations (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5750912A (en) * 1996-01-18 1998-05-12 Yamaha Corporation Formant converting apparatus modifying singing voice to emulate model voice
US5963620A (en) * 1995-07-18 1999-10-05 Jetstream Communications, Inc. Integrated communications control device for a small office configured for coupling within a scalable network including multiple simultaneous call capability
US6195358B1 (en) * 1997-03-15 2001-02-27 International Business Machines Corp. Internet telephony signal conversion
US6456964B2 (en) * 1998-12-21 2002-09-24 Qualcomm, Incorporated Encoding of periodic speech using prototype waveforms
US20020184004A1 (en) * 2001-05-10 2002-12-05 Utaha Shizuka Information processing apparatus, information processing method, recording medium, and program
US6502073B1 (en) * 1999-03-25 2002-12-31 Kent Ridge Digital Labs Low data transmission rate and intelligible speech communication
US20030083884A1 (en) * 2001-10-26 2003-05-01 Gilad Odinak Real-time display of system instructions
US6574489B1 (en) * 1998-06-26 2003-06-03 Nec Corporation Incoming call notification method and device for a multimode radio device
US20040034528A1 (en) * 2002-06-12 2004-02-19 Canon Kabushiki Kaisha Server and receiving terminal
US20040053646A1 (en) * 2000-12-22 2004-03-18 Jun Noguchi Radio mobile terminal communication system
US6745024B1 (en) * 2000-01-10 2004-06-01 Qualcomm Incorporated System and method for preparing and sending an electronic mail communication using a wireless communications device
US20040203448A1 (en) * 2002-03-28 2004-10-14 Nec Corporation Method for changing voice coding mode, communication system, communication network and communication terminal
US6845359B2 (en) * 2001-03-22 2005-01-18 Motorola, Inc. FFT based sine wave synthesis method for parametric vocoders
US20050058075A1 (en) * 2003-09-12 2005-03-17 Gorday Robert M. Apparatus and method for mixed-media call formatting
US20050201291A1 (en) * 2004-03-11 2005-09-15 Ixi Mobile (R&D) Ltd. Power management system and method for a wireless communications device
US6950799B2 (en) * 2002-02-19 2005-09-27 Qualcomm Inc. Speech converter utilizing preprogrammed voice profiles
US20060095265A1 (en) * 2004-10-29 2006-05-04 Microsoft Corporation Providing personalized voice front for text-to-speech applications
US20060122840A1 (en) * 2004-12-07 2006-06-08 David Anderson Tailoring communication from interactive speech enabled and multimodal services
US20060136226A1 (en) * 2004-10-06 2006-06-22 Ossama Emam System and method for creating artificial TV news programs
US20060149558A1 (en) * 2001-07-17 2006-07-06 Jonathan Kahn Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile
US7136811B2 (en) * 2002-04-24 2006-11-14 Motorola, Inc. Low bandwidth speech communication using default and personal phoneme tables
US20070036086A1 (en) * 2005-08-09 2007-02-15 Sbc Knowledge Ventures, L.P. System and method of providing communications based on a predetermined device status
US20070116194A1 (en) * 2005-10-27 2007-05-24 International Business Machines Corporation Communications involving devices having different communication modes
US7233655B2 (en) * 2001-10-03 2007-06-19 Accenture Global Services Gmbh Multi-modal callback
US20070143103A1 (en) * 2005-12-21 2007-06-21 Cisco Technology, Inc. Conference captioning
US7251152B2 (en) * 2004-08-26 2007-07-31 Infineon Technologies Ag Memory circuit having memory cells which have a resistance memory element
US7251252B2 (en) * 2001-06-13 2007-07-31 Qwest Communications International Inc. Negotiated cell delivery capability
US7305342B2 (en) * 2001-05-10 2007-12-04 Sony Corporation Text-to-speech synthesis system and associated method of associating content information
US7464034B2 (en) * 1999-10-21 2008-12-09 Yamaha Corporation Voice converter for assimilation by frame synthesis with temporal alignment

Family Cites Families (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1984004458A1 (en) * 1983-05-06 1984-11-22 Matthew Pollack Monoclonal antibodies reactive with endotoxin core
US5917405A (en) * 1993-06-08 1999-06-29 Joao; Raymond Anthony Control apparatus and methods for vehicles
JP3781213B2 (en) * 1996-06-20 2006-05-31 ソニー株式会社 E-mail system, computer apparatus and incoming call notification method
US5786768A (en) * 1997-04-16 1998-07-28 Patrick Plastics Inc. Clock radio gas detector apparatus and method for alerting residents to hazardous gas concentrations
US6178170B1 (en) * 1997-05-13 2001-01-23 Sprint Communications Company, L. P. System and method for transporting a call
US5983190A (en) * 1997-05-19 1999-11-09 Microsoft Corporation Client server animation system for managing interactive user interface characters
US6366651B1 (en) * 1998-01-21 2002-04-02 Avaya Technology Corp. Communication device having capability to convert between voice and text message
US6094681A (en) * 1998-03-31 2000-07-25 Siemens Information And Communication Networks, Inc. Apparatus and method for automated event notification
US7009990B1 (en) * 1998-04-23 2006-03-07 Nortel Networks Limited Methods and systems for a telephone e-mail interface
US6173250B1 (en) * 1998-06-03 2001-01-09 At&T Corporation Apparatus and method for speech-text-transmit communication over data networks
CN1160700C (en) 1998-10-02 2004-08-04 国际商业机器公司 System and method for providing network coordinated conversational services
JP3842913B2 (en) * 1998-12-18 2006-11-08 富士通株式会社 Character communication method and character communication system
US6775360B2 (en) * 2000-12-28 2004-08-10 Intel Corporation Method and system for providing textual content along with voice messages
KR20020091920A (en) * 2001-06-01 2002-12-11 엘지전자 주식회사 A calling system by voice and charater in a mobile phone
US7085259B2 (en) * 2001-07-31 2006-08-01 Comverse, Inc. Animated audio messaging
JP3733322B2 (en) * 2001-11-21 2006-01-11 キヤノン株式会社 Multimodal document receiving apparatus, multimodal document transmitting apparatus, multimodal document transmitting / receiving system, control method therefor, and program
US7483832B2 (en) 2001-12-10 2009-01-27 At&T Intellectual Property I, L.P. Method and system for customizing voice translation of text to speech
US20040098266A1 (en) 2002-11-14 2004-05-20 International Business Machines Corporation Personal speech font
US20050048992A1 (en) * 2003-08-28 2005-03-03 Alcatel Multimode voice/screen simultaneous communication device
US20050266829A1 (en) * 2004-04-16 2005-12-01 Lg Elcectronics, Inc. Speech-to-text messaging system and method
WO2005109661A1 (en) * 2004-05-10 2005-11-17 Sk Telecom Co., Ltd. Mobile communication terminal for transferring and receiving of voice message and method for transferring and receiving of voice message using the same
KR100617826B1 (en) * 2004-05-11 2006-08-28 삼성전자주식회사 Mobile terminal and method for converting voice message to text message
JP4025355B2 (en) 2004-10-13 2007-12-19 松下電器産業株式会社 Speech synthesis apparatus and speech synthesis method
US7333014B2 (en) * 2004-11-04 2008-02-19 International Business Machines Corporation Notifying users of device events in a networked environment
US7296740B2 (en) * 2004-11-04 2007-11-20 International Business Machines Corporation Routing telecommunications to a user in dependence upon location
US7324476B2 (en) * 2004-11-04 2008-01-29 International Business Machines Corporation Establishing user accounts for RFID-based telecommunications routing
EP1694044B1 (en) * 2005-02-17 2008-06-04 Sony Ericsson Mobile Communications AB Method and communication system for transferring messages containing audio and/or voice information
US7917178B2 (en) * 2005-03-22 2011-03-29 Sony Ericsson Mobile Communications Ab Wireless communications device with voice-to-text conversion
US7394405B2 (en) * 2005-06-01 2008-07-01 Gm Global Technology Operations, Inc. Location-based notifications
US20100030557A1 (en) 2006-07-31 2010-02-04 Stephen Molloy Voice and text communication system, method and apparatus
EP2119208A1 (en) * 2007-01-09 2009-11-18 Spinvox Limited Selection of a link in a received message for speaking reply, which is converted into text form for delivery
US8914290B2 (en) * 2011-05-20 2014-12-16 Vocollect, Inc. Systems and methods for dynamically improving user intelligibility of synthesized speech in a work environment

Patent Citations (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5963620A (en) * 1995-07-18 1999-10-05 Jetstream Communications, Inc. Integrated communications control device for a small office configured for coupling within a scalable network including multiple simultaneous call capability
US5750912A (en) * 1996-01-18 1998-05-12 Yamaha Corporation Formant converting apparatus modifying singing voice to emulate model voice
US6195358B1 (en) * 1997-03-15 2001-02-27 International Business Machines Corp. Internet telephony signal conversion
US6574489B1 (en) * 1998-06-26 2003-06-03 Nec Corporation Incoming call notification method and device for a multimode radio device
US6456964B2 (en) * 1998-12-21 2002-09-24 Qualcomm, Incorporated Encoding of periodic speech using prototype waveforms
US6502073B1 (en) * 1999-03-25 2002-12-31 Kent Ridge Digital Labs Low data transmission rate and intelligible speech communication
US7464034B2 (en) * 1999-10-21 2008-12-09 Yamaha Corporation Voice converter for assimilation by frame synthesis with temporal alignment
US6745024B1 (en) * 2000-01-10 2004-06-01 Qualcomm Incorporated System and method for preparing and sending an electronic mail communication using a wireless communications device
US7292878B2 (en) * 2000-12-22 2007-11-06 Nec Corporation Radio mobile terminal communication system
US20040053646A1 (en) * 2000-12-22 2004-03-18 Jun Noguchi Radio mobile terminal communication system
US6845359B2 (en) * 2001-03-22 2005-01-18 Motorola, Inc. FFT based sine wave synthesis method for parametric vocoders
US20020184004A1 (en) * 2001-05-10 2002-12-05 Utaha Shizuka Information processing apparatus, information processing method, recording medium, and program
US7305342B2 (en) * 2001-05-10 2007-12-04 Sony Corporation Text-to-speech synthesis system and associated method of associating content information
US7251252B2 (en) * 2001-06-13 2007-07-31 Qwest Communications International Inc. Negotiated cell delivery capability
US20060149558A1 (en) * 2001-07-17 2006-07-06 Jonathan Kahn Synchronized pattern recognition source data processed by manual or automatic means for creation of shared speaker-dependent speech user profile
US7233655B2 (en) * 2001-10-03 2007-06-19 Accenture Global Services Gmbh Multi-modal callback
US20030083884A1 (en) * 2001-10-26 2003-05-01 Gilad Odinak Real-time display of system instructions
US7406421B2 (en) * 2001-10-26 2008-07-29 Intellisist Inc. Systems and methods for reviewing informational content in a vehicle
US6950799B2 (en) * 2002-02-19 2005-09-27 Qualcomm Inc. Speech converter utilizing preprogrammed voice profiles
US20040203448A1 (en) * 2002-03-28 2004-10-14 Nec Corporation Method for changing voice coding mode, communication system, communication network and communication terminal
US7136811B2 (en) * 2002-04-24 2006-11-14 Motorola, Inc. Low bandwidth speech communication using default and personal phoneme tables
US20040034528A1 (en) * 2002-06-12 2004-02-19 Canon Kabushiki Kaisha Server and receiving terminal
US7519042B2 (en) * 2003-09-12 2009-04-14 Motorola, Inc. Apparatus and method for mixed-media call formatting
US20050058075A1 (en) * 2003-09-12 2005-03-17 Gorday Robert M. Apparatus and method for mixed-media call formatting
US20050201291A1 (en) * 2004-03-11 2005-09-15 Ixi Mobile (R&D) Ltd. Power management system and method for a wireless communications device
US7251152B2 (en) * 2004-08-26 2007-07-31 Infineon Technologies Ag Memory circuit having memory cells which have a resistance memory element
US20060136226A1 (en) * 2004-10-06 2006-06-22 Ossama Emam System and method for creating artificial TV news programs
US20060095265A1 (en) * 2004-10-29 2006-05-04 Microsoft Corporation Providing personalized voice front for text-to-speech applications
US20060122840A1 (en) * 2004-12-07 2006-06-08 David Anderson Tailoring communication from interactive speech enabled and multimodal services
US20070036086A1 (en) * 2005-08-09 2007-02-15 Sbc Knowledge Ventures, L.P. System and method of providing communications based on a predetermined device status
US20070116194A1 (en) * 2005-10-27 2007-05-24 International Business Machines Corporation Communications involving devices having different communication modes
US20070143103A1 (en) * 2005-12-21 2007-06-21 Cisco Technology, Inc. Conference captioning

Cited By (178)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9247383B1 (en) 2001-10-18 2016-01-26 Iwao Fujisaki Communication device
US8744515B1 (en) 2001-10-18 2014-06-03 Iwao Fujisaki Communication device
US8064964B1 (en) 2001-10-18 2011-11-22 Iwao Fujisaki Communication device
US8068880B1 (en) 2001-10-18 2011-11-29 Iwao Fujisaki Communication device
US7945236B1 (en) 2001-10-18 2011-05-17 Iwao Fujisaki Communication device
US7945286B1 (en) 2001-10-18 2011-05-17 Iwao Fujisaki Communication device
US7945287B1 (en) 2001-10-18 2011-05-17 Iwao Fujisaki Communication device
US7945256B1 (en) 2001-10-18 2011-05-17 Iwao Fujisaki Communication device
US7949371B1 (en) 2001-10-18 2011-05-24 Iwao Fujisaki Communication device
US7996037B1 (en) 2001-10-18 2011-08-09 Iwao Fujisaki Communication device
US8731540B1 (en) 2001-10-18 2014-05-20 Iwao Fujisaki Communication device
US8290482B1 (en) 2001-10-18 2012-10-16 Iwao Fujisaki Communication device
US8024009B1 (en) 2001-10-18 2011-09-20 Iwao Fujisaki Communication device
US9026182B1 (en) 2001-10-18 2015-05-05 Iwao Fujisaki Communication device
US9154776B1 (en) 2001-10-18 2015-10-06 Iwao Fujisaki Communication device
US9197741B1 (en) 2001-10-18 2015-11-24 Iwao Fujisaki Communication device
US8750921B1 (en) 2001-10-18 2014-06-10 Iwao Fujisaki Communication device
US8200275B1 (en) 2001-10-18 2012-06-12 Iwao Fujisaki System for communication device to display perspective 3D map
US8805442B1 (en) 2001-10-18 2014-08-12 Iwao Fujisaki Communication device
US8086276B1 (en) 2001-10-18 2011-12-27 Iwao Fujisaki Communication device
US9537988B1 (en) 2001-10-18 2017-01-03 Iwao Fujisaki Communication device
US9883021B1 (en) 2001-10-18 2018-01-30 Iwao Fujisaki Communication device
US8538486B1 (en) 2001-10-18 2013-09-17 Iwao Fujisaki Communication device which displays perspective 3D map
US8538485B1 (en) 2001-10-18 2013-09-17 Iwao Fujisaki Communication device
US8498672B1 (en) 2001-10-18 2013-07-30 Iwao Fujisaki Communication device
US9883025B1 (en) 2001-10-18 2018-01-30 Iwao Fujisaki Communication device
US10284711B1 (en) 2001-10-18 2019-05-07 Iwao Fujisaki Communication device
US8150473B1 (en) 2001-10-18 2012-04-03 Iwao Fujisaki Communication device
US10805451B1 (en) 2001-10-18 2020-10-13 Iwao Fujisaki Communication device
US10425522B1 (en) 2001-10-18 2019-09-24 Iwao Fujisaki Communication device
US8165638B1 (en) 2001-10-18 2012-04-24 Iwao Fujisaki Communication device
US8229512B1 (en) 2003-02-08 2012-07-24 Iwao Fujisaki Communication device
US8682397B1 (en) 2003-02-08 2014-03-25 Iwao Fujisaki Communication device
US8425321B1 (en) 2003-04-03 2013-04-23 Iwao Fujisaki Video game device
US8430754B1 (en) 2003-04-03 2013-04-30 Iwao Fujisaki Communication device
US8241128B1 (en) 2003-04-03 2012-08-14 Iwao Fujisaki Communication device
US8311578B1 (en) 2003-09-26 2012-11-13 Iwao Fujisaki Communication device
US8195228B1 (en) 2003-09-26 2012-06-05 Iwao Fujisaki Communication device
US8233938B1 (en) 2003-09-26 2012-07-31 Iwao Fujisaki Communication device
US10560561B1 (en) 2003-09-26 2020-02-11 Iwao Fujisaki Communication device
US8244300B1 (en) 2003-09-26 2012-08-14 Iwao Fujisaki Communication device
US10805444B1 (en) 2003-09-26 2020-10-13 Iwao Fujisaki Communication device
US8260352B1 (en) 2003-09-26 2012-09-04 Iwao Fujisaki Communication device
US10547724B1 (en) 2003-09-26 2020-01-28 Iwao Fujisaki Communication device
US10805443B1 (en) 2003-09-26 2020-10-13 Iwao Fujisaki Communication device
US10547722B1 (en) 2003-09-26 2020-01-28 Iwao Fujisaki Communication device
US8295880B1 (en) 2003-09-26 2012-10-23 Iwao Fujisaki Communication device
US8301194B1 (en) 2003-09-26 2012-10-30 Iwao Fujisaki Communication device
US8781526B1 (en) 2003-09-26 2014-07-15 Iwao Fujisaki Communication device
US8320958B1 (en) 2003-09-26 2012-11-27 Iwao Fujisaki Communication device
US8326357B1 (en) 2003-09-26 2012-12-04 Iwao Fujisaki Communication device
US8326355B1 (en) 2003-09-26 2012-12-04 Iwao Fujisaki Communication device
US8331984B1 (en) 2003-09-26 2012-12-11 Iwao Fujisaki Communication device
US8331983B1 (en) 2003-09-26 2012-12-11 Iwao Fujisaki Communication device
US8335538B1 (en) 2003-09-26 2012-12-18 Iwao Fujisaki Communication device
US10547725B1 (en) 2003-09-26 2020-01-28 Iwao Fujisaki Communication device
US8340720B1 (en) 2003-09-26 2012-12-25 Iwao Fujisaki Communication device
US8346303B1 (en) 2003-09-26 2013-01-01 Iwao Fujisaki Communication device
US8346304B1 (en) 2003-09-26 2013-01-01 Iwao Fujisaki Communication device
US8351984B1 (en) 2003-09-26 2013-01-08 Iwao Fujisaki Communication device
US8364202B1 (en) 2003-09-26 2013-01-29 Iwao Fujisaki Communication device
US8364201B1 (en) 2003-09-26 2013-01-29 Iwao Fujisaki Communication device
US8380248B1 (en) 2003-09-26 2013-02-19 Iwao Fujisaki Communication device
US8391920B1 (en) 2003-09-26 2013-03-05 Iwao Fujisaki Communication device
US10547723B1 (en) 2003-09-26 2020-01-28 Iwao Fujisaki Communication device
US8417288B1 (en) 2003-09-26 2013-04-09 Iwao Fujisaki Communication device
US10547721B1 (en) 2003-09-26 2020-01-28 Iwao Fujisaki Communication device
US8229504B1 (en) 2003-09-26 2012-07-24 Iwao Fujisaki Communication device
US8165630B1 (en) 2003-09-26 2012-04-24 Iwao Fujisaki Communication device
US8160642B1 (en) 2003-09-26 2012-04-17 Iwao Fujisaki Communication device
US8442583B1 (en) 2003-09-26 2013-05-14 Iwao Fujisaki Communication device
US8150458B1 (en) 2003-09-26 2012-04-03 Iwao Fujisaki Communication device
US8447353B1 (en) 2003-09-26 2013-05-21 Iwao Fujisaki Communication device
US8447354B1 (en) 2003-09-26 2013-05-21 Iwao Fujisaki Communication device
US10237385B1 (en) 2003-09-26 2019-03-19 Iwao Fujisaki Communication device
US10805445B1 (en) 2003-09-26 2020-10-13 Iwao Fujisaki Communication device
US8121641B1 (en) 2003-09-26 2012-02-21 Iwao Fujisaki Communication device
US8532703B1 (en) 2003-09-26 2013-09-10 Iwao Fujisaki Communication device
US10805442B1 (en) 2003-09-26 2020-10-13 Iwao Fujisaki Communication device
US8095181B1 (en) 2003-09-26 2012-01-10 Iwao Fujisaki Communication device
US8095182B1 (en) 2003-09-26 2012-01-10 Iwao Fujisaki Communication device
US9596338B1 (en) 2003-09-26 2017-03-14 Iwao Fujisaki Communication device
US8090402B1 (en) 2003-09-26 2012-01-03 Iwao Fujisaki Communication device
US11184469B1 (en) 2003-09-26 2021-11-23 Iwao Fujisaki Communication device
US8064954B1 (en) 2003-09-26 2011-11-22 Iwao Fujisaki Communication device
US8055298B1 (en) 2003-09-26 2011-11-08 Iwao Fujisaki Communication device
US9077807B1 (en) 2003-09-26 2015-07-07 Iwao Fujisaki Communication device
US8041371B1 (en) 2003-09-26 2011-10-18 Iwao Fujisaki Communication device
US8010157B1 (en) 2003-09-26 2011-08-30 Iwao Fujisaki Communication device
US8694052B1 (en) 2003-09-26 2014-04-08 Iwao Fujisaki Communication device
US8712472B1 (en) 2003-09-26 2014-04-29 Iwao Fujisaki Communication device
US7996038B1 (en) 2003-09-26 2011-08-09 Iwao Fujisaki Communication device
US11184470B1 (en) 2003-09-26 2021-11-23 Iwao Fujisaki Communication device
US11184468B1 (en) 2003-09-26 2021-11-23 Iwao Fujisaki Communication device
US11190632B1 (en) 2003-09-26 2021-11-30 Iwao Fujisaki Communication device
US8774862B1 (en) 2003-09-26 2014-07-08 Iwao Fujisaki Communication device
US8781527B1 (en) 2003-09-26 2014-07-15 Iwao Fujisaki Communication device
US8295876B1 (en) 2003-11-22 2012-10-23 Iwao Fujisaki Communication device
US8238963B1 (en) 2003-11-22 2012-08-07 Iwao Fujisaki Communication device
US9955006B1 (en) 2003-11-22 2018-04-24 Iwao Fujisaki Communication device
US8224376B1 (en) 2003-11-22 2012-07-17 Iwao Fujisaki Communication device
US9674347B1 (en) 2003-11-22 2017-06-06 Iwao Fujisaki Communication device
US8554269B1 (en) 2003-11-22 2013-10-08 Iwao Fujisaki Communication device
US9554232B1 (en) 2003-11-22 2017-01-24 Iwao Fujisaki Communication device
US8565812B1 (en) 2003-11-22 2013-10-22 Iwao Fujisaki Communication device
US8121635B1 (en) 2003-11-22 2012-02-21 Iwao Fujisaki Communication device
US11115524B1 (en) 2003-11-22 2021-09-07 Iwao Fujisaki Communication device
US9325825B1 (en) 2003-11-22 2016-04-26 Iwao Fujisaki Communication device
US9094531B1 (en) 2003-11-22 2015-07-28 Iwao Fujisaki Communication device
US8081962B1 (en) 2004-03-23 2011-12-20 Iwao Fujisaki Communication device
US8121587B1 (en) 2004-03-23 2012-02-21 Iwao Fujisaki Communication device
US8195142B1 (en) 2004-03-23 2012-06-05 Iwao Fujisaki Communication device
US8270964B1 (en) 2004-03-23 2012-09-18 Iwao Fujisaki Communication device
US9948890B1 (en) 2005-04-08 2018-04-17 Iwao Fujisaki Communication device
US8433364B1 (en) 2005-04-08 2013-04-30 Iwao Fujisaki Communication device
US8208954B1 (en) 2005-04-08 2012-06-26 Iwao Fujisaki Communication device
US9549150B1 (en) 2005-04-08 2017-01-17 Iwao Fujisaki Communication device
US9143723B1 (en) 2005-04-08 2015-09-22 Iwao Fujisaki Communication device
US10244206B1 (en) 2005-04-08 2019-03-26 Iwao Fujisaki Communication device
US9940923B2 (en) 2006-07-31 2018-04-10 Qualcomm Incorporated Voice and text communication system, method and apparatus
US9185657B1 (en) 2007-05-03 2015-11-10 Iwao Fujisaki Communication device
US8825090B1 (en) 2007-05-03 2014-09-02 Iwao Fujisaki Communication device
US9092917B1 (en) 2007-05-03 2015-07-28 Iwao Fujisaki Communication device
US9396594B1 (en) 2007-05-03 2016-07-19 Iwao Fujisaki Communication device
US8825026B1 (en) 2007-05-03 2014-09-02 Iwao Fujisaki Communication device
US9596334B1 (en) 2007-08-24 2017-03-14 Iwao Fujisaki Communication device
US8676273B1 (en) * 2007-08-24 2014-03-18 Iwao Fujisaki Communication device
US9232369B1 (en) * 2007-08-24 2016-01-05 Iwao Fujisaki Communication device
US10148803B2 (en) 2007-08-24 2018-12-04 Iwao Fujisaki Communication device
US8639214B1 (en) 2007-10-26 2014-01-28 Iwao Fujisaki Communication device
US8676705B1 (en) 2007-10-26 2014-03-18 Iwao Fujisaki Communication device
US9082115B1 (en) 2007-10-26 2015-07-14 Iwao Fujisaki Communication device
US8472935B1 (en) 2007-10-29 2013-06-25 Iwao Fujisaki Communication device
US9094775B1 (en) 2007-10-29 2015-07-28 Iwao Fujisaki Communication device
US8755838B1 (en) 2007-10-29 2014-06-17 Iwao Fujisaki Communication device
US9139089B1 (en) 2007-12-27 2015-09-22 Iwao Fujisaki Inter-vehicle middle point maintaining implementer
US20090276214A1 (en) * 2008-04-30 2009-11-05 Motorola, Inc. Method for dual channel monitoring on a radio device
US8856003B2 (en) * 2008-04-30 2014-10-07 Motorola Solutions, Inc. Method for dual channel monitoring on a radio device
US8543157B1 (en) 2008-05-09 2013-09-24 Iwao Fujisaki Communication device which notifies its pin-point location or geographic area in accordance with user selection
US8584388B1 (en) 2008-05-09 2013-11-19 Iwao Fujisaki Firearm
US9241060B1 (en) 2008-06-30 2016-01-19 Iwao Fujisaki Communication device
US8340726B1 (en) 2008-06-30 2012-12-25 Iwao Fujisaki Communication device
US10503356B1 (en) 2008-06-30 2019-12-10 Iwao Fujisaki Communication device
US11112936B1 (en) 2008-06-30 2021-09-07 Iwao Fujisaki Communication device
US9060246B1 (en) 2008-06-30 2015-06-16 Iwao Fujisaki Communication device
US10175846B1 (en) 2008-06-30 2019-01-08 Iwao Fujisaki Communication device
US9049556B1 (en) 2008-07-02 2015-06-02 Iwao Fujisaki Communication device
US9326267B1 (en) 2008-07-02 2016-04-26 Iwao Fujisaki Communication device
US8452307B1 (en) 2008-07-02 2013-05-28 Iwao Fujisaki Communication device
US20110015930A1 (en) * 2009-07-08 2011-01-20 Intelepeer, Inc. Unified communication system
US20110092187A1 (en) * 2009-10-15 2011-04-21 Larry Miller Configurable phone with interactive voice response engine
US8611876B2 (en) * 2009-10-15 2013-12-17 Larry Miller Configurable phone with interactive voice response engine
US20110112834A1 (en) * 2009-11-10 2011-05-12 Samsung Electronics Co., Ltd. Communication method and terminal
US9788349B2 (en) 2011-09-28 2017-10-10 Elwha Llc Multi-modality communication auto-activation
US9762524B2 (en) 2011-09-28 2017-09-12 Elwha Llc Multi-modality communication participation
US9503550B2 (en) * 2011-09-28 2016-11-22 Elwha Llc Multi-modality communication modification
US9002937B2 (en) 2011-09-28 2015-04-07 Elwha Llc Multi-party multi-modality communication
US9477943B2 (en) 2011-09-28 2016-10-25 Elwha Llc Multi-modality communication
US20130080917A1 (en) * 2011-09-28 2013-03-28 Royce A. Levien Multi-Modality communication modification
US9699632B2 (en) 2011-09-28 2017-07-04 Elwha Llc Multi-modality communication with interceptive conversion
US9906927B2 (en) 2011-09-28 2018-02-27 Elwha Llc Multi-modality communication initiation
US9794209B2 (en) 2011-09-28 2017-10-17 Elwha Llc User interface for multi-modality communication
CN103108092A (en) * 2011-11-15 2013-05-15 希姆通信息技术(上海)有限公司 Communication terminal and information transmitting method between communication terminals
US8423366B1 (en) * 2012-07-18 2013-04-16 Google Inc. Automatically training speech synthesizers
US9767802B2 (en) * 2013-08-29 2017-09-19 Vonage Business Inc. Methods and apparatus for conducting internet protocol telephony communications
US20150100315A1 (en) * 2013-08-29 2015-04-09 Vonage Network, Llc Methods and apparatus for conducting internet protocol telephony communications
US10733974B2 (en) * 2014-01-14 2020-08-04 Interactive Intelligence Group, Inc. System and method for synthesis of speech from provided text
US20150199956A1 (en) * 2014-01-14 2015-07-16 Interactive Intelligence Group, Inc. System and method for synthesis of speech from provided text
US9911407B2 (en) * 2014-01-14 2018-03-06 Interactive Intelligence Group, Inc. System and method for synthesis of speech from provided text
US20180144739A1 (en) * 2014-01-14 2018-05-24 Interactive Intelligence Group, Inc. System and method for synthesis of speech from provided text
WO2015119492A1 (en) * 2014-02-05 2015-08-13 Alvarez Heinemeyer Marco System, method and computer-readable medium for a language interface
US20160171982A1 (en) * 2014-12-10 2016-06-16 Honeywell International Inc. High intelligibility voice announcement system
US9558747B2 (en) * 2014-12-10 2017-01-31 Honeywell International Inc. High intelligibility voice announcement system
US20160372114A1 (en) * 2015-06-18 2016-12-22 Airbus Operations Gmbh Announcement signaling on board an aircraft
US10460730B2 (en) * 2015-06-18 2019-10-29 Airbus Operations Gmbh Announcement signaling on board an aircraft
US20180096334A1 (en) * 2016-10-03 2018-04-05 Paypal, Inc. Voice activated remittances
US11449868B2 (en) 2016-10-03 2022-09-20 Paypal, Inc. Voice activated remittances
US10805774B2 (en) 2016-12-19 2020-10-13 Samsung Electronics Co., Ltd. Methods and apparatus for managing control data

Also Published As

Publication number Publication date
WO2008016949A2 (en) 2008-02-07
WO2008016949A3 (en) 2008-03-20
KR101136769B1 (en) 2012-04-19
US9940923B2 (en) 2018-04-10
JP5149292B2 (en) 2013-02-20
JP2010505134A (en) 2010-02-18
KR20090036141A (en) 2009-04-13
US20160111082A1 (en) 2016-04-21
CN101496096B (en) 2012-11-07
CN101496096A (en) 2009-07-29
EP2047459A2 (en) 2009-04-15

Similar Documents

Publication Publication Date Title
US9940923B2 (en) Voice and text communication system, method and apparatus
EP3090531B1 (en) Smart bluetooth headset for speech command
JP4071631B2 (en) Method and apparatus for interoperability between voice transmission systems during voice inactivity
CN101510917B (en) Silent call method for mobile terminal
JP2006295931A (en) Apparatus and method for switching voice codec of mobile terminal
US8498667B2 (en) System and method for mixing audio with ringtone data
JPH05160773A (en) Voice communication equipment
ATE521962T1 (en) PREPROCESSING OF DIGITAL AUDIO DATA FOR MOBILE AUDIO CODECS
KR101367722B1 (en) Method for communicating voice in wireless terminal
JP4983417B2 (en) Telephone device having conversation speed conversion function and conversation speed conversion method
JP2001272998A (en) Communication method and wireless call connection device
KR20050029728A (en) Identification and exclusion of pause frames for speech storage, transmission and playback
EP1159738B1 (en) Speech synthesizer based on variable rate speech coding
CN111385780A (en) Bluetooth audio signal transmission method and device
CN101322423B (en) System and method for non-cascade operation signal transmission
JP5177476B2 (en) Wireless communication terminal, wireless communication system, and wireless communication program
JP2005222410A (en) On-vehicle handsfree mail apparatus
JP2979859B2 (en) Digital mobile radio equipment
JP3885989B2 (en) Speech complementing method, speech complementing apparatus, and telephone terminal device
KR100684944B1 (en) Apparatus and method for improving the quality of a voice data in the mobile communication
JP2974427B2 (en) Voice communication system and voice communication device
KR100760905B1 (en) Method of processing audio signals for improving the quality of output audio signal which is transferred to subscriber?s terminal over network and audio signal pre-processing apparatus of enabling the method
KR20010066405A (en) Telephone conversation recording or reproducing apparatus in portable radio telephone
KR20060076900A (en) Method for notifying short message while playing music of mobile terminal
JP2016127367A (en) Telephone conversation device, telephone conversation system and telephone conversation method

Legal Events

Date Code Title Description
AS Assignment

Owner name: QUALCOMM INCORPORATED A DELWARE CORPORATION,CALIFO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MOLLOY, STEPHEN;EL-MALEH, KHALED HELMI;SIGNING DATES FROM 20061005 TO 20061006;REEL/FRAME:018412/0958

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION