The disclosure relates to communications and, more particularly, to a voice and text communication system, method and apparatus.
A cellular phone may include an audio capture device, such as a microphone and/or speech synthesizer, and an audio encoder to generate audio packets or frames. The phone may use communication protocol layers and modules to transmit packets across a wireless communication channel to a network or another communication device.
One aspect relates to an apparatus comprising a vocoder, a speech to text conversion engine, a text to speech conversion engine, and a user interface. The vocoder is operable to convert speech signals into packets and convert packets into speech signals. The speech to text conversion engine is operable to convert speech to text. The text to speech conversion engine is operable to convert text to speech. The user interface is operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the speech to text conversion engine, a second mode enables the text to speech conversion engine, and a third mode enables the speech to text conversion engine and the text to speech conversion engine.
Another aspect relates to an apparatus comprising: a vocoder operable to convert speech signals into packets and convert packets into speech signals; a speech to text conversion engine operable to convert speech to text; a user interface operable to receive a user selection of a mode from among a plurality of modes, wherein a first mode enables the vocoder, and a second mode enables the speech to text conversion engine; and a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
Another aspect relates to a network apparatus comprising: a vocoder operable to convert packets into speech signals; a speech to text conversion engine operable to convert speech to text; a selection unit operable to switch between first and second modes, wherein the first mode enables the vocoder, and a second mode enables the vocoder and the speech to text conversion engine; and a transceiver operable to wirelessly transmit encoded speech packets and text packets to a communication network.
Another aspect relates to a method comprising: receiving encoded speech packets; converting the received encoded speech packets into speech signals; and receiving a user selection of a mode from among a plurality of modes, wherein a first mode enables speech to text conversion, a second mode enables text to speech conversion, and a third mode enables speech to text and text to speech conversion.
The details of one or more embodiments are set forth in the accompanying drawings and the description below.
Receiving a call on a mobile device in a meeting, airplane, train, theater, restaurant, church or other place may be disruptive to others. It may be much less disruptive if a user could select another mode on the mobile device to receive the call and/or respond to the call. In one mode, the device receives the call and converts speech/voice signals to text without requiring the caller on the other end to input text.
The first communication device 100 comprises a voice coder (vocoder) 102 and a transceiver 104. The first communication device 100 may include other components in addition to or instead of the components shown in
The vocoder 102 may include an encoder to encode speech signals into packets and a decoder to decode packets into speech signals. The vocoder 102 may be any type of vocoder, such as an enhanced variable rate coder (EVRC), Adaptive Multi-Rate (AMR), Fourth Generation vocoder (4GV), etc. Vocoders are described in co-assigned U.S. Pat. Nos. 6,397,175, 6,434,519, 6,438,518, 6,449,592, 6,456,964, 6,477,502, 6,584,438, 6,678,649, 6,691,084, 6,804,218, 6,947,888, which are hereby incorporated by reference.
The transceiver 104 may wirelessly transmit and receive packets containing encoded speech.
The network 110 may represent one or more base stations, base station controllers (BSCs), mobile switching centers (MSCs), etc. If the first device 100 is a landline phone, then network 110 may include components in a plain old telephone service (POTS) network. The network 110 comprises a vocoder 112 and a transceiver 114. The network 110 may include other components in addition to or instead of the components shown in
The second communication device 120 may represent or be implemented in a wireless communication device, a personal digital assistant (PDA), a handheld device, a laptop computer, a desktop computer, a digital camera, a digital recording device, a network-enabled digital television, a mobile phone, a cellular phone, a satellite telephone, a camera phone, a terrestrial-based radiotelephone, a direct two-way communication device (sometimes referred to as a “walkie-talkie”), a camcorder, etc.
The second communication device 120 comprises a transceiver 124, a speech and text unit 140, a speaker 142, a display 128, a user input interface 130, e.g., a keypad, and a microphone 146. The speech and text unit 140 comprises a vocoder 122, a speech to text conversion engine 126, a controller 144, a text to speech conversion engine 132, and a voice synthesizer 134. The speech and text unit 140 may include other components in addition to or instead of the components shown in
One or more of the components or functions in the speech and text unit 140 may be integrated into a single module, unit, component, or software. For example, the speech to text conversion engine 126 may be combined with the vocoder 122. The text to speech conversion engine 132 may be combined with the vocoder 122, such that text is converted into encoded speech packets. The voice synthesizer 134 may be combined with the vocoder 122 and/or the text to speech conversion engine 132.
The speech to text conversion engine 126 may convert voice/speech to text. The text to speech conversion engine 132 may convert text to speech. The controller 144 may control operations and parameters of one or more components in the speech and text unit 140.
The device 120 may provide several modes of communication for a user to receive calls and/or respond to calls, as shown in the table below and in
In a normal mode (blocks 202 and 210), the user of the second device 120 receives a call from the first device 100, listens to speech from the speaker 142, and speaks into the microphone 146.
In a second mode (blocks 204 and 212), the user of the second device 130 may listen to speech from the first device 100, such as using an ear piece, headset, or headphones, but not talk. Instead, the user of the second device 130 may type on the keypad 130 or use a writing stylus to enter handwritten text on the display 128. The display 128 or the text to speech conversion engine 132 may have a module that recognizes handwritten text and characters. The device 120 may (a) send the text to the first device 100 or (b) convert the text to speech with the text to speech conversion engine 132.
The voice synthesizer 134 may synthesize the speech to produce personalized speech signals to substantially match the user's natural voice. The voice synthesizer 134 may include a memory that stores characteristics of the user's voice, such as pitch. A voice synthesizer is described in co-assigned U.S. Pat. No. 6,950,799, which is incorporated by reference. Another voice synthesizer is described in co-assigned U.S. patent application Ser. No. 11/398,364, which is incorporated by reference.
The vocoder 122 encodes the speech into packets. There may or may not be a short delay. In one configuration, other than a short time delay, communication with the second device 120 may appear seamless to the user of the first device 100. If the user of the second device 120 is in a meeting, the conversation may be more message-based than seamless.
In third and fourth modes (blocks 206, 208, 214 and 216), the device 120 receives a call, and the speech to text conversion engine 126 converts speech/voice signals to text for display on the display 128. In one configuration, the third and fourth modes may allow the user of the first device 100 to continue talking and not require the user of the first device 100 to switch to a text input mode. The speech to text conversion engine 126 may include a voice recognition module to recognize words and sounds to convert them to text.
In the third mode, the device 120 allows the user to speak into the microphone 146, which passes speech to the vocoder 122 to encode into packets.
In the fourth mode, the user of the second device 130 may type on the keypad 130 or use a writing stylus to enter handwritten text on the display 128. The device 120 may (a) send the text to the first device 100 or (b) convert the text to speech with the text to speech conversion engine 132. The voice synthesizer 134 may synthesize the speech to produce personalized speech signals to substantially match the user's natural voice. The vocoder 122 encodes the speech into packets.
In the second and fourth modes, if the second device 120 is set to convert text to speech and synthesize speech, there may be a time delay between when the second device 120 accepts a call from the first device 100 and when the first device 100 receives speech packets. The second device 120 may be configured to play a pre-recorded message by the user to inform the first device 100 that the user of the second device 120 is in a meeting and will respond using text to speech conversion.
The second and fourth modes may provide one or more advantages, such as transmitting speech without background noise, no need or reduced need for echo cancellation, no need or reduced need for noise suppression, faster encoding, less processing, etc.
The network 110A in
Thus, the network 110A in
For example, the first device 100 in
If the second device 120 does not have the unit 140, the second device 120 can signal (in-band for example) a desired mode to the network 110A and thus ask the network 110A to convert between speech and text, i.e., do the functions of unit 140.
Personalized speech synthesis may be done in the network 110A. As described above, the unit 140 in
An example of transmitting packets for second or fourth modes from the second device 120 to the network 110A is described. The second device 120 transmits to the network 110A at the beginning of using these “text modes” (second or fourth modes) user pre-stored voice packets (N packets) plus a mode of operation (1, 2, 3, or 4; request to do the conversion). The second device 120 may then send text packets.
A combination of the two configurations (
One or more components and features described above may be implemented in a push to talk (PTT) or push to read communication device. A PTT device allows a user to push a button on the device and talk, while the device converts speech to text and transmits text packets to a network or directly to another communication device. PTT communication is “message based,” rather than continuous, such as a standard voice call. A time period over which a user holds down the PTT button on the device may nicely frame the message that is then converted to text, etc.
The device 120 may have a dedicated memory for storing instructions and data, as well as dedicated hardware, software, firmware, or combinations thereof. If implemented in software, the techniques may be embodied as instructions on a computer-readable medium such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), FLASH memory, or the like. The instructions cause one or more processors to perform certain aspects of the functionality described in this disclosure.
The techniques described in this disclosure may be implemented within a general purpose microprocessor, digital signal processor (DSP), application specific integrated circuit (ASIC), field programmable gate array (FPGA), or other equivalent logic devices. For example, the speech and text unit 140 and associated components and modules, may be implemented as parts of an encoding process, or coding/decoding (CODEC) process, running on a digital signal processor (DSP) or other processing device. Accordingly, components described as modules may form programmable features of such a process, or a separate process.
The speech and text unit 140 may have a dedicated memory for storing instructions and data, as well as dedicated hardware, software, firmware, or combinations thereof. If implemented in software, the techniques may be embodied as instructions executable by one or more processors. The instructions may be stored on a computer-readable medium such as random access memory (RAM), read-only memory (ROM), non-volatile random access memory (NVRAM), electrically erasable programmable read-only memory (EEPROM), FLASH memory, magnetic or optical data storage device, or the like. The instructions cause one or more processors to perform certain aspects of the functionality described in this disclosure.
Various embodiments have been described. These and other embodiments are within the scope of the following claims.
The present Application for Patent is a Divisional of patent application Ser. No. 11/497,011 entitled “Enabling multiple modes in a communication device” filed Jul. 31, 2006, pending, and assigned to the assignee hereof and hereby expressly incorporated by reference herein.
Number | Name | Date | Kind |
---|---|---|---|
5750912 | Matsumoto | May 1998 | A |
5786768 | Chan | Jul 1998 | A |
5917405 | Joao | Jun 1999 | A |
5963620 | Frankel | Oct 1999 | A |
5983190 | Trower, II | Nov 1999 | A |
6094681 | Shaffer | Jul 2000 | A |
6173250 | Jong | Jan 2001 | B1 |
6195358 | Bowater | Feb 2001 | B1 |
6237027 | Namekawa | May 2001 | B1 |
6366651 | Griffith | Apr 2002 | B1 |
6456964 | Manjunath | Sep 2002 | B2 |
6502073 | Guan | Dec 2002 | B1 |
6574489 | Uriya | Jun 2003 | B1 |
6745024 | Dejaco | Jun 2004 | B1 |
6775360 | Davidson | Aug 2004 | B2 |
6795822 | Matsumoto | Sep 2004 | B1 |
6845359 | Ramabadran | Jan 2005 | B2 |
6950799 | Bi | Sep 2005 | B2 |
7009990 | Adams | Mar 2006 | B1 |
7136811 | Tirpak | Nov 2006 | B2 |
7203199 | Duree | Apr 2007 | B1 |
7233655 | Gailey | Jun 2007 | B2 |
7251152 | Roehr | Jul 2007 | B2 |
7251252 | Norby | Jul 2007 | B2 |
7292878 | Noguchi | Nov 2007 | B2 |
7305342 | Shizuka | Dec 2007 | B2 |
7406421 | Odinak | Jul 2008 | B2 |
7464034 | Kawashima | Dec 2008 | B2 |
7519042 | Gorday | Apr 2009 | B2 |
8682360 | Lessing | Mar 2014 | B2 |
8989713 | Doulton | Mar 2015 | B2 |
9697818 | Hendrickson | Jul 2017 | B2 |
20020184004 | Shizuka | Dec 2002 | A1 |
20030035412 | Wang | Feb 2003 | A1 |
20030083884 | Odinak | May 2003 | A1 |
20030097265 | Sakai | May 2003 | A1 |
20040034528 | Sakai | Feb 2004 | A1 |
20040053646 | Noguchi | Mar 2004 | A1 |
20040098266 | Hughes et al. | May 2004 | A1 |
20040111271 | Tischer | Jun 2004 | A1 |
20040203448 | Nagata | Oct 2004 | A1 |
20050058075 | Gorday | Mar 2005 | A1 |
20050201291 | Gluck | Sep 2005 | A1 |
20050255837 | Kwon | Nov 2005 | A1 |
20050266829 | Tran | Dec 2005 | A1 |
20060091205 | Agrawal | May 2006 | A1 |
20060092015 | Agrawal | May 2006 | A1 |
20060092874 | Agrawal | May 2006 | A1 |
20060095265 | Chu | May 2006 | A1 |
20060122840 | Anderson | Jun 2006 | A1 |
20060136213 | Hirose et al. | Jun 2006 | A1 |
20060136226 | Emam | Jun 2006 | A1 |
20060149558 | Kahn | Jul 2006 | A1 |
20060217159 | Watson | Sep 2006 | A1 |
20060273930 | Godden | Dec 2006 | A1 |
20070036086 | Walter | Feb 2007 | A1 |
20070043574 | Coffman et al. | Feb 2007 | A1 |
20070116194 | Agapi | May 2007 | A1 |
20070143103 | Asthana | Jun 2007 | A1 |
20100030557 | Molloy | Feb 2010 | A1 |
Number | Date | Country |
---|---|---|
1511286 | Mar 2005 | EP |
2007505564 | Mar 2007 | JP |
20020091920 | Dec 2002 | KR |
2005109661 | Nov 2005 | WO |
Entry |
---|
Written Opinion—PCT/US2007/074902, International Search Authority—European Patent Office—dated Jan. 31, 2008 (3 pp.). |
International Search Report—PCT/US07/074902—International Search Authority, European Patent Office, dated Jan. 31, 2008 (7 pp.). |
Number | Date | Country | |
---|---|---|---|
20160111082 A1 | Apr 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11497011 | Jul 2006 | US |
Child | 14981298 | US |