This application is based on and hereby claims priority to German Application No. 10 2006 006 551.4 filed on Feb. 13, 2006, the contents of which are hereby incorporated by reference.
Described below are a method and a system for providing speech dialogue applications on mobile terminals, in particular on mobile terminals in vehicles.
In speech dialogue systems a distinction can be made between command and control speech dialogue applications and systems for information dialogues. Command and control systems serve the purpose of inputting speech commands, for example for voice operation of a navigation system inside a vehicle. Speech dialogue systems are not used to input commands to control a device but enable the user to search for information on any desired subject, for example hotel information. In speech dialogue applications the information provided always has to be up to date. Depending on the application it is therefore necessary that the corresponding speech dialogue applications can be constantly, for example hourly or daily, updated.
In mobile terminals, as used in vehicles, up to now only speech dialogue applications have been provided for inputting speech control commands which are updated by data carriers. For example, the voice operation for a navigation system inside a vehicle can be loaded or updated from a CD. Known mobile terminals establish a link to a data network, for example the internet, via a GMS network or a WLAN network. Owing to its low bandwidth, however, a GMS link is not suitable for transmitting speech dialogue applications for the retrieval of information as such speech dialogue applications include very large volumes of data. A WLAN link is not suitable owing to its short range.
Therefore, described below are a method and a system for providing speech dialogue applications for the retrieval of information on mobile terminals by performing the following operations:
One advantage of the method is that the speech dialogue applications can be made available at the same time to numerous mobile terminals.
A further advantage of the method is that the speech dialogue applications can be updated easily and in a very close time-frame.
In an embodiment of the method the speech dialogue application exhibits background system data.
In an embodiment of the method each speech dialogue application exhibits a time stamp showing the point in time when it was produced.
In an embodiment of the method the description language is an XML description language.
In an embodiment of the method each speech dialogue application exhibits a name for its identification.
In an embodiment of the method the background system data are programmed in Java.
In a further embodiment of the method the background system data are also transmitted with an associated speech dialogue application by digital radio to the mobile terminals.
In an alternative embodiment of the method, after receipt of a speech dialogue application by the mobile terminal the background system data are retrieved via a bidirectional mobile radio interface of the mobile terminal.
In an embodiment of the method the speech dialogue applications are produced in a server which is connected to a data network.
In an embodiment of the method the data network is formed by the internet.
In an embodiment of the method the produced speech dialogue applications are stored in a first directory of the server and copied from the first directory of the server to a second directory of the server when the respective speech dialogue application is recognized as being complete.
In an embodiment of the method the speech dialogue applications copied to the second directory of the server are transmitted by a radio transmitter to a radio receiver of the mobile terminal at regular intervals.
In an embodiment of the method the speech dialogue applications received by the radio receiver of the mobile terminal are stored in a first directory of the mobile terminal and then copied from the first directory of the mobile terminal to a second directory of the mobile terminal when the respective time stamp of the speech dialogue application indicates that the speech dialogue application is younger than a corresponding speech dialogue application stored in the second directory of the mobile terminal.
In an embodiment of the method the speech dialogue applications are generated by the server automatically from local internet sites.
In an embodiment of the method the speech dialogue applications are transmitted by the radio transmitter to the radio receivers of the mobile terminals in its transmission area across a certain local radio reception frequency.
In an embodiment of the method each local radio reception frequency is assigned internet addresses of various local internet sites from which the server produces local speech dialogue applications automatically for transmission to the mobile terminals in the transmission area of the radio transmitter.
Also described below is a system for providing speech dialogue applications for mobile terminals with:
In an embodiment of the system the server is connected to a data network.
In an embodiment of the system the data network is formed by the internet.
In an embodiment of the system each mobile terminal exhibits a speech dialogue machine for processing the received speech dialogue applications and a background system with stored information dialogue data.
In an embodiment of the system the speech dialogue machine exhibits a speech interpretation unit for interpreting data which are output by an automatic speech recognition unit.
In an embodiment of the system the speech dialogue machine exhibits an output unit for outputting data to a speech synthesis unit.
In an embodiment of the system the mobile terminal exhibits a transmitter-receiver unit for a mobile telephone connection.
In an embodiment of the system the mobile terminal is installed in an associated vehicle.
Also described below is a mobile terminal with:
In an embodiment the mobile terminal is located in a vehicle.
These and other aspects and advantages will become more apparent and more readily appreciated from the following description of the exemplary embodiments with reference to the accompanying drawings of which:
Reference will now be made in detail to exemplary embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout.
As can be seen from
The radio transmitter 4 transmits the speech dialogue applications received from the server 2 to mobile terminals 7 which are inside its transmission area. For this purpose the mobile terminals 7 are in each case connected to a reception antenna 8 which receives the digital radio signal from the radio transmitter 4. The mobile terminals 7 are preferably located in a vehicle 9, for example in a motor vehicle, in a train, on a ship or in an airplane.
Via lines 18 the dialogue machine 12 of the mobile terminal 7 is also connected with a background system 19 (BGS: Background System), in which information dialogue data are stored. The background system data of the background system 19 are for example programmed in Java or in C.
The speech recognition system 13 receives an analogue speech signal from the microphone 14. This speech signal is digitized for example by a soundcard and then converted into a frequency spectrum by Fourier transformation. The frequency spectrum is then compared with the content of a database and the symbol of the acoustically most similar reference vector is passed on. Recognition takes place for example by hidden Markov models (HMM). Using a language model, the probability of certain word combinations is then determined in order to exclude incorrect hypotheses. For this purpose either a grammar model or a trigram statistic is used. A bigram or trigram statistic stores the occurrence probability of word combinations from two or three words.
In addition to the speech dialogue flow description and the language models, in an embodiment a speech dialogue application also includes background system data (BGS: Background System). These BGS data are for example programmed in Java and contain up-to-date information data on certain subjects.
In an embodiment each speech dialogue application additionally contains a time stamp which shows the point in time when it was produced.
In a first embodiment of the method the background system data are transmitted together with the language model and the speech dialogue flow description within a speech dialogue application by digital radio by the radio transmitter 4 to the radio receiver 10 within the mobile terminal 7.
In an alternative embodiment the speech dialogue application transmitted by radio does not contain the background system data but an address provided instead, for example an IP address. After the mobile terminal 7 has received the speech dialogue application from the digital radio transmitter 4 it creates a bidirectional transmission channel to a base station of a data network by a separate data link, in particular a mobile telephone link.
For example, the user 17 is asked by the mobile terminal 7 whether he/she would like to receive information about leisure-time activities in the transmission area. If the user 17 answers in the affirmative, he/she can for example enquire about musical activities available. The user can ask, for example, whether any jazz concerts are taking place in Cologne in the evening.
The dialogue machine 12 extracts reference words from this, such as for example “jazz”, “concert”, “Cologne” and generates a search enquiry to the background system 19, for example: Search (music=“jazz”; town/city=“Cologne”; time=“evening”).
The background system 19 conducts a search process and answers the search enquiry by sending an output data record to the dialogue machine 12.
After speech synthesis the dialogue machine 12 gives the following answer to the user 17: “Jazz is being played in Cologne this evening starting 20.00 hrs in the Domizil Club at Luxemburgerstrasse 117”.
In an embodiment of the system the server 2 generates background information data automatically on the basis of internet documents from the internet 6. To this end, the server 2 for example evaluates a given group of internet homepages relating to the city of Cologne. The local radio transmitter 4 transmits the speech dialogue application to those mobile terminals whose radio receivers are in its transmission area. The transmission takes place on a certain local radio reception frequency f. In an embodiment certain internet addresses of various local internet sites are assigned to each local radio reception frequency. From these local internet sites the server 2 produces local speech dialogue applications for transmission in the corresponding local transmission area.
In the method described herein it is particularly important that the speech dialogue applications provided are always up to date and for example can be updated daily or hourly. The speech dialogue application is updated as shown in
The server 2 contains a first directory A and a second directory B. In a step S1 the server 2 checks whether there is a newly produced speech dialogue application in its directory A.
If this is the case the server 2 further checks in a step S2 by an index file whether the speech dialogue application is complete.
If this is also the case the server 2 further checks in a step S3 whether there is a speech dialogue application with the same name in its directory B.
If this is the case a check is made in a step S4 whether the two speech dialogue applications with the same name are identical.
If this is the case the procedure reverts to step S1. If there are not two speech dialogue applications with the same name in the two directories A, B or the two speech applications are not identical, the server 2 copies the newly produced speech dialogue application from its directory A to its directory B in a step S5. The copied speech dialogue application is transmitted to the radio transmitter 4 for example by FTP and is transmitted from there in a step S6 to all the mobile terminals 7 in the transmission area.
As can be seen from
In a step S8 it is checked whether there are any new speech dialogue applications in the directory C.
If this is the case it is checked in a step S9 whether the received speech dialogue application is complete.
If the speech dialogue application is complete, in a step S10 it is checked whether in a further directory D of the mobile terminal 7 a speech dialogue application of the same name exists.
If this is the case, in a step S11 it is checked whether the speech dialogue application in the directory C is younger than the speech dialogue application in the other directory D. This check is made using the time stamp provided in the speech dialogue application. If the speech dialogue application in the reception directory C is younger than the speech dialogue application in directory D the updated speech dialogue application is copied from directory C to directory D in step S12 and the old speech dialogue application is preferably deleted. The procedure shown in
The speech dialogue data are transmitted continuously by the radio transmitter 4, so that the transmitted speech dialogue applications are available complete on the mobile terminal 7 at a given time.
In addition, it is ensured that the relatively time-consuming process of producing the language models only takes place once per updated speech dialogue application.
The method described above ensures that the user can conduct speech dialogues on up-to-date subjects with his/her mobile terminal 7 without a continuous WLAN link having to exist. Furthermore, the method ensures that a knowledge status exists which is updated daily or hourly. The mobile terminal 7 may be any mobile terminal, for example a vehicle unit or a PDA.
The system also includes permanent or removable storage, such as magnetic and optical discs, RAM, ROM, etc. on which the process and data structures of the present invention can be stored and distributed. The processes can also be distributed via, for example, downloading over a network such as the Internet. The system can output the results to a display device, printer, readily accessible memory or another computer on a network.
A description has been provided with particular reference to exemplary embodiments thereof and examples, but it will be understood that variations and modifications can be effected within the spirit and scope of the claims which may include the phrase “at least one of A, B and C” as an alternative expression that means one or more of A, B and C may be used, contrary to the holding in Superguide v. DIRECTV, 358F3d 870, 69 USPQ2d 1865 (Fed. Cir. 2004).
Number | Date | Country | Kind |
---|---|---|---|
10 2006 006 551 | Feb 2006 | DE | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP2006/067997 | 10/31/2006 | WO | 00 | 6/22/2010 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2007/093236 | 8/23/2007 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
6246672 | Lumelsky | Jun 2001 | B1 |
6721633 | Funk et al. | Apr 2004 | B2 |
7010263 | Patsiokas | Mar 2006 | B1 |
7277696 | Preiss et al. | Oct 2007 | B2 |
8051369 | Zirngibl et al. | Nov 2011 | B2 |
8195468 | Weider et al. | Jun 2012 | B2 |
8279844 | Hammond | Oct 2012 | B1 |
20020184373 | Maes | Dec 2002 | A1 |
20050043067 | Odell et al. | Feb 2005 | A1 |
20060029109 | Moran | Feb 2006 | A1 |
20070136069 | Veliu et al. | Jun 2007 | A1 |
20090019061 | Scannell, Jr. | Jan 2009 | A1 |
Number | Date | Country |
---|---|---|
1 679 695 | Jul 2006 | EP |
2005006752 | Jan 2005 | WO |
Entry |
---|
S. Goronzy et al.; “The Dynamic, Multi-lingual Lexicon in SmartKom”; Proceedings of Eurospeech 2003; pp. 1937-1940. |
R. Malaka et al.; “SmartKom Mobile—Intelligent Ubiquitous User Interaction”; Proceedings of IUI '04; Jan. 2004; pp. 310-312. |
N. Reithinger et al.; “SmartKom—Adaptive and Flexible Multimodal Access to Multiple Applications”; 5th International Conference on Multimodal Interfaces; 2003; vol. Conf. 5; pp. 101-108. |
International Search Report for Application No. PCT/EP2006/067997; mailed Jan. 30, 2007. |
Number | Date | Country | |
---|---|---|---|
20100267345 A1 | Oct 2010 | US |