The invention relates to an information terminal and voice operation method, and more specifically, an information terminal capable of operating with a voice input and a voice operation method.
An information terminal that can be operated by a voice input is known. In a certain voice recognition/response type mobile phone, a user can arbitrarily perform a telephone calling function, a mail function, etc. by a voice operation.
In a recent mobile phone, a user can install an arbitrary application in the mobile phone freely. In such a case, similar applications may be installed in plural, and a following problem occurs.
Even if a voice input saying “Start camera” is performed as a voice operation, for example, since there are applications concerning a camera in plural, the mobile phone cannot determine which application should be performed.
Therefore, a primary object of the invention is to provide a novel information terminal and voice operation method.
Another object of the invention is to provide an information terminal and voice operation method, having high convenience of a voice operation.
A first aspect of the invention is an information terminal that an operation by a voice input is possible, comprising: a storage module operable to store a plurality of applications and a use history of each of the applications; an acquisition module operable to acquire specific information for specifying an application to be performed based on an input voice; a narrowing-down module operable to narrow down, based on the use history, the specific information that is acquired; and a performing module operable to perform an application based on a result that is narrowed down by the narrowing-down module.
A second aspect of the invention is a voice operation method in an information terminal that comprises a storage module operable to store a plurality of applications and a use history of each of the applications, and an operation by a voice input is possible, a processor of the information terminal performing: acquiring specific information for specifying an application to be performed based on an input voice; narrowing-down, based on the use history, the specific information that is acquired; and performing an application based on a result that is narrowed down.
According to the invention, it is possible to increase convenience of a voice operation.
The above described objects and other objects, features, aspects and advantages of the invention will become more apparent from the following detailed description of the invention when taken in conjunction with the accompanying drawings.
With referring to
A display 14 such as a liquid crystal, organic EL, etc. called a display module is provided on a main surface (front surface) of the housing 12. A touch panel 16 is provided on the display 14.
A first speaker 18 is housed in the housing 12 at one end of a longitudinal direction on a side of the main surface, and a microphone 20 is housed at the other end in the longitudinal direction on the side of the main surface.
As hardware keys that constitute an input operating module together with the touch panel 16, a call key 22a, an end key 22b and a menu key 22c are provided on the main surface of the housing 12, in this embodiment.
A lens aperture 24 that communicates with a camera module 52 (see
For example, a user can input a telephone number by performing a touch operation by the touch panel 16 to a dial key (not shown) displayed on the display 14, and start a telephone conversation by operating the call key 22a. If operating the end key 22b, the telephone conversation can be ended. In addition, by long-depressing the end key 22b, it is possible to turn on/off a power supply of the mobile phone 10.
If operating the menu key 22c, a menu screen is displayed on the display 14, and in such a state, by performing a touch operation by means of the touch panel 16 to software keys, menu icons, etc. being displayed on the display 14, it is possible to perform a desired function.
Furthermore, although details will be described later, if a camera function is performed, the camera module 52 is started and a preview image (through image) corresponding to a photographic subject is displayed on the display 14. Then, the user can image the photographic subject by turning the rear surface that the lens aperture 24 is provided to the photographic subject and performing an imaging operation.
Furthermore, a plurality of applications are installed in the mobile phone 10. First, a standard camera and an AR (Augmented Reality) camera are installed as applications of a camera system. The standard camera is an application that is pre-installed in the mobile phone 10 and saves an image in response to an imaging operation. The AR camera is an application that is arbitrarily installed by a user and displays information while superposed on a through image.
Furthermore, as application of an email system, an E-mail, an SMS (Short Message Service) and an MMS (Multimedia Message Service) are installed.
Furthermore, applications such as a browser, an address book, a schedule, time, a music player, a video player, etc. are also installed, and the user can arbitrarily start such an application.
With referring to
The wireless communication circuit 32 is wirelessly connected with a network 100 (communication network, telephone network). A server 102 is connected with the network 100 via a wire or wirelessly.
The processor 30 is in charge of entire control of the mobile phone 10. The processor 30 includes an RTC 30a that outputs date and time information. A whole or part of a program that is set in advance in the flash memory 46 is, in use, developed or loaded into the RAM 48 that functions as a storing module, and the processor 30 operates in accordance with the program developed in the RAM 48. In addition, the RAM 48 is further used as a working area or buffer area for the processor 30.
The input device 42 includes the hardware keys 22a-22c shown in
The wireless communication circuit 32 is a circuit for sending and receiving a radio wave for a telephone conversation, a mail, etc. via an antenna 34. In this embodiment, the wireless communication circuit 32 is a circuit for performing a wireless communication with a CDMA system. For example, if the user designates an outgoing call (telephone call) using the input device 42, the wireless communication circuit 32 performs telephone call processing under instructions from the processor 30 and outputs a telephone call signal via the antenna 34. The telephone call signal is sent to a telephone at the other end of line through a base station and a communication network. Then, when incoming call processing is performed in the telephone at the other end of line, a communication-capable state is established and the processor 30 performs telephonic communication processing.
The microphone 20 shown in
The display driver 44 is connected to the display 14 shown in
The touch panel 16 shown in
In the embodiment, the touch panel 16 is of an electrostatic capacitance system that detects a change of an electrostatic capacitance between electrodes, which occurs when an object such as a finger is in close to a surface of the touch panel 16. The touch panel 16 detects that one or more fingers are brought into contact with the touch panel 16, for example. Therefore, the touch panel 16 is also called a pointing device. The touch panel control circuit 50 functions as a detecting module, and detects a touch operation within a touch-effective range of the touch panel 16, and outputs coordinate data indicative of a position of the touch operation to the processor 30. That is, the user inputs to the mobile phone 10 an operation position, an operation direction and so on through a touch operation to the surface of the touch panel 16. In addition, the touch operation in this embodiment includes a tap operation, a long-tap operation, a flick operation, a slide operation, etc.
The camera module 52 includes a control circuit, a lens, an image sensor, etc. The processor 30 starts the control circuit and the image sensor if an operation for performing a camera function is performed. Then, if image data based on a signal that is output from the image sensor is input to the processor 30, a preview image according to a photographic subject is displayed on the display 14.
Furthermore, the mobile phone 10 has a voice recognition function that recognizes a voice that is input to the microphone 20, an utterance function that outputs a voice message based on a database of synthesized voices and a voice operation function using these functions. Then, the voice operation function of this embodiment is supports a voice input of a natural language.
If a user inputs a voice saying “Call the home” to a mobile phone 10 that the voice operation function is performed, a voice of the user is recognized by the voice recognition function. Furthermore, the mobile phone 10 outputs a response message saying “Call the home?” based on a recognized voice by the utterance function. At this time, if the user replies by saying “Call”, the mobile phone 10 reads the telephone number that is registered as the home from an address book, and call to that telephone number. If the voice operation function is thus performed, the user can operate the mobile phone 10 without performing a touch operation to the touch panel 16. Then, it becomes for the user easy to grasp a state of the mobile phone 10 by hearing the contents of voice guidance (response messages).
Specifically describing, when a user performs a voice input and thus voice recognition processing is started, a feature amount of the user (hereinafter, merely called a user feature amount) is derived from an input voice and compared with each feature amount that is read from the local database 332. Each comparison result of the user feature amount and each feature amount is calculated as a likelihood, and a feature amount corresponding to the largest likelihood is specified. Then, a character string corresponding to the feature amount that is specified is read from the local database 332, and the character string thus read becomes a recognition result. If a user performs a voice input and a character string that is read based on a user feature amount of an input voice is “camera”, for example, a recognition result becomes “camera”.
However, when the largest likelihood is equal to or less than a predetermined value, that is, when an input voice is not registered in the local database, an input voice may be sent to the server 102 to perform voice recognition processing by the server 102. Then, a result of the voice recognition performed by the server 102 is returned to the mobile phone 10. Thus, it is possible to shorten a time until a result of the voice recognition is obtained by performing a part of the voice recognition processing to an input voice using the local database in the mobile phone 10. Furthermore, a burden of the voice recognition processing imposed on the server 102 is also reducible.
In addition, the character string indicating a date and time, that is, time information is acquired from the RTC 30a. Furthermore, the use history data may be called a user log.
For example, the application of “standard camera” that a category is classified into “camera” is started seven (7) times within one week, and the application of “AR camera” is started once within one week. Furthermore, “E-mail” and “MMS” that categories are classified into “mail” are started four (4) times within one week, respectively, and “SMS” is started three (3) times within one week.
With reference to
Here, if a voice operation function is performed, a voice operation icon VI is displayed in the status display area 70 as shown in
Therefore, this embodiment deals with an ambiguous voice input based on the use frequency of each application. Specifically, based on the use frequency of each application recorded on the application table, a result of a voice input is narrowed down.
For example, since “camera” is included in the recognition result of voice recognition when a user performs a voice input saying “Use camera” as shown in
Then, when search results are in plural, the search results are narrowed down based on the use frequency corresponding to each application. Here, since the use frequency of “standard camera” is “7” and the use frequency of “AR camera” is “1”, the search term is narrowed down only to “standard camera”. Therefore, the mobile phone 10 starts “standard camera” after outputting a voice message saying “Starting camera”.
With reference to
Thus, it is possible to increase convenience of a voice operation by narrowing down a search result based on the use history of the user.
Next, a description will be made about a case where applications that are narrowed down are in plural. With reference to
With reference to
Thus, when the search result cannot be narrowed down, it is possible to make the user select an application that the user wish to use by displaying the candidate list.
Furthermore, when an application name is designated by a voice input of the user, an application corresponding to a recognition result is performed. In addition, if the application is terminated within a predetermined time period (15 seconds, for example), a candidate list is displayed based on a second candidate in the recognition result of the voice recognition.
For example, in the recognition result of the voice recognition, when a character string corresponding to the feature amount with the highest likelihood is “SMS” and a character string corresponding to the feature amount with a second highest likelihood is “MMS”, a recognition result becomes as “SMS”, and therefore, an SMS is performed. In this state, if the SMS is terminated within the predetermined time period, “MMS” with a second highest likelihood in the recognition result of the voice recognition is re-acquired as a search term. If a search term is re-acquired, the search term is re-searched in the application table, and an application name of “MMS” is re-acquired as a search result, here. When an application name is re-acquired as a search result, applications of a category that the applications belong is displayed as a candidate list. That is, a candidate list comprising “E-mail”, “SMS” and “MMS” is displayed on the display 14.
Furthermore, if a search result is not acquirable as a result of searching by the search term based on a voice input, that is, if the application corresponding to the search term is not registered in the application table, a browser function is performed. If the browser function is performed, a predetermined search engine site is connected, and a search term is searched in the search engine site. Then, a result that is searched with the search engine site is displayed on the display 14. That is, even if performing a voice input of a word that is not registered in the application table, it is possible to provide information based on the search term to the user.
In addition, a candidate list may be displayed even if the use frequencies of all the applications in the search result are the same value. Furthermore, in other embodiments, even if a difference of the use frequencies of respective applications is equal to or less than a predetermined value (“1”, for example), a candidate list may be displayed.
Furthermore, a voice operation function is performed if the menu key 22c is long-depressed. However, in other embodiments, a software key (icon) for performing a voice operation function may be displayed on the display 14.
Furthermore, a voice saying “No”, “Other” or the like is input at a time that the application is performed, the application being performed is ended. Furthermore, in other embodiments, after the application is ended, the voice operation function may be performed again.
Although the feature of the embodiment is outlined in the above, in the following, the embodiment will be described in detail using a memory map shown in
With reference to
The program storage area 302 is stored with a use history record program 310 for recording a use history, a voice operation program 312 for operating the mobile phone 10 with a voice input, a voice recognition program 314 for recognizing an input voice, etc. In addition, programs for performing respective applications, etc. are also included in the program storage area 302.
Subsequently, the data storage area 304 of the RAM 48 is provided with a voice recognition buffer 330, and stored with a local database 332, use history data 334 and an application table 336. In addition, the data storage area 302 is provided also with an erroneous determination counter 338.
In the voice recognition buffer 330, data of a voice that a voice input is performed and a result of the voice recognition are temporarily stored. The local database 332 is a database of a format shown in
The erroneous determination counter 338 is a counter for counting a time period after an application is performed by a voice operation. If initialized, the erroneous determination counter 338 starts counting, and expires if a predetermined time period (15 seconds, for example) elapses. Therefore, the erroneous determination counter 340 may be called an erroneous determination timer.
The data storage area 304 is stored with data of a character string that is stored by a copy or cut-out, image data that is displayed in the standby state, etc., and provided with counters and flags necessary for an operation of the mobile phone 10.
The processor 30 processes a plurality of tasks including use history record processing shown in
With reference to
Subsequently, the processor 30 records a use history in a step S7. That is, the date and time and the application name that are acquired in the above-mentioned steps S3 and S5 are recorded in the application table 336 in association with each other. In addition, after the processing of the step S7 is ended, the processor 30 returns to the processing of the step S1.
Subsequently, the processor 30 determines, in a step S25, whether a voice is input. That is, it is determined whether a voice that the user utters is received by the microphone 20. If “NO” is determined in the step S25, that is, if a voice is not input, the processor 30 repeats the processing of the step S25. If “YES” is determined in the step S25, that is, if a voice is input, the processor 30 performs voice recognition processing in a step S27. That is, a user feature amount is derived from an input voice, and a likelihood with each feature amount is evaluated, and a character string corresponding to a feature amount with the highest likelihood is regarded as a recognition result.
Subsequently, the processor 30 extracts a search term from the recognition result in a step S29. For example, a character string of “camera” is extracted from the recognition result of the voice input as a search term. Subsequently, the processor 30 performs a search based on the search term in a step S31. That is, it is determined whether the search term is included in the application table. Then, if the search term corresponds to either among character strings recorded in the application table, a search result is obtained based on a corresponding character string.
Subsequently, with reference to
Furthermore, if “YES” is determined in the step S33, that is, if the search result is “camera”, for example, and thus corresponds to the category of “camera” of the application table, the processor 30 acquires the contents of the category corresponding to the search result in a step S35. For example, “standard camera” and “AR camera” included in the category of “camera” are acquired. In addition, the processor 30 that performs the processing in the step S35 functions as an acquisition module.
Subsequently, the processor 30 determines, in a step S37, whether a plurality of applications are included. That is, the processor 30 determines whether a plurality of applications are included in the contents of the category acquired in the step S35. If “NO” is determined in the step S37, that is, if a plurality of applications are not included in the contents of the category acquired, the processor 30 proceeds to processing of a step S49.
Furthermore, if “YES” is determined in the step S37, that is, if a plurality of applications are included, the processor 30 performs narrowing-down processing in a step S39. That is, based on the use histories corresponding to the plurality of applications, an application with the most use history is selected. Then, a selected application becomes a result of the narrowing-down. In addition, the processor 30 that performs the processing in the step S39 functions as a narrowing-down module.
Subsequently, the processor 30 determines, in a step S41, whether a result of the narrowing-down is only one. That is, the processor 30 determines whether the number of the applications narrowed down based on the use history is one (1). If “YES” is determined in the step S41, that is, if the application narrowed down is only “standard camera”, for example, the processor 30 proceeds to processing of a step S49.
Furthermore, if “NO” is determined in the step S41, that is, if the applications narrowed down are “E-mail” and “MMS”, for example, the processor 30 displays a candidate list in a step S43. As shown in
Subsequently, the processor 30 determines, in a step S45, whether an application is selected. That is, it is determined whether an arbitrary application is selected based on the candidate list being displayed. Specifically, the processor 30 determines whether a touch operation is performed to an arbitrary performing key AK in the candidate list being displayed. If “NO” is determined in the step S45, that is, if no application is selected, the processor 30 repeats the processing of the step S45. On the other hand, if “YES” is determined in the step S45, that is, if a touch operation is performed to the first performing key AK1 corresponding to “E-mail”, for example, the processor 30 performs a selected application in a step S47. The function of an E-mail is performed in a step S47, for example. Then, if the processing of the step S47 is ended, the processor 30 terminates the voice operation processing.
Furthermore, if the number of the applications included in the category of the search result is one (1) or if the applications narrowed down by the narrowing-down processing is one (1), the processor 30 performs the application in a step S49. If the application that is narrowed down is “standard camera”, for example, the processor 30 performs a standard camera. Then, if the processing of the step S49 is ended, the processor 30 terminates the voice operation processing.
In addition, the processor 30 that performs the processing in the steps S47 and S49 functions as a performing module.
With reference to
Subsequently, the processor 30 performs the application in a step S55. The SMS is performed based on the application name (“SMS”) that is acquired, for example. Subsequently, the processor 30 initializes the erroneous determination timer in a step S57. That is, in order to measure a time period after the application is performed, the erroneous determination counter 338 is initialized.
Subsequently, the processor 30 determines, in a step S59, whether the erroneous determination timer expires. That is, it is determined whether the predetermined time period elapses after the application is performed. If “NO” is determined in the step S59, that is, if the predetermined time period does not elapse after the application is performed, the processor 30 determines, in a step S61, whether an end is instructed. That is, the processor 30 determines whether there is any voice input or an input operation that ends the application that is performed. If “NO” is determined in the step S61, that is, if an operation that ends the application that is performed is not performed, the processor 30 returns to the processing of the step S59. Furthermore, if “YES” is determined in the step S59, that is, if the predetermined time period elapses after the application is performed, the processor 30 terminates the voice operation processing.
If “YES” is determined in the step S61, that is, if “NO” is input by a voice, for example, the processor 30 re-acquires a recognition result in a step S63. In the step S63, first, the application that is performed is ended. Next, a second candidate in the recognition result of the voice recognition is acquired from the voice recognition buffer 330. Subsequently, the process proceeds to the processing of the step S43, and the processor 30 displays a candidate list. When a recognition result that is re-acquired is “MMS”, for example, the application included in the category that the MMS is classified is displayed on the display 14 as a candidate list in a step S43.
Furthermore, if the search result is not an application name, that is, if the search term is not included in the application table, the processor 30 performs a browser function in a step S65, and connects it to a search engine site in a step S67. In addition, the processor 30 that performs the processing in the step S65 functions as a browser function performing module, and the processor 30 that performs the processing in the step S67 functions as a search module.
Subsequently, the processor 30 searches the search term in the search engine site in a step S69, and displays a web page in a step S71. If the search term is “dinner”, for example, a site containing a character string of “dinner” is searched with the search engine site, and a web page indicating a search result thereof is displayed on the display 14. Then, if the processing of the step S71 is ended, the processor 30 terminates the voice operation processing. In addition, the processor 30 that performs the processing of the step S71 functions as a web page display module.
In the second embodiment, when a browser function is performed by a voice operation, a web page is displayed based on a browsing frequency of a web page of a user. In addition, since basic structure of a mobile phone 10 is approximately the same as that of the first embodiment, a detailed description thereof is omitted.
Next, a case where a browser function is performed by a voice input will be described. With reference to
With reference to
In addition, when searching a search term with a web page, if a search form is prepared in the page, a search result is acquired using the search form. On the other hand, when a search form is not provided, a link that corresponds to a search term is specified by searching a character string, and a web page of a link destination is acquired as a search result.
In the above, the feature of the second embodiment is outlined. In the following, the second embodiment will be described in detail using a memory map shown in
In the data storage area 304 of the RAM 48 of the second embodiment, browsing history data 340 and a URL table 342 are stored. The browsing history data 340 is data of a format shown in
If a browser function is performed in a step S65, a web page with a high browsing frequency is connected by the processor 30 in a step S91. That is, the URL table 342 is read, and a web page corresponding to a URL with the highest browsing frequency is connected. In the step S91, the web page corresponding to “http://sports.***.com/” is connected based on the URL table 342 shown in
Subsequently, the processor 30 searches the search term in the web page being connected in a step S93. If the search terms are “baseball” and “game result”, for example, these search terms are searched using a search form, etc. in the web page being connected.
Subsequently, the processor 30 displays the web page in a step S71. As shown in
In addition, it is possible to arbitrarily combine the first embodiment and the second embodiment with each other and easy to conceive such combination, a detailed description thereof is omitted here.
Furthermore, a category of an application may include “game”, “map”, etc. besides “camera” and “mail”.
Furthermore, when the mobile phone 10 further comprises a GPS circuit and a GPS antenna and thus can perform positioning of a current position, position information may be included in the use history of application. Then, when narrowing-down the search result, this position information may be used. Specifically, after narrowing-down to an application(s) having been performed within a predetermined range from a current position among a plurality of applications, the applications are further narrowed down based on the use history. For example, in a case where an application of a standard camera is mainly used in own home, but an AR camera is mainly used out of the home, if “camera” is performed by a voice operation function outside the home, the AR camera comes to be performed automatically.
Furthermore, in other embodiments, the mobile phone 10 may display a selection screen of two applications on the display 14 when an AR camera and a standard camera are obtained as a result of the narrowing-down processing to the specific information. In such a case, the AR camera is displayed at a higher rank position outside the home while the standard camera is displayed at a position of a lower rank of the AR camera. On the other hand, in own home, the standard camera is displayed at a higher rank position while the AR camera is displayed at a position of a lower rank of the standard camera.
Furthermore, in other embodiments, a color and/or size of a character string indicating an application name may be changed without displaying an application name at a higher rank position.
By processing in such a way, even if a plurality of candidates are displayed, the user can recognize easily which application is an application that should be mainly used in a specific place. That is, the user can easily select the application that is mainly used in the specific place.
Although the mobile phone 10 performs the primary voice recognition processing by providing the local database (dictionary for voice recognition) in the mobile phone 10 and the secondary voice recognition processing is performed by the server 102 in the above-mentioned embodiment, in other embodiments, only the mobile phone 10 may perform the voice recognition processing or only the server 102 may perform the voice recognition processing.
Furthermore, when the mobile phone 10 supports a gaze input, the mobile phone 10 may be operated by a gaze operation in addition to a key operation and a touch operation.
The programs used in the embodiments may be stored in an HDD of the server for data distribution, and distributed to the mobile phone 10 via the network. The plurality of programs may be stored in a storage medium such as an optical disk of CD, DVD, BD or the like, a USB memory, a memory card, etc. and then, such the storage medium may be sold or distributed. In a case where the programs downloaded via the above-described server or storage medium are installed to a portable terminal having the structure equal to the structure of the embodiments, it is possible to obtain advantages equal to the advantages according to the embodiments.
The specific numerical values mentioned in this specification are only examples, and changeable appropriately in accordance with the change of product specifications.
It should be noted that reference numerals inside the parentheses and the supplements show one example of a corresponding relationship with the embodiments described above for easy understanding of the invention, and do not limit the invention.
An embodiment is an information terminal that an operation by a voice input is possible, comprising: a storage module operable to store a plurality of applications and a use history of each of the applications; an acquisition module operable to acquire specific information for specifying an application to be performed based on an input voice; a narrowing-down module operable to narrow down, based on the use history, the specific information that is acquired; and a performing module operable to perform an application based on a result that is narrowed down by the narrowing-down module.
In this embodiment, the information terminal (10: reference numeral exemplifying a portion or module corresponding in the embodiment, and so forth) can be operated by a voice input, and is installed with a plurality of applications. The storage module (48) is a storage media such as a RAM and a ROM, for example, and stores programs of the applications being installed and use histories of the application that the user uses etc. If a user performs a voice input, a recognition result by voice recognition processing is obtained for the input voice. Then, a search term is extracted from the recognition result. When the search term is extracted, an application that can be performed is searched. The acquisition module (30, S35) acquires a result that is thus searched as specific information for specifying the application to be performed. The narrowing-down module (30, S39) narrows down the specific information based on the use history of the application that the user used, for example. The performing module (30, S47, S49) performs an application based on a result that is thus narrowed down.
According to the embodiment, it is possible to increase the convenience of the voice operation by narrowing-down the specific information based on the use history of the user.
A further embodiment further comprises a display module that displays the result that is narrowed down by the narrowing-down module, wherein the performing module performs an application based on a result that is selected when a selection operation is performed to the result that is narrowed down.
In the further embodiment, the display module (30, S43) displays the result that is narrowed down. Then, if the selection operation is performed to the result, the performing module performs an application based on the selection result.
In a still further embodiment, the display module displays results when there are a plurality of results that are narrowed down by the narrowing-down module.
In the still further embodiment, the display module displays a plurality of applications that are narrowed down as a candidate list when the results narrowed down are in plural. Then, the performing module performs an application based on a result of selection if a selection operation is performed to either one among the applications being displayed.
According to the further embodiment and the still further embodiment, when the specific information cannot be narrowed down, it is possible to make a user select an application to be used by displaying the candidate list.
In a yet further embodiment, the display module does not display a result when the result that is narrowed down by the narrowing-down module is one, and the performing module performs an application based on the result that is narrowed down by the narrowing-down module.
A yet still further embodiment further comprises a browsing module that performs a browser function connected to a network when the acquisition module cannot acquire the specific information; a search module that searches a search term based on an input voice using the network connected by the browser function; and a web page display module that displays a web page that is searched by the search module.
In the yet still further embodiment, the information terminal can perform the browser function connected to the network (100). The browsing module (30, S65) performs the browser function when the specific information cannot be acquired. If the browser function is performed, the search module (30, S67) searches the search term based on the input voice with a search engine site that is connected via the network, for example. The web page display module (30, S71) displays the web page that is thus searched.
According to the yet still further embodiment, even if a voice input of the language that is not registered in an application table is performed, it is possible to provide information to a user.
In a further embodiment, a browsing history of a web page is included in the use history, and the web page display module displays a web page based on the browsing history.
In the further embodiment, if the user browses a web page, the browsing history of the web page is recorded. If the browser function is performed by the browsing module, a web page with the highest browsing frequency is connected, and the search term is searched in that web page. Then, the web page display module displays the web page of a result that is thus searched.
According to the further embodiment, it is possible to provide specific information based on the browsing frequency of the web page by the user.
The other embodiment is a voice operation method in an information terminal (10) that comprises a storage module (48) operable to store a plurality of applications and a use history of each of the application, and can be operated by a voice input, a processor (30) of the information terminal performing: acquiring (S35) specific information for specifying an application to be performed based on a voice that is input; narrowing down (S39), based on the use history, the specific information that is acquired; and performing (S47, S49) an application based on a result that is narrowed down.
According to the other embodiment, it is possible to increase convenience of the voice operation by narrowing down the specific information based on a user use history.
Number | Date | Country | Kind |
---|---|---|---|
2012-211731 | Sep 2012 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2013/074975 | 9/17/2013 | WO | 00 |