This application claims priority to Japanese Patent Application No. 2018-038508 filed on Mar. 5, 2018, the contents of which are incorporated herein by reference.
The disclosure relates to a technique for inputting characters using speech input.
Patent Literature 1 describes a technique for generating a phrase list through speech input.
Patent Literature 1: Japanese Unexamined Patent Application Publication No. 2014-164260
However, the structure described in Patent Literature 1 converts an input speech into a string of characters, which is then output directly. In this case, the user needs to correctly pronounce characters to be input, but may sometimes forget some characters and fail to input those characters.
One or more aspects are directed to a technique for efficient speech input.
A character input device according to one or more aspects includes a speech input unit that receives an input speech, a controller that detects a substitution voice included in the input speech received by the speech input unit, and generates a predictive suggestion that includes a part replaced with the substitution voice and corresponds to the input speech, and a suggestion output unit that outputs the predictive suggestion.
This structure outputs a predictive suggestion that replaces characters generated based on a substitution voice with other characters for speech input, thus improving user convenience.
The controller included in this character input device may retrieve a predictive suggestion that completes a phrase corresponding to the input speech.
This structure outputs a predictive suggestion based on context in which a substitution voice is used.
The controller included in the character input device may display a plurality of predictive suggestions generated based on speech input data in an order of closer association with the speech input data.
This structure outputs predictive suggestions in the order of closer association with speech input data.
The character input device may include a storage to store the substitution voice that is customizable.
This structure allows customization for each user, thus improving convenience.
The technique according to one or more aspects enables efficient speech input.
Embodiments will now be described with reference to the drawings.
An embodiment will be described first with reference to
The character input device 10 includes an operation unit 100, a controller 200, a storage 300, and an output unit 400. The operation unit 100 includes a key operation unit 110 and a speech input unit 120. The output unit 400 includes a suggestion output unit 410 and an accepted-character output unit 420. The storage 300 stores, for example, dictionary data and data such as the history of inputs from a user. The suggestion output unit 410 displays normal suggestions retrieved from the dictionary data and predictive suggestions generated based on speech data.
When the key operation unit 110 is used by the user, the operation unit 100 activates the speech input unit 120. The speech input unit 120 is used by the user for speech input.
A string of characters to be input by speech (hereafter, a speech input string) may include a substitution voice. A substitution voice, which may be a bleep, serves as a wildcard. More specifically, the user substitutes a bleep for characters that the user has forgotten or does not know during speech input.
The speech input unit 120 outputs a speech input string to the controller 200. The controller 200 determines whether the speech input string includes a substitution voice. When the speech input string includes a substitution voice, or a bleep, the controller 200 searches the storage 300 for predictive suggestions that replace the bleep to complete the phrase.
The controller 200 outputs the predictive suggestions to the suggestion output unit 410. The suggestion output unit 410 allows the user to select an intended string of characters from the suggestions, and outputs the string to the accepted-character output unit 420.
The above structure outputs predictive suggestions from the storage 300 when the user has forgotten a string of characters to be input. In other words, the above structure allows the user to input characters efficiently using speech input, thus improving user convenience.
An example structure will be described in more detail with reference to
As shown in
When the speech input switching button 101 is pressed by the user, the operation unit 100 activates the speech input unit 120. The speech input unit 120 receives a speech including “inscrutable are,” “bleep,” and “heaven” from the user.
The speech input unit 120 then outputs a string of characters including “inscrutable are,” “bleep,” and “heaven” to the controller 200. The controller 200 detects the “bleep” used as a substitution voice in the string.
The controller 200 then searches the storage 300 for a character string that can be associated with the string “inscrutable are bleep heaven” using the “bleep” as a wildcard. More specifically, the controller 200 searches the storage 300 for a predictive suggestion that replaces the “bleep” to complete the phrase. Thus, the controller 200 outputs a search result such as “Inscrutable are the ways of Heaven.” to the suggestion output unit 410.
When the user selects the result “Inscrutable are the ways of Heaven,” the suggestion output unit 410 can output the intended string “Inscrutable are the ways of Heaven” to the accepted-character output unit 420.
An operation for enabling speech input using a substitution voice will be described in detail with reference to
The speech input unit 120 receives speech input (S101).
The controller 200 recognizes the input speech (S102).
The controller 200 determines whether the speech input string includes a substitution voice (S103).
When the speech input string includes a substitution voice (Yes in S103), the controller 200 searches the storage 300 for predictive suggestions, and outputs the suggestions to the suggestion output unit 410 (S104).
The suggestion output unit 410 displays the predictive suggestions (S105).
When the user selects an intended string of characters from the predictive suggestions, the suggestion output unit 410 outputs the string to the accepted-character output unit 420 (S106).
When the speech input string includes no substitution voice (No in S103), the controller 200 searches the storage 300 for suggestions, and outputs the suggestions to the suggestion output unit 410 (S114).
The suggestion output unit 410 displays the suggestions (S105).
When the user selects an intended string of characters from the suggestions, the suggestion output unit 410 outputs the string to the accepted-character output unit 420 (S106).
The above structure can search the storage 300 using a substitution voice included in an input string as a wildcard. In other words, the structure can retrieve predictive suggestions, and output the suggestions including a string of characters intended by the user.
Although the above example structure uses a speech input string including a single substitution voice, the same processing as in the example can be performed for a speech input string including multiple substitution voices.
Further, for example, a substitution voice can be used to replace a speech not to be heard by other people during speech input, thus protecting the user's privacy and enabling efficient character input.
A substitution voice is customizable by the user, and can be stored in the storage 300. This further improves user convenience.
A specific structure of a character input device according to an embodiment will now be described with reference to
A character input device 10A according to a second embodiment is included in a car navigation system. The character input device 10A according to a second embodiment has the same basic structure as in a first embodiment, which will not be described.
As shown in
The speech input unit 120 then outputs a string of characters including “bleep” and “museum” to the controller 200. The controller 200 detects the “bleep” used as a substitution voice in the string.
The controller 200 then searches the storage 300 for strings of characters that can be associated with the string “bleep museum” using the “bleep” as a wildcard. The controller 200 retrieves strings of characters that are closely associated with character strings previously input by the user, or searches the dictionary data.
Thus, the controller 200 outputs search results such as “AA National Museum” and “BB Museum” to the suggestion output unit 410. The user may select “AA National Museum” to start the navigation.
The above structure with the character input device 10A included in a car navigation system also receives a string of characters including a substitution voice, and predicts the part replaced with the substitution voice. Thus, the structure can output suggestions including a string of characters intended by the user.
In embodiments described above, a string of characters is output when a suggestion or a predictive suggestion displayed on the suggestion output unit 410 is selected. However, a suggestion or a predictive suggestion that is associated most closely with an input speech may be displayed without being displayed on the suggestion output unit 410.
Number | Date | Country | Kind |
---|---|---|---|
JP2018-038508 | Mar 2018 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
5734749 | Yamada et al. | Mar 1998 | A |
7461059 | Richardson | Dec 2008 | B2 |
8249873 | Blair | Aug 2012 | B2 |
8825484 | Yamada | Sep 2014 | B2 |
9799327 | Chan | Oct 2017 | B1 |
10854181 | Joshi | Dec 2020 | B2 |
20020007275 | Goto et al. | Jan 2002 | A1 |
20020184035 | Hartley | Dec 2002 | A1 |
20070088549 | Mowatt | Apr 2007 | A1 |
20070213983 | Ramsey | Sep 2007 | A1 |
20070288241 | Cross | Dec 2007 | A1 |
20110170674 | Egami | Jul 2011 | A1 |
20120084075 | Yamada | Apr 2012 | A1 |
20140379325 | Houache | Dec 2014 | A1 |
20150003735 | Feng et al. | Jan 2015 | A1 |
20160379638 | Basye | Dec 2016 | A1 |
20170242832 | Li | Aug 2017 | A1 |
20190043490 | Rivlin | Feb 2019 | A1 |
20190378501 | Mithra | Dec 2019 | A1 |
Number | Date | Country |
---|---|---|
2816489 | Dec 2014 | EP |
H07-191986 | Jul 1995 | JP |
H11-95793 | Apr 1999 | JP |
2000-330984 | Nov 2000 | JP |
2002-91492 | Mar 2002 | JP |
2014-164260 | Sep 2014 | JP |
Entry |
---|
Notice of Reasons for Refusal(Japanese Office Action = JPOA) dated Mar. 2, 2021 in a counterpart Japanese patent application(Japan Patent Application No. 2018-038508). |
Number | Date | Country | |
---|---|---|---|
20190272826 A1 | Sep 2019 | US |