This application is based on and incorporates herein by reference Japanese Patent Application No. 2004-333660 filed on Nov. 17, 2004.
The present invention relates to a speech interaction apparatus and a speech interaction method.
Conventionally, an apparatus in which information is interactively provided to a user has been suggested (for example, see patent documents 1 and 2). A navigation apparatus disclosed in the patent document 1 actively provides information in accordance with situation by means of a speech interaction. In this apparatus, it is determined whether a condition for starting an information interaction is satisfied in accordance with, e.g., navigation information about a movement of a mobile body and user specific information about a usage history of a mobile body, and a speech is outputted to a user when this condition is satisfied.
In a navigation apparatus disclosed in the patent document 2, a language dictionary for recognition and a sound dictionary for recognition are changed in accordance with a workload of a user. When the workload is assumed to be high, each of a query response and a confirmation response is a single sentence. When the workload is low, a query response including multiple sentences or a query response including an implicit confirmation is used.
Patent document 1: JP-2003-329477A (U.S. Pat. No. 6,865,481 B2)
Patent document 2: JP-2003-108191A
In the navigation apparatus of the patent document 1, a condition for a speech interaction is not set in consideration of a drive route, e.g., set by a user. Accordingly, particularly when this navigation apparatus is used as, e.g., a guide book, a user may get the impression that a speech interaction starts suddenly. Additionally, an interaction starts even when a user gets tired, e.g., on a return route of a drive route and thus does not want to interact.
The navigation apparatus of the patent document 2 considers generation of a speech and a response when a user queries a speech interaction apparatus. However, a case where information is pushed, namely, a case where information is actively provided without a query of the user, is not considered. Accordingly, when there are many pieces of information provided from the speech interaction apparatus to the user, the user may not respond to the provided information because the information is pushed without considering a workload of the user.
Further, in both apparatuses of the patent documents 1 and 2, a current interaction cannot be suspended, and cannot be restarted at an appropriate timing.
The present invention is achieved in view of the above problems. It is an object of the present invention to provide a speech interaction method and speech interaction apparatus for achieving a user-friendly speech interaction. It is another object of the present invention to provide a speech interaction method and speech interaction apparatus in which a speech interaction can be executed while maintaining safety during a drive.
To achieve the above objects, a speech interaction apparatus according to a first aspect of the present invention is provided with the following. The speech interaction apparatus starts pushing information to a user and executes a speech interaction about the information in a case where (i) an interaction starting condition for starting pushing the information is set and (ii) the interaction starting condition is satisfied. In the speech interaction apparatus, an interaction policy setting unit is included for setting, in consideration of a drive route intended by the user, an interaction policy of certain information which satisfies an interaction starting condition. Furthermore, a speech interaction unit is included for pushing the certain information and executing a speech interaction about the certain information, in accordance with the interaction policy set by the interaction policy setting unit.
The speech interaction apparatus of the present invention not only pushes information when an interaction starting condition is satisfied, but also sets an interaction policy of information which satisfies an interaction starting condition in consideration of a drive route intended by the user. Accordingly, for example, an interaction policy can be set so that information is pushed at an appropriate timing in accordance with a drive route. Therefore, a user does not get the impression that a speech starts suddenly. As a result, a user-friendly speech interaction can be achieved.
As another aspect of the present invention, a speech interaction apparatus is provided with the following. The speech interaction apparatus starts pushing information to a user and executes a speech interaction about the information. In the speech interaction apparatus, an object information detecting unit is included for detecting object information about at least one object of a sign, a board, and a building. A speech interaction unit is included for pushing certain information about the detected at least one object and executing a speech interaction about the certain information, when the object information detecting unit detects the object information.
Accordingly, a speech interaction of information about, e.g., signs, boards, and buildings detected by an object information detecting unit can be executed even in a case where the information is not included in previously provided information.
As another aspect of the present invention, a speech interaction apparatus is provided with the following. A speech interaction unit is included for pushing information to a user and for executing a speech interaction about the information. An interaction policy setting unit is included for setting an interaction policy that, when a speech for suspending a speech interaction under execution by the speech interaction unit and a speech for specifying a timing for restarting the speech interaction are made by a user, the speech interaction under execution is suspended, and the suspended speech interaction is restarted at the specified timing. Here, the speech interaction unit executes a speech interaction in accordance with an interaction policy set by the interaction policy setting unit.
Accordingly, for example, when a user cannot determine a specific request about an interaction item, a speech interaction can be prevented from being forcibly executed. Additionally, at a timing intended by the user, a speech interaction can be restarted. As a result, a specific request which could not be determined when a speech interaction was suspended can be determined at the restarting.
As another aspect of the present invention, a speech interaction apparatus is provided with the following. A speech interaction unit is included for pushing information to a user and executing a speech interaction about the information. An interaction policy setting unit is included for setting an interaction policy that, in a case where information to be pushed includes a plurality of items or sentences, when a predetermined time passes after one item or sentence is pushed, an item or sentence following the pushed item or sentence is pushed. Here, the speech interaction unit executes a speech interaction in accordance with an interaction policy set by the interaction policy setting unit.
For example, when information about a characteristic, charge, business hours, and so on of a tourist facility is successively pushed, a user may not memorize the pushed information. Additionally, for example, when an interaction with the user is executed using a synthesized speech, the user may not hear successively pushed items and sentences because the synthesized speech is sometimes unclear. To avoid the above problems, information is pushed at specific intervals.
As another aspect of the present invention, a speech interaction apparatus is provided with the following. A speech interaction unit is included for pushing information to a user and executing a speech interaction about the information. An interaction policy setting unit is included for setting an interaction policy that a speech interaction of information which introduces a speech interaction function is executed at a preset timing. Here, the speech interaction unit executes a speech interaction in accordance with an interaction policy set by the interaction policy setting unit.
Accordingly, for example, at an approach of an expressway, a speech interaction function (for example, guiding to a nearest service area) in a case where a user feels tired and sleepy can be introduced.
As another aspect of the present invention, a speech interaction apparatus is provided with the following. The speech interaction apparatus starts pushing information to a user and executes a speech interaction about the information in a case where (i) an interaction starting condition for starting pushing the information is set and (ii) the interaction starting condition is satisfied. In the speech interaction apparatus, a usage history managing unit is included for setting an interaction starting condition relating to a usage history with respect to information, and for managing the usage history of the information based on an execution of a speech interaction about the information; further, a speech interaction unit is included for pushing certain information having no usage history and for executing a speech interaction about the certain information.
Accordingly, when a user moves on the same drive route as a previous route, information about the previous route can be prevented from being pushed.
As another aspect of the present invention, a speech interaction apparatus is provided with the following. A speech interaction unit is included for pushing information to a user and executing a speech interaction about the information. Further, an interaction policy setting unit is included for setting an interaction policy as follows: in a case where there is no speech of response to suggestion information even when a response waiting time passes after the speech interaction unit pushes the suggestion information to a user, a speech interaction of an additional explanation for the suggestion information or a speech interaction of different suggestion information from the suggestion information is executed. Here, the speech interaction unit executes a speech interaction in accordance with an interaction policy set by the interaction policy setting unit.
When there is no speech of response to suggestion information, it is assumed that a user cannot determine on the suggestion because provided information is insufficient, or that the user requests another suggestion because the suggestion information is not appropriate. Therefore, an additional explanation of the suggestion information and/or a speech interaction of different suggestion information are executed, so that a user can easily determine on the suggestion, or a different suggestion intended by the user can be made.
The above and other objects, features, and advantages of the present invention will become more apparent from the following detailed description made with reference to the accompanying drawings. In the drawings:
A speech interaction apparatus and speech interaction method of the present invention are explained below in reference to the drawings. In this embodiment, an example in which the speech interaction apparatus and speech interaction method of the present invention are used as a function of a navigation apparatus mounted in a subject vehicle such as an automobile is explained.
The control circuit 8 is structured as a normal computer, and includes a known CPU, ROM, RAM, and I/O, and, e.g., a bus line which connects these components. A program executed by the control circuit 8 is written in the ROM. In accordance with the program, the CPU and so on execute predetermined operations.
The position detector 1 has a geomagnetic sensor 2 for detecting a traveling direction of the subject vehicle, a gyroscope 3 for detecting an angular speed around a vertical direction of the subject vehicle, and a distance sensor 4 for detecting a movement distance of the subject vehicle, and a GPS receiver 5 for GPS (Global Positioning System) for detecting a current position of the subject vehicle in accordance with radio waves from satellites. Since these sensors have errors having different characteristics, the sensors complement each other.
In accordance with an accuracy of each sensor, the position detector 1 may have part of the above sensors. Additionally, the position detector 1 may have, e.g., a steering rotation sensor (not shown) and a vehicle speed sensor (not shown) for detecting a speed from a rotation speed of each wheel.
The map data input device 6 inputs map data including road data, index data, and drawing data. To store the map data, storage media such as CD-ROMs and DVD-ROMs and rewritable storage media such as memory cards and hard disk are used. Link data and node data structuring road data, index data, and drawing data are explained below.
First, a link means a road between nodes. Each road on the map is divided at nodes such as intersections, branches, and meeting points. By connecting the links, roads are structured. The link data includes an identification number (link ID) for identifying a link, a link length showing a length of a link, coordinates (latitudes, longitudes) of a start and end of a link, a name of a road, a type of a road (such as an expressway, a toll road, a national road, and a prefecture road), a road width, a time required for traveling a link (traveling time by walk and traveling time using an automobile), and so on.
The node data includes a node ID which is a unique number provided to each node where each road on the map intersects, branches, and meets, a node coordinate (latitude, longitude), a connection link ID in which link IDs of all links connected to a node are described, and an attribute and name of a node (for example, a name of an intersection of a general road and a name of an interchange (IC), a junction (JCT) of an expressway or toll road), the attribute showing whether a node is a branch, a junction, or an intersection.
The index data is used for retrieving, e.g., buildings, and includes, e.g., names, types, addresses, phone numbers, positions (latitudes, longitudes) of buildings, roads, and points where roads intersect, branch, and meet.
The drawing data is used for displaying maps, and includes display characters, display positions (latitudes, longitudes), polygon data, and marks (such as landmarks), with respect to landforms (e.g., seas, lakes, ponds, and mountains), facilities, roads, and points where roads branch, meet, and intersect.
The operation switch group 7 uses, e.g., a touch switch integrated with the display device 10 or a mechanical switch, and is used for various inputs.
The guide point DB 9 stores guidance information used for speech interactions. In the guidance information, as shown in
As shown in
Interaction starting conditions of information (g) about a function introduction are the same as those of the pieces of the information (a) to (d) shown in
The display device 10 includes, e.g., a liquid crystal display. The display device 10 displays on its screen (i) a vehicle position mark corresponding to a current position of the subject vehicle inputted from the position detector 1, and (ii) a road map around the subject vehicle generated using map data inputted from the map data input device 6.
The transmitter-receiver 11 communicates with the outside of the subject vehicle, and receives update data of guidance information to be stored in the guide point DB 9 from an external information center. The transmitter-receiver 11 is connected, via beacons mounted on roads and FM broadcast stations, to a VICS sensor for receiving road traffic information provided from a VICS (Vehicle Information and Communication System) (registered trademark) center, and transmits this road traffic information to the control circuit 8.
Road traffic information received by this VICS sensor includes, e.g., traffic congestion information such as a traffic congestion of each link and a required traveling time of each link, regulation information about roads closed due to accidents and construction works and about closures of interchanges of expressways, and traffic density information showing traffic density (high, middle, low). The received road traffic information is processed in the control circuit 8. For example, the traffic congestion information and regulation information can be displayed over a road map displayed on the screen of the display device 10.
The speech input microphone 12 receives a speech inputted by a user, converts it to speech signals, and outputs them to the control circuit 8. When a speech is inputted, by pressing, e.g., a button of a talk switch (not shown), an input trigger signal is transmitted to the control circuit 8. The control circuit 8 receives the input trigger signal to start receiving a speech inputted through the speech input microphone 12.
The speaker 13 is used for providing various information to a user of the navigation apparatus by means of a speech. The speaker 13 may be, e.g., a speaker mounted in the subject vehicle or a speaker installed in the navigation apparatus.
The route setting unit 100 sets a destination, visiting points (or passing points) before arrival at the destination, and a starting point (such as a home position and a current position) in accordance with operations of, e.g., the remote controller and the operation switch group 7, or with, e.g., a speech input from a speech input microphone 12.
The route calculating unit 101 automatically calculates an optimum route of a drive route (from the starting point to the destination through the visiting points) set by the route setting unit 100. The Dijkstra method is one of known methods for setting an optimum route.
The route calculating unit 101 outputs route guidance information about a calculated route to the interaction scenario generating unit 107. The route calculating unit 101 can calculate a drive route including an outward route and a return route. The outward route is from the starting point to the destination through the visiting points. The return route is from the destination to the starting point.
The GPS receiving unit 102 receives information about, e.g., a current position and a current time. The road traffic information acquiring unit 103 acquires road traffic information transmitted from the transmitter-receiver 11.
The speech recognizing unit 104 includes, e.g., an analog-to-digital conversion circuit, a recognition program processing unit, a sound model storing unit, and a recognition dictionary storing unit. These components are not shown. The analog-to-digital conversion circuit receives a speech signal inputted via the speech input microphone 12, and converts this signal to a digitalized signal. The converted digital speech signal is transmitted to the recognition program processing unit.
The recognition program processing unit converts a digital signal to recognition words by use of the sound model storing unit and the recognition dictionary storing unit. For example, the recognition program processing unit analyzes a speech (reading of recognition words) corresponding to a digital speech signal by means of, e.g., the known Hidden Markov Model. This analyzed reading of the recognition words is checked with recognition words stored in the recognition dictionary storing unit, and the most certain recognition words are extracted. These extracted words are outputted to the speech interaction unit 106 as the speech recognition result.
The speech synthesizing unit 105 receives a message (output message) that is to be outputted via the speaker 13, from the speech interaction unit 106. The unit 105 then generates a synthesized speech corresponding to the content of the message.
The speech interaction unit 106 is a control unit for executing an interaction with a user. The speech interaction unit 106 acquires, from the interaction scenario generating unit 107, an interaction scenario appropriate for a speech recognition result from the speech recognizing unit 104. The speech interaction unit 106 outputs an output message in accordance with the acquired interaction scenario to the speech synthesizing unit 105.
Even when the speech recognizing unit 104 does not output a speech recognition result, the speech interaction unit 106 acquires an interaction scenario from the scenario generating unit 107 in accordance with an interaction policy set by the scenario generating unit 107. Then, the speech interaction unit 106 outputs an output message in accordance with the scenario to the speech synthesizing unit 105.
The interaction scenario generating unit 107 actively provides (hereinafter called “pushes”) information to a user, generates an interaction scenario for executing a speech interaction about the information, and sets the interaction policy. Guidance information needed for the interaction is acquired from the guide point DB 9.
The navigation apparatus of this embodiment including the above structure pushes information to a user, and executes a speech interaction process which executes a speech interaction about the information. The navigation apparatus not only pushes information when the above interaction starting condition is satisfied, but also sets an interaction policy of information which satisfies an interaction starting condition in consideration of a drive route intended by the user.
The navigation apparatus pushes information which satisfies an interaction starting condition in accordance with this set interaction policy, and executes a speech interaction about the information. This results in achieving the user-friendliness and safety.
Next, a feature of the navigation apparatus of this embodiment is explained. The navigation apparatus of this embodiment executes an interaction policy setting process for setting an interaction policy in accordance with previously set rules (rules 1 to 9). The rules 1 to 4 are in accordance with a drive route. The rule 5 is in accordance with a convenience of a user. The rule 6 is for the case when many pieces of information are pushed. The rule 7 is in consideration of user-friendliness. The rule 8 is about a usage history of each user. The rule 9 is for the case when a user does not respond. (Rules 1 to 4: in accordance with a drive route) The rules 1 to 4 are in accordance with a drive route intended by a user, and include the rule when an expressway is used on an outward route (rule 1), the rule for a point where, e.g., a sign, a board, or a building is visible (rule 2), the rule on a return route (rule 3), and the rule when there are a plurality of pieces of information for a point (rule 4).
Rule 1: when an expressway is used on an outward route
The rule 1 is that, when an expressway is used on an outward route, information which can be previously processed is pushed during a drive on the expressway. Since a workload of a user on an expressway is usually lower than a workload on a general road, a speech interaction is easily executed on the expressway. Therefore, on an expressway where a workload is relatively low, information which can be previously processed is pushed, so that a speech interaction can be executed while maintaining safety during a drive.
On a general road where there are few signals and thus a traffic flow is smooth, a workload is low as well as on an expressway. In this sense, the expressway in this embodiment includes a general road where a workload is low.
An example of this rule 1 is as follows. A drive route is first grasped in accordance with route guidance information. Premises of the rule are that the drive route includes an outward route and a return route, that the outward route includes an expressway, and that there is guidance information which satisfies the interaction starting conditions shown in
For example, as shown in
When an expressway includes a section where a traffic density is high and a workload is high, only information in which a priority of equal to or over a certain value (for example, equal to or over two) is set as the interaction starting condition may be placed. Accordingly, a speech interaction in a state that a workload is high can be avoided to the utmost. The information may be placed on a section except a section where a workload is high, or on a section where a workload is low on a general road used on an outward route.
For instance, a road section where a traffic density shown by the above road traffic information is high may be specified as the section where a workload of a user is high.
Rule 2: at a point where a sign, a board, a building, or the like is visible
The rule 2 is that, at a point where a sign, a board, or a building is visible, the related information is pushed. At a point where a user can see a sign, a board, or a building, information about that is pushed, so that the user does not get the impression that a speech interaction is suddenly started. The information is pushed at such an appropriate timing, so that signs, boards, buildings, and so on are effectively advertised.
In detail, in this rule 2, as well as in the rule 1, a drive route is first grasped. Guidance information which satisfies the interaction starting condition of the point range shown in
At a point where a sign “2 km for Okazaki IC” is visible, interactions about sightseeing such as “The Okazaki Castle is famous in Okazaki” and “A specialty of Okazaki is Hatcho Miso” are executed. At a point where a board “Kawasa Chikuwa” is visible, interactions about noted local products such as “A specialty around here is Chikuwa” or about advertisement are executed. (Here, “Miso” is a Japanese word indicating fermented soybean pastes, while “Chikuwa” is a Japanese word indicating a fish sausage.)
Buildings of the rule 2 may include incomplete buildings, e.g., under construction. Since a user tends to be strongly interested in incomplete buildings, e.g., under construction, information about incomplete buildings is pushed to enhance availability of the speech interaction apparatus.
As described above, the information about buildings includes incomplete buildings, e.g., under construction, so that, for example, in the guidance information (e) of
As shown in
Even if information about a building under construction has a usage history, the information having an expiration date may be pushed. Accordingly, information about an incomplete building in which a user was previously interested can be repeatedly pushed. The above related information is such that the latest information about an incomplete building is acquired from the outside via the transmitter-receiver 11, and pushed.
For example, in the guidance information (f) of
Rule 3: a rule on a return route
The rule 3 is that the number of interactions is made minimum, and that the interactions are made short. Since a user often gets tired on a return route (or homeward route), the user does not actively interact. When an outward route and return route are the same, an interaction of the same information as that for the outward route does not need to be executed on the return route.
Therefore, with respect to the return route, among information having no usage history, low-priority information about, e.g., sightseeing, specialties, and native products is not pushed, but only high-priority information (equal to or over a certain value) of information is pushed. Accordingly, an effective speech interaction can be executed.
Rule 4: when there are multiple pieces of information at a point
The rule 4 is that when there are multiple pieces of information at a point, information is pushed at time intervals. When a user receives many pieces of information at once, the user often cannot respond to them. In a case where there are multiple pieces of information which satisfy an interaction starting condition of the same point range on a drive route, information is pushed as follows. When a predetermined time passes after an interaction about one piece of information is pushed, another piece of information is pushed, as long as an interaction starting condition of each piece of information is satisfied.
As described above, instead of providing many pieces of information to a user at once, when a predetermined time passes after an interaction about one piece of information, another piece of information is pushed, so that the user can grasps provided information. Information may be pushed at equal time intervals as long as a response starting condition is satisfied. As a result, a time during which no interaction is executed can be prevented from being prolonged.
For example, as shown in
As described above, by setting interaction policies of the rules 1 to 4 in accordance with a drive route, guidance information shown in
Rule 5: in accordance with convenience of a user
The rule 5 is that a current interaction can be suspended, and a timing for restarting the interaction can be specified. When an interaction is forcibly executed at the time when a user cannot determine a specific request about an interaction, the user can be discouraged from using a speech interaction.
When a speech for suspending a current speech interaction is made, the speech interaction is suspended. Accordingly, for example, a speech interaction can be prevented from being forcibly executed at the time when a user cannot determine a specific request about an interaction.
When a speech about a timing for restarting a speech interaction is made at the same time as a speech for suspending the speech interaction, the suspended speech interaction can be restarted at the specified timing. Accordingly, a speech interaction can be restarted at a timing intended by a user. As a result, a request which could not be determined when a speech interaction was suspended can be determined when the speech interaction is restarted.
In the example shown in
Rule 6: when there are many pieces of information to be pushed
The rule 6 is that, when there are many pieces of information to be pushed to a user, speeches are made at time intervals. For example, when information about characteristics, fee, and business hours of a sightseeing facility is successively pushed, a user may not memorize the pushed information. For example, when a user interacts with a synthesized speech, the user may not hear the successively pushed items and sentences because the synthesized speech is not clear.
In a case where information to be pushed includes multiple items or sentences, when a predetermined time (for example, thirty seconds in the example of
Rule 7: in consideration of user-friendliness
The rule 7 is that a speech interaction function is introduced at an appropriate timing. For example, as information for introducing functions shown in
Rule 8: about a usage history of each user
The rule 8 is that a usage history of each user is managed, and information having a usage history is not pushed to a user who uses a navigation apparatus. In a case where a user moves on the same drive route as the previous one, when the previously pushed information is pushed again, the user may feel it to be annoying.
Therefore, a usage history of each user is managed, and information having no usage history is pushed to a user who uses the navigation apparatus. Accordingly, when a user moves on the same drive route as the previous one, previously pushed information can be prevented from being pushed again.
To manage a usage history of each user, a user needs to be specified. To specify a user, known user certification processes using a face image, living body information, speech information, a password, and so on may be executed.
In the rule 8, for example, since traffic information and weather information always change, usage histories of the information are not managed. Namely, a usage history corresponding to time varying information which always varies is not managed. Accordingly, the time varying information can be pushed repeatedly.
Rule 9: rule when a user does not respond
The rule 9 is that when a user does not respond to a suggestion, an additional explanation of the suggestion (or another suggestion) is made. Namely, when there is no speech in response to suggestion information, it is assumed that a user cannot determine the suggestion because provided information is insufficient, or requests another suggestion because the suggestion is not appropriate.
In such a case, an additional explanation of the suggestion or a speech interaction of another suggestion information are made. A user thereby easily determines the suggestion. Alternatively, another suggestion is thereby made which is assumed to follow the intention of user.
For example, like the content of the interaction shown in the example 1 of
As described above, the interaction policy setting processes for setting interaction policies in accordance with the rules 1 to 9 are divided into the rules which can be set after a drive route is determined and the rules which are set in accordance with, e.g., a workload and speech content of a user during a speech interaction process.
Next, interaction policy setting processes in the navigation apparatus of this embodiment are explained using flowcharts shown in FIGS. 11 to 14.
First, an interaction policy setting process in accordance with the rules 1 to 4, 7, and 8 is explained. In Step 10 shown in
For example, in a case of the pieces of the information (a) to (d), the pieces of the information (b) to (d) are placed on a drive route in Step 10, as shown in
In Step 20 shown in
Next, an interaction policy setting process in accordance with the rule 5 is explained. This process is executed during a speech interaction with a user. First, in Step 30 shown in
In Step 40, a current interaction is suspended, and a timing for restarting the interaction is placed to a timing specified by a user. In Step 50, it is determined whether the interaction is completed. When the determination is positive, the process is finished. When the determination is negative, the process goes to S30 to repeat the above-described process.
Next, an interaction policy setting process in accordance with the rule 6 is explained. This process is executed during an interaction with a user. First, in Step 60 shown in
When the determination is positive, the process is finished. When the determination is negative, a time interval for reading speeches is set in Step 80 in accordance with a workload of a user. In Step 90, the process waits until the set time in Step 80 passes. When the time passes, the process goes to Step 60 to repeat the above-described process.
Lastly, an interaction policy setting process in accordance with the rule 9 is explained. This process is executed during an interaction with a user. First, in Step 100 shown in
The navigation apparatus of this embodiment sets an interaction policy of information which satisfies an interaction starting condition in consideration of a drive route intended by a user. In accordance with the set interaction policy, information which satisfies an interaction starting condition is pushed to execute a speech interaction about the information.
Accordingly, for example, since information can be pushed at an appropriate timing in accordance with a drive route, a user does not get the impression that a speech is suddenly started. As a result, a user-friendly speech interaction can be achieved.
Alternative 1
For example, since a traffic congestion and traffic regulation suddenly happen, an interaction policy needs to be set for dynamic information such as road traffic information about the traffic congestion and traffic regulation when the dynamic information is acquired. Therefore, in an interaction policy setting process of this alternative, when dynamic information such as road traffic information is acquired by the road traffic information acquiring unit 103 as a dynamic information acquiring unit, an interaction policy including the acquired dynamic information is reset. Accordingly, a speech interaction about the dynamic information such as the traffic congestion information can be achieved.
Alternative 2
The rule 2 explained in this embodiment “at a point where a sign, a board, or a building is visible, the related information is pushed” is as follows. As well as in the rule 1, a drive route is first grasped, and guidance information which satisfies an interaction starting condition of a point range shown in
On the other hand, in this alternative, an object information detecting unit is provided for detecting information about at least one object of a sign, a board, and a building. An interaction policy is set so that, when the object information detecting unit detects object information, information about the object is pushed and a speech interaction about the information is executed.
Accordingly, a speech interaction of information about, e.g., a sign, a board, and a building, which is not previously provided, can be executed. As well as in the embodiment, buildings may include incomplete buildings even in this alternative. Accordingly, information about the incomplete buildings can be detected.
To detect a point where an object such as a sign, a board, or a building is visible, an image recognition process apparatus for detecting objects by means of an image recognition process may be used to determine whether the object is visible. Without the image recognition process apparatus, when information about a position of an object and about a speech interaction of the object is detected, it can be determined from a current position and a position of the object whether the object is placed on a point where the object is visible. Additionally, information required for a speech interaction can be detected.
Further, information about the object can be acquired from the outside via the transmitter-receiver 11. By use of the various apparatuses, information about the object can be detected. The information about the object acquired via the transmitter-receiver 11 may be acquired using interactive communications such as a digital broadcast. The information about the object may be transmitted from signs, boards, and buildings themselves.
It will be obvious to those skilled in the art that various changes may be made in the above-described embodiments of the present invention. However, the scope of the present invention should be determined by the following claims.
Number | Date | Country | Kind |
---|---|---|---|
2004-333660 | Nov 2004 | JP | national |