The present disclosure relates to an information processing apparatus and an information processing method. Background
Conventionally, an information processing apparatus that executes various types of information processing according to utterance content of a user via an interactive voice user interface (UI) is known. Such an information processing apparatus includes, for example, a game system such as an online Role-Playing Game (RPG) capable of progressing a game according to a voice command uttered by the user (See, for example, Patent Literature 1) .
Patent Literature 1: Japanese Patent No. 6673513
However, in the above-described conventional technology, there is still room for further improvement in assigning a uniquely identifiable call to a display element such as an object for which general-purpose voice recognition is difficult.
Specifically, for example, in the RPG or the like, a unique name is set to an object such as a monster appearing as a character, but such a name is usually not a general phrase. For this reason, a general-purpose voice recognition engine cannot perform voice recognition by converting the name of the monster into text, for example.
Note that such a problem can be solved by registering the name of a monster or the like in dictionary information used by the voice recognition engine, but it is usual that such unknown phrases such as proper nouns continue to increase. For this reason, it is not realistic to update the dictionary information in accordance with an increase in the phrase in terms of cost.
Furthermore, even when the name of a monster or the like can be recognized by voice, if the user does not know the name in the first place, the user does not know how to specify a certain monster, for example.
Therefore, the present disclosure proposes an information processing apparatus and an information processing method capable of assigning a uniquely identifiable call to a display element for which general-purpose voice recognition is difficult.
According to the present disclosure, an information processing apparatus includes an acquisition unit that acquires a feature value related to a display element that is a target of a voice command uttered by a user, and a determination unit that determines a call of the display element on the basis of the feature value acquired by the acquisition unit such that the display element is uniquely specified with another display element other than the display element.
According to the present disclosure, an information processing method includes acquiring a feature value related to a display element that is a target of a voice command uttered by a user, and determining a call of the display element on the basis of the feature value acquired by the acquiring such that the display element is uniquely specified with another display element other than the display element.
Hereinafter, embodiments of the present disclosure will be described in detail with reference to the drawings. In each of the following embodiments, the same parts are denoted by the same reference numerals, and redundant description will be omitted.
In addition, in the present specification and the drawings, a plurality of components having substantially the same functional configuration may be distinguished by attaching different hyphenated numerals after the same reference numerals. For example, a plurality of configurations having substantially the same functional configuration are distinguished as a terminal device 10-1 and a terminal device 10-2 as necessary. However, in a case where it is not particularly necessary to distinguish each of a plurality of components having substantially the same functional configuration, only the same reference numeral is attached. For example, in a case where it is not necessary to particularly distinguish the terminal device 10-1 and the terminal device 10-2, they are simply referred to as the terminal device 10.
In addition, the present disclosure will be described according to the following item order.
In the present embodiment described below, a case where an information processing system 1 according to an embodiment is a game system that provides an online RPG service capable of progressing a game via a voice UI will be described as a main example.
First,
Furthermore, on the game screen, for example, an operation object of an online chat function represented as “Notification UI” or the like is displayed.
The user can progress the game by uttering a voice command including the call of the object, for example, while viewing the game screen.
Note that, although various objects are usually given proper nouns in terms of game settings, these are not general phrases, and thus cannot be recognized by a general-purpose voice recognition engine. Therefore, in order to call a proper noun in the game setting as a call in the voice command, the proper noun needs to be registered in the dictionary information of the voice recognition engine.
However, even when the proper noun is registered in the dictionary information, if the user does not know the proper noun in the first place, the user does not know what utterance can be used to designate the target object.
Therefore, in the information processing method according to the embodiment of the present disclosure, a feature value regarding the object that can be the target of the voice command uttered by the user is acquired, and the call of the object is determined such that the object is uniquely specified with another object other than the object on the basis of the acquired feature value. Note that the object mentioned here corresponds to an example of a “display element” presented to the user. In addition, the feature value corresponds to a static or dynamic value indicating a feature of the display element, such as a property value or a state value to be described later.
Specifically, in the information processing method according to the embodiment, the call that can uniquely specify each object is determined using attribute information assigned as static metadata to each object and analysis information obtained as a result of image analysis of the game screen being displayed.
More specifically, as illustrated in
Such property values may overlap, for example, for the same type, but all property values of a plurality of objects being displayed do not coincide with each other. Therefore, in the information processing method according to the embodiment, as illustrated in
For example,
By determining the call in this manner, the user can use a voice command designating an object by utterance as illustrated in
Note that a pronoun (hereinafter, referred to as a “distance reserved word”) including distance nuances such as “this” in the second line of
In addition, pronouns (hereinafter, referred to as a “time-series reserved word”) including time-series nuances such as “him” in the third line and “it” in the fifth line in
Furthermore, an adjective or the like (hereinafter, referred to as a “positional reserved word”) including positional nuances such as “left” in the fourth line of
As described above, in the information processing method according to the embodiment, the feature value related to the display element that can be the target of the voice command uttered by the user is acquired, and the call of the display element is determined such that the display element is uniquely specified with another display element other than the display element on the basis of the acquired feature value.
Therefore, according to the information processing method according to the embodiment, it is possible to assign a uniquely identifiable call to an object for which general-purpose voice recognition is difficult.
Hereinafter, a configuration example of the information processing system 1 to which the information processing method according to the above-described embodiment is applied will be described more specifically.
The terminal device 10 is a device used by each user, includes a voice UI, and executes various types of information processing according to utterance content of the user via the voice UI. In the present embodiment, the terminal device 10 executes the online RPG and progresses the game according to the voice command uttered by the user.
The terminal device 10 is a desktop personal computer (PC), a notebook PC, a tablet terminal, a mobile phone, a personal digital assistant (PDA), or the like. Furthermore, the terminal device 10 may be, for example, a robot that interacts with the user, a wearable terminal worn by the user, a navigation device mounted on a vehicle, or the like.
The server device 100 is a server device that provides an online RPG service to each terminal device 10 via the network N. The server device 100 collects a progress status of the game transmitted from each terminal device 10.
Furthermore, the server device 100 can assign a common call (hereinafter, referred to as a “common call”) to the same object simultaneously viewed by a plurality of users on the basis of the collected progress status or the like. Such an example will be described later in the description of the “fifth call determination processing” using
Next,
In other words, each component illustrated in
In the description using
As illustrated in
The terminal device 10 includes a communication unit 11, a storage unit 12, and a control unit 13. The communication unit 11 is realized by, for example, a network interface card (NIC) or the like. The communication unit 11 is connected to the server device 100 in a wireless or wired manner via the network N, and transmits and receives information to and from the server device 100.
The storage unit 12 is realized by, for example, a semiconductor memory element such as a random access memory (RAM), a read only memory (ROM), or a flash memory, or a storage device such as a hard disk or an optical disk. In the example illustrated in
The recognition model 12a is a model group for voice recognition in automatic voice recognition (ASR) processing to be described later, meaning understanding in natural language understanding (NLU) processing, dialogue recognition in interactive game execution processing, and the like, and is generated by the server device 100 as a learning model group using a machine learning algorithm such as deep learning, for example. The recognition model 12a corresponds to the general-purpose voice recognition engine described above.
The object information DB 12b is a database of information regarding each object displayed on the game screen, and includes attribute information of each object described above.
The reserved word information DB 12c is a database of information regarding reserved words, and includes definition information of each reserved word such as the above-described distance reserved word, time-series reserved word, and positional reserved word.
The control unit 13 is a controller, and is implemented by, for example, a central processing unit (CPU), a micro processing unit (MPU), or the like executing various programs stored in the storage unit 12 using a RAM as a work area. Furthermore, the control unit 13 can be realized by, for example, an integrated circuit such as an application specific integrated circuit (ASIC) or a field programmable gate array (FPGA).
The control unit 13 includes a voice recognition unit 13a, a meaning understanding unit 13b, an interactive game execution unit 13c, an acquisition unit 13d, a call determination unit 13e, and a transmission/reception unit 13f, and realizes or executes a function and an action of information processing described below.
The voice recognition unit 13a performs the ASR processing on the voice data input from the voice input unit 2, and converts the voice data into text data. Furthermore, the voice recognition unit 13a outputs the converted text data to the meaning understanding unit 13b.
The meaning understanding unit 13b performs meaning understanding processing such as NLU processing on the text data converted by the voice recognition unit 13a, and outputs a processing result to the interactive game execution unit 13c.
The interactive game execution unit 13c executes the game on the basis of the processing result of the meaning understanding unit 13b. Specifically, the interactive game execution unit 13c generates image information and voice information to be presented to the user on the basis of the processing result of the meaning understanding unit 13b.
In addition, the interactive game execution unit 13c presents the generated image information to the user via the display unit 3, performs voice synthesis processing on the generated voice information, and presents the generated voice information to the user via the voice output unit 4 to advance the game.
The acquisition unit 13d acquires attribute information including a property value that is an attribute value of each object from the object information DB 12b. In addition, the acquisition unit 13d appropriately acquires image information being presented to the user from the interactive game execution unit 13c.
In addition, the acquisition unit 13d performs image analysis on the acquired image information, and acquires a dynamic state value of each object being displayed. In addition, the acquisition unit 13d outputs the acquired state value of each object to the call determination unit 13e.
The call determination unit 13e executes call determination processing of determining the call of each object so that each object is uniquely specified on the basis of the attribute value and/or the state value of each object acquired by the acquisition unit 13d. Here, the call determination unit 13e can execute first call determination processing to fourth call determination processing. Specific contents of these processes will be described later with reference to
In addition, the call determination unit 13e appropriately outputs the determined call of each object to the interactive game execution unit 13c, and the interactive game execution unit 13c causes the game to proceed while specifying each object on the basis of the call determined by the call determination unit 13e.
The transmission/reception unit 13f transmits the progress status of the game output by the interactive game execution unit 13c to the server device 100 via the communication unit 11 as needed. In addition, the transmission/reception unit 13f receives the common call transmitted from the server device 100 via the communication unit 11, and appropriately outputs the common call to the interactive game execution unit 13c. The interactive game execution unit 13c causes the game to proceed while specifying each object on the basis of the common call received by the transmission/reception unit 13f.
Next, a configuration example of the server device 100 will be described.
As illustrated in
Similarly to the storage unit 12 described above, the storage unit 102 is realized by, for example, a semiconductor memory element such as a RAM, a ROM, or a flash memory, or a storage device such as a hard disk or an optical disk. In the example illustrated in
The object information DB 102a is similar to the object information DB 12b described above. The reserved word information DB 102b is similar to the reserved word information DB 12c described above.
Similarly to the control unit 13 described above, the control unit 103 is a controller, and is implemented by, for example, a CPU, an MPU, or the like executing various programs stored in the storage unit 102 using a RAM as a work area. Furthermore, similarly to the control unit 13 described above, the control unit 103 can be realized by, for example, an integrated circuit such as an ASIC or an FPGA.
The control unit 103 includes a collection unit 103a, a game progress control unit 103b, an acquisition unit 103c, a common call determination unit 103d, and a transmission unit 103e, and realizes or executes a function and an action of information processing described below.
The collection unit 103a collects the progress status of the game from each terminal device 10 via the communication unit 101 and outputs the progress status to the game progress control unit 103b. The game progress control unit 103b controls the progress of the game in each terminal device 10 via the communication unit 101 on the basis of the progress status collected by the collection unit 103a.
When the common call determination unit 103d determines the common call, the acquisition unit 103c acquires the attribute information including the attribute value of each object from the object information DB 102a. Furthermore, the acquisition unit 103c appropriately acquires image information being presented to each user from the game progress control unit 103b.
Furthermore, the acquisition unit 103c performs image analysis on the acquired image information, and acquires a dynamic state value of each object being displayed to each user from the analysis information. In addition, the acquisition unit 13d outputs the acquired state value of each object to the common call determination unit 103d.
On the basis of the attribute value and/or the state value of each object acquired by the acquisition unit 103c, the common call determination unit 103d executes fifth call determination processing of determining a common call so that each object is uniquely specified between users. Specific content of the fifth call determination processing will be described later with reference to
In addition, the common call determination unit 103d appropriately outputs the determined common call to the game progress control unit 103b, and the game progress control unit 103b controls the progress of the game while specifying each object common between the users on the basis of the common call determined by the common call determination unit 103d.
In addition, the common call determination unit 103d outputs the determined common call to the transmission unit 103e. The transmission unit 103e transmits the common call determined by the common call determination unit 103d to the corresponding terminal device 10 via the communication unit 101.
Next, a specific example of the call determination processing executed by the call determination unit 13e will be described with reference to
In the first call determination processing, the property values of the respective objects are compared, uniqueness is secured by using the non-overlapping property values, and the call of the target object is determined.
Specifically, as illustrated in
Here, in a case where there is no overlap (Step S102, No), the call determination unit 13e generates the call of the object using the property value (Step S103). On the other hand, in a case where there is the overlap (Step S102, Yes), the call determination unit 13e determines whether or not there is the next property value in the target object (Step S104).
Here, in a case where there is the next property value (Step S104, Yes), the call determination unit 13e repeats the processing from Step S101. In addition, in a case where there is no next property value (Step S104, No), the call determination unit 13e proceeds to another algorithm in the call determination processing.
More specifically, as illustrated in
Furthermore, as illustrated in
Note that, in
Furthermore, in
Next,
In the second call determination processing, a call is determined by assigning a time-series reserved word on the basis of a time-series change in a display object, a UI event, or the like. Here, the time-series reserved word is, for example, “It”, “Him”, “Her”, “Them”, or the like.
Specifically, as illustrated in
Here, in a case where there is a display change or an occurrence of a UI event (Step S201, Yes), the call determination unit 13e determines whether the time-series reserved word cannot be assigned (Step S202).
When the assignment to the time-series reserved word is possible (Step S202, No), the call determination unit 13e performs the assignment to the time-series reserved word (Step S203). When the assignment to the time-series reserved word is impossible (Step S202, Yes), the call determination unit 13e repeats the processing from Step S201.
More specifically, as illustrated in
Furthermore, as illustrated in
Furthermore, as illustrated in
Furthermore, as illustrated in
By this second call determination processing, each object can be uniquely specified by an appropriate pronoun according to a time-series change.
Next,
In the third call determination processing, uniqueness is secured by the positional reserved word from the positional relationship of each object, and the call of each object is determined. Here, the positional reserved word is, for example, “left”, “right”, “upper”, “lower”, or the like.
Specifically, as illustrated in
Here, in a case where there is an expressible object (Step S302, Yes), the call determination unit 13e determines the call by, for example, the positional reserved word and the object type (Step S303). Meanwhile, in a case where there is no expressible object (Step S302, No), the call determination unit 13e proceeds to another algorithm in the call determination processing.
More specifically, as illustrated in
Then, if the expression is possible, the call is determined using the object type and the positional reserved word. In the example of
In addition, since the area “upper” cannot be uniquely expressed, the algorithm shifts to another algorithm.
Note that, in the example of
Next,
In the fourth call determination processing, the uniqueness is secured by the distance reserved word on the basis of the spatial distance relationship from the predetermined point of each object or the temporal distance relationship from the current time point, and the call of each object is determined. Here, the distance reserved word is, for example, “This”, “That”, or the like. “It” already mentioned as the time-series reserved word may be used as the distance reserved word.
Specifically, as illustrated in
Here, in a case where there is an expressible object (Step S402, Yes), the call determination unit 13e determines the call by “This” or “That” (Step S403). Meanwhile, in a case where there is no expressible object (Step S402, No), the call determination unit 13e proceeds to another algorithm in the call determination processing.
More specifically, as illustrated in
Then, in the fourth call determination processing, it is determined whether or not the object can be uniquely expressed using the distance reserved word in each area.
Then, when it is expressible, the area name “This” or “That” of the corresponding area is assigned as the call. In the example of
Furthermore, as illustrated in
Next, a specific example of the common call determination processing will be described with reference to
In the fifth call determination processing, the server device 100 determines the common call so that the call is made by necessary players so that the call does not deviate between the players in the online chat or the like.
Specifically, as illustrated in
The description returns to
Furthermore, even users who are not in the same group at the current time point may be given the same call as much as possible to users who are in the same group at a certain high frequency or more. In addition, an object displayed only for some users may also be processed as a determination target of the common call.
In addition, since the same monsters and items in the screen are displayed without depending on the player, the same monsters and items may be subjected to the common call as shared objects and integrated.
Furthermore, it is preferable that the Notification notice or the like displayed to each individual user is not a target of the integration processing as a sharing prohibited personal object. In addition, the same applies to the fact that the call already assigned as the call to the individual object is not used as the common call.
By the way, the target range of the call assignment may be set according to the importance level of each object, for example. Furthermore, in such a target range, for example, priority may be determined according to the importance level of each object, and the order of assignment may be set. Furthermore, for example, the importance level may be recalculated on the basis of a change instruction by a voice command of the user, and the target range may be appropriately changed.
Next, these specific examples will be described.
In a case of setting the target range of the call assignment, as illustrated in
Here, the importance level is, for example, a spatial distance from a predetermined reference point P. The importance level is calculated to be higher as the distance is shorter, for example.
Then, it is determined whether there is a reference point change instruction by the user (Step S603). Here, when there is the change instruction (Step S603, Yes), the call determination unit 13e updates the importance level according to the change instruction (Step S604). When there is no change instruction (Step S603, No), the process proceeds to Step S605.
More specifically, for example, in the case of a spatial reference point change instruction, it is assumed that the importance level of each object being displayed is calculated based on the distance from the reference point P as illustrated in the upper part of
Then, here, it is assumed that a voice command “look farther to the left” is uttered from the user. Then, as illustrated in the lower part of
The reference point change instruction can also be applied to, for example, a temporal reference point (for example, the current time point). As illustrated in
The description returns to
Note that, in Step S605, the priority is set by, for example, sorting by importance level. The target range is set by a predetermined threshold, a number limit, or the like with respect to the importance level.
Then, it is determined whether or not the call determination within the target range has been completed (Step S607), and in a case where the call determination has been completed (Step S607, Yes), the processing ends. In addition, in a case where the processing has not been completed (Step S607, No), the target range is reset by changing the threshold, the number limit, or the like (Step S608), and the processing from Step S606 is repeated.
Meanwhile, the call determination processing described so far may be appropriately connected or may be appropriately combined. In the case of connection, the order may be statically fixed or may be dynamically changed according to the game situation.
Next, a specific example of such a case will be described.
As illustrated in
The example illustrated in
In addition, as illustrated in
Here, in a case where there is no overlap (Step S802, No), the call determination unit 13e generates the call of the object using the property value (Step S803). Meanwhile, in a case where there is the overlap (Step S802, Yes), the call determination unit 13e determines whether or not it can be uniquely expressed by “This” or “That” + property value (Step S804).
Here, in a case where expression is possible (Step S804, Yes), the call determination unit 13e determines the call by “This” or “That” + the property value (Step S805). Meanwhile, when the expression cannot be expressed (Step S804, No), the call determination unit 13e determines whether the target object has the next property value (Step S806).
Here, in a case where there is the next property value (Step S806, Yes), the call determination unit 13e repeats the processing from Step S801. In addition, in a case where there is no next property value (Step S806, No), the call determination unit 13e proceeds to another algorithm.
Note that, in
Furthermore, for example, in a combination of the positional reserved word + This/That, the call example is “This left monster”, “That left monster”, or the like. Furthermore, for example, in a combination of the property value + the positional reserved word + This/That, the call example is “This left red monster”, “That left red monster”, or the like.
Meanwhile, the call determined by each call determination processing described so far can be presented to the user by being displayed in the game screen. Such display examples are illustrated in
Note that the voice UI screens in
Furthermore, as illustrated in
In this manner, by visualizing what other users see, it is possible to easily determine availability at the time of communication such as online chat.
Furthermore, as illustrated in
Note that the case where the information processing system 1 according to the embodiment is the game system that provides an online RPG service has been described as a main example heretofore, but the present embodiment is not limited thereto, and can be applied to various other use cases.
As illustrated in
Furthermore, as illustrated in
Furthermore, as illustrated in
Furthermore, as illustrated in
For example, in a case where the vehicle is an autonomous driving vehicle and the user desires to follow and travel another vehicle visually recognized from the AR navigation system, or the like, it is possible to utter a voice command such as “following the red car that has just run” as illustrated in
Furthermore, in addition to this, the present invention may be applied to voice operation on an object in an AR space or a virtual reality (VR) space, communication with another user, or the like.
Among the processes described in the above embodiments, all or a part of the processes described as being performed automatically can be performed manually, or all or a part of the processes described as being performed manually can be performed automatically by a known method. In addition, the processing procedure, specific name, and information including various data and parameters illustrated in the document and the drawings can be arbitrarily changed unless otherwise specified. For example, the various types of information illustrated in each figure are not limited to the illustrated information.
In addition, each component of each device illustrated in the drawings is functionally conceptual, and is not necessarily physically configured as illustrated in the drawings. That is, a specific form of distribution and integration of each device is not limited to the illustrated form, and all or a part thereof can be functionally or physically distributed and integrated in an arbitrary unit according to various loads, usage conditions, and the like. For example, voice recognition unit 13a and meaning understanding unit 13b illustrated in
Furthermore, each function executed by the control unit 13 of the terminal device 10 illustrated in
In addition, the above-described embodiments can be appropriately combined in an area in which the processing contents do not contradict each other. In addition, the order of each Step illustrated in the sequence diagram or the flowchart of the present embodiment can be changed as appropriate.
The information device such as the terminal device 10 and the server device 100 according to the above-described embodiment is realized by a computer 1000 having a configuration as illustrated in
The CPU 1100 operates on the basis of a program stored in the ROM 1300 or the HDD 1400, and controls each unit. For example, the CPU 1100 develops a program stored in the ROM 1300 or the HDD 1400 in the RAM 1200, and executes processing corresponding to various programs.
The ROM 1300 stores a boot program such as a basic input output system (BIOS) executed by the CPU 1100 when the computer 1000 is activated, a program depending on hardware of the computer 1000, and the like.
The HDD 1400 is a computer-readable recording medium that non-transiently records a program executed by the CPU 1100, data used by the program, and the like. Specifically, the HDD 1400 is a recording medium that records an information processing program according to the present disclosure as an example of a program data 1450.
The communication interface 1500 is an interface for the computer 1000 to connect to an external network 1550 (for example, the Internet). For example, the CPU 1100 receives data from another device or transmits data generated by the CPU 1100 to another device via the communication interface 1500.
The input/output interface 1600 is an interface for connecting a input/output device 1650 and the computer 1000. For example, the CPU 1100 receives data from an input device such as a keyboard and a mouse via the input/output interface 1600. In addition, the CPU 1100 transmits data to an output device such as a display, a speaker, or a printer via the input/output interface 1600. Furthermore, the input/output interface 1600 may function as a media interface that reads a program or the like recorded in a predetermined recording medium (medium). The medium is, for example, an optical recording medium such as a digital versatile disc (DVD) or a phase change rewritable disk (PD), a magneto-optical recording medium such as a magneto-optical disk (MO), a tape medium, a magnetic recording medium, a semiconductor memory, or the like.
For example, in a case where the computer 1000 functions as the terminal device 10 according to the embodiment, the CPU 1100 of the computer 1000 executes the information processing program loaded on the RAM 1200 to implement the functions of the voice recognition unit 13a, the meaning understanding unit 13b, the interactive game execution unit 13c, the acquisition unit 13d, the call determination unit 13e, the transmission/reception unit 13f, and the like. In addition, the HDD 1400 stores the information processing program according to the present disclosure and data in the storage unit 12. Note that the CPU 1100 reads the program data 1450 from the HDD 1400 and executes the program data, but as another example, these programs may be acquired from another device via the external network 1550.
As described above, according to an embodiment of the present disclosure, the terminal device 10 (corresponding to an example of an “information processing apparatus”) includes the acquisition unit 13d that acquires the feature value regarding the object (corresponding to an example of the“ display element”) that can be the target of the voice command uttered by the user, and a call determination unit 13e (corresponding to an example of the “determination unit”) that determines the call of the object such that the object is uniquely specified with another object other than the object on the basis of the feature value acquired by the acquisition unit 13d. As a result, it is possible to assign a uniquely identifiable call to an object for which general-purpose voice recognition is difficult.
Although the embodiments of the present disclosure have been described above, the technical scope of the present disclosure is not limited to the above-described embodiments as it is, and various modifications can be made without departing from the gist of the present disclosure. In addition, components of different embodiments and modifications may be appropriately combined.
Furthermore, the effects of each embodiment described in the present specification are merely examples and are not limited, and other effects may be provided.
Note that the present technology can also have the following configurations.
(1) An information processing apparatus comprising:
(2) The information processing apparatus according to (1), wherein the acquisition unit acquires a state value of the display element acquired from an analysis result of an image including the display element and/or an attribute value set in the display element as the feature value.
(3) The information processing apparatus according to (1) or (2),
wherein the determination unit compares a first feature value that is the feature value of the display element with a second feature value that is the feature value of another display element corresponding to the first feature value, and determines the call of the display element so that the first feature value is included when the first feature value has uniqueness from the second feature value.
(4) The information processing apparatus according to (3), wherein the determination unit sequentially searches the first feature values and compares the first feature value with the second feature value when the display element has a plurality of the first feature values, and determines the call of the display element such that the first feature value is included when the first feature value has uniqueness from the second feature value.
(5) The information processing apparatus according to any one of (1) to (4), wherein the determination unit determines whether or not the call of the display element has uniqueness by assigning a time-series reserved word to the call of the display element when a change in the feature value of the display element or occurrence of an event related to the display element is detected, and determines the time-series reserved word as the call of the display element when the call has uniqueness.
(6) The information processing apparatus according to (5), wherein the determination unit assigns a pronoun to the call of the display element when the display element is an element relating to a message transmitted and received among a plurality of the users.
(7) The information processing apparatus according to (6), wherein when the display element is an element related to a partner user of the message, the determination unit assigns a personal pronoun according to genders or the number of the partner users to the call of the display element.
(8) The information processing apparatus according to any one of (1) to (7),
(9) The information processing apparatus according to (8),
(10) The information processing apparatus according to (8) or (9),
(11) The information processing apparatus according to any one of (1) to (10),
(12) The information processing apparatus according to (11),
wherein the acquisition unit sets the distance from the predetermined reference point of the display element as a spatial distance or a temporal distance.
(13) The information processing apparatus according to any one of (1) to (12),
(14) The information processing apparatus according to any one of (1) to (13),
wherein the determination unit determines priority and a target range for determining a call of the display element based on an importance level of each of a plurality of the display elements calculated from a predetermined reference point, and determines the call the display element in order according to the priority for the target range.
(15) The information processing apparatus according to (14),
wherein when change of the reference point is instructed from the user, the determination unit recalculates the importance level according to the change and changes the priority and the target range according to the recalculated importance level.
(16) The information processing apparatus according to (15),
wherein when a spatial change of the reference point is instructed from the user, the determination unit recalculates the importance level according to the spatial change.
(17) The information processing apparatus according to (15) or (16),
wherein when a change in which the reference point is temporally past is instructed from the user, the determination unit acquires the importance level in a past image according to the change in which the reference point is temporally past.
(18) The information processing apparatus according to any one of (15) to (17),
wherein the determination unit resets the target range when the calls of all the display elements in the target range are not uniquely determined.
(19) The information processing apparatus according to any one of (1) to (18),
wherein the display element is an object to be presented to the user.
(20) An information processing method comprising:
(21) A computer-readable recording medium storing a program for realizing, by a computer,
1
2
3
4
10
11
12
12
a
12
b
12
c
13
13
a
13
b
13
c
13
d
13
e
13
f
100
101
102
102
a
102
b
103
103
a
103
b
103
c
103
d
103
e
Number | Date | Country | Kind |
---|---|---|---|
2020-078461 | Apr 2020 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2021/014991 | 4/9/2021 | WO |