The present invention relates to a gameplay operation learning apparatus, a gameplay operation learning method, and a recording medium.
In various games including board games such as Go and Shogi and computer games such as fighting games and shooting games, computers may control characters and the like.
One of the techniques used for such control by computer is described in, for example, Patent Literature 1. Patent Literature 1 describes a learning apparatus for a fighting game that includes a storing unit for storing various programs and data, and a control unit for controlling the motions of a plurality of characters appearing in the fighting game based on the state of an operation by an input operation unit and the programs stored in the storing unit. According to Patent Literature 1, the control unit collects operation data related to tricks performed by the characters in response to operations by the input operation unit and screen state data related to screen display at predetermined timings, and executes a learning program to write the screen state data collected at the predetermined timings into a learning data storing unit. Then, the control unit optimizes the weight of a learning result by performing a deep learning calculation process based on the screen state data stored in the learning data storing unit.
In order to perform more proper learning, such as more human-like or more like a learning target, it is desirable to perform imitation learning as described in Patent Literature 1 rather than reinforcement learning in which a learner learns through his/her own action, but a large number of histories of play data and the like are necessary for properly performing imitation learning. However, in the case of simply collecting play data corresponding to operations by players as described in Patent Literature 1, it is difficult to collect enough play data necessary for imitation learning. As a result, a problem occurs that it is difficult to perform learning for getting close to a learning target, for example, it is difficult to learn computer player's operations that are closer to human operations.
Accordingly, an object of the present invention is to provide a gameplay operation learning apparatus, a gameplay operation learning method and a recording medium that can solve the problem that it may be difficult to perform learning for getting close to a learning target.
In order to achieve the object, a gameplay operation learning apparatus as an aspect of the present disclosure includes: an acquiring means that acquires play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; a learning means that generates a game player model for outputting an action of the learning target in response to input of a second play state based on the play data and the label; and an output means that outputs the game player model.
Further, a gameplay operation learning method as another aspect of the present disclosure is a gameplay operation learning method by an information processing apparatus, including: acquiring play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; and generating a game player model for outputting an action of the learning target in response to input of a second play state based on the play data and the label.
Further, a recording medium as another aspect of the present disclosure is a non-transitory computer-readable recording medium on which a program is recorded, and the program includes instructions for causing an information processing apparatus to realize processes to: acquire play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; and generate a game player model for outputting an action of the learning target in response to input of a second play state based on the play data and the label.
According to the respective configurations as described above, it is possible to provide a learning apparatus, a learning method and a recording medium that can favorably learn so as to get close to a learning target, for example, so as to bring computer player's operations closer to human operations.
A first example embodiment of the present disclosure will be described with reference to
In the first example embodiment of the present disclosure, the learning apparatus 100 (gameplay operation learning apparatus) that performs machine learning based on play data in various games including board games such as Go and Shogi and computer games such as fighting games and shooting games will be described. As shown in
The learning apparatus 100 is an information processing apparatus that performs machine learning based on the play data of a game acquired from an external device and the like. The game may include board games such as Go and Shogi, computer games such as fighting games and shooting games, and any other games. For example, the learning apparatus 100 is a server apparatus and the like. The learning apparatus 100 may be a single information processing apparatus, or may be realized, for example, on the cloud.
The communication I/F unit 110 includes a data communication circuit. The communication I/F unit 110 performs data communication with an external device and the like connected via a communication line.
The storing unit 120 is a storage device such as a hard disk or a memory. The storing unit 120 stores processing information necessary for a variety of processing by the operation processing unit 130 and a program 123. The program 123 is read and executed by the operation processing unit 130 to realize various processing units. The program 123 is loaded in advance from an external device or a recording medium via a data input/output function such as the communication I/F unit 110 and stored into the storing unit 120. Major information stored in the storing unit 120 includes, for example, input data 121 and a neural network 122.
The input data 121 includes play data indicating an action taken by a player in a game, the state of the game, and so forth. The input data 121 is acquired for learning from an external device or the like via the communication I/F unit 110 or the like.
The attribute indicates information corresponding to a player who plays a game, the characteristics of the player, and the like, such as the type and proficiency of the player.
Play data has attributes corresponding to the characteristics of a player who plays a game as illustrated above. Meanwhile, the attributes may be those corresponding to the characteristics of play data such as a specific action is taken many times within a predetermined time period, instead of the characteristics of a player or along with the characteristics of a player.
Further, a label indicating whether or not to be a learning target is assigned in advance to play data by an external device, for example. Specifically, for example, a success case label that is a first label is assigned to play data having an attribute to be a learning target, and a failure case label that is a second label different from the first label is assigned to play data having an attribute different from an attribute to be a learning target. A failure case label may be assigned to play data having an attribute that is contrary to an attribute to be a learning target, rather than a simply different attribute.
As an example, in a case where a success case label is assigned to play data of an advanced player, a failure case label is assigned to play data having an attribute different from that of an advanced player, such as an attribute of an intermediate player or a beginner. In a case where a success case label is assigned to play data of an advanced player, a failure case label may be assigned to play data having an attribute opposite that of an advanced player, such as an attribute of a beginner. In a case where a success case label is assigned to play data having a specific person attribute indicating a specific person such as professional A, a failure case label may be assigned to play data that does not have a specific person attribute indicating a specific person. In the respective examples described above, a failure case label may be assigned to play data having a player attribute indicating that the player is not a human but AI. By assigning a failure case label to play data having a player attribute indicating AI, it becomes possible to perform a machine learning process so as to become away from an AI-like, namely, non-human-like action. For example, by assigning a success case label to play data having a specific person attribute indicating a specific person and assigning a failure case label to play data having a player attribute indicating AI, it becomes possible to update the weight value of the neural network 122 and so forth so as to get closer to play data of the specific person and become away from a non-human-like action.
In the case of this example embodiment, an attribute to be a learning target may be identified by any means. In addition, instead of assigning a label in advance, the learning apparatus 100 may be configured to assign a label to play data based on information showing an attribute acquired in addition to play data, information showing an attribute to be a learning target, and the like.
Further, play data included by the input data 121 shows an action taken by a player in a game, the state of a game, and so forth. For example, play data includes state information showing a game state in a game (first play state), behavior information indicating an action taken by a player in the state, and so forth.
As an example,
Further, as another example,
Thus, the input data 121 includes play data corresponding to a game to be a learning target. Meanwhile, the input data 121 may include play data in individual scenes, respectively, or may include play data as time-series data in which states and behaviors are chained as shown in
The neural network 122 is subjected to a machine learning process using the input data 121 and the like that is training data so as to output behavior information corresponding to state information and the like when play data including the state information is input. In other words, the neural network 122 is subjected to a machine learning process so as to output an action of a learning target in response to input of a second play state.
The operation processing unit 130 has an arithmetic logic unit such as a CPU (Central Processing Unit) and a peripheral circuit thereof. The operation processing unit 130 reads the program 123 from the storage unit 120 and executes the program 123 to make the abovementioned hardware and the program 123 cooperate with each other and realize various processing units. Major processing units realized by the operation processing unit 130 include, for example, an acquiring unit 131, a learning unit 132, and an output unit 133.
The acquiring unit 131 acquires play data and so forth from an external device or the like. For example, the acquiring unit 131 acquires, in addition to play data, information indicating the attribute of the play data, and so forth. Moreover, the acquiring unit 131 stores the acquired play data and so forth as the input data 121 into the storing unit 120.
Further, the acquiring unit 131 can acquire information indicating an attribute to be a learning target, and so forth. For example, the acquiring unit 131 may acquire, in addition to play data, information indicating an attribute to be a learning target, and so forth, or may acquire information indicating an attribute to be a learning target, and so forth, at a different timing from that of play data.
The learning unit 132 performs machine learning for outputting an action of a learning target in response to input of the second play state based on the input data 121 that includes play data including the first play state and an action and a label. For example, the learning unit 132 inputs the input data 121 that is training data into the neural network 122. Then, the learning unit 132 updates a weight value and the like of the neural network 122 so as to get close to play data to which a success case label is assigned and become away from play data to which a failure case label is assigned. For example, the learning unit 132 repeats the above processing using a number of training data to generate a game player model that is a created model corresponding to an attribute to be a learning target. The learning unit 132 may perform a machine learning process using a known means.
As an example, for example, as shown in
Further, the learning unit 132 may be configured to generate training data by assigning a label to play data based on information indicating an attribute to be a learning target acquired by the acquiring unit 131. For example, the learning unit 132 can assign a success case label to play data having an attribute to be a learning target, and assign a failure case label to play data having an attribute different from the attribute to be the learning target. The learning unit 132 may assign a failure case label to play data having an attribute contrary to the attribute to be the learning target among play data having an attribute different from the attribute to be the learning target. Which attribute and which attribute are contrary to each other may be determined in advance or may be determined by the learning unit 132 by any means, for example.
The output unit 133 outputs a game player model and the like that is the result of learning by the learning unit 132. For example, the output unit 133 can output the above gameplay model and the like to an external device or the like via the communication I/F unit 110 and the like.
The above is an example of the configuration of the learning apparatus 100. Subsequently, an example of the operation of the learning apparatus 100 will be described with reference to
The learning unit 132 inputs the input data 121 that is training data to the neural network 122. Then, the learning unit 132 updates the weight value and the like of the neural network 122 so as to get close to play data to which a success case label is assigned and become away from play data to which a failure case label is assigned. For example, the learning unit 132 performs a machine learning process based on the input data 121 as described above (step S102). Meanwhile, the processing at step S101 and the processing at step S102 do not need to be necessarily consecutive.
Thus, the learning apparatus 100 has the learning unit 132. With such a configuration, the learning unit 132 can perform a machine learning process based on the input data 121 including play data having a specific attribute to be a learning target and play data having an attribute different from the learning target. That is to say, the learning unit 132 can perform a machine learning process using both play data to which a label indicating a learning target is assigned and play data to which a label indicating not a learning target is assigned. As a result, compared with the case of performing machine learning based on play data having a specific attribute to be a learning target alone, it is possible to perform machine learning based on more play data. Consequently, for example, even if it is difficult to collect enough play data having a specific attribute, it is possible to properly perform learning to get close to a learning target.
Meanwhile, the configuration of the learning apparatus 100 is not limited to the case illustrated in
The audio information acquiring unit 134 acquires audio information representing the voice of a specific person. Then, the audio information acquiring unit 134 stores the acquired audio information as audio information 124 into the storing unit 120. For example, when the acquiring unit 131 acquires play data, the audio information acquiring unit 134 acquires information representing voice having the same specific person attribute as the play data. In a case where information indicating an attribute to be a learning target, and the like, at a different timing from the play data, the audio information acquiring unit 134 may acquire the information representing voice at the timing when the information indicating an attribute to be a learning target, and the like, is acquired.
In a case where the audio information 124 is included in the storing unit 120, when outputting a game player model that is the result of learning by the learning unit 132, and the like, the output unit 133 can output the audio information 124 corresponding to a learning target in addition to the game player model. Consequently, an external device or the like receiving the audio information 124 can use the result of learning by the learning unit 132 and also output audio based on the audio information 124, for example. As a result, for example, it is possible to provide a communication experience in an external device or the like as if a player is playing with a player imitated by AI.
Next, a second example embodiment of the present invention will be described with reference to
In the second example embodiment of the present disclosure, the learning system 200 is described that includes a learning apparatus 500 having the same function as the learning apparatus 100 described in the first example embodiment. As will be described later, in this example embodiment, the learning apparatus 500 performs machine learning so as to get close to play data of a specific person such as a professional like an e-sports player, a YouTuber, or an entertainer, by using the same method as the learning apparatus 100 described in the first example embodiment.
The customer device 300 is an information processing apparatus on which a player plays a game. For example, the customer device 300 may be any information processing apparatus such as a video game apparatus that runs a video game, a personal computer, and a tablet device.
When a player plays a game, the play data acquiring unit 310 acquires play data showing an action taken by the player in the game, the state of the game, and so forth. The play data acquiring unit 310 may acquire the play data at predetermined intervals, or may acquire the play data when a predetermined condition is satisfied, such as when the player takes an action. Moreover, the play data acquiring unit 310 may acquire the play data as time-series data in which states and behaviors are chained. The play data acquired by the play data acquiring unit 310 may be stored in the memory unit included by the customer device 300.
The transmitting unit 320 transmits the play data acquired by the play data acquiring unit 310 to the server apparatus 400. The transmitting unit 320 may transmit, in addition to the play data, information indicating the attribute of the player stored in advance in the customer device 300, and the like, to the server apparatus 400. For example, the transmitting unit 320 can transmit the play data and the like to the server apparatus 400 at any timing.
The use instructing unit 330 instructs the server apparatus 400 to allow use of a game player model and the like corresponding to the result of learning corresponding to a specific person attribute indicating a specific person. In other words, the use instructing unit 330 transmits a use instruction indicating that a request for transmission of model information and the like necessary for allowing use of a game player model in the customer device 300 to the server apparatus 400. For example, in response to input from a player operating the customer device 300, the use instructing unit 330 instructs the server apparatus 400 to allow use of a game player model and the like indicated by the input from the player.
The server apparatus 400 is an information processing apparatus that accumulates play data and accumulates game player models. Moreover, the server apparatus 400 accepts a learning instruction to instruct the learning apparatus 500 to perform learning corresponding to a specific person attribute indicating a specific person, and transmits a game player model or model information and the like for using a game player model to the customer device 300 in response to an instruction from the customer device 300. The server apparatus 400 may be a single information processing apparatus, or may be realized on the cloud, for example.
The communication I/F unit 410 includes a data communication circuit and the like. The communication I/F unit 410 performs data communication with an external device and the like connected via a communication line.
The storing unit 420 is a storage device such as a hard disk or a memory. The storing unit 420 stores processing information necessary for a variety of processing in the operation processing unit 430 and a program 423. The program 423 is loaded and executed by the operation processing unit 430 to realize various processing units. The program 423 is loaded in advance from an external device or a recording medium via a data input/output function such as the communication I/F unit 410, and stored in the storing unit 420. Major information stored in the storing unit 420 includes, for example, play data information 421 and created model information 422. The storing unit 420 may store information corresponding to the audio information 124 described in the first example embodiment.
The play data information 421 includes play data received from the customer device 300. For example, in the play data information 421, play data and an attribute corresponding to the play data are associated with each other and stored. The details of the play data and the attribute may be the same as in the first example embodiment.
The created model information 422 includes a game player model that is a created model having been created by performing a machine learning process in the learning apparatus 500. For example, in the created model information 422, a game player model and information indicating an attribute to be a learning target when the game player model is created are associated with each other.
The operation processing unit 430 has an arithmetic logic unit such as a CPU and a peripheral circuit thereof. The operation processing unit 430 loads the program 423 from the storing unit 420 and executes the program 423 to make the above hardware and the program 423 cooperate with each other and realize various processing units. Major processing units realized by the operation processing unit 430 include, for example, a play data receiving unit 431, a creation instruction transmitting and receiving unit 432, a created model receiving unit 433, a use instruction accepting unit 434, an output unit 435, and a charging unit 436.
The play data receiving unit 431 receives play data and information indicating an attribute from the customer device 300. Moreover, the play data receiving unit 431 stores the received information as play data information 421 into the storing unit 420.
The creation instruction transmitting and receiving unit 432 receives an instruction for creating a game player model from an external device such as the customer device 300. For example, the creation instruction transmitting and receiving unit 432 receives a game player model creation instruction in addition to a specific person attribute that is an attribute to be a learning target.
Further, when receiving the creation instruction, the creation instruction transmitting and receiving unit 432 identifies play data having a specific person attribute to be a learning target with reference to the play data information 421. Moreover, the creation instruction transmitting and receiving unit 432 identifies play data to which a failure case label is to be assigned with reference to the play data information 421. As described in the first example embodiment, play data to which a failure case label is to be assigned may be play data having an attribute contrary to that of play data to which a success case is assigned. Then, the creation instruction transmitting and receiving unit 432 transmits the identified play data and a game player model creation instruction to the learning apparatus 500.
Assignment of a success case label and a failure case label may be performed by the creation instruction transmitting and receiving unit 432, or may be performed by the learning apparatus 500. Moreover, the play data may be transmitted to the learning apparatus 500 in advance. In this case, the creation instruction transmitting and receiving unit 432 may omit the identification of the play data and the transmission process.
The created model receiving unit 433 receives a game player model that is a created model created in response to a creation instruction transmitted by the creation instruction transmitting and receiving unit 432, from the learning apparatus 500. That is to say, the created model receiving unit 433 receives, from the learning apparatus 500, a game player model created based on play data having an attribute to be a learning target and play data having an attribute different from that of the learning target. For example, the created model receiving unit 433 receives a game player model and information indicating an attribute having been a learning target at the time of creating the game player model. Moreover, the created model receiving unit 433 stores the variety of information having been received as the created model information 422 into the storing unit 420.
The use instruction accepting unit 434 accepts a use instruction from the customer device 300.
When the use instruction accepting unit 434 accepts a use instruction from the customer device 300, the output unit 435 identifies a game player model corresponding to the use instruction with reference to the created model information 422. Then, the output unit 435 transmits model information and the like necessary for using the identified game player model to the customer device 300. That is to say, the output unit 435 transmits, to the customer device 300, model information necessary for using a game player model that is a created model created based on play data having an attribute to be a learning target and play data having an attribute different from the learning target. The model information may be the game player model itself, or may be, for example, allowance information for allowing the customer device 300 to use the game player model by accessing the server apparatus 400 or the like. The allowance information may have a predetermined time limit. The output unit 435 may be configured to transmit or make available, in addition to the game player model, the audio information 124 having a matching attribute.
The charging unit 436 performs a charging process on the customer device 300 or the like.
Further, the charging unit 436 can be configured to pay a model use fee to an external device such as the customer device 300 having transmitted a game player model creation instruction in accordance with the number of availabilities of game player model. For example, the charging unit 436 may be configured to check the number of availabilities of game player model at predetermined intervals to check whether or not a model use fee is to be paid. The model use fee may vary so as to become higher as the number of uses of game player model increases, up to the upper limit of a predetermined amount, for example.
Further, the charging unit 436 can pay a model provision cost to the learning apparatus 500, for example, when the created model receiving unit 433 receives a game player model from the learning apparatus 500, or when the creation instruction transmitting and receiving unit 432 transmits a game player model creation instruction and so forth to the learning apparatus 500. The model provision fee may be, for example, a predetermined amount. Moreover, the charging unit 436 may pay the learning apparatus 500 an additional use fee corresponding to the number of availabilities of game player model, the number of game player model creation instructions, and the like. The additional use fee may vary so as to become higher as the number of availabilities of game player model, the number of game player model creation instructions, and the like increases, for example.
As shown in
The above is an example of the configuration of the server apparatus 400. The server apparatus 400 may be connected to a reinforcement learning apparatus that creates AI by performing reinforcement learning that a learner learns through his/her own actions. Moreover, the server apparatus 400 may be configured to receive play data between AIs received from the reinforcement learning apparatus as play data having a player attribute “AI”. In this case, the server apparatus 400 may be configured to always identify play data having the player attribute “AI” as play data to which a failure case is assigned.
The learning apparatus 500 has the same configuration as the learning apparatus 100 described in the first example embodiment. In the case of this example embodiment, the learning apparatus 500 mainly performs machine learning so as to get close to play data having a specific person attribute. Moreover, the learning apparatus 500 performs machine learning so as to become away from play data having a failure case label.
The above is an example of the configuration of the learning system 200. Subsequently, an example of the operation of the server apparatus 400 will be described with reference to
The creation instruction transmitting and receiving unit 432 identifies play data having a specific person attribute to be a learning target with reference to the play data information 421. Moreover, the creation instruction transmitting and receiving unit 432 identifies play data to which a failure case label is to be assigned with reference to the play data information 421. Then, the creation instruction transmitting and receiving unit 432 transmits the identified play data and the game player model creation instruction to the learning apparatus 500 (step S202). Meanwhile, the creation instruction transmitting and receiving unit 432 may be configured to transmit the gameplay model creation instruction and so forth to the learning apparatus 500 on the condition of reception of a registration fee by the charging unit 436.
The created model receiving unit 433 receives a game player model that is a created model created in response to the creation instruction transmitted by the creation instruction transmitting and receiving unit 432, from the learning apparatus 500 (step S203). For example, the created model receiving unit 433 receives the game player model, and information indicating the attribute having been the learning target at the time of creation of the game player model. Moreover, the created model receiving unit 433 stores the variety of information having been received as the created model information 422 into the storing unit 420.
Further, referring to
Thus, the server apparatus 400 is configured to provide a game player model created based on play data having a specific attribute and play data having an attribute different from the abovementioned attribute. With such a configuration, it is possible to provide a customer with a gaming experience closer to a specific individual or more natural motion.
The configuration of the learning system 200 is not limited to the case illustrated in this example embodiment. For example, in this example embodiment, a case where play data is accumulated in the server apparatus 400 has been illustrated. However, play data may be accumulated in a place other than the server apparatus 400, such as the learning apparatus 500. In this case, the server apparatus 400 may only output model information without acquiring or accumulating play data. Moreover, the customer device 300, the server apparatus 400, or the like may have a function as the learning apparatus 500. Thus, the learning system 200 may employ various modified examples that have the same function as a whole system.
Next, a third example embodiment of the present invention will be described with reference to
The gameplay operation learning apparatus 600 is an information processing apparatus that performs a machine learning process based on play data to which a label indicating whether or not to be a learning target is assigned.
Further, the gameplay operation learning apparatus 600 can realize functions as an acquiring means 621, a learning means 622 and an output means 623 shown in
The acquiring means 621 acquires play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target.
The learning means 622 generates a game player model for outputting an action of a learning target in response to input of a second play state based on the play data and the label.
The output means 623 outputs the game player model.
Thus, the gameplay operation learning apparatus 600 has the learning means 622. With such a configuration, the learning means 622 can generate a game player model for outputting an action of a learning target in response to input of a second play state based on play data and a label. That is to say, the learning means 622 can perform a machine learning process using both play data to which a label indicating a learning target is assigned and play data to which a label indicating not a learning target is assigned. As a result, the learning means 622 can perform machine learning based on more play data compared with the case of performing machine learning simply based on play data having a specific attribute to be a learning target. Consequently, even if it is difficult to collect enough play data having a specific attribute, it is possible to properly perform learning for getting close to a learning target.
The gameplay operation learning apparatus 600 described above can be realized by installation of a predetermined program into an information processing apparatus such as the gameplay operation learning apparatus 600. Specifically, a program as another aspect of the present invention is a program for causing an information processing apparatus such as the gameplay operation learning apparatus 600 to realize processes to: acquire play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; and generate a game player model for outputting an action of a learning target in response to input of a second play state based on the play data and the label.
Further, a gameplay operation learning method executed by an information processing apparatus such as the gameplay operation learning apparatus 600 described above is a method by an information processing apparatus such as the gameplay operation learning apparatus 600, including: acquiring play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; and generating a game player model for outputting an action of a learning target in response to input of a second play state based on the play data and the label.
Inventions of a program, a computer-readable recording medium on which a program is recorded and a game play operation learning method having the above configurations can achieve the object of the present invention described above because they have the same actions and effects as the game play operation learning apparatus 600.
Next, a fourth example embodiment of the present invention will be described with reference to
The game player model use provision apparatus 700 can have the same hardware configuration as the gameplay operation learning apparatus 600 described in the third example embodiment. Moreover, the game player model use provision apparatus 700 can realize functions as an accepting means 721 and an output means 722 shown in
The accepting means 721 accepts a use instruction from an external device. A use instruction is an instruction for allowing an external device to use a game player model having learned an action of a learning target in a second play state. For example, a game player model having learned in advance based on play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target.
The output means 722 outputs, in response to the use instruction accepted by the accepting means 721, model information for using the game player model indicated by the use instruction.
Thus, the game player model use provision apparatus 700 has the output means 722. With such a configuration, the output means 722 can output a game player model created by machine learning using both play data to which a label indicating a learning target is assigned and play data to which a label indicating not a learning target is assigned. As a result, it is possible to provide a customer with a game experience closer to a specific individual or attribute and closer to natural motion.
The game player model use provision apparatus 700 described above can be realized by installation of a predetermined program into an information processing apparatus such as the game player model use provision apparatus 700. Specifically, a program as another aspect of the present invention is a program for causing an information processing apparatus such as the game player model use provision apparatus 700 to realize processes to: accept a use instruction for allowing use of a game player model that has learned an action of a learning target in a second play state based on play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; and output, in response to the use instruction, model information for using the game player model.
Further, a game player model use provision method executed by an information processing apparatus such as the game player model use provision apparatus 700 described above is a method by an information processing apparatus such as the game player model use provision apparatus 700, including: accepting a use instruction for allowing use of a game player model that has learned an action of a learning target in a second play state based on play data including a first play state in a game and an action taken by a player in the first play state, and a label indicating whether or not to be a learning target; and outputting, in response to the use instruction, model information for using the game player model.
Inventions of a program, a computer-readable recording medium on which a program is recorded and a game player model use provision method that have the above configurations can achieve the object of the present invention described above because they have the same actions and effects as the game player model use provision apparatus 700.
The whole or part of the example embodiments disclosed above can also be described as the following supplementary notes. Below, the outline of a gameplay operation learning apparatus, a gameplayer model use provision apparatus and so forth in the present invention will be described. Meanwhile, the present invention is not limited to the following configurations.
A gameplay operation learning apparatus comprising:
The gameplay operation learning apparatus according to Supplementary Note 1, wherein:
The gameplay operation learning apparatus according to Supplementary Note 1 or 2, wherein:
The gameplay operation learning apparatus according to Supplementary Note 2 or 3, wherein
The gameplay operation learning apparatus according to any one of Supplementary Notes 2 to 4, wherein:
The gameplay operation learning apparatus according to any one of Supplementary Notes 2 to 5, wherein
The gameplay operation learning apparatus according to any one of Supplementary Notes 1 to 6, comprising
The gameplay operation learning apparatus according to any one of Supplementary Notes 1 to 7, wherein
A gameplay operation learning method by an information processing apparatus, the method comprising:
A non-transitory computer-readable recording medium on which a program is recorded, the program comprising instructions for causing an information processing apparatus to realize processes to:
A game player model use provision apparatus comprising:
The game player model use provision apparatus according to Supplementary Note 11, comprising
The game player model use provision apparatus according to Supplementary Note 12, comprising
The game player model use provision apparatus according to Supplementary Note 13, wherein
The game player model use provision apparatus according to any one of Supplementary Notes 11 to 14, wherein
The game player model use provision apparatus according to any one of Supplementary Notes 11 to 15, wherein
The game player model use provision apparatus according to any one of Supplementary Notes 11 to 16, wherein
The game player model use provision apparatus according to any one of Supplementary Notes 11 to 17, wherein
A game player model use provision method by an information processing apparatus, the method comprising:
A non-transitory computer-readable recording medium on which a program is recorded, the program comprising instructions for causing an information processing apparatus to realize processes to:
Although the invention of this application has been described with reference to the respective example embodiments, the invention of this application is not limited to the example embodiments described above. The configurations and details of the invention of this application can be changed in various manners that can be understood by one skilled in the art within the scope of the present invention.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2021/033373 | 9/10/2021 | WO |