PREDICTION APPARATUS, PREDICTION METHOD AND PROGRAM

Information

  • Patent Application
  • 20240220771
  • Publication Number
    20240220771
  • Date Filed
    June 10, 2021
    3 years ago
  • Date Published
    July 04, 2024
    a month ago
  • CPC
    • G06N3/042
    • G06N7/01
  • International Classifications
    • G06N3/042
    • G06N7/01
Abstract
Provided is a prediction apparatus, including a parameter estimation unit configured to estimate a parameter of an intensity function indicating a probability of which an opinion expression is performed by each user, based on information dissemination sequence information indicating a history of opinion expressions made by a plurality of users and network information indicating relationships between the plurality of users; and a prediction unit configured to predict an opinion expression of each user based on the estimated parameter.
Description
TECHNICAL FIELD

The present invention relates to a prediction apparatus, a prediction method and a program.


BACKGROUND ART

In recent years, communication via social media has become more common with the penetration of smartphones. In these social media, information is spread via social networking consisting of, for example, friends and peers. The information dissemination mechanism can be modeled using a probability model. The most representative mechanism is a Hawkes process. The Hawkes process is a type of point process. The point process is a model for describing the number of occurrences of an event in a minute interval, and the event is information transmission in a social network such as a tweet. An occurrence probability of an event at any time is modeled using a function called an “intensity function.” The Hawkes process is a point process that describes bursty information dissemination (a phenomenon in which information is disseminated at an extremely rapid pace in a short period of time).


Social media is widely used as a place where users express their opinions on various topics such as news, political issues, and products. It is known that an opinion of each user may be influenced by their friends in an opinion dissemination process. Each user updates their opinions on each topic by learning from posts of their friends. For example, Non Patent Literature 1 discloses a point process model in which the influence of friends is taken into consideration in the opinion dissemination process. This approach is to describe an intensity function using the linear differential equation, thereby describing the transition of a user's opinion influenced by friends.


CITATION LIST
Non Patent Literature





    • Non Patent Literature 1: Abir, De, et al., Learning and Forecasting Opinion Dynamics in Social Networks. (2016).





SUMMARY OF INVENTION
Technical Problem

The conventional approach described above is characterized by describing an intensity function of a point process using the linear differential equation, thereby building up an information dissemination model. However, since this approach assumes that interaction between the users has a linearity, it is not possible to implement highly accurate prediction of opinion expressions in consideration of complex interaction.


An object of the disclosed technology is to improve prediction accuracy of opinion expressions.


Solution to Problem

According to the disclosed technology, provided is a prediction apparatus, including a parameter estimation unit configured to estimate a parameter of an intensity function indicating a probability of which an opinion expression is performed by each user, based on information dissemination sequence information indicating a history of opinion expressions made by a plurality of users and network information indicating relationships between the plurality of users; and a prediction unit configured to predict an opinion expression of each user based on the estimated parameter.


Advantageous Effects of Invention

It is possible to improve accuracy of prediction for opinion expressions.





BRIEF DESCRIPTION OF DRAWINGS


FIG. 1 is a functional configuration diagram of a prediction device.



FIG. 2 is a diagram illustrating one example of information dissemination sequence information.



FIG. 3 is a flowchart illustrating one example of a flow of learning processing.



FIG. 4 is a flowchart illustrating one example of a flow of prediction processing.



FIG. 5 is a diagram illustrating a hardware configuration example of a computer.





DESCRIPTION OF EMBODIMENTS

Hereinafter, an embodiment (present embodiment) of the present invention will be described with reference to the drawings. The embodiment described below is merely an illustrative example, and the embodiment to which the present invention is applied is not limited to the following embodiment.


Outline of Present Embodiment

Similarly to the related art, a prediction device (prediction apparatus) according to the present embodiment describes a temporal change of interaction between users by a differential equation, and describes a temporal evolution of the differential equation by a graph neural network (GNN). Specifically, the prediction device captures the social network as a graph with a user as a vertex and a relationship between users as a side, and obtains a latent vector of a node using the GNN. A temporal change of the latent vector thus obtained is described using a differential equation. By modeling an intensity function of a point process using this differential equation, a dissemination process of information (opinion) is modeled, and transmission of near-future information (opinion) is predicted.


(Exemplified Functional Configuration of Prediction Device)


FIG. 1 is a functional configuration diagram of the prediction device (prediction apparatus). A prediction device 10 (prediction apparatus) includes an operation unit 3, a parameter estimation unit 4, a parameter storage unit 5, a prediction unit 6, and an output unit 7.


The operation unit 3 accepts various operations on data of an information dissemination sequence storage device 1 and a network information storage device 2. The various operations include registration, correction, and deletion of the stored information. An input device of the operation unit 3 is, for example, a keyboard, a mouse, or a touchscreen. The operation unit 3 is implemented by, for example, a device driver of an input device such as a mouse or control software of a menu screen.


The information dissemination sequence storage device 1 stores history information of information dissemination that can be analyzed by the device, reads the information dissemination sequence information according to a request from the device, and transmits the information to the information dissemination prediction device. The information dissemination sequence is, for example, data indicating an occurrence history of crimes, a history of financial transactions, and an occurrence history of demonstrations or strikes, and is represented by a time series. The information dissemination sequence to be analyzed is represented as follows.











(


u
i

,

y
i

,

t
i


)

}


i
=
1

I




[

Eq
.

1

]







Here, ui denotes a user, yi denotes an opinion on a specific topic, ti denotes a time, and I denotes the number of items of data. The number of users is denoted by U. It is assumed that a binary value yi∈{0,1} is given as the opinion on the specific topic. For example, during the US presidential election, a dataset is generated by collecting political tweets and extracting posts including a Republican-related tag as yi=0 and posts including a Democrat-related tag as yi=1. The type of opinion is Y=2 herein. The information dissemination sequence storage device 1 is, for example, a web server that holds web pages, or a database server including a database.


The network information storage device 2 stores network information that can be analyzed by the device, reads the network information according to a request from the device, and transmits the information to the information dissemination prediction device. The network information is an adjacency matrix of a social network consisting of, for example, friends. The social network is represented by a graph G=(V, E) with a user as a vertex and a relationship between users as a side. V is a set of vertices (users) and


E is a set of sides. The network information is an adjacency matrix of the graph G, and represented as follows.









G




U
×
U






[

Eq
.

2

]







Here, for example, a binary value representing a friendship (following-follower relationship) on the social media. In a case where there is any relationship (edges) between users (nodes) i and j, an element Gi,j in the i-row j-column of a matrix G is set to 1. In a case where there is no relationship (edge) between the users (nodes) i and j, Gi,j is set to 0.


The parameter estimation unit 4 learns a parameter of an intensity function representing an occurrence probability of an event on the basis of information stored in the information dissemination sequence storage device 1 and the network information storage device 2. The event is an opinion expression in the social network such as a tweet.


The parameter storage unit 5 stores a set of optimal parameters obtained by the parameter estimation unit 4. The device used for storage may be any device that can store and restore the estimated set of parameters. For example, the set of parameters is stored in a database or a specific area of a general-purpose storage device (memory or hard disk device) provided in advance.


The prediction unit 6 performs simulation of a point process on the basis of the estimation result of the parameter estimation unit 4 and calculates a probability of opinion expressions of each user. There are a plurality of methods for simulating the point process; the prediction unit 6 herein can apply, for example, a method called “thinning” (Reference Literature [1]).


The output unit 7 outputs the result obtained by the prediction unit 6. The concept “output” includes displaying data on a display device, printing by a printer, sound output, and transmission to an external device. The output unit 7 may or may not include an output device such as a display or a speaker. The output unit 7 is implemented by, for example, driver software of the output device.


(Specific Example of Information Dissemination Sequence Information)


FIG. 2 is a diagram illustrating one example of the information dissemination sequence information. Information dissemination sequence information 101 includes “user”, “opinion”, and “time” as items.


A value of the item “user” is an identifier for identifying each user. A value of the item “opinion” is a binary value indicating a posted opinion. A value of the item “time” is a value indicating a time when the opinion was posted.


(Exemplified Operation of the Prediction Device) Exemplified operation of the prediction device 10 will be described with reference to the drawings. The prediction device 10 initiates the learning processing in response to a user operation or periodically.



FIG. 3 is a flowchart illustrating one example of a flow of the learning processing. The prediction device 10 acquires an information dissemination sequence (step S11). The parameter estimation unit 4 estimates a parameter (step S12). Specifically, the parameter estimation unit 4 estimates a parameter of an intensity function designed according to procedures a general point process model. The intensity function is a function representing a probability that information is transmitted per unit time. For example, a probability that a user u expresses an opinion y at a time t is represented by Equation (1) using an intensity function λu(t,y).









[

Eq
.

3

]











λ
u

(

t
,
y

)

=



P
u
*

(

y

t

)




λ
u
*

(
t
)






(
1
)







Here, t is a time and y is an opinion. For better understanding, an example in which the intensity function is decomposed into a term P*u(y|t) depending on the opinion y and a term λ*u(t) depending only on the time t has been described. However, the intensity function may be represented by another equation. In addition, λ*u(t) is modeled using the intensity function of the Hawkes process as shown in Equation (2).









[

Eq
.

4

]











λ
u
*

(
t
)

=


μ
u

+





j
:

t
j


<
t




α

u
,

u
j





k

(

t
-

t
j


)








(
2
)







Here, μu is referred to as a “background rate” and represents an occurrence probability of an event independent of a past event. The constant μu that does not change over time is used for better understanding, but μu may vary depending on time. k(·) is a trigger function used in the conventional Hawkes process model, and is modeled using, for example, an exponential attenuation function, a Weibull distribution, or a gamma distribution. A different background rate is assumed for each user u. P*u(y|t) is a function representing a temporal change of the opinion the user u shows, and is modeled using a differential equation. Specifically, the latent state of the differential equation is represented as follows.










X

(
t
)





U
×
d






[

Eq
.

5

]







A time evolution is described as Equation (3).









[

Eq
.

6

]












dX

(
t
)

dt

=

f

(


X

(
t
)

,
G
,
W

)


,


X

(
0
)

=

X
0






(
3
)







Here, d is the number of dimensions of the latent state, and f(·) is a function that describes the temporal evolution of the differential equation. X0 is an initial value of a latent state X(t), and W is a parameter of the GNN.









G




U
×
U






[

Eq
.

7

]









    • is network information and represents, for example, a friendship between users.





In order to model a complicated opinion transition due to the influence of other users, f(·) is described using the graph neural network (GNN). In a case where the opinion y is represented by a binary value, the probability P*u(y=k) that the user u has the opinion y=k at the time t is described as shown in Equation (4) using a softmax function.












[
0032
]






[

Eq
.

8

]














P
u
*

(

y
=

k


X

(
t
)



)

=

softmax
(




X

(
u
)


(
t
)



W
k


+

b
k


)





(
4
)









    • where














X

(
u
)


(
t
)




d





[

Eq
.

9

]









    • is a vector obtained by extracting a component related to the user u from the latent state X(t).














W
k




d


,


b
k








[

Eq
.

10

]









    • is a parameter of the softmax function for each opinion k.





When the information dissemination sequence up to a time T is given, the likelihood of this model is represented by Equation (5).









[

Eq
.

11

]










=





i
=
1

I






u
=
1

U



[


log




P
u
*

(


y

i
+
1




X

(

t
i

)


)


+

log




λ
u
*

(

t

i
+
1


)


-




t
i


t

i
+
1






λ
u
*

(
τ
)


d

τ



]



=




i
=
1

I






u
=
1

U




[


log




P
u
*

(


y

i
+
1




X

(

t
i

)


)


+

log
(


μ
u

+





j
:

t
j


<

t

i
+
1






α

u
,

u
j





k

(


t

i
+
1


-

t
j


)




)

-


μ
u

(


t

i
+
1


-

t
i


)

-


α

u
,

u

i
+
1




(


K

(


t

i
+
1


-

t
i


)

-

K

(
0
)


)


]

.








(
5
)







K(·) is an integral of the trigger function k(·), and an analytical solution is obtained for several trigger functions k(·) such as an exponential decay function, a Weibull distribution, and a gamma distribution.


The parameter estimation unit 4 estimates the following parameters: a parameter W of f(·) which is GNN that minimizes a likelihood L, an initial value X0 of the latent state X(t), a parameter {Wk,bk} of P*u(·), a parameter of the kernel function k(·), and a parameter p of the intensity function={μ1, . . . , μu}.









A
=


(

α

i
,
j


)





U
×
U







[

Eq
.

12

]







The parameter estimation unit 4 may use any method for optimizing the parameter. Since the likelihood of Equation (5) can be differentiated for all the parameters, the parameter estimation unit 4 can optimize the likelihood by using a steepest descent method (gradient method) such as error back propagation.


The parameter storage unit 5 stores the estimated parameters (step S13).


Furthermore, the prediction device 10 initiates prediction processing in response to, for example, a user operation.



FIG. 4 is a flowchart illustrating one example of a flow of the prediction processing. The prediction device 10 acquires the information dissemination sequence (step S21). The prediction unit 6 reads the parameters estimated by the learning process (step S22). Subsequently, the prediction unit 6 performs a point process simulation using the read parameters and predicts transmission of information in the near future (step S23). The output unit outputs the prediction result (step S24).


(Hardware Configuration Example According to Present Embodiment)

The prediction device 10 can be implemented, for example, by causing a computer to execute a program describing processing content described in the present embodiment. Note that the “computer” may be a physical machine, or may be a virtual machine in a cloud. In a case where a virtual machine is used, “hardware” described herein is virtual hardware.


The program is recorded in a computer-readable recording medium (such as a portable memory) so that the program can be stored and distributed. In addition, the program can also be provided through a network such as the Internet or an electronic mail.



FIG. 5 is a diagram illustrating a hardware configuration example of the computer. The computer in FIG. 5 includes a drive device 1000, an auxiliary storage device 1002, a memory device 1003, a CPU 1004, an interface device 1005, a display device 1006, an input device 1007, and an output device 1008, which are connected to each other by a bus B.


The program for performing processing in the computer is provided through a recording medium 1001 such as a CD-ROM or a memory card, for example. When the recording medium 1001 storing the program is set in the drive device 1000, the program is installed on the auxiliary storage device 1002 from the recording medium 1001 via the drive device 1000. However, the program is not necessarily installed from the recording medium 1001, and may be downloaded from another computer via a network. The auxiliary storage device 1002 stores the installed program, and also stores necessary files and data.


In a case where an instruction to start the program is issued, the memory device 1003 reads and stores the program from the auxiliary storage device 1002. The CPU 1004 achieves the functions related to the device, according to the program stored in the memory device 1003. The interface device 1005 is used as an interface for connecting to a network. The display device 1006 displays a graphical user interface (GUI) or the like by the program. The input device 1007 includes a keyboard and mouse, buttons, a touch panel, or the like, and is used to input various operation instructions. The output device 1008 outputs operation results. Note that the computer may include a graphics processing unit (GPU) or a tensor processing unit (TPU) instead of the CPU 1004, and may include a GPU or a TPU in addition to the CPU 1004. In that case, for example, the processing may be shared and executed such that the GPU or TPU executes processing requiring special operation such as a neural network and the CPU 1004 executes other processing.


Effects of the Present Embodiment

According to the prediction device 10 of the present embodiment, it is possible to learn a complicated influence between users by modeling an opinion transition using GNN. Therefore, it is possible to improve accuracy of prediction for opinion expressions.


REFERENCE LITERATURE



  • [1]:OGATA, Yosihiko. On Lewis' simulation method for point processes. IEEE Transactions on Information Theory, 1981, 27.1: 23-31.



Summary of Embodiment

In the present specification, at least the prediction device (prediction apparatus), the prediction estimation method, and the program described in the following items are described.


(Item 1)

A prediction apparatus, comprising:

    • a parameter estimation unit configured to estimate a parameter of an intensity function indicating a probability of which an opinion expression is performed by each user, based on information dissemination sequence information indicating a history of opinion expressions made by a plurality of users and network information indicating relationships between the plurality of users; and
    • a prediction unit configured to predict an opinion expression of each user based on the estimated parameter.


(Item 2)

The prediction apparatus as set forth in Item 1, wherein the network information is an adjacency matrix of graph neural networks (GNNs), which represents a temporal evolution of a differential equation representing temporal change in interaction between users.


(Item 3)

The prediction apparatus as set forth in Item 1 or 2, wherein the parameter estimation unit is configured to estimate a parameter that minimizes a likelihood of which the opinion expression is performed by means of gradient method.


(Item 4)

The prediction apparatus as set forth in any one of Items 1 to 3, wherein the prediction unit is configured to calculate a probability of which the opinion expressions by simulation of a point process based on the estimated parameter.


(Item 5)

A prediction method executed by a prediction apparatus, the method comprising steps of:

    • estimating a parameter of an intensity function indicating a probability of which an opinion expression is performed by each user, based on information dissemination sequence information indicating a history of opinion expressions made by a plurality of users and network information indicating relationships between the plurality of users; and
    • predicting an opinion expression of each user based on the estimated parameter.


(Item 6)

A program for causing a computer to function as each unit in the prediction apparatus as set forth in any one of Items 1 to 4.


Although the present embodiment has been described above, the present invention is not limited to such a particular embodiment, and various modifications and changes can be made within the scope of the gist of the present invention described in accompanying claims.


REFERENCE SIGNS LIST






    • 1 Information dissemination sequence storage device


    • 2 Network information storage unit


    • 3 Operation unit


    • 4 Parameter estimation unit


    • 5 Parameter storage unit


    • 6 Prediction unit


    • 7 Output unit


    • 10 Prediction device


    • 1000 Drive device


    • 1001 Recording medium


    • 1002 Auxiliary storage device


    • 1003 Memory device


    • 1004 CPU


    • 1005 Interface device


    • 1006 Display device


    • 1007 input device


    • 1008 Output device




Claims
  • 1. A prediction apparatus, comprising: a processor; anda memory that includes instructions, which when executed, cause the processor to execute:estimating a parameter of an intensity function indicating a probability of which an opinion expression is performed by each user, based on information dissemination sequence information indicating a history of opinion expressions made by a plurality of users and network information indicating relationships between the plurality of users; andpredicting an opinion expression of each user based on the estimated parameter.
  • 2. The prediction apparatus according to claim 1, wherein the network information is an adjacency matrix of graph neural networks (GNNs), which represents a temporal evolution of a differential equation representing temporal change in interaction between users.
  • 3. The prediction apparatus according to claim 1, wherein the estimating includes estimating a parameter that minimizes a likelihood of which the opinion expression is performed by means of gradient method.
  • 4. The prediction apparatus according to claim 1, wherein the predicting includes calculating a probability of the opinion expression by simulation of a point process based on the estimated parameter.
  • 5. A prediction method executed by a computer in a prediction apparatus, the method comprising: estimating a parameter of an intensity function indicating a probability of which an opinion expression is performed by each user, based on information dissemination sequence information indicating a history of opinion expressions made by a plurality of users and network information indicating relationships between the plurality of users; andpredicting an opinion expression of each user based on the estimated parameter.
  • 6. A non-transitory computer-readable recording medium having computer-readable instructions stored thereon, which when executed, cause a computer including a memory and a processor to execute the estimating and the predicting of the prediction apparatus according to claim 1.
PCT Information
Filing Document Filing Date Country Kind
PCT/JP2021/022203 6/10/2021 WO