This patent application claims the benefit and priority of Chinese Patent Application No. 202211602476.3 filed with the China National Intellectual Property Administration on Dec. 11, 2022, the disclosure of which is incorporated by reference herein in its entirety as part of the present application.
The present disclosure belongs to application in the field of deep learning and graph neural networks, and particularly relates to a multi-view hyperbolic-hyperbolic graph representation learning method.
A graph neural network processes graph data through deep learning, and is widely used in natural language processing, recommendation systems, biomedical treatment, etc. Although existing study on the graph neural network has achieved good results, graph information is learnt only through single-view information in most cases. For a given downstream task, topology of an underlying graph is unknown beforehand, and describing a relation between nodes only with a single view inevitably results in information loss to some extent. Thus, how to use multi-view information to learn an effective representation of the node is to be studied.
Existing study on multi-view graph representation learning is limited to an Euclidean space. An existing hyperbolic graph neural network excessively relies on a tangent space for neighborhood aggregation. However, the tangent space only locally approximates points in the hyperbolic space, and does not strictly follow mathematical meaning of hyperbolic geometry. As a result, a graph structure and properties in the hyperbolic space cannot be well preserved. A number of graphs in real world, such as protein interaction networks and social networks, tend to exhibit scale-free or hierarchical structures. Embedding such graphs into the Euclidean space results in distortion to a large extent, and thus it is difficult to express hierarchical information of networks. In contrast, the hyperbolic geometry has natural advantages in capturing this hierarchical structure. Therefore, constructing a graph convolution neural network under a multi-view structure is proposed.
In view of the above problem, the present disclosure provides a multi-view hyperbolic-hyperbolic graph representation learning method, which solves a problem of high distortion caused by the Euclidean space, and learns multi-view information. A main objective of the present disclosure is to fully explore graph information of different views and learn a more accurate node representation using a multi-view structure; completely establish basic operations of a graph in the hyperbolic space by using a hyperbolic-hyperbolic graph neural network based on characteristics of a hyperbolic structure, and transmit information in a form of low distortion to obtain a better node representation for node classification and link prediction tasks.
The present disclosure is implemented by the following technical solutions.
A multi-view hyperbolic-hyperbolic graph representation learning method is provided, which includes:
where D is a degree matrix of the graph, A is the adjacency matrix of the graph, α is a parameter, and In is a n-order identity matrix; and
where xi and xj are eigenvectors of the node i and the node j respectively; and
where indicates the Lorentz model, E indicates the Euclidean space, xE∈
n×d is an Euclidean feature of a node, and x
∈
n×(d+1) is a hyperbolic feature of a node;
where W is a learnable transformation matrix, Ŵ is an orthogonal submatrix, I is an identity matrix, and is a hyperbolic embedding representation of node i in layer l;
where is the Klein model,
→
and
→
are identical transformations between the Lorentz model and the Klein model, and hil,
is hyperbolic embedding of node i under the Lorentz model after neighbor aggregation; and
where →
and
→
are identical transformations between the Lorentz model and the Poincare model; and
where k,
is a graph embedding representation of view k,
is an importance score of a node, di is a degree of node i, and hik,is a node representation of node i on view k;
where cat indicates a concatenating operation, v indicates a view number, and v,
indicates a hyperbolic graph representation of the view v;
where s indicates an attention score vector obtained by the MLP layer of the Lorentz model, and f1 and f2 indicate the two linear layers; and
where sk is the attention score of view k, hjk,is the hyperbolic node embedding of view k on a jth dimension, and cj
is the hyperbolic node embedding after attention weighting on a jth dimension.
Compared with the prior art, the present disclosure has the beneficial effects:
The method provided in the present disclosure is a multi-view hyperbolic-hyperbolic graph representation learning method. In the present disclosure, multi-view learning is provided, such that the node representations are more accurate, and the problem of representation capability limitation caused by information difference between a single-view network and a target adjacency matrix is solved. Moreover, with geometric characteristics of hyperbolic geometry, low distortion embedding of the node is achieved under the condition that all graph operations are carried out in the hyperbolic space, and the deviation caused by the existing hyperbolic model operation depending on tangent space is solved. The present disclosure is also widely applied to downstream tasks, is suitable for various network structures and scenes, can be applied to link prediction tasks of a community network, a citation network, a recommendation network, etc., and can also be applied to node classification and graph classification tasks of a protein structure graph, a molecular graph, etc., such that data analysis and information mining are carried out, and the present disclosure has great significance for graph machine learning and actual business.
A multi-view hyperbolic-hyperbolic graph representation learning method of the present disclosure will be further described in detail below with reference to the accompanying drawings.
As shown in
The multi-view hyperbolic-hyperbolic graph representation learning method specifically includes as follows.
A multi-view construction module (step 101): graph data can include topological information and node feature information in a graph, and compared with an ideal structure of an underlying network, there can be some deviation. By means of multiple topological structures with different perspectives, the deviation can be naturally reduced, and more accurate node representations can be learned. Therefore, on the basis of an adjacency matrix, views are constructed based on a topological structure and node features respectively. For the topological structure, a diffusion matrix is constructed by using an adjacency matrix-based diffusion method, so as to reflect a global structure of a network. For the node features, a probability of a connecting edge of two nodes on their feature similarity is measured by using a cosine similarity method, and a connecting edge is constructed for two nodes based on a certain threshold.
A node feature mapping module (step 102): for graph structure data with a scale-free property, representation capability of Euclidean space is extremely limited. High distortion is produced when the graph is embedded. However, representation capability of hyperbolic geometric space increases exponentially with a radius, and the hyperbolic geometric space is extremely suitable for embedding of such a network. Therefore, a hyperbolic graph convolution network is used to model the graph data. According to properties of hyperbolic geometry, the node features are mapped to the hyperbolic space through exponential mapping. A Lorentz model in hyperbolic models is used:
where xE∈n×d is an Euclidean feature of the node, and x
∈
n×(d+1) is a hyperbolic feature of the node.
A hyperbolic-hyperbolic graph convolution module (step 103): an existing hyperbolic graph operation is often performed in a tangent space, and the tangent space is only a local approximation of the hyperbolic space. In order to minimize the deviation of graph information in a propagation process of a neural network, the node representation is always embedded in the hyperbolic model, and the hyperbolic-hyperbolic graph convolution module is defined. The module is mainly divided into three parts:
and a submatrix with orthogonality is used as a linear transformation matrix:
where W is a learnable transformation matrix, Ŵ is an orthogonal submatrix, I is an identity matrix, and is a hyperbolic representation of node i in layer l;
where is the Klein model, p
→
and p
→
are identical mapping between the Lorentz model and the Klein model, and hil,
is a hyperbolic embedding of node i under the Lorentz model after neighbor aggregation; and
where →
and
→
are identical transformations between the Lorentz model and the Poincare model, and σ is an activation function Relu.
A hyperbolic attention fusion module: three views and node features are input into the hyperbolic-hyperbolic graph convolution layer, so as to obtain the hyperbolic node embeddings of different views. In order to better fuse consistency information in different views, the hyperbolic attention fusion module is defined. The module is mainly divided into four parts:
where jk,
is a graph embedding representation of view k on the jth dimension, wi=di/Σi=1N di is a node importance score, di is a degree of node i, and hi,jk,
is a node representation of node i on view k on the jth dimension.
where cat indicates a concatenating operation, v indicates a view number, and v,
indicates the hyperbolic graph representation on view v.
where s indicates an attention score vector obtained through the Lorentz MLP layer, and f1 and f2 indicate two linear layers and an activation layer.
where sk is the attention score of view k, hjk,is the hyperbolic node embedding of view k, and cj
is a hyperbolic node embedding representation after attention weighting on the jth dimension.
| Number | Date | Country | Kind |
|---|---|---|---|
| 202211602476.3 | Dec 2022 | CN | national |