Hostname: page-component-cd9895bd7-gbm5v Total loading time: 0 Render date: 2024-12-23T19:11:18.933Z Has data issue: false hasContentIssue false

Graph representation learning: a survey

Published online by Cambridge University Press:  28 May 2020

Fenxiao Chen*
Affiliation:
University of Southern California, Los Angeles, CA90089, USA
Yun-Cheng Wang
Affiliation:
University of Southern California, Los Angeles, CA90089, USA
Bin Wang
Affiliation:
University of Southern California, Los Angeles, CA90089, USA
C.-C. Jay Kuo
Affiliation:
University of Southern California, Los Angeles, CA90089, USA
*
Corresponding author: Fenxiao Chen Email: [email protected]

Abstract

Research on graph representation learning has received great attention in recent years since most data in real-world applications come in the form of graphs. High-dimensional graph data are often in irregular forms. They are more difficult to analyze than image/video/audio data defined on regular lattices. Various graph embedding techniques have been developed to convert the raw graph data into a low-dimensional vector representation while preserving the intrinsic graph properties. In this review, we first explain the graph embedding task and its challenges. Next, we review a wide range of graph embedding techniques with insights. Then, we evaluate several stat-of-the-art methods against small and large data sets and compare their performance. Finally, potential applications and future directions are presented.

Type
Original Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Author(s), 2020 published by Cambridge University Press in association with Asia Pacific Signal and Information Processing Association

I. INTRODUCTION

Research on graph representation learning has gained more and more attention in recent years since most real-world data can be represented by graphs conveniently. Examples include social networks [Reference Bourigault, Lagnier, Lamprier, Denoyer and Gallinari1], linguistic (word co-occurrence) networks [Reference Globerson, Chechik, Pereira and Tishby2], biological networks [Reference Theocharidis, Van Dongen, Enright and Freeman3], and many other multimedia domain-specific data. Graph representation allows the relational knowledge of interacting entities to be stored and accessed efficiently [Reference Angles and Gutierrez4]. Analysis of graph data can provide significant insights into community detection [Reference Gargi, Lu, Mirrokni and Yoon5], behavior analysis [Reference Roggen, Wirz, Tröster and Helbing6], and other useful applications such as node classification [Reference Bhagat, Cormode and Muthukrishnan7], link prediction [Reference Liben-Nowell and Kleinberg8], and clustering [Reference Goyal and Ferrara9]. Various graph embedding techniques have been developed to convert the raw graph data into a high-dimensional vector while preserving intrinsic graph properties. This process is also known as graph representation learning. With a learned graph representation, one can adopt machine-learning tools to perform downstream tasks conveniently.

Obtaining an accurate representation of a graph is challenging in three aspects. First, finding the optimal embedding dimension of representation [Reference Yan, Xu, Zhang and Zhang10] is not an easy task [Reference Yan, Xu, Zhang, Zhang, Yang and Lin11]. A representation of a higher dimension tends to preserve more information of the original graph at the cost of more storage requirement and computation time. A representation of a lower dimension is more resource efficient. It may reduce noise in the original graph as well. However, there is a risk of losing some critical information from the original graph. The dimension choice depends on the input graph type as well as the application domain [Reference Shaw and Jebara12]. Second, choosing the proper graph property to embed is an issue of concern if a graph has a plethora of properties. Graph characteristics can be reflected by node features, link structures,meta-data information, etc. Determining what kind of information is more useful is application dependent. For example, we focus more on the link structure in the friend recommendation application but more on the node feature in advertisement recommendation applications. For the former, the task is mainly grouping users into different categories to provide the most suitable service/goods. Third, many graph embedding methods have been developed in the past. It is desired to have some guidelines in selecting a suitable embedding method for a target application. In this paper, we will mainly focus on node prediction and vertex classification, which are widely used in real-world applications. We intend to provide an extensive survey on graph embedding methods with the following three contributions in mind:

  1. We would like to offer new comers in this field a global perspective with insightful discussion and an extensive reference list. Thus, a wide range of graph embedding techniques, including the most recent graph representation models, are reviewed.

  2. To shed light on the performance of different embedding methods, we conduct extensive performance evaluation on both small and large data sets in various application domains. To the best of our knowledge, this is the first survey paper that provides a systematic evaluation of a rich set of graph embedding methods in domain-specific applications.

  3. We provide an open-source Python library, called the Graph Representation Learning Library (GRLL), to readers. It offers a unified interface for all graph embedding methods discussed in this paper. This library covers the largest number of graph embedding techniques until now.

The rest of this paper is organized as follows. We first state the problem as well as several definitions in Section II. Then, traditional and emerging graph embedding methods are reviewed in Section III. Next, we conduct extensive performance evaluation on a large number of embedding methods against different data sets in different application domains in Section V. The application of the learned graph representation and the future research directions are discussed in Sections VI and VII, respectively. Finally, concluding remarks are given in Section VIII.

II. DEFINITION AND PRELIMINARIES

A) Notations

A graph, denoted by $G=(\mathcal {V},\mathcal {E})$, consists of vertices, $\mathcal {V} = \{v_1,v_2,\ldots ,v_{\vert \mathcal {V}\vert }\}$, and edges, $\mathcal {E} = \{e_{i,j}\}$, where an edge $e_{i,j}$ connects vertex $v_i$ to vertex $v_j$. Graphs are usually represented by an adjacency matrix or a derived vector space representation [Reference Ding, He, Zha, Gu and Simon13]. The adjacency matrix, A, of graph G contains non-negative weights associated with each edge, $a_{ij} \geq 0$. If $v_i$ and $v_j$ are not directly connected to one another, $a_{ij} = 0$. For undirected graphs, $a_{ij} = a_{ji}$ for all $1 \leq i \leq j \leq \vert \mathcal {V}\vert$.

Graph representation learning (or graph embedding) aims to map each node to a vector where the distance characteristics among nodes is preserved. Mathematically, for graph $G=(\mathcal {V},\mathcal {E})$, we would like to find a mapping:

\[ f:v_i \rightarrow x_i\in \mathbb{R}^{d},\]

where $d \ll \vert \mathcal {V}\vert$, and $X_i = \{x_1,x_2,\ldots ,x_d\}$ is the embedded (or learned) vector that captures the structural properties of vertex $v_i$.

The first-order proximity [Reference Cavallari, Zheng, Cai, Chang and Cambria14] in a network is the pairwise proximity between vertices. For example, in weighted networks, the weights of the edges are the first-order proximity between vertices. If there is no edge observed between two vertices, the first-order proximity between them is 0. If two vertices are linked by an edge with a high weight, they should be close to each other in the embedding space. This objective can be obtained by minimizing the distance between the joint probability distribution in the vector space and the empirical probability distribution of the graph. If we use the KL-divergence [Reference Goldberger, Gordon and Greenspan15] to calculate the distance, the objective function is given by:

(1)\begin{equation} O_1 = -\sum_{(i,j) \in E} w_{ij}\log p_1(\mathbf{v_i},{\mathbf{v_j})},\end{equation}

where

\[ p_1(\mathbf{v_i},\mathbf{v_j)} = \frac{1}{1+\exp(-\mathbf{v_i}^{\rm T} \cdot \mathbf{v_j})}\]

and $\mathbf {v_i} \in \mathbb {R}^{d}$ is the low-dimensional vector representation of vertex $v_i$ and $w_{ij}$ is the edge weight between node i and j. $\mathbf {v_i}$,$\mathbf {v_j}$ are the embeddings for node $v_i$ and $v_j$.

The second-order proximity [Reference Zhou, Liu, Liu, Liu and Gao16] is used to capture the two-step relationship between two vertices. Although there is no direct edge between two vertices of the second-order proximity, their representation vectors should be close in the embedded space if they share similar neighborhood structures.

The objective function of the second-order proximity can be defined as:

(2)\begin{equation} O_2 = -\sum_{(i,j) \in E}w_{ij}\log p_2(\mathbf{v_i}\vert \mathbf{v_j}),\end{equation}

where

(3)\begin{equation} p_2(v_j\vert v_i) = \frac{\exp(\mathbf{u_j^{\,'}}^{\rm T} \cdot \mathbf{u_i})}{\sum_{k = 1}^{\vert V\vert } \exp(\mathbf{u_k^{\,'}}^{\rm T} \cdot \mathbf{u_i})},\end{equation}

and $\mathbf {u_j^{\,'}} \in \mathbb {R}^{d}$ is the vector representation of vertex $v_j$ when it is treated as a specific context for vertex $v_i$.

Graph sampling is used to simplify graphs [Reference Anis, Gadde and Ortega17]. Sometimes, even if whole graph is known, we need to use sampling to obtain a smaller graph. If the graph is unknown, then sampling is regarded as a way to explore the graph. Commonly used techniques can be categorized into two types:

  1. Negative sampling [Reference Mikolov, Sutskever, Chen, Corrado and Dean18, Reference Xu, Feng, Huang and Zhao19]:

    Negative sampling is proposed as an alternative to the hierarchical computation of the softmax. Computing softmax is expensive since the optimization requires the summation over the entire set of vertices. It is computationally expensive for large-scale networks. Negative sampling is developed to address this problem. It helps distinguish the neighbors from other nodes by sampling multiple negative samples according to the noise distribution. In the training process, correct surrounding neighbors are positive examples in contrast to a set of sampled negative examples (usually noise).

  2. Edge sampling [Reference Leskovec and Faloutsos20, Reference Ribeiro and Towsley21]:

    In the training stage, it is difficult to choose an appropriate learning rate in graph optimization when the difference between edge weights is large. To address this problem, one solution is to use edge sampling that unfolds weighted edges into several binary edges at the cost of increased memory. An alternative is treating weighted edges as binary ones with their sampling probabilities proportional to the weights. This treatment would not modify the objective function.

B) Graph embedding input

Graph embedding methods take a graph as the input, where the graph can be a homogeneous graph, a heterogeneous graph, a graph with/without auxiliary information, or a constructed graph [Reference Cai, Zheng and Chang22]. They are detailed below:

  1. Homogeneous graphs refer to graphs whose nodes and edges belong to the same type. All nodes and edges of homogeneous graphs are treated equally.

  2. Heterogeneous graphs contain different edge types to represent different relationships among different entities or categories. For example, their edges can be directed or undirected. Heterogeneous graphs typically exist in community-based question answering (cQA) sites, multimedia networks and knowledge graphs. Most social network graphs are directed graphs [Reference Tang, Qu, Wang, Zhang, Yan and Mei23]. Only the basic structural information of input graphs is provided in real-world applications.

  3. Graphs with auxiliary information [Reference Gilbert and Levchenko24, Reference Taylor, Black and Caley25] are those that have labels, attributes, node features, information propagation, etc. A label indicates node's category. Nodes with different labels should be embedded further away than those with the same label. An attribute is a discrete or continuous value that contains additional information about the graph rather than just the structural information. Information propagation indicates dynamic interaction among nodes, such as post sharing or “retweet” while Wikipedia [Reference Yang, Tang and Cohen26], DBpedia [Reference Bizer, Lehmann, Kobilarov, Auer, Becker, Cyganiak and Hellmann27], Freebase [Reference Bollacker, Evans, Paritosh, Sturge and Taylor28], etc.

  4. Graphs constructed from non-relational data are assumed to lie in a low dimensional manifold. For this kind of graph, inputs are usually represented by feature matrix, $X \in \mathbb {R}^ {{\vert V\vert } \times {N}}$, where each row $X_i$ is a N-dimensional feature vector for the ith training instance. A similarity matrix, denoted by S, can be constructed by computing the similarity between $X_i$ and $X_j$ for graph classifications. An illustration of graph input and output is shown in Fig. 1.

Fig. 1. Illustration of graph representation learning input and output.

C) Graph embedding output

The output of a graph embedding method is a set of vectors representing the input graph. Based on the need for specific application, different information or aspect of the graphs can be embedded. Graph embedding output could be node embedding, edge embedding, hybrid embedding, or whole-graph embedding. The preferred output form is application-oriented and task-driven, as elaborated below:

  1. Node embedding represents each node as a vector, which would be useful for node clustering and classification. For node embedding, nodes that are close in the graph are embedded closer together in the vector representations. Closeness can be first-order proximity, second-order proximity or other similarity calculation.

  2. Edge embedding aims to map each edge into a vector. It is useful for predicting whether a link exists between two nodes in a graph. For example, knowledge graph embedding can be used for knowledge graph entity/relation prediction.

  3. Hybrid embedding is the combination of different types of graph components such as node and edge embedding. Hybrid embedding is useful for semantic proximity search and sub-graphs learning. It can also be used for graph classification based on graph kernels. Substructure or community embedding can also be done by aggregating individual node and edge embedding inside it. Sometimes, better node embedding is learned by incorporating hybrid embedding methods.

  4. Whole graph embedding is usually done for small graphs such as proteins and molecules. These smaller graphs are represented as one vector, and two similar graphs are embedded to be closer. Whole-graph embedding facilitates graph classification tasks by providing a straightforward and efficient solution in computing graph similarities.

D) Overview of graph embedding ideas

The study of graph embedding can be traced back to 1900s when people questioned whether all planar graphs with n vertices have a straight line embedding in an $n_k\times n_k$ grid. This problem was solved in [Reference de Fraysseix, Pach and Pollack30] and [Reference Fáry31]. The same result for convex maps was proved in [Reference Stein32]. More analytic work on the embedding method and time/space complexity of such a method were studied in [Reference Chrobak and Payne33] and [Reference De Fraysseix, Pach and Pollack34]. However, a more general approach is needed since most real-world graphs are not planer. A large number of methods have been proposed since then.

We provide an overview of various graph embedding ideas below:

  1. Dimensionality reduction:

    In early 2000s, graph embedding is achieved by dimensionality reduction. For a graph with n nodes, each of which is of dimension D. These embedding methods aim to embed nodes into a d-dimensional vector space, where $d \ll D$. They are called classical methods and reviewed in Section A. Dimensionality reduction is not very scalable, more advanced methods are needed for graph representation learning.

  2. Random walk:

    One can trace a graph by starting random walks from random initial nodes so as to create multiple paths. These paths reveal the context of connected vertices. The randomness of these walks gives the ability to explore the graph, capture the global and local structural information by walking through neighboring vertices. Later on, probability models like skip-gram and bag-of-word are performed on the randomly sampled paths to learn node representations. The random walk based methods will be discussed in Section III, B.

  3. Matrix factorization:

    By leverage the sparsity of real-world networks, one can apply the matrix factorization technique that finds an approximation matrix for the original graph. This idea is elaborated in Section III, C.

  4. Neural networks:

    Neural network models such as convolution neural network (CNN) [Reference Krizhevsky, Sutskever and Hinton35], recursive neural networks (RNN) [Reference Mikolov, Karafiát, Burget, Černockỳ and Khudanpur36] and their variants have been widely adopted in graph embedding. This topic will be described in Section III, D.

  5. Large graphs:

    Some large graphs are difficult to embed since CNN and RNN models do not scale well with the numbers of edges and nodes. New embedding methods are designed targeting at large graphs. They become popular due to their efficiency. This topic is reviewed in Section III, E.

  6. Hypergraphs:

    Most social networks are hypergraphs. As social networks get more attention in recent years, hypergraph embedding becomes a hot topic, which will be presented in Section III, F.

  7. Attention mechanism:

    The attention mechanism can be added to existing embedding models to increase embedding accuracy, which will be examined in Section III, E.

An extensive survey on graph embedding methods will be conducted in the next section.

III. Graph embedding methods

A) Dimension-reduction-based methods

Classical graph embedding methods aim to reduce the dimension of high-dimensional graph data into a lower dimensional representation while preserving the desired properties of the original data. They can be categorized into linear and nonlinear two types. The linear methods include the following:

  1. 1 Principal component analysis (PCA) [Reference Jolliffe and Cadima37]:

    The basic assumption for PCA is that that principal components that are associated with larger variances represent the important structure information while those smaller variances represent noise. Thus, PCA computes the low-dimensional representation that maximizes the data variance. Mathematically, it first finds a linear transformation matrix $W \in \mathbb {R}^{D\times d}$ by solving

    (4)\begin{align} W =\text{argmax}\, \text{Tr} (W^T \mbox{Cov} (X)W), \quad d=1, 2, \ldots, D,\end{align}
    This equation aims to find the weight vector which extracts the maximum variance from this new data matrix, where Tr denotes the trace of a matrix, $\mbox {Cov} (X)$ denotes the co-variance of data matrix X. D is the original dimension of the data, and d is the dimension the data are reduced to. It is well known that the principal components are orthogonal and they can be solved by eigen decomposition of the co-variance of data matrix [Reference Umeyama38].
  2. 2 Linear discriminant analysis (LDA) [Reference Ye, Janardan and Li39]:

    The basic assumption for LDA is that each class is Gaussian distributed. Then, the linear projection matrix, $W \in \mathbb {R}^{D\times d}$, can be obtained by maximizing the ratio between the inter-class scatter and intra-class scatters. The maximization problem can be solved by eigen decomposition and the number of low dimension d can be obtained by detecting a prominent gap in the eigen-value spectrum.

  3. 3 Multidimensional scaling (MDS) [Reference Robinson and Bennett40]:

    MDS is a distance-preserving manifold learning method. It preserves spatial distances. MDS derives a dissimilarity matrix D, where $D^{i,j}$ represents the dissimilarity between points i and j, and produces a mapping in a lower dimension to preserve dissimilarities as much as possible.

The above-mentionedthree methods are referred to as “subspace learning” [Reference Yan, Xu, Zhang, Zhang, Yang and Lin11] under linear assumption. However, linear methods might fail if the underlying data are highly non-linear [Reference Saul, Weinberger, Ham, Sha and Lee41]. Then, non-linear dimensionality reduction (NLDR) [Reference DeMers and Cottrell42] can be used for manifold learning. The objective is to learn the nonlinear topology automatically. The NLDR methods include the following:

  1. Isometric feature mapping (Isomap) [Reference Samko, Marshall and Rosin43]:

    The Isomap finds low-dimensional representation that most accurately preserves the pairwise geodesic distances between feature vectors in all scales as measured along the sub-manifold from which they were sampled. Isomap first constructs neighborhood graph on the manifold, then it computes the shortest path between pairwise points. Finally, it constructs low-dimensional embedding by applying MDS.

  2. Locally linear embedding (LLE) [Reference Roweis and Saul44]:

    LLE preserves the local linear structure of nearby feature vectors. LLE first assigns neighbors to each data point. Then it computes the weights $W^{i,j}$ that best linearly reconstruct the features, $X_i$, from its neighbors. Finally, it compute the low-dimensional embedding that best reconstructed by $W^{i,j}$. Besides NLDR, kernel PCA is another dimensionality reduction technique that is comparable to Isomap, LLE.

  3. Kernel methods [Reference Harandi, Sanderson, Shirazi and Lovell45]:

    Kernel extension can be applied to algorithms that only need to compute the inner product of data pairs. After replacing the inner product with kernel function, data are mapped implicitly from the original input space to a higher dimensional space. Then linear algorithms are applied in the new feature space. The benefit of kernel trick is that data which are not linearly separable in the original space could be separable in new high dimensional space. Kernel PCA is often used for NLDR with polynomial or Gaussian kernels.

B) Random-walk-based methods

Random-walk-based methods sample a graph with a large number of paths by starting walks from random initial nodes. These paths indicate the context of connected vertices. The randomness of walks gives the ability to explore the graph and capture both the global and the local structural information by walking through neighboring vertices. After the paths are built, probability models such as skip-gram [Reference Guthrie, Allison, Liu, Guthrie and Wilks46] and bag-of-words [Reference Zhang, Jin and Zhou47] can be performed on these randomly sampled paths to learn the node representation:

  1. 1 DeepWalk [Reference Perozzi, Al-Rfou and Skiena48]:

    DeepWalk is the most popular random-walk-based graph embedding method. In DeepWalk, a target vertex, $v_i$, is said to belong to a sequence $S = \{ v_1, \ldots ,v_{\vert s\vert }\}$ sampled by random walks if $v_i$ can reach any vertex in S within a certain number of steps. The set of vertices, $V_s = \{ v_{i-t}, \ldots ,v_{i-1},v_{i+1}, \ldots , v_{i+t}\}$, is the context of center vertex $v_i$ with a window size of t (t is the sampling window size, which is a hyper parameter that can be tuned). DeepWalk aims to maximize the average logarithmic probability of all vertex context pairs in random walk sequence S. It can be written as:

    (5)\begin{equation} \frac{1}{\vert S\vert }\sum_{i=1}^{\vert S\vert } \sum_{-t\leq j \leq t,j\neq 0} \log p(v_{i+j}\vert v_i),\end{equation}
    where $p(v_j\vert v_i)$ is calculated using the softmax function. It is proven in [Reference Yang, Liu, Zhao, Sun and Chang49] that DeepWalk is equivalent to factorizing a matrix
    (6)\begin{equation} M = W^{\rm T} \times H,\end{equation}
    each entry in $M\in \mathbb {R}^{\vert V\vert \times \vert V\vert }$, $M_{ij}$, is the logarithm of the average probability that vertex $v_i$ can reach vertex $v_j$ in a fixed number of steps. $W\in \mathbb {R}^{k\times \vert V\vert }$ is the vertex representation. The information in $H \in \mathbb {R}^{k\times \vert V\vert }$ is rarely utilized in the classical DeepWalk model.
  2. 2 node2vec [Reference Grover and Leskovec50]:

    node2vec is a modified version of DeepWalk. In DeepWalk, sampled sequences are based onthe depth-first sampling (DFS) strategy. They consist of neighboring nodes sampled at increasing distances from the source node sequentially. However, if the contextual sequences are sampled by the DFS strategy alone, only a few vertices close to the source node will be sampled. Consequently, the local structure will be easily overlooked. In contrast with the DFS strategy, the breadth-first sampling (BFS) strategy will explore neighboring nodes with a restricted maximum distance to the source node while the global structure may be neglected. As a result, node2vec proposes a probability model in which the random walk has a certain probability, ${1}/{p}$, to revisit nodes being traveled before. Furthermore, it uses an in-out parameter q to control the ability to explore the global structure. When the return parameter p is small, the random walk may get stuck in a loop and capture the local structure only. When in-out parameter q is small, the random walk is more similar to a DFS strategy and capable of preserving the global structure in the embedding space.

C) Matrix-factorization-based methods

Matrix-factorization-based embedding methods, also called graph factorization (GF) [Reference Ahmed, Shervashidze, Narayanamurthy, Josifovski and Smola51], was the first one to achieve graph embedding in $O(\vert E\vert )$ time for node embedding tasks. To obtain the embedding, GF factorizes the adjacency matrix of a graph. It corresponds to a structure-preserving dimensionality reduction process. There are several variations as summarized below:

  1. 1 Graph Laplacian eigenmaps [Reference Belkin and Niyogi52]:

    This technique minimizes a cost function to ensure that points close to each other on the manifold are mapped close to each other in the low-dimensional space to preserve local distances.

  2. 2 Node proximity matrix factorization [Reference Singh and Gordon53]:

    This method approximates node proximity in a low-dimensional space via matrix factorization by minimizing the following objective function:

    (7)\begin{equation} \min\vert W-Y{Y^c}^{T}\vert ,\end{equation}
    where Y is the node embedding and $Y^c$ is the embedding for the context nodes. W is the node proximity matrix, which can be derived by several methods. One way to obtain W is to use equation (6).
  3. 3 Text-associated DeepWalk (TADW) [Reference Yang, Liu, Zhao, Sun and Chang49]:

    TADW is an improved DeepWalk method for text data. It incorporates the text features of vertices in network representation learning via matrix factorization. Recall that the entry, $m_{ij}$, of matrix $M \in \mathbb {R}^{\vert V\vert \times \vert V\vert }$ denotes the logarithm of the average probability that vertex $v_i$ randomly walks to vertex $v_j$. Then, TADW factorizes Y into three matrices:

    (8)\begin{equation} Y = W^{\rm T} \times H \times T,\end{equation}
    where $W\in \mathbb {R}^{k\times \vert V\vert }$, $H\in \mathbb {R}^{k\times f_t}$ and $T \in \mathbb {R}^{f_t\times \vert V\vert }$ is the text feature matrix. In TADW, W and HT are concatenated as the representation for vertices.
  4. 4 Homophily, structure, and content augmented (HSCA) network [Reference Zhang, Yin, Zhu and Zhang54]:

    The HSCA model is an improvement upon the TADW model. It uses skip-gram and hierarchical Softmax to learn a distributed word representation. The objective function for HSCA can be written as

    (9)\begin{align} &\underset{W,H}\min\left(\vert \vert M-W^T H T\vert \vert _F^2+\frac{\lambda}{2} (\vert \vert W\vert \vert _F^2 +\vert \vert H\vert \vert _F^2)\right.\nonumber\\ &\quad\left.\vphantom{\frac{\lambda}{2}}+ \mu(R_1(W)+R_2(H))\right),\end{align}
    where $\vert \vert .\vert \vert _2$ is the matrix $l_2$ norm and $\vert \vert .\vert \vert _F$ is the matrix Frobenius form. In equation (9), the first term aims to minimize the matrix factorization error of TADW. The second term imposes the low-rank constraint on W and H and uses ${\lambda }$ to control the trade-off. The last regularization term enforces the structural homophily between connected nodes in the network. The conjugate gradient (CG) [Reference Møller55] optimization technique can be used to update W and H. We may consider another regularization term to replace the third term; namely,
    (10)\begin{equation} R(W,H) = \frac{1}{4} \sum_{i=1,j = 1}^{\vert V\vert } A_{i,j} \Bigg \vert \Bigg \vert \begin{bmatrix} w_i \\ Ht_i \end{bmatrix} - \begin{bmatrix} w_j \\ Ht_j \end{bmatrix} \Bigg \vert \Bigg \vert _2^2.\end{equation}
    This term will make connected nodes close to each other in the learned network representation [Reference Chen, Wang and Kuo56].
  5. 5 GraRep [Reference Cao, Lu and Xu57]:

    GraRep aims to preserve the high order proximity of graphs in the embedding space. While the random-walk based methods have a similar objective, their probability model and objective functions used are difficult to explain how the high order proximity is preserved. GraRep derives a k-th order transition matrix, $A^k$, by multiplying the adjacency matrix to itself k times. The transition probability from vertex w to vertex c is the entry in the w-th row and c-th column of the k-th order transition matrix. Mathematically, it can be written as

    (11)\begin{equation} p_k(c\vert w) = A^k_{w,c}.\end{equation}
    With the transition probability defined in equation (11), the loss function is defined by the skip-gram model and negative sampling. To minimize the loss function, the embedding matrix can be expressed as
    (12)\begin{equation} Y^k_{i,j} = W_i^k \cdot C_j^k = \log\left(\frac{A^k_{i,j}}{\sum_{t}A^k_{t,j}}\right) - \log(\beta),\end{equation}
    where β is a constant ${\lambda }/{N}$, λ is the negative sampling parameter, and N is the number of vertices. The embedding matrix, W, can be obtained by factorizing matrix Y in (12).
  6. 6 HOPE [Reference Ou, Cui, Pei, Zhang and Zhu58]:

    HOPE preserves asymmetric transitivity in approximating the high order proximity, where asymmetric transitivity indicates a specific correlation among directed graphs. Generally speaking, if there is a directed edge from u to v, it is likely that there is a directed edge from v to u as well. Several high order proximities such as the Katz index [Reference Katz59], the Rooted Page-Rank, the Common Neighbors, and the Adamic-Adar were experimented in [Reference Ou, Cui, Pei, Zhang and Zhu58]. The embedding, $v_i$, for node i can be obtained by factorizing the proximity matrix, S, derived from these proximities. To factorize S, SVD is adopted, and only the top-k eigenvalues are chosen.

D) Neural-network-based methods

Neural network models became popular again since 2010. Being inspired by the success of RNNs and CNNs, researchers attempt to generalize and apply them to graphs. Natural language processingmodels often use the RNNs to find a vector representation for words. The Word2Vec [Reference Mikolov, Chen, Corrado and Dean60] and the skip-gram models [Reference Mikolov, Sutskever, Chen, Corrado and Dean18] aim to learn the continuous feature representation of words by optimizing a neighborhood preserving likelihood function. Following this idea, one can adopt a similar approach for graph embedding, leading to the Node2Vec method [Reference Grover and Leskovec50]. Node2Vec utilizes random walks [Reference Spitzer61] with a bias to sample the neighborhood of a target node and optimizes its representation using stochastic gradient descent. Another family of neural-network-based embedding methods adopt CNN models. The input can be paths sampled from a graph or the whole graph itself. Some use the original CNN model designed for the Euclidean domain and reformat the input graph to fit it. Others generalize the deep neural model to non-Euclidean graphs.

Several neural-network-based methods based graph embedding methods are presented below:

  1. 1 Graph convolutional network (GCN) [Reference Kipf and Welling62]:

    GCN allows end-to-end learning of the graph with arbitrary size and shape. This model uses convolution operator on the graph and iteratively aggregates the embedding of neighbors for nodes. This approach is widely used for semi-supervised learning on graph-structured data. It is based on an efficient variant of convolutional neural networks that operate directly on graphs. GCN learns hidden layer representations that encode both local graph structure and features of nodes. In the first step of the GCN, a node sequence will be selected. The neighborhood nodes will be assembled, then the neighborhood might be normalized to impose the order of the graph, then convolutional layers will be used to learn the representation of nodes and edges. The propagation rule used is:

    (13)\begin{equation} f(H^{(l)},A) = \sigma ({D}^{-({1}/{2})}\hat{A}{D}^{-({1}/{2})}H^{(l)}W^{(l)}),\end{equation}
    where A is the adjacency matrix, with enforced self-loops to include the node features of itself, $\hat {A} = A + I$, I is the identity matrix. $\hat {D}$ is the diagonal node degree matrix of $\hat {A}$. Under spectral graph theory of CNNs on graphs, GCN is equivalent to Graph Laplacian in the non-Euclidean domain [Reference Defferrard, Bresson and Vandergheynst63]. The decomposition of eigenvalues for the normalized graph Laplacian data can also be used for tasks such as classification and clustering. GCN usually only uses two convolutional layers and why it works is not well explained. One recent work showed that GCN model is a special form of Laplacian smoothing [Reference Field64, Reference Li, Han and Wu65]. This is the reason that GCN works. Using more than two convolutional layers will lead to over-smoothing, therefore making the features of nodes similar to each other and more difficult to separate from each other.
  2. 2 Signed graph convolutional network (SGCN) [Reference Derr, Ma and Tang66]:

    Most GCNs operate on unsigned graphs, however, many links in real-world have negative links. To solve this problem, signed GCNs aims to learn graph representation with the additional signed link information. Negative links usually contain semantic information that is different from positive links, also the principles are inherently different from positive links. The signed network will have a different representation as $G = (\mathcal {V}, \mathcal {E^+}, \mathcal {E^-})$, where the signs of the edges are differentiated. The aggregation for positive and negative links are different. Each layer will have two representations, one for the balanced user where the number of negative links is even. One for the unbalanced user where the number of negative links is odd. The hidden states are:

    (14)\begin{align} h_i^{B(1)}&= \sigma \left(W^{B(l)}\left[\sum_{j \in N_i^+}\frac{h_j^{(0)}}{\vert N_i^+\vert },h_i^{(0)}\right]\right),\end{align}
    (15)\begin{align} h_i^{U(1)} &= \sigma \left(W^{U(l)}\left[\sum_{j \in N_i^-}\frac{h_j^{(0)}}{\vert N_i^-\vert },h_i^{(0)}\right]\right),\end{align}
    where σ is the non-linear activation function, $W^{B(l)}$ and $W^{U(l)}$ are the linear transformation matrices for balanced and unbalanced sets.
  3. 3 Variational graph auto-encoders (VGAE) [Reference Kingma and Welling67]:

    VGAE uses an autoencoder minimizes the reconstruction error of the input and the output using an encoder and a decoder. The encoder maps input data to a representation space. Then, it is further mapped to a reconstruction space that preserves the neighborhood information. VGAE uses GCN as the encoder and an inner product decoder to embed graphs.

  4. 4 GraphSAGE [Reference Hamilton, Ying and Leskovec68]:

    GraphSAGE uses a sample and aggregate method to conduct inductive node embedding. It uses node features such as text attributes, node profiles, etc. GraphSAGE trains a set of aggregation functions that integrate features of local neighborhood and pass it to the target node $v_i$. Then, the hidden state of node $v_i$ is updated by:

    (16)\begin{align} h_i^{(k+1)} = ReLU\left(W^{(k)} h_i^{(k)}, \sum_{n \in N(i)} (ReLU(Q^{(k)} h_n^{(k)}))\right),\end{align}
    where $h_i^{(0)} = X_i$ is the initial node attributes and $\sum (\cdot )$ denotes a certain aggregator function, e.g. average, LSTM, max-pooling, etc.
  5. 5 Structural deep network embedding (SDNE) [Reference Wang, Cui and Zhu69]:

    SDNE learns a low-dimensional network-structure-preserving representation by considering both the first-order and the second-order proximities between vertexes using CNNs. To achieve this objective, it adopts a semi-supervised model to minimize the following objective function:

    (17)\begin{align} {\vert \vert (\hat{X}-X) \odot B\vert \vert }_F^2 + \alpha \sum_{i,j = 1}^{n} s_{i,j} \vert \vert y_i-y_j\vert \vert _2^2 + v L_{reg},\end{align}
    where $L_{reg}$ is an $L_2$-norm regularizing term to avoid over-fitting, S is the adjacency matrix, and B is the bias matrix.

E) Large graph embedding methods

Some large graphs are difficult to embed using the methods mentioned previously. Classical dimension reduction-based methods cannot capture the higher order proximity of large graphs, therefore cannot generate accurate representation. Most matrix factorization-based methods cannot take in large graph all at once, or has a high run time complexity. For example, graph Laplacian eigenmaps has the time complexity of $O(\vert E\vert d^2)$, making these methods not suitable for embedding large graphs. Random-walk based methods, such as DeepWalk, need to update parameters using back propagation (BP) [Reference Hecht-Nielsen70], which is hardware demanding and time consuming. To address the scalability issue, several embedding methods targeting at large graphs have been proposed recently. They are examined in this subsection.

  1. LGCL [Reference Gao, Wang and Ji29]:

    For each feature dimension, every node in the LGCL method selects a fixed number of features from its neighboring nodes with value ranking. Figure 2 serves as an example. Each node in this figure has a feature vector of dimension n=3. For the target node (in orange), the first feature component of its six neighbors takes the values of 9, 6, 5, 3, 0, and 0. If we set the window size to k=4, then the four largest values (i.e. 9, 6, 5, and 3) are selected. The same process is repeated for the two remaining features. By including the feature vector of the target node itself, we obtain a data matrix of dimension $(k + 1) \times n$. This results in a grid-like structure. Then, the traditional CNN can be conveniently applied so as to generate the final feature vector. To embed large-scale graphs, a sub-graph selection method is used to reduce the memory and resource requirements. As shown in Fig. 3, it begins with $N_{init} = 3$ randomly sampled nodes (in red) that are located in the center of the figure. At the first iteration, the BFS is used to find all first-order neighboring nodes of initial nodes. Among them, $N_m = 5$ nodes (in blue) are randomly selected. At the next iteration, $N_m = 7$ nodes (in green) are randomly selected. After two iterations, 15 nodes are selected as a sub-graph that serves as the input to LGCL.

  2. Graph partition neural networks (GPNN) [Reference Motsinger, Lee, Mellick and Ritchie71]:

    GPNN extends graph neural networks (GNNs) to embed extremely large graphs. It alternates between local (propagate information among nodes) and global information propagation (messages among sub-graphs). This scheduling method can avoid deep computational graphs required by sequential schedules. The graph partition is done using a multi-seed flood fill algorithm, where nodes with large out-degrees are sampled randomly as seeds. The sub-graphs grow from seeds using flood fill, which reaches out unassigned nodes that are direct neighbors of the current sub-graph.

  3. LINE [Reference Tang, Qu, Wang, Zhang, Yan and Mei23]:

    LINE is used to embed graphs of an arbitrary type such as undirected, directed, and weighted graphs. It utilizes negative sampling to reduce optimization complexity. This is especially useful in embedding networks containing millions of nodes and billions of edges. It is trained to preserve the first- and second-order proximities, separately. Then, the two embeddings are merged to generate a vector space to better represent the input graph. One way to merge two embeddings is to concatenate embedding vectors trained by two different objective functions at each vertex.

Fig. 2. Illustration of a LGCL method [Reference Gao, Wang and Ji29].

Fig. 3. Illustration of the sub-graph selection process [Reference Gao, Wang and Ji29].

F) Hyper-graph embedding

As research on social network embedding proliferates, a simple graph is not powerful enough to represent the information in social networks. The relationship of vertices in social networks is far more complicated than the vertex-to-vertex edge relationship. Different from traditional graphs, edges in hyper-graphs may have a degree larger than two. All related nodes are connected by a hyper-edge to form a super-node. Mathematically, an unweighted hyper-graph is defined as follows. A hyper-graph, denoted by $G=(\mathcal {V},\mathcal {E})$, consists of a vertex set $\mathcal {V} = \{v_1,v_2,\ldots ,v_n\}$, and a hyper-edge set, $\mathcal {E} =\{e_1,e_2,\ldots ,e_m\}$. A hyper-edge, e, is said to be incident with a vertex v if $v \in e$. When $v \in e$, the incidence function $h(v,e) = 1$. Otherwise, $h(v, e) = 0$. The degree of a vertex v is defined as $d(v) = \sum _{e \in E , v \in e} h(v, e).$ Similarly, the degree of a hyper-edge e is defined as $d(e) = \sum _{v \in V} h(v, e).$ A hyper-graph can be represented by an incidence matrix H of dimension $\vert \mathcal {V}\vert \times \vert \mathcal {E}\vert$ with entries $h(v, e)$.

Hyper-edges possess the properties of edges and nodes at the same time. As an edge, hyper-edges connect multiple nodes that are closely related. A hyper-edge can also be seen as a super-node. For each pair of two super-nodes, their connection is established by shared incident vertices. As a result, hyper-graphs can better indicate the community structure in the network data. These unique characteristics of hyper-edges make hyper-graphs more challenging. An illustration of graph and hyper-graph structures is given in Fig. 4. It shows how to express a hyper-graph in table form. The hyper-edges, which are indecomposable [Reference Tu, Cui, Wang, Wang and Zhu72], can express the community structure of networks. Furthermore, properties of graphs and hyper-graphs are summarized and compared in Table 1. Graphs and hyper-graphs conversion techniques have been developed. Examples include clique expansion and star expansion. Due to the indecomposibility of hyper-edges, conversion from a hyper-graph to a graph will result in information loss.

Fig. 4. The architecture of HGNN [Reference Feng, You, Zhang, Ji and Gao73].

Table 1. Comparison of properties of graphs and hyper-graphs

Hyper-graph representation learning provides a good tool for social network modeling, and it has been a hot research topic nowadays. On the one hand, hyper-graph modeling can be used for many applications that are difficult to achieve using other methods. For example, multi-modal data can be better represented using hyper-graphs than traditional graph representation. On the other hand, hyper-graphs can be viewed as a variant of simple graphs. Many graph embedding methods could be applied onto the hyper-graphs with minor modifications. There are embedding methods proposed for simple graphs and they can be applied to hypergraphs as well as reviewed below:

  1. 1 Spectral hyper-graph embedding [Reference Zhou, Huang and Schölkopf74]:

    Hyper-graph embedding can be treated as a k-way partitioning problem and solved by optimizing a combinatorial function. It can be further converted to a real-valued minimization problem by normalizing the hyper-graph Laplacian. Its solution is any lower dimension embedding space spanned by orthogonal eigen vectors of the hyper-graph Laplacian, Δ, with the k smallest eigenvalues.

  2. 2 Hyper-graph neural network (HGNN) [Reference Feng, You, Zhang, Ji and Gao73]:

    Inspired by the spectral convolution on graphs in GCN [Reference Kipf and Welling62], HGNN applies the spectral convolution to hyper-graphs. By training the network through a semi-supervised node classification task, one can obtain the node representation at the end of convolutional layers. The architecture is depicted in Fig. 5. The hyper-graph convolution is derived from the hyper-graph Laplacian, Δ, which is a positive semi-definite matrix. Its eigen vectors provide certain basic functions while its associated eigenvalues are the corresponding frequencies. The spectral convolution in each layer is carried out via:

    (18)\begin{equation} f(X, W, \Theta) = \sigma(D_v^{-1/2} H W D_e^{-1} H^T D_v^{{-1}/{2}} X \Theta),\end{equation}
    where X is the hidden embedding in each layer, Θ is the filter response, and $D_v$ and $D_e$ are diagonal matrices with entries being the degree of the vertices and the hyper-edges, respectively.
  3. 3 Deep hyper-network embedding (DHNE) [Reference Tu, Cui, Wang, Wang and Zhu72]:

    DHNE aims to preserve the structural information of hyper-edges with a deep neural auto-encoder. The auto-encoder first embed each vertex to a vector in a lower dimensional latent space and then reconstruct it to the original incidence vector. In the process of encoding and decoding, the second-order proximity is preserved to learn global structural information. The first order proximity is preserved in the embedding space by defining an N-tuple-wise similarity function. That is, if N nodes are in the same hyper-edge, the similarity of these nodes in the embedding space should be high. Based on similarity, one can predict whether N nodes are connected by a single hyper-edge. However, the N-tuple-wise similarity function should be non-linear; otherwise, it will lead to contradicted predictions. The local information of a hyper-graph can be preserved by shortening the distance of connected vertices in the embedding space.

Fig. 5. Illustration of graph and hypergraph structures [Reference Feng, You, Zhang, Ji and Gao73].

G) Attention graph embedding

Attention mechanisms can be used to allow the learning process to focus on parts of a graph that are more relevant to a specific task. One advantage of applying attention to graphs is to avoid the noisy part of a graph so as to increase the signal-to-noiseratio [Reference Lee, Rossi, Kim, Ahmed and Koh75] in information processing. Attention-based node embedding aims to assign an attention weight, $\alpha _i \in [0,1]$, to the neighborhood nodes of a target node t, where $\sum _{i \in N(t)} \alpha _i = 1$ and $N(t)$ denotes the set of neighboring nodes of t.

  1. 1 Graph attention networks (GAT) [Reference Veličković, Cucurull, Casanova, Romero, Lio and Bengio76]:

    GAT utilizes masked self-attentional layers to limit the shortcomings of prior graph convolutional based methods. They aim to compute the attention coefficients

    (19)\begin{equation} \alpha _{ij} = \frac{\exp(LeakyReLU(\overrightarrow{a}^T[W \overrightarrow{h_i}\vert \vert W \overrightarrow{h_j}]))}{ \sum_{k \in N_i} \exp(LeakyReLU(\overrightarrow{a}^T[W \overrightarrow{h_i}\vert \vert W \overrightarrow{h_j}]},\end{equation}
    where W is the weight matrix for the initial linear transformation, then the transformed information on each neighbor's feature are concatenated to obtain the new hidden state, which will be passed through a LeakyReLu activation function, which is one of the most commonly used rectifiers. The above attention mechanism is a single-layer feed-forward neural network parameterized by the above weight vector.
  2. 2 AttentionWalks [Reference Abu-El-Haija, Perozzi, Al-Rfou and Alemi77, Reference Abu-El-Haija, Perozzi, Al-Rfou and Alemi78]:

    Generally speaking, one can use the random walk to find the context of the node. For a graph, G, with corresponding transition matrix T and window size c the parameterized conditional expectation after a k-step walk can be expressed as:

    (20)\begin{equation} E[D\vert q_1, q_2,\ldots,q_c] = I_n \sum_{k=1}^{c} q_k T^k,\end{equation}
    where $I_n$ is the size-n identity matrix, $q_k$, $1 \leq i \leq c$, are the trainable weights, D is the walk distribution matrix whose entry $D_{uv}$ encodes the number of times node u is expected to visit node v. The trainable weights are used to steer the walk toward a broader neighborhood or restrict it within a smaller neighborhood. Following this idea, AttentionWalks adopts an attention mechanism to guide the learning procedure. This mechanism suggests which part of the data to focus on during the training process. The weight parameters are called the attention parameters in this case.
  3. 3 Attentive graph-based recursive neural network (AGRNN) [Reference Xu, Wang, Xu and Qu79]:

    AGRNN applies attention to a graph-based recursive neural network (GRNN) [Reference Xu, Wang, Xu and Qu80] to make the model focus on vertices with more relevant semantic information. It builds sub-graphs to construct recursive neural networks by sampling a number of k-step neighboring vertices. AGRNN finds a soft attention, $\alpha _r$, to control how neighbor information should be passed to the target node. Mathematically, we have:

    (21)\begin{equation} \alpha_r = \mbox{Softmax} (x^{\rm T} W^{(a)} h_r ),\end{equation}
    where $x_k$ is the input, $W^{(a)}$ is the weight to learn, and $h_r$ is the hidden state of the neighbors. The aggregated representation from all neighbors is used as the hidden state of the target vertex:
    (22)\begin{equation} h_k = \sum\nolimits_{v_r \in N(v_k)} \alpha_r h_r,\end{equation}
    where $N(v_k)$ denotes the set of neighboring nodes of vertex $v_k$. Although attention has been proven to be useful in improving some neural network models, it does not always increase the accuracy of graph embedding [Reference Chen, Wang and Kuo81].

H) Others

  1. 1 GraphGAN [Reference Wang, Wang, Wang, Zhao, Zhang, Zhang, Xie and Guo82]:

    GraphGAN employs both generative and discriminative models for graph representation learning. It adopts adversarial training and formulates the embedding problem as a mini-max game, borrowing the idea from the generative adversarial network (GAN) [Reference Goodfellow, Pouget-Abadie, Mirza, Xu, Warde-Farley, Ozair, Courville and Bengio83]. To fit the true connectivity distribution $p_{true}(v\vert v_c)$ of vertices connected to target vertex $v_c$, GraphGAN models the connectivity probability among vertices in a graph with a generator, $G(v\vert v_c;\theta _G)$, to generate vertices that are most likely connected to $v_c$. A discriminator $D(v,v_c;\theta _D)$ outputs the edge probability between v and $v_c$ to differentiate the vertex pair generated by the generator from the ground truth. The final vertex representation is determined by alternately maximizing and minimizing the value function $V(G,D)$ as:

    (23)\begin{align} &\min_{\theta_G}\max_{\theta_D} V(G,D)\nonumber\\ &\quad V(G,D) = \sum_{c=1}^V (\mathbb{E}_{v \sim p_{true} (\cdot\vert v_c)[\log D(v,v_c;\theta_D)])}\nonumber\\ &\qquad+\mathbb{E}_{v \sim G (\cdot\vert v_c;\theta_G)}[\log( 1-D(v,v_c;\theta_D)]).\end{align}
  2. 2 GenVector [Reference Yang, Tang and Cohen26]:

    GenVector leverages large-scale unlabeled data to learn large social knowledge graphs. This is a weakly supervised problem and can be solved by unsupervised techniques with a multi-modal Bayesian embedding model. GenVector can serve as a generative model in applications. For example, it uses latent discrete topic variables to generate continuous word embeddings, graph-based user embeddings, and integrates the advantages of topic models and word embeddings.

IV. COMPARISON OF DIFFERENT METHODS and APPLICATIONS

Classical dimensionality reduction methods have been widely used in graph embedding. They are mathematically transparent, yet most of them cannot represent the high order proximity in graphs. DeepWalk-based methods do not attempt to embed the whole graph but sample the neighborhood information of each node statistically. On the one hand, they can capture the long-distance relationship among nodes. On the other hand, the global information of a graph may not be fully preserved due to sampling. Matrix factorization methods find the graph representation based on the statistics of pairwise similarities. Such methods can outperform deep-learning method built upon random walk methods, where only a local context window is used. However, matrix factorization can be inefficient and unscalable for large graphs. This is because proximity matrix construction as well as eigen decomposition demand higher computational and storage complexities. Moreover, factorization methods such as LLE, Laplacian Eigenmaps and Graph Factorization only conserve the first order proximity. The LLE method has a time complexity of $O(Ed^2)$ while the GF has a time complexity of $O(Ed)$, where d is the number of dimensions. In contrast, random walks usually have a time complexity $O(\vert V\vert d)$. Moreover, factorization based method usually such as LLE, Laplacian Eigenmaps, and Graph Factorization only conserves first order proximity while DeepWalk-based methods can conserve second order proximity.

Deep learning architectures are mostly built upon neural networks. Such models rely heavily on modern GPU to optimize the parameters. They can preserve higher order proximities. Most state-of-the-art applications are based on such models. However, they are mathematically intractable and difficult to interpret. Besides, the training of model parameters with BP demands high time complexity. Large graph embedding methods such as LGCL and GPNN can handle larger graphs. They are suitable for the embedding of social networks, which contain thousands or millions of nodes. However, there are several limitations on these methods. First, these methods have a high time complexity. Second, graphs are dynamic in nature. Social network graphs and citation graphs in academic database are changing from time to time, and the graph structure keep growing over time. However, these methods usually work on static graphs. Third, they demand the pre-processing of the raw input data. It is desired to develop scalable embedding techniques.

Hyper graph embedding can be used to model more complex data and more dynamic networks, they are powerful in representing the information of social networks. However, they are more difficult to implement. Other methods provide alternatives for graph embedding but have not been used widely. Most of them are still in the “proof-of-concept” stage. Kernel methods convert a graph into a single vector. The resulting vector can facilitate graph level analytical tasks such as graph classification. They are more efficient than deep network models since they only need to enumerate the desired atomic substructures in a graph. However, they have a redundant substructure in the graph representation. Also, the embedding dimension can grow exponentially. Generative models leverage the information from different aspects of the graph such as graph structure, node attribute, etc. in a unified model. However, modeling observations based on the assumption of certain distributions is difficult to justify. Furthermore, generative models demand a large amount of training data to fit the data, which might not work well for small graphs.

All the methods mentioned above are summarized in Table 2.

Table 2. Summary of different graph embedding methods

V. EVALUATION

We study the evaluation of various graph representation methods in this section. Evaluation tasks and data sets will be discussed in Sections A and B, respectively. Then, evaluation results will be presented and analyzed in Section C.

A) Evaluation tasks

The two most common evaluation tasks are vertex classification and link prediction. We use vertex classification to compare different graph embedding methods and draw insights from the obtained results.

  1. Vertex classification:

    Vertex classification aims to assign a class label to each node in a graph based on the information learned from other labeled nodes. Intuitively, similar nodes should have the same label. For example, closely-related publication may be labeled as the same topic in the citation graph while individuals of the same gender, similar age, and shared interests may have the same preference in social networks. Graph embedding methods embed each node into a low-dimensional vector. Given an embedded vector, a trained classifier can predict the label of a vertex of interest, where the classifier can be support vector machine (SVM) [Reference Golub and Reinsch84], logistic regression [Reference Wang, Cui, Wang, Pei, Zhu and Yang85], kNN (k nearest neighbors) [Reference Le and Lauw86], etc. The vertex label can be obtained in an unsupervised or semi-supervised way. Node clustering is an unsupervised method that groups similar nodes together. It is useful when labels are unavailable. The semi-supervised method can be used when part of the data are labeled. The F1 score is used for evaluation in binary-class classification, while the micro-F1 score is used in multi-class classification. Since accurate vertex representations contribute to high classification accuracy, vertex classification can be used to measure the performance of different graph embedding methods.

  2. Link prediction [Reference Gao, Denoyer and Gallinari87]:

    Link prediction aims to infer the existence of relationship or interaction among pairs of vertices in a graph. The learned representation should help infer the graph structure, especially when some links are missing. For example, links might be missing between two users and link prediction can be used to recommend friends in social networks. The learned representation should preserve the network proximity and the structural similarity among vertices. The information encoded in the vector representation for each vertex can be used to predict missing links in incomplete networks. The link prediction performance can be measured by the area under curveor the receiver operating characteristiccurve. A better representation should be able to capture the connections among vertices better.

We describe the benchmark graph data sets and conduct experiments in vertex classification on both small and large data sets in the following subsections.

B) Evaluation data sets

Citation data sets such as Citeseer [Reference Giles, Bollacker and Lawrence88], Cora [Reference Cabanes, Grouazel, Schuckmann, Hamon, Turpin, Coatanoan, Paris, Guinehut, Boone and Ferry89], and PubMed [Reference Canese and Weis90] are examples of small data sets. They can be represented as directed graphs in which edges indicates author-to-author or paper-to-paper citation relationship and text attributes of paper content at nodes.

First, we describe several representative citation data sets below:

  1. Citeseer [Reference Giles, Bollacker and Lawrence88]:

    It is a citation index data set containing academic papers of six categories. It has 3312 documents and 4723 links. Each document is represented by a 0/1-valued word vector indicating the absence/presence of the corresponding word from a dictionary of 3703 words. Thus, the text attributes of a document is a binary-valued vector of 3703 dimensions.

  2. Cora [Reference Cabanes, Grouazel, Schuckmann, Hamon, Turpin, Coatanoan, Paris, Guinehut, Boone and Ferry89]:

    It consists of 2708 scientific publications of seven classes. The graph has 5429 links that indicate citation relations between documents. Each document has text attributes that are expressed by a binary-valued vector of 1433 dimensions.

  3. Wikipedia [Reference Cucerzan91]:

    The Wikipedia is an online encyclopedia created and edited by volunteers around the world. The data set is a word co-occurrence network constructed from the entire set of English Wikipedia pages. This data contains 2405 nodes, 17981 edges and 19 labels.

Next, we present several commonly used large graph data sets below:

  1. BlogCatalog [Reference Tang and Liu92]:

    It is a network of social relationships of bloggers listed in the BlogCatalog website. The labels indicate blogger's interests inferred from the meta-data provided by bloggers. The network has 10 312 nodes, 333 983 edges and 39 labels.

  2. YouTube [Reference Wattenhofer, Wattenhofer and Zhu93]:

    It is a social network of YouTube users. This graph contains 1 157 827 nodes, 4 945 382 edges and 47 labels. The labels represent groups of users who enjoy common video genres.

  3. Flickr [Reference Sigurbjörnsson and Van Zwol94]:

    It is an online photo management and sharing data set. It contains 80 513 nodes, 5 899 882 edges and 195 labels.

Finally, the parameters of the above-mentioned data sets are summarized in Table 3.

Table 3. Summary of representative graph data sets

C) Evaluation results and analysis

Since evaluations were often performed independently on different data sets under different settings in the past, it is difficult to draw a concrete conclusion on the performance of various graph embedding methods. Here, we compare the performance of graph embedding methods using a couple of metrics under the common setting and analyze obtained results. In addition, we provide an open-source Python library, called the graph representation learning library (GRLL), to readers in the Github. It offers a unified interface for all graph embedding methods that were experimented in this work. To the best of our knowledge, this library covers the most significant number of graph embedding techniques until now.

1) Vertex classification

We compare vertex classification accuracy of seven graph embedding methods on Cora, Citeseer, and Wiki. We used the default hyper-parameter setting provided by each graph embedding method. For the classifier, we adopt linear regression for all methods (except for GCN since it is a semi-supervised algorithm). We randomly split samples equally into the training and the testing sets (i.e. 50 and 50%). The default embedding size is 128. The vertex classification results are shown in Table 4. DeepWalk and node2vec offer the highest accuracy for Cora and Wiki, respectively. The graph convolutional methods (i.e. GCN) yields the best accuracies in all three data sets because it uses multiple layers of graph convolution to propagate information between nodes. The connectivity information is therefore interchanged and is useful for vertex classification. The random-walk-based methods (e.g. DeepWalk, node2vec, and TADW) also get superior performance since they are able to capture the contextual information in graphs. In Wiki, data set, the node features are not given so it's initialized as an identity matrix. Therefore, TADW couldn't get comparable performance as other methods, demonstrating that good initial attributes for nodes is important for vertex classification. DeepWalk and node2vec are preferred among random-walk-based methods since TADW usually demands initial node features and much more memory. However, if the initial node attributes are presented, TADW could get better results than simple random-walk based methods.

Table 4. Performance comparison of nine common graph embedding methods in vertex classification on Cora, Citeseer, and Wiki

2) Visualization

Visualizing the embedding is useful to understand how well the embedding methods learn from the graph structure and node attributes. To be considered as good representation, vectors for nodes in the same class should have shorter distance and higher similarity to each other. Meanwhile, vectors for nodes from different classes should be as separable as possible so the downstream machine learning models will be able to obtain better performance. To visualize high-dimensional vectors, we adopt a widely-used dimension reduction algorithms, t-SNE [Reference Maaten and Hinton95], to transform the high-dimensional vectors to a two-dimensional space for visualization. We visualized three major branches of graph embedding learning methods, which are random-walk-based (DeepWalk), structural-preservation-based (LINE), and neural-network-based methods (SDNE). We choose Cora as the data set to be tested. The results are shown in Fig. 6. The random-walk-based method provides separable inter-class representation by capturing the contextual information in the graphs. Nodes that don't appear in the same context tend to be separated. However, the intra-class vectors are not clustered well due to the randomly sampled paths. The structural-preservation-based method provides the most compact intra-class clusters due to preservation of the first and second order proximity. The graph structure can be reflected the most by this category of methods. The representation vectors from the neural-network-based method are mainly located on a high-dimensional manifold in the latent space. The arrangement of the embedding is often uneven and biased.The authors in [Reference Wang, Chen, Wang and Kuo96] discussed this effect and proposed a post-processing scheme to solve the problem. Consequently, the representation vectors from the neural-network-based method are nearly not linearly-separable.

Fig. 6. t-SNE visualization of different embedding methods on Cora. The seven different colors of points represent different classes of the nodes.

3) Clustering quality

We compare various graph embedding methods by examining their clustering quality in terms of the macro and micro-F1 scores. The K-means++ algorithm is adopted for the clustering task. Since the results of K-means++ clustering are dependent upon seed initialization, we perform 10 consecutive runs and report the best result. We tested them on three large graph data sets (i.e. YouTube, Flickr, and BlogCatalog). The experimental results are shown in Table 6. YouTube and Flickr contain more than millions of nodes and edges and we can only run DeepWalk, node2vec, and LINE on them with the 24G RAM limit as reported in the table. We see that DeepWalk and node2vec provide the best results. They are both random-walk based methods with different sampling schemes. Also, they demand less memory as compared with others. In general, random walk with the skip-gram model is a good baseline for unsupervised graph embedding. GraRep offers a comparable graph embedding quality for BlogCatalog. However, its memory requirement is significant so that it is not suitable for large graphs.

4) Time complexity

Time complexity is an essential factor to consider, which is especially true for large graphs. The time complexity of three embedding methods (DeepWalk, node2vec, and LINE) against three data sets (YouTube, Flickr, and Wiki) is compared in Table 5. Three embedding methods all apply negative sampling [Reference Mikolov, Sutskever, Chen, Corrado and Dean18] to reduce the time complexity. We observe that the training time of DeepWalk is significantly lower than node2vec and LINE for larger graph data sets such as YouTube and Flickr. DeepWalk is an efficient graph embedding method with high accuracy by considering embedding quality as well as training complexity.

Table 5. Comparison of time used in training (s)

Table 6. Comparison of clustering quality of six graph embedding methods in terms of macro- and micro-F1 scores against three large graph data sets

5) Influence of embedding dimensions

As the embedding dimension decreases, less information about the input graph is preserved so that the performance drops. However, some methods will be affected by the decrease of embedding size than others. We show the node classification accuracy as a function of the embedding dimension for the Cora and Citeseer data set in Fig. 7. We compare eight graph embedding methods (DeepWalk, node2vec, GraRep, HOPE, SDNE, LINE, GF, and TADW) and their embedding dimensions vary from 4, 8, 16, 32, 64 to 128. The experimental results are shown in Fig. 7. We see that the performance of the random-walk based embedding methods (Deep-Walk and node2vec) degrades slowly. Only about 10% drop in performance when the embedding dimension size drops from 128 to 4. An exception is TADW, which requires initial node features to obtain the final embedding. The initial attributes are usually sparse and high dimensional, transforming them to a lower-dimensional embedding space will cause great information loss. In contrast, the performance of the structure preserving methods (LINE and GraRep) drops significantly (as much as 45%) when the embedding size goes from 128 to 4. One explanation is that the structural preserving methods are compressing the adjacency matrix into a lower dimensional space. When the compression ratio is higher, the distortion will be higher as well. Random-walk based methods obtain embedding vectors by selecting paths from the input graph randomly. Yet, the relationship between nodes is still preserved when the embedding dimension is small. SDNE adopts the auto-encoder architecture to preserve the information of the input graph so that its performance remains about the same regardless of the embedding dimension.

Fig. 7. The node classification accuracy as a function of the embedding size.

6) Influence of training sample ratio

By the training sample ratio, we mean the percentages of total node samples that are used for the training purpose. When the ratio is high, the classifier could be overfitted. On the other hand, if the ratio is too low, the offered information may not be sufficient for the training purpose. Such analysis is classifier dependent, and we adopt a simple linear regression classifier from the python sklearn toolkit in the experiment. The node classification accuracy as a function of the training sample ratio for the Cora and Citeseer data set is shown in Fig. 8. Most methods have consistent performance for the training data ratio between 0.2 and 0.8 except for the deep learning based methods (SDNE). Its accuracy drops when the training data ratio is low. It needs a higher ratio of training data. GCN uses multiple graph convolutional layers to propagate information between neighboring nodes so it could get good performance with smaller training set. However, when the training ratio becomes higher, the accuracy for GCN would not increase much or even drop due to over-fitting.

Fig. 8. The node classification accuracy as a function of the embedding size.

VI. EMERGING APPLICATIONS

Graphs offer a powerful modeling tool and find a wide range of applications. Since many real-world data have some certain relationships among entities, they can be conveniently modeled by graphs. Multi-modal data can also be embedded into the same space through graph representation learning and, as a result, the information from different domains can be represented and analyzed in one common setting.

In this section, we examine three emerging areas that benefit from graph embedding techniques.

  1. Community detection:

    Graph embedding can be used to predict the label of a node given a fraction of labeled node [Reference Ding, He, Zha, Gu and Simon13, Reference Chen97Reference Zhang and Wu99]. Thus, it has been widely used for community detection [Reference Dourisboure, Geraci and Pellegrini100, Reference Newman101]. In social networks, node labels might be gender, demography, or religion. In language networks, documents might be labeled with categories or keywords. Missing labels can be inferred from labeled nodes and links in the network. Graph embedding can be used to extract node features automatically based on the network structure and predict the community that a node belongs to. Both vertex classification [Reference Girvan and Newman102] and link prediction [Reference De Bacco, Power, Larremore and Moore103, Reference Zheleva, Getoor, Golbeck and Kuter104] can facilitate community detection [Reference Fortunato105, Reference Yang, Jin, Chi and Zhu106].

  2. Recommendation system:

    Recommendation is an important function in social networks and advertising platforms [Reference Hamilton, Ying and Leskovec107Reference Zhang, Yuan, Lian, Xie and Ma109]. Besides the structure, content and label data [Reference Kermarrec, Leroy and Trédan110], some networks contain spatial and temporal information. For example, Yelp may recommend restaurants based on a user's location and preference. Spatial-temporal embedding [Reference Zhang, Zhang, Yuan, Peng, Zheng, Hanratty, Wang and Han111] is an emerging topic in mobile applications.

  3. Graph compression and coarsening:

    By graph compression (graph simplification), we refer to a process of converting one graph to another, where the latter has a smaller number of edges. Graph coarsening can be used for compression, which is a method used to partition a graph in multiple stages. Graph coarsening is usually used in the first stage for graph embedding. It works by collapsing pairs of nodes and edges using a determinate criteria. Graph coarsening is useful because the size of the graphs may be too big to be partitioned. Graph compression aims to store a graph more efficiently and run graph analysis algorithms faster. For example, a graph is partitioned into bipartite cliques and replaced by trees to reduce the edge number in [Reference Feder and Motwani112]. Along this line, one can also aggregate nodes or edges for faster processing with graph coarsening [Reference Liang, Gurukar and Parthasarathy113], where a graph is converted into smaller ones repeatedly using a hybrid matching technique to maintain its backbone structure. The structural equivalence matching (SEM) method [Reference Green and Yang114] and the normalized heavy edge matching method (NHEM) [Reference Karypis and Kumar115] are two examples.

  4. Biomedical application:

    Graph representation learning can be used to for biomedical data analysis. For example, brain network data can be modeled through the graph, with the brain activities as signals residing on the nodes [Reference Zhang, Yin, Zhu and Zhang116]. The embedding methods can be used for studying the structures and functions of brains when under different stimuli. Some frameworks have been proposed for studying Alzheimer's disease [Reference Hu, Cheng, Sepulcre, Johnson, Fakhri, Lu and Li117], and brain reaction to magnetoencephalographysignals [Reference Liu, Nejati and Cheung118].

  5. Other application domain:

    Graph learning can also be applied to meteorology for studying networks of weather stations. The authors in [Reference Chepuri, Liu, Leus and Hero119, Reference Dong, Thanou, Frossard and Vandergheynst120] aim to capture the relationship among different weather stations in terms of their attitude. It is still in the stage of proof of concept and more research is needed. Graph embedding can also be used in traffic flow inference [Reference Thanou, Dong, Kressner and Frossard121], Internet news propagation study [Reference Baingana and Giannakis122], inter-region political relationship [Reference Dong, Thanou, Frossard and Vandergheynst120], human robotics interaction [Reference Li, Wang, Zhu and Huang123],ontologies of concepts [Reference Lake and Tenenbaum124], etc.

VII. FUTURE RESEARCH DIRECTIONS

Graph representation learning is a well-motivated topic. It is an effective way to convert graph data into a low dimensional space [Reference Dong, Thanou, Rabbat and Frossard125, Reference Gao, Hu, Tang, Liu and Guo126] in which important feature information are well preserved. Graph analytic can provide researchers with a deeper understanding of the data with the help of efficient graph embedding techniques. Based on the discussion above, we would like to discuss several future research opportunities in graph embedding.

  1. Deep graph embedding:

    GCN [Reference Kipf and Welling62] and some of its adaptation [Reference Li, Wang, Zhu and Huang123] has drawn great attention due to its superior performance. However, the number of graph convolutional layers is typically not higher than two. When there are more graph convolutional layers in cascade, the performance drops significantly. It was argued in [Reference Li, Han and Wu65] that each GCN layer corresponds to graph Laplacian smoothing since node features are propagated in the spectral domain. When a GCN is deeper, the graph Laplacian is over-smoothed and the corresponding node features become obscure. Each layer of GCN usually only learns one-hop information, and two GCN layers learn the first- and second-order proximity in the graph. It is difficult for a shallow structure to learn global information. One solution to fix this problem is to conduct the convolution in the spatial domain. For example, one can convert graph data into grid-structure data as proposed in [Reference Gao, Wang and Ji29]. Then, the graph representation can be learned using multiple CNN layers. Another way to address the problem is to down-sample graphs and merge similar nodes. Such a graph coarsening idea was adopted by [Reference Gao, Chen and Ji127Reference Ying, You, Morris, Ren, Hamilton and Leskovec129] to build deep GCNs. Then, we can build a hierarchical network structure, which allows us to learn both local and global graph data in a hierarchical manner.

  2. Dynamic graph embedding:

    Social graphs, such as graphs in Twitter, are always changing. Another example is graphs of mobile users whose location information is changing along with time. To learn the representation of dynamic graphs is an important research topic and it finds applications in real-time and interactive processes such as the optimal travel path planning in a city at traffic hours. Hyper-graphs is a good option for modeling such dynamics in graphs. Embed the time sequence into each node can also be useful; for example, long-short-term-memory (LSTM) can be used on each vertex to incorporate sequential changes.

  3. Scalability of graph embedding:

    With the rapid growth of social networks, which contain millions and billions of nodes and edges, We expect to see graphs of a larger scale and higher diversity. How to embed enormous graph data efficiently and accurately is still an open problem. Deep neural network models have the state-of-the-art performance. However, these methods suffer the low efficiency problem. They rely on modern GPU to find the optimal parameters. Better paradigms are needed for processing large-scale graphs. One possibility is to use a feedforward machine-learning design to process the graph without BP. Another option is to use better graph coarsening or partitioning method to preprocess the data.

  4. Interpretability of graph embedding:

    Most state-of-the-art graph embedding methods are built upon CNNs, which are trained with BP to determine their model parameters. However, the training complexity is very high. Some research was performed to lower the training complexity such as quickprop [Reference Fahlman130]. However, training model parameters iteratively using BP is still time consuming and hardware demanding. In addition, CNNs are mathematically intractable. Very recently, some researchers have tried to explain the interpretability of neural network models [Reference Kuo131, Reference Zhang, Nian Wu and Zhu132]. The authors in [Reference Kuo, Zhang, Li, Duan and Chen133] attempt to explain CNNs using an interpretable and feedforward (FF) design without any BP. The work in [Reference Kuo, Zhang, Li, Duan and Chen133] adopts a FFdata-centric approach to network parameters of the current layer based on data statistics from the output of the previous layer in a one-pass manner. It would be worthwile to apply FF machine-learning methods to graph embedding tasks. An interpretable design as an alternative to advanced neural network architectures can shed light on current graph embedding-related machine-learning research.

VIII. CONCLUSION

A comprehensive survey of the literature on graph representation learning techniques was conducted in this paper. We examined various graph embedding techniques that convert the input graph data into a low-dimensional vector representation while preserving intrinsic graph properties. Besides classical graph embedding methods, we covered several new topics such as neural-network-based embedding methods, hypergraph embedding and attention graph embedding methods. Furthermore, we conducted an extensive performance evaluation of several stat-of-the-art methods against small and large data sets. For experiments conducted in our evaluation, an open-source Python library, called the GRLL, was provided to readers. Finally, we presented some emerging applications and future research directions. We hope our work can inspire more follow-up work in graph embedding.

Fenxiao (Jessica) Chen received her Bachelor's degree in General Engineering from Harvey Mudd College (HMC), California, USA in May, 2014. Since Jan 2017, she joined Media Communication Lab (MCL) at University of Southern California (USC) as a Ph.D. student, supervised by Prof. C.-C. Jay Kuo. Her research interests include natural language processing (NLP), machine learning and graph embedding.

Yun Cheng Wang received his B.S. degree in electrical engineering from National Taiwan University, Taipei, Taiwan, in 2018, and M.S. degree in electrical and computer engineering from University of Southern California (USC), Los Angeles, in 2019. He joined Media Communications Lab (MCL) in August, 2018, supervised by Prof. C.-C. Jay Kuo. His research interests include graph representation learning (GRL) and deep learning.

Bin Wang received his Bachelor's degree in Electronic Information Engineering from University of Electronic Science and Technology of China (UESTC), Chengdu, China in June, 2017. Since July 2017, he joined Media Communication Lab (MCL) at University of Southern California (USC) as a Ph.D. student, supervised by Prof. C.-C. Jay Kuo. His research interests include natural language processing (NLP), image processing and machine learning.

C.-C. Jay Kuo (F'99) received the B.S. degree in electrical engineering from the National Taiwan University, Taipei, Taiwan, in 1980, and the M.S. and Ph.D. degrees in electrical engineering from the Massachusetts Institute of Technology, Cambridge, in 1985 and 1987, respectively. He is currently the Director of the Multimedia Communications Laboratory and a Distinguished Professor of electrical engineering and computer science at the University of Southern California, Los Angeles. His research interests include multimedia computing and machine learning. Dr. Kuo is a Fellow of the American Association for the Advancement of Science (AAAS) and The International Society for Optical Engineers (SPIE).

References

REFERENCES

[1]Bourigault, S.; Lagnier, C.; Lamprier, S.; Denoyer, L.; Gallinari, P.: Learning social network embeddings for predicting information diffusion. in Proceedings of the 7th ACM International Conference on Web Search and Data Mining, ACM, 2014, 393402.CrossRefGoogle Scholar
[2]Globerson, A.; Chechik, G.; Pereira, F.; Tishby, N.: Euclidean embedding of co-occurrence data. J. Mach. Learn. Res., 8 (Oct) (2007), 22652295.Google Scholar
[3]Theocharidis, A.; Van Dongen, S.; Enright, A.J.; Freeman, T.C.: Network visualization and analysis of gene expression data using biolayout express 3D. Nat. Protoc., 4 (10) (2009), 1535.CrossRefGoogle ScholarPubMed
[4]Angles, R.; Gutierrez, C.: Survey of graph database models. ACM Comput. Surveys (CSUR), 40 (1) (2008), 1.CrossRefGoogle Scholar
[5]Gargi, U.; Lu, W.; Mirrokni, V.; Yoon, S.: Large-scale community detection on YouTube for topic discovery and exploration. in Fifth International AAAI Conference on Weblogs and Social Media, 2011.Google Scholar
[6]Roggen, D.; Wirz, M.; Tröster, G.; Helbing, D.: Recognition of crowd behavior from mobile sensors with pattern analysis and graph clustering methods. arXiv preprint arXiv:1109.1664, 2011.Google Scholar
[7]Bhagat, S.; Cormode, G.; Muthukrishnan, S.: Node classification in social networks. In Social Network Data Analytics, Springer, 2011, 115148,CrossRefGoogle Scholar
[8]Liben-Nowell, D.; Kleinberg, J.: The link-prediction problem for social networks. J. Am. Soc. Inf. Sci. Technol., 58 (7) (2007), 10191031.CrossRefGoogle Scholar
[9]Goyal, P.; Ferrara, E.: Graph embedding techniques, applications, and performance: A survey. Knowl. Based Syst., 151, (2018), 7894.CrossRefGoogle Scholar
[10]Yan, S.; Xu, D.; Zhang, B.; Zhang, H.-J.: Graph embedding: a general framework for dimensionality reduction. In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05), Vol. 2, IEEE, 2005, 830837.Google Scholar
[11]Yan, S.; Xu, D.; Zhang, B.; Zhang, H.-J.; Yang, Q.; Lin, S.: Graph embedding and extensions: A general framework for dimensionality reduction. IEEE Trans. Pattern Anal. Mach. Intell., 1 (2007), 4051.CrossRefGoogle Scholar
[12]Shaw, B.; Jebara, T.: Structure preserving embedding. In Proceedings of the 26th Annual International Conference on Machine Learning, ACM, 2009, 937944.CrossRefGoogle Scholar
[13]Ding, C.H.; He, X.; Zha, H.; Gu, M.; Simon, H.D.: A min-max cut algorithm for graph partitioning and data clustering. in Proceedingsof the 2001 IEEE International Conference on Data Mining, IEEE, 2001, 107114.Google Scholar
[14]Cavallari, S.; Zheng, V.W.; Cai, H.; Chang, K.C.-C.; Cambria, E.: Learning community embedding with community detection and node embedding on graphs. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, ACM, 2017, 377386.CrossRefGoogle Scholar
[15]Goldberger, J.; Gordon, S.; Greenspan, H.: An efficient image similarity measure based on approximations of kl-divergence between two Gaussian mixtures. In Null, IEEE, 2003, 487.CrossRefGoogle Scholar
[16]Zhou, C.; Liu, Y.; Liu, X.; Liu, Z.; Gao, J.: Scalable graph embedding for asymmetric proximity. In Thirty-First AAAI Conference on Artificial Intelligence, 2017.Google Scholar
[17]Anis, A.; Gadde, A.; Ortega, A.: Efficient sampling set selection for bandlimited graph signals using graph spectral proxies. IEEE Trans. Signal Process., 64 (14) (2016), 37753789.CrossRefGoogle Scholar
[18]Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J.: Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems, 2013, 31113119.Google Scholar
[19]Xu, K.; Feng, Y.; Huang, S.; Zhao, D.: Semantic relation classification via convolutional neural networks with simple negative sampling. arXiv preprint arXiv:1506.07650, 2015.Google Scholar
[20]Leskovec, J.; Faloutsos, C.: Sampling from large graphs. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2006, 631636.CrossRefGoogle Scholar
[21]Ribeiro, B.; Towsley, D.: Estimating and sampling graphs with multidimensional random walks. In Proceedings of the 10th ACM SIGCOMM Conference on Internet Measurement, ACM, 2010, 390403.CrossRefGoogle Scholar
[22]Cai, H.; Zheng, V.W.; Chang, K.C.-C.: A comprehensive survey of graph embedding: problems, techniques, and applications. IEEE Trans. Knowl. Data. Eng., 30 (9) (2018), 16161637.CrossRefGoogle Scholar
[23]Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q.: LINE: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, 2015, 10671077. International World Wide Web Conferences Steering Committee.CrossRefGoogle Scholar
[24]Gilbert, A.C.; Levchenko, K.: Compressing network graphs. in Proceedings of the LinkKDD Workshop at the 10th ACM Conference on KDD, Vol. 124, 2004.Google Scholar
[25]Taylor, P.; Black, A.W.; Caley, R.: Heterogeneous relation graphs as a formalism for representing linguistic information. Speech Commun., 33 (1-2) (2001), 153174.CrossRefGoogle Scholar
[26]Yang, Z.; Tang, J.; Cohen, W.: Multi-modal Bayesian embeddings for learning social knowledge graphs. arXiv preprint arXiv:1508.00715, 2015b.Google Scholar
[27]Bizer, C.; Lehmann, J.; Kobilarov, G.; Auer, S.; Becker, C.; Cyganiak, R.; Hellmann, S.: Dbpedia – a crystallization point for the web of data. Web Semant.: Sci., Services Agents World Wide Web, 7 (3) (2009), 154165.CrossRefGoogle Scholar
[28]Bollacker, K.; Evans, C.; Paritosh, P.; Sturge, T.; Taylor, J.: Freebase: a collaboratively created graph database for structuring human knowledge. in Proceedings of the 2008 ACM SIGMOD International Conference on Management of Data, ACM, 2008, 12471250.CrossRefGoogle Scholar
[29]Gao, H.; Wang, Z.; Ji, S.: Large-scale learnable graph convolutional networks. in Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ACM, 2018, 14161424.CrossRefGoogle Scholar
[30]de Fraysseix, H.; Pach, J.; Pollack, R.: Small sets supporting Fary embeddings of planar graphs. In Proceedings of the Twentieth Annual ACM Symposium on Theory of Computing, ACM, 1988, 426433.CrossRefGoogle Scholar
[31]Fáry, I.: On straight-line representation of planar graphs. Acta Sci. Math., 11, (1948), 229233.Google Scholar
[32]Stein, S.K.: Convex maps. Proc. Am. Math. Soc., 2 (3) (1951), 464466.CrossRefGoogle Scholar
[33]Chrobak, M.; Payne, T.H.: A linear-time algorithm for drawing a planar graph on a grid. Inf. Process. Lett., 54 (4) (1995), 241246.CrossRefGoogle Scholar
[34]De Fraysseix, H.; Pach, J.; Pollack, R.: How to draw a planar graph on a grid. Combinatorica, 10 (1) (1990), 4151.CrossRefGoogle Scholar
[35]Krizhevsky, A.; Sutskever, I.; Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, 2012, 10971105.Google Scholar
[36]Mikolov, T.; Karafiát, M.; Burget, L.; Černockỳ, J.; Khudanpur, S.: Recurrent neural network based language model. In Eleventh Annual Conference of the International Speech Communication Association, 2010.CrossRefGoogle Scholar
[37]Jolliffe, I.T.; Cadima, J: Principal component analysis: a review and recent developments. Philos. Trans. A Math. Phys. Eng Sci., 374 (2065) (2016), 20150202.CrossRefGoogle ScholarPubMed
[38]Umeyama, S.: An eigen decomposition approach to weighted graph matching problems. IEEE Trans. Pattern Anal. Mach. Intell., 10 (5) (1988), 695703.CrossRefGoogle Scholar
[39]Ye, J.; Janardan, R.; Li, Q.: Two-dimensional linear discriminant analysis. In Advances in Neural Information Processing Systems, 2005, 15691576.Google Scholar
[40]Robinson, S.L.; Bennett, R.J.: A typology of deviant workplace behaviors: A multidimensional scaling study. Acad. Manage. J., 38 (2) (1995), 555572.Google Scholar
[41]Saul, L.K.; Weinberger, K.Q.; Ham, J.H.; Sha, F.; Lee, D.D.: Spectral methods for dimensionality reduction. Semisupervised learning, (2006), 293308.Google Scholar
[42]DeMers, D.; Cottrell, G.W.: Non-linear dimensionality reduction. in Advances in Neural Information Processing Systems, 1993, 580587.Google Scholar
[43]Samko, O.; Marshall, A.D.; Rosin, P.L.: Selection of the optimal parameter value for the isomap algorithm. Pattern Recognit. Lett., 27 (9) (2006), 968979.CrossRefGoogle Scholar
[44]Roweis, S.T.; Saul, L.K.: Nonlinear dimensionality reduction by locally linear embedding. Science, 290 (5500) (2000), 23232326.CrossRefGoogle ScholarPubMed
[45]Harandi, M.T.; Sanderson, C.; Shirazi, S.; Lovell, B.C.: Graph embedding discriminant analysis on Grassmannian manifolds for improved image set matching. In CVPR 2011, IEEE, 2011, 27052712.CrossRefGoogle Scholar
[46]Guthrie, D.; Allison, B.; Liu, W.; Guthrie, L.; Wilks, Y.: A closer look at skip-gram modelling. In LREC, 2006, 12221225.Google Scholar
[47]Zhang, Y.; Jin, R.; Zhou, Z.-H.: Understanding bag-of-words model: a statistical framework. Int. J. Mach. Learn. Cybernetics, 1 (1-4) (2010), 4352.CrossRefGoogle Scholar
[48]Perozzi, B.; Al-Rfou, R.; Skiena, S.: Deepwalk: online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2014, 701710.CrossRefGoogle Scholar
[49]Yang, C.; Liu, Z.; Zhao, D.; Sun, M.; Chang, E.: Network representation learning with rich text information. In Twenty-Fourth International Joint Conference on Artificial Intelligence, 2015a.Google Scholar
[50]Grover, A.; Leskovec, J.: node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2016, 855864.CrossRefGoogle Scholar
[51]Ahmed, A.; Shervashidze, N.; Narayanamurthy, S.; Josifovski, V.; Smola, A.J.: Distributed large-scale natural graph factorization. In Proceedings of the 22nd International Conference on World Wide Web, pp. 3748. ACM 2013.CrossRefGoogle Scholar
[52]Belkin, M.; Niyogi, P.: Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput., 15 (6) (2003), 13731396.CrossRefGoogle Scholar
[53]Singh, A.P.; Gordon, G.J.: Relational learning via collective matrix factorization. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2008, 650658.CrossRefGoogle Scholar
[54]Zhang, D.; Yin, J.; Zhu, X.; Zhang, C.: Homophily, structure, and content augmented network representation learning. In Data Mining (ICDM), 2016 IEEE 16th International Conference on, 2016a, IEEE, 609618.CrossRefGoogle Scholar
[55]Møller, M.F.: A scaled conjugate gradient algorithm for fast supervised learning. Neural Netw., 6 (4) (1993), 525533.CrossRefGoogle Scholar
[56]Chen, F.; Wang, B.; Kuo, C.-C.J.: Deepwalk-assisted graph PCA (DGPCA) for language networks. in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2019, 29572961.CrossRefGoogle Scholar
[57]Cao, S.; Lu, W.; Xu, Q.: Grarep: Learning graph representations with global structural information. in Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, 2015, ACM, 891900CrossRefGoogle Scholar
[58]Ou, M.; Cui, P.; Pei, J.; Zhang, Z.; Zhu, W.: Asymmetric transitivity preserving graph embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2016, 11051114.CrossRefGoogle Scholar
[59]Katz, L.: A new status index derived from sociometric analysis. Psychometrika, 18 (1) (1953), 3943.CrossRefGoogle Scholar
[60]Mikolov, T.; Chen, K.; Corrado, G.; Dean, J.: Efficient estimation of word representations in vector space. arXiv preprint arXiv:1301.3781, 2013a.Google Scholar
[61]Spitzer, F.: Principles of Random Walk, Springer Science & Business Media, Vol. 34, 2013.Google Scholar
[62]Kipf, T.N.; Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.Google Scholar
[63]Defferrard, M.; Bresson, X.; Vandergheynst, P.: Convolutional neural networks on graphs with fast localized spectral filtering. in Advances in Neural Information Processing Systems, 2016, 38443852.Google Scholar
[64]Field, D.A.: Laplacian smoothing and delaunay triangulations. Commun. Appl. Numer. Methods, 4 (6) (1988), 709712.CrossRefGoogle Scholar
[65]Li, Q.; Han, Z.; Wu, X.-M.: Deeper insights into graph convolutional networks for semi-supervised learning. In Thirty-Second AAAI Conference on Artificial Intelligence, 2018a.Google Scholar
[66]Derr, T.; Ma, Y.; Tang, J.: Signed graph convolutional networks. in 2018 IEEE International Conference on Data Mining (ICDM), IEEE, 2018, 929934.CrossRefGoogle Scholar
[67]Kingma, D.P.; Welling, M.: Auto-encoding variational Bayes. arXiv preprint arXiv:1312.6114, 2013.Google Scholar
[68]Hamilton, W.; Ying, Z.; Leskovec, J.: Inductive representation learning on large graphs. In Advances in Neural Information Processing Systems, 2017a, 10241034.Google Scholar
[69]Wang, D.; Cui, P.; Zhu, W.: Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2016, 12251234.CrossRefGoogle Scholar
[70]Hecht-Nielsen, R.: Theory of the backpropagation neural network. In Neural Networks for Perception, Elsevier, 1992, 6593.CrossRefGoogle Scholar
[71]Motsinger, A.A.; Lee, S.L.; Mellick, G.; Ritchie, M.D.: GPNN: Power studies and applications of a neural network method for detecting gene-gene interactions in studies of human disease. BMC Bioinf., 7 (1) (2006), 39–11.CrossRefGoogle ScholarPubMed
[72]Tu, K.; Cui, P.; Wang, X.; Wang, F.; Zhu, W.: Structural deep embedding for hyper-networks. CoRR, abs/1711.10146, 2017.Google Scholar
[73]Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y.: Hypergraph Neural Networks. arXiv preprint arXiv:1809.09401, 2018.Google Scholar
[74]Zhou, D.; Huang, J.; Schölkopf, B.: Learning with hypergraphs: Clustering, classification, and embedding. In Advances in Neural Information Processing Systems, 2007, 16011608.Google Scholar
[75]Lee, J.B.; Rossi, R.A.; Kim, S.; Ahmed, N.K.; Koh, E.: Attention models in graphs: a survey. arXiv preprint arXiv:1807.07984, 2018.Google Scholar
[76]Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y.: Graph attention networks. arXiv preprint arXiv:1710.10903, 2017.Google Scholar
[77]Abu-El-Haija, S.; Perozzi, B.; Al-Rfou, R.; Alemi, A.: Watch your step: learning graph embeddings through attention. arXiv preprint arXiv:1710.09599 2017.Google Scholar
[78]Abu-El-Haija, S.; Perozzi, B.; Al-Rfou, R.; Alemi, A.A.: Watch your step: learning node embeddings via graph attention. In Advances in Neural Information Processing Systems, pp. 91809190 2018.Google Scholar
[79]Xu, Q.; Wang, Q.; Xu, C.; Qu, L.: Attentive graph-based recursive neural network for collective vertex classification. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, ACM, 2017a, 24032406.CrossRefGoogle Scholar
[80]Xu, Q.; Wang, Q.; Xu, C.; Qu, L.: Collective vertex classification using recursive neural network. arXiv preprint arXiv:1701.06751, 2017b.Google Scholar
[81]Chen, F.; Wang, B.; Kuo, C.-C.J.: Graph-based deep-tree recursive neural network (DTRNN) for text classification. arXiv preprint arXiv:1809.01219, 2018.Google Scholar
[82]Wang, H.; Wang, J.; Wang, J.; Zhao, M.; Zhang, W.; Zhang, F.; Xie, X.; Guo, M.: Graphgan: graph representation learning with generative adversarial nets. In Thirty-Second AAAI Conference on Artificial Intelligence, 2018b.Google Scholar
[83]Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y.: Generative adversarial nets. In Advances in Neural Information Processing Systems, 2014, 26722680.Google Scholar
[84]Golub, G.H.; Reinsch, C.: Singular value decomposition and least squares solutions. In Linear Algebra, Springer, 1971, 134151.CrossRefGoogle Scholar
[85]Wang, X.; Cui, P.; Wang, J.; Pei, J.; Zhu, W.; Yang, S.: Community preserving network embedding. In Thirty-First AAAI Conference on Artificial Intelligence, 2017.Google Scholar
[86]Le, T.M.; Lauw, H.W.: Probabilistic latent document network embedding. In 2014 IEEE International Conference on Data Mining, IEEE, 2014, 270279.CrossRefGoogle Scholar
[87]Gao, S.; Denoyer, L.; Gallinari, P.: Temporal link prediction by integrating content and structure information. in Proceedings of the 20th ACM International Conference on Information and knowledge Management, 2011, ACM, 11691174.CrossRefGoogle Scholar
[88]Giles, C.L.; Bollacker, K.D.; Lawrence, S.: Citeseer: An automatic citation indexing system. In ACM DL, 1998, 8998.CrossRefGoogle Scholar
[89]Cabanes, C.; Grouazel, A.; Schuckmann, K. v.; Hamon, M.; Turpin, V.; Coatanoan, C.; Paris, F.; Guinehut, S.; Boone, C.; Ferry, N.: The Cora dataset: validation and diagnostics of in-situ ocean temperature and salinity measurements. Ocean Sci., 9 (1) (2013), 118.CrossRefGoogle Scholar
[90]Canese, K.; Weis, S.: Pubmed: the bibliographic database. in The NCBI Handbook [Internet]. 2nd edition. National Center for Biotechnology Information (US), 2013.Google Scholar
[91]Cucerzan, S.: Large-scale named entity disambiguation based on Wikipedia data. in Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL), 2007, 708716.Google Scholar
[92]Tang, J.; Liu, H.: Unsupervised feature selection for linked social media data. In Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2012, 904912.CrossRefGoogle Scholar
[93]Wattenhofer, M.; Wattenhofer, R.; Zhu, Z.: The YouTube social network. In Sixth International AAAI Conference on Weblogs and Social Media, 2012.Google Scholar
[94]Sigurbjörnsson, B.; Van Zwol, R.: Flickr tag recommendation based on collective knowledge. In Proceedings of the 17th International Conference on World Wide Web, ACM, 2008, 327336.CrossRefGoogle Scholar
[95]Maaten, L. v. d.; Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res., 9 (NOV) (2008), 25792605.Google Scholar
[96]Wang, B.; Chen, F.; Wang, A.; Kuo, C.-C.J.: Post-processing of word representations via variance normalization and dynamic embedding. arXiv preprint arXiv:1808.06305, 2018a.Google Scholar
[97]Chen, D.: A novel clustering algorithm for graphs. In 2009 International Conference on Artificial Intelligence and Computational Intelligence, Vol. 4, IEEE, 2009, 279283CrossRefGoogle Scholar
[98]Yuruk, N.; Mete, M.; Xu, X.; Schweiger, T.A.: Ahscan: Agglomerative hierarchical structural clustering algorithm for networks. In 2009 International Conference on Advances in Social Network Analysis and Mining, IEEE, 2009, 7277.CrossRefGoogle Scholar
[99]Zhang, Y.; Wu, B.: Finding community structure via rough k-means in social network. In 2015 IEEE International Conference on Big Data (Big Data), 2015, IEEE, 23562361.CrossRefGoogle Scholar
[100]Dourisboure, Y.; Geraci, F.; Pellegrini, M.: Extraction and classification of dense communities in the web. in Proceedings of the 16th International Conference on World Wide Web, ACM, 2007, 461470.CrossRefGoogle Scholar
[101]Newman, M.E.: Detecting community structure in networks. Eur. Phys. J. B, 38 (2) (2004), 321330.CrossRefGoogle Scholar
[102]Girvan, M.; Newman, M.E.: Community structure in social and biological networks. Proc. Natl. Acad. Sci., 99 (12) (2002), 78217826.CrossRefGoogle ScholarPubMed
[103]De Bacco, C.; Power, E.A.; Larremore, D.B.; Moore, C.: Community detection, link prediction, and layer interdependence in multilayer networks. Phys. Rev. E., 95 (4) (2017), 042317.CrossRefGoogle ScholarPubMed
[104]Zheleva, E.; Getoor, L.; Golbeck, J.; Kuter, U.: Using friendship ties and family circles for link prediction. In International Workshop on Social Network Mining and Analysis, Springer, 2008, 97113.CrossRefGoogle Scholar
[105]Fortunato, S.: Community detection in graphs. Phys. Rep., 486 (3-5) (2010), 75174.CrossRefGoogle Scholar
[106]Yang, T.; Jin, R.; Chi, Y.; Zhu, S.: Combining link and content for community detection: a discriminative approach. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2009, 927936.CrossRefGoogle Scholar
[107]Hamilton, W.L.; Ying, R.; Leskovec, J.: Representation learning on graphs: methods and applications. arXiv preprint arXiv:1709.05584, 2017b.Google Scholar
[108]Ying, R.; He, R.; Chen, K.; Eksombatchai, P.; Hamilton, W.L.; Leskovec, J.: Graph convolutional neural networks for web-scale recommender systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ACM, 2018a, 974983.CrossRefGoogle Scholar
[109]Zhang, F.; Yuan, N.J.; Lian, D.; Xie, X.; Ma, W.-Y.: Collaborative knowledge base embedding for recommender systems. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ACM, 2016b, 353362.CrossRefGoogle Scholar
[110]Kermarrec, A.-M.; Leroy, V.; Trédan, G.: Distributed social graph embedding. In Proceedings of the 20th ACM international conference on Information and Knowledge Management, ACM, 2011, 12091214.CrossRefGoogle Scholar
[111]Zhang, C.; Zhang, K.; Yuan, Q.; Peng, H.; Zheng, Y.; Hanratty, T.; Wang, S.; Han, J.: Regions, periods, activities: Uncovering urban dynamics via cross-modal representation learning. In Proceedings of the 26th International Conference on World Wide Web, 2017, 361370. International World Wide Web Conferences Steering Committee.CrossRefGoogle Scholar
[112]Feder, T.; Motwani, R.: Clique partitions, graph compression and speeding-up algorithms. J. Comput. Syst. Sci., 51 (2) (1995), 261272.CrossRefGoogle Scholar
[113]Liang, J.; Gurukar, S.; Parthasarathy, S.: Mile: a multi-level framework for scalable graph embedding. arXiv preprint arXiv:1802.09612, 2018.Google Scholar
[114]Green, S.B.; Yang, Y.: Reliability of summed item scores using structural equation modeling: an alternative to coefficient alpha. Psychometrika, 74 (1) (2009), 155167.CrossRefGoogle Scholar
[115]Karypis, G.; Kumar, V.: Multilevel k-way partitioning scheme for irregular graphs. J. Parallel. Distrib. Comput., 48 (1) (1998), 96129.CrossRefGoogle Scholar
[116]Zhang, D.; Yin, J.; Zhu, X.; Zhang, C.: Network representation learning: a survey. IEEE Transactions on Big Data, 2018a.Google Scholar
[117]Hu, C.; Cheng, L.; Sepulcre, J.; Johnson, K.A.; Fakhri, G.E.; Lu, Y.M.; Li, Q.: A spectral graph regression model for learning brain connectivity of Alzheimer's disease. PLoS ONE, 10 (5) (2015), e0128136.CrossRefGoogle ScholarPubMed
[118]Liu, R.; Nejati, H.; Cheung, N.-M.: Joint estimation of low-rank components and connectivity graph in high-dimensional graph signals: application to brain imaging. arXiv preprint arXiv:1801.02303, 2018.Google Scholar
[119]Chepuri, S.P.; Liu, S.; Leus, G.; Hero, A.O.: Learning sparse graphs under smoothness prior. in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, 2017, 65086512.CrossRefGoogle Scholar
[120]Dong, X.; Thanou, D.; Frossard, P.; Vandergheynst, P.: Learning Laplacian matrix in smooth graph signal representations. IEEE Trans. Signal. Process., 64 (23) (2016), 61606173.CrossRefGoogle Scholar
[121]Thanou, D.; Dong, X.; Kressner, D.; Frossard, P.: Learning heat diffusion graphs. IEEE Trans. Signal Inform. Proc. Over Netw., 3 (3) (2017), 484499.CrossRefGoogle Scholar
[122]Baingana, B.; Giannakis, G.B.: Tracking switched dynamic network topologies from information cascades. IEEE Trans. Signal. Process., 65 (4) (2016), 985997.CrossRefGoogle Scholar
[123]Li, R.; Wang, S.; Zhu, F.; Huang, J.: Adaptive graph convolutional neural networks. In Thirty-Second AAAI Conference on Artificial Intelligence, 2018b.Google Scholar
[124]Lake, B.; Tenenbaum, J.: Discovering structure by learning sparse graphs, 2010.Google Scholar
[125]Dong, X.; Thanou, D.; Rabbat, M.; Frossard, P.: Learning graphs from data: A signal representation perspective. IEEE Signal. Process. Mag., 36 (3) (2019), 4463.CrossRefGoogle Scholar
[126]Gao, X.; Hu, W.; Tang, J.; Liu, J.; Guo, Z: Optimized skeleton-based action recognition via sparsified graph regression. in Proceedings of the 27th ACM International Conference on Multimedia, 2019b, 601610.CrossRefGoogle Scholar
[127]Gao, H.; Chen, Y.; Ji, S.: Learning graph pooling and hybrid convolutional operations for text representations. arXiv preprint arXiv:1901.06965, 2019a.Google Scholar
[128]Hu, F.; Zhu, Y.; Wu, S.; Wang, L.; Tan, T.: Semi-supervised node classification via hierarchical graph convolutional networks. arXiv preprint arXiv:1902.06667, 2019.Google Scholar
[129]Ying, Z.; You, J.; Morris, C.; Ren, X.; Hamilton, W.; Leskovec, J.: Hierarchical graph representation learning with differentiable pooling. In Advances in Neural Information Processing Systems, 2018b, 48004810.Google Scholar
[130]Fahlman, S.E.: An empirical study of learning speed in back-propagation networks. 1988.Google Scholar
[131]Kuo, C.-C.J.: Understanding convolutional neural networks with a mathematical model. J. Vis. Commun. Image Represent., 41, (2016), 406413.CrossRefGoogle Scholar
[132]Zhang, Q.; Nian Wu, Y.; Zhu, S.-C.: Interpretable convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018b, 88278836.CrossRefGoogle Scholar
[133]Kuo, C.-C.J.; Zhang, M.; Li, S.; Duan, J.; Chen, Y.: Interpretable convolutional neural networks via feedforward design. arXiv preprint arXiv:1810.02786, 2018.Google Scholar
Figure 0

Fig. 1. Illustration of graph representation learning input and output.

Figure 1

Fig. 2. Illustration of a LGCL method [29].

Figure 2

Fig. 3. Illustration of the sub-graph selection process [29].

Figure 3

Fig. 4. The architecture of HGNN [73].

Figure 4

Table 1. Comparison of properties of graphs and hyper-graphs

Figure 5

Fig. 5. Illustration of graph and hypergraph structures [73].

Figure 6

Table 2. Summary of different graph embedding methods

Figure 7

Table 3. Summary of representative graph data sets

Figure 8

Table 4. Performance comparison of nine common graph embedding methods in vertex classification on Cora, Citeseer, and Wiki

Figure 9

Fig. 6. t-SNE visualization of different embedding methods on Cora. The seven different colors of points represent different classes of the nodes.

Figure 10

Table 5. Comparison of time used in training (s)

Figure 11

Table 6. Comparison of clustering quality of six graph embedding methods in terms of macro- and micro-F1 scores against three large graph data sets

Figure 12

Fig. 7. The node classification accuracy as a function of the embedding size.

Figure 13

Fig. 8. The node classification accuracy as a function of the embedding size.