[论文阅读笔记]2018_WWW_Dual Graph Convolutional Networks for Graph-Based Semi-Supervised Classification—(The World Wide Web Conference, 2018.04.23)-- Chenyi Zhuang, Qiang Ma

论文下载地址: https://doi.org/10.1145/3178876.3186116
发表期刊:IW3C2 (International World Wide Web Conference Committee)
Publish time: 2018.04
作者单位: Kyoto University, Kyoto, Japan
数据集:

  • 3 个citation network datasets

    • Citeseer
    • Cora
    • Pubmed
  • 1个knowledge graph dataset
    • NELL
  • Simplified NELL(作者自己根据NELL改的)

代码:

  • DGCN: https://github.com/ZhuangCY/Coding-NN
  • GCN: https://github.com/tkipf/gcn
  • PLANETOID: https://github.com/kimiyoung/planetoid
  • DeepWalk: https://github.com/phanein/deepwalk

其他人写的文章

  • 《Dual Graph Convolutional Networks for Graph-Based Semi-Supervised Classification》论文理解
  • Dual Graph Convolutional Networks for Graph-Based Semi-Supervised Classification
  • 点互信息(PMI)和正点互信息(PPMI)

简要概括创新点:(有很多实现的细节,值得去读)

  • (1) n this paper, we have proposed a Dual Graph Convolutional Net-work method for graph-based semi-supervised learning. (我们提出了一种基于图的半监督学习的对偶图卷积网络方法)
  • (2) In addition to the often-considered local consistency, our method uses the PPMI to encode the global consistency.(除了经常考虑的局部一致性之外,我们的方法还使用PPMI对全局一致性进行编码)
  • (3) Our work provides a solution for combining the prior knowledge learned from different views of raw data. (我们的工作为结合从原始数据的不同视图中获得的先验知识提供了一种解决方案。)
  • (4) C o n v A Conv_A ConvA是前人的工作, C o n v P Conv_P ConvP是本文的创新
  • (5) 为了缝合2个部分,提出了一个author设计的Loss函数
  • (6) 把NLP中的PPMI拿过来,缝合。为了缝合好,设计了很多细节,包括Random Walk得到频率矩阵 F F F,再由 F F F得到矩阵 P P P.
  • (7) adjacency matrix A A A 和 matrix P P P本质是diffusion
  • (8) 自己定义好 context, 和semantic information

Abstract

  • (1) we present a simple and scalable semi-supervised learning method(半监督学习) for graph-structured data in which only a very small portion of the training data are labeled.

  • (2) In particular, a dual graph convolutional neural network(双图卷积神经网络) method is devised to jointly consider the two essential assumptions of semi-supervised learning:

    • (1) local consistency (这一部分,是前人的工作)
    • (2)and global consistency.(本文创新的着力点)
  • (2) Accordingly, two convolutional neural networks are devised to embed the local-consistency-based and global-consistency-based knowledge, respectively。

  • (3) Given the different data transformations from the two networks, we then introduce an unsupervised temporal loss function(就是为了缝合2部分神经网络,而设计的) for the ensemble

KeyWords

  • Graph convolutional networks,
  • Semi-supervised learning,
  • Graph diffusion,
  • Adjacency matrix,
  • Pointwise mutual information

1 Introduction

(第一部分:前人已经做的工作)

  • (1) in this paper, we propose a new general semi-supervised learning algorithm that can be applied to different kinds of graphs

  • (2) Conventionally, graph-based semi-supervised learning could be defined through the following loss function

    where L 0 \mathcal{L}_0 L0 denotes the supervised loss with respect to the labeled data and L r e g \mathcal{L}_{reg} Lreg denotes the regularizer with respect to the graph structure

  • (3) ) Using an explicit graph-based regularizer L r e g \mathcal{L}_{reg} Lreg, the formulation of Eq. (1) smooths the label information in L 0 \mathcal{L}_0 L0 over the whole graph. (式(1)的公式在整个图上平滑了 L 0 \mathcal{L}_0 L0中的标签信息???)

    • For instance, a graph Laplacian regularization term is typically used in L r e g \mathcal{L}_{reg} Lreg [3] [5] [33], which relies on the prior assumption that connected nodes in the graph are likely to share the same label. (即图中的连通节点可能共享相同的标签)
    • However, this assumption might restrict the modeling capacity, as graph edges need not encode the node similarity, but could instead contain additional information.(种假设可能会限制建模的能力,因为图边不需要编码节点的相似性)
  • (4) In Eq. (2), instead of using an explicit graph-based regularizer in the loss function, a convolutional function Conv is derived to encode the graph structure directly.

  • (5) In approximate terms, the structure encoding Conv could be conducted in two domains:

    • (i) the graph vertex domain [2] [32]; and
    • (ii) the graph spectral domain































the structure encoding Conv







the graph vertex domain







the graph spectral domain






  • (6) 前人的不足----->没有考虑global consistency
    However, by using Eqs. (1) and (2), most of the related work have only considered the local consistency of a graph for knowledge embedding. To sufficiently embed the graph knowledge, we find that the global consistency of a graph has not been well investigated yet

(第二部分:在前人的基础上,本文做的工作和创新)

  • (7) Hence, in this paper, we propose a dual graph convolutional neural network method to jointly take both of them into consideration.
    The form of the loss function in the proposed strategy is
  • (8) Using the graph adjacency matrix and positive pointwise mutual information matrix, the two convolutional networks encode the local and global structure information
  • (9) Corresponding to the two essential assumptions in semi-supervised learning [32],
    • ConvA embeds the local-consistency-based knowledge (i.e., nearby data points are likely to have the same label),
    • whereas ConvP embeds the global-consistency-based knowledge (i.e., data points that occur in similar contexts tend to have the same label).









































2 assumption in semi-supervised learning







ConvA embeds the local-consistency-based knowledge







i.e., nearby data points are likely to have the same label







ConvP embeds the global-consistency-based knowledge







i.e., data points that occur in similar contexts tend to have the same label






  • (10) The output of either C o n v A Conv_A ConvA or C o n v P Conv_P ConvP is then used for supervised learning, e.g., L 0 ( C o n v A ) \mathcal{L}_0(Conv_A) L0(ConvA).
  • (11) However, to give better predictions, an ensemble-oriented regularizer L r e g ( C o n v A , C o n v P ) \mathcal{L}_{reg}(Conv_A, Conv_P) Lreg(ConvA,ConvP) for these transformations is derived.
    • By minimizing the difference between predictions from different transformations of an input sample, the regularizer combines the opinions of C o n v A Conv_A ConvA and C o n v P Conv_P ConvP .
    • Accordingly, L r e g ( C o n v A , C o n v P ) \mathcal{L}_{reg}(Conv_A, Conv_P) Lreg(ConvA,ConvP) is an unsupervised loss function

(第三部分:本文Contribution的一个Summary)

  • (12) In addition to the graph adjacency matrix-based convolution ConvA,

    • we propose a new convolutional neural network C o n v P Conv_P ConvP that depends on the positive pointwise mutual information (PPMI) matrix. (依赖于正点互信息矩阵PPMI)
    • Unlike C o n v A Conv_A ConvA, which embeds local-consistency-based knowledge,
    • we employ a random walk to construct a PPMI matrix that further embeds semantic information(语义信息), i.e., global-consistency-based knowledge.
  • (13) In addition to the supervised learning on a small portion of labeled training data (i.e., L 0 L_0 L0 in Eq. (3)), (除了对一小部分标注的训练数据(即式(3)中的L0)进行有监督学习外)
    • an unsupervised loss function (i.e., L r e g \mathcal{L}_{reg} Lreg in Eq. (3)) is proposed as a kind of regularizer to combine the output of different convolved data transformations.

2 Background to Graph-based Semi-supervised Learning(是前人的工作,知识基础)

  • (1) Semi-supervised learning considers the general problem of learning from labeled and unlabeled data. (半监督学习是干嘛的)

    • Given a set of data points X = x 1 , . . . , x l , x l + 1 , . . . , x n \mathcal{X} = {x_1, ..., x_l , x_{l + 1}, ..., x_n} X=x1,...,xl,xl+1,...,xn
    • and a set of labels C = { 1 , . . . , c } C = \{1, ...,c\} C={1,...,c},
    • the first l l l points have labels { y 1 , . . . , y l } ∈ C \{y1, ...,yl \} \in C {y1,...,yl}C and the remaining points are unlabeled.
    • The goal is to predict the labels of the unlabeled points
  • (2) In addition to the labeled and unlabeled points, graph-based semisupervised learning
    • also involves a given graph, denoted as an n × n n \times n n×n matrix A A A. Each entry a i , j ∈ A a_{i,j} \in A ai,jA indicates the similarity between data points x i x_i xi and x j x_j xj
    • the similarity can be derived by calculating the distances among data points [33], or may be explicitly given by structured data, such as knowledge graphs [29], citation graphs [14], hyperlinks between documents [20], and so on
  • (3) Therefore, the key problem of graph-based semi-supervised learning concerns how to embed the additional information of the graph for better label prediction.
  • (4) In approximate terms, we classify the different graph knowledge embeddings into two groups,
    • i.e., explicit and implicit graph-based semi-supervised learning































    knowledge embeddings







    explicit graph-based semi-supervised learning







    implicit graph-based semi-supervised leraning






2.1 Explicit Graph-based Semi-Supervised Learning

  • (1) Explicit graph-based semi-supervised learning uses a graph-based regularizer (i.e., L r e g \mathcal{L}_{reg} Lreg in Eq. (1)) to incorporate the information in the graph.

    • Eq. (4) presents an instance of the graph Laplacian regularizer,
    • where f ( ⋅ ) f (·) f() is the label prediction function (e.g., a neural network).
    • The unnormalized graph Laplacian is defined as∆ = A− D,
    • where A A A is the adjacency matrix and D D D is a diagonal matrix with each entry defined as d i , j = ∑ j a i , j d_{i,j} = \sum_j{a_{i,j}} di,j=jai,j.
  • (2) 前人的相似的工作

    • a label propagation algorithm based on Gaussian Random Fields
    • a PageRank-like algorithm is proposed to account for both local and global consistency in the graph
    • a sampling-based method. Instead of using the graph Laplacian∆, they derived a random walk-based sampling algorithm to obtain the positive and negative contexts for each data point
    • a feed-forward neural network method was then used fo knowledge embedding

2.2 Implicit Graph-Based Semi-Supervised Learning

2.2.1 Convolution in the vertex domain(点域的卷积)

  • (1) a data point xi will usually be transformed in a diffusive manner. A simple example of a k k k-hop localized linear transform is

    where b i , j b_{i,j} bi,j are some weights for filtering and N ( i , k ) \mathcal{N}(i,k) N(i,k) denotes the set of neighbors connected to x i x_i xi by a path of k k k or fewer edges.
  • (2) Such a convolution is built on the idea of a diffusion kernel, which can be thought of as a measure of the level of connectivity between any two nodes in a grap
    • diffusion-based convolution (基于扩散的卷积)

2.2.2 Convolution in the spectral domain(谱域的卷积)

  • (1) We first consider the most simple situation of scalar(标量) x i x_i xi.

    • In this context, the input X ∈ R n × 1 X\in R^{n\times 1} XRn×1 is considered as a signal defined on the graph with n n n nodes
  • (2) As shown in Eq. (6), the spectral convolution on a graph can then be defined as the multiplication of the signal X X X with a filter g θ = d i a g ( θ ) g_{\theta} = diag(\theta) gθ=diag(θ) parametrized by θ ∈ R n \theta \in R^n θRn in the graph Fourier domain(图傅里叶域)

    • where U U U is the matrix of eigenvectors(特征向量) of the normalized Laplacian matrix(归一化拉普拉斯矩阵) △ = I n − D − 1 2 A D − 1 2 \bigtriangleup = I_n - D^{-\frac{1}{2}} AD^{-\frac{1}{2}} =InD21AD21, which plays the role of the Fourier transform
    • when x i x_i xi has more than one feature, we can regard X X X as a signal with multiple input channels.(当有多个特征,就看作是多通道)
    • Eq.(6) is computationally expensive for large graphs.

2.2.3 Realtion between the two domains

  • (3) when the filter function g θ g_{\theta} gθ is approximated as an order- k k k polynomial, the spectral convolution can be interpreted as a k-hop diffusion convolution.
  • (4) Thus, by approximating g θ g_{\theta} gθ as an order-1 Chebyshev polynomial(切比雪夫多项式), after some derivation, it is equivalent to a 1-hop diffusion.
  • (5) In addition to adjacency matrix-based convolution, we further calculate a PPMI matrix for encoding the semantic information during the convolution.

3 Dual Graph Convolutional Networks(本文设计的模型)

3.1 Problem Definition and an Example

  • (1) the input of our model includes

    • a set of data points X = { x 1 , . . . , x l , x l + 1 , . . . , x n } \mathcal{X} =\{x_1, ..., x_l, x_{l+1}, ..., x_n\} X={x1,...,xl,xl+1,...,xn}
    • the labels {y_1, …, y_l} for the first l l l points,
    • and the graph structure.
  • (2) Assuming that each point has at most k k k features, the dataset is denoted as a matrix X ∈ R n × k X\in R^{n\times k} XRn×k
  • (3) the graph structure is represented by the adjacency matrix A ∈ R n × n A\in R^{n\times n} ARn×n


3.2 Local Consistency Convolution: C o n v A Conv_A ConvA

3.2这一部分是前人的工作,是一个基础,找到问题,然后提出解决的思路。解决的思路,就是3.3(本文的创新部分)

  • (1) we formulate the graph-structure-based convolution C o n v A Conv_A ConvA as a type of feed-forward neural network. Given the input feature matrix X X X and adjacency matrix A A A, the output of the i i i-th hidden layer of the network Z ( i ) Z^{(i)} Z(i) is defined as:

    • A ~ = A + I n \tilde{A}=A+I_n A~=A+In , where I n ∈ R n × n I_n\in R^{n\times n} InRn×n is the identity matrix, is the adjacency matrix with self-loops and D ~ i , i = ∑ j A ~ i , j \tilde{D}_{i, i} = \sum_j\tilde{A}_{i,j} D~i,i=jA~i,j
    • Accordingly, D ~ − 1 2 A ~ D ~ − 1 2 \tilde{D}^{-\frac{1}{2}} \tilde{A} \tilde{D}^{-\frac{1}{2}} D~21A~D~21 is the normalized adjacency matrix.
    • Z ( i − 1 ) Z^{(i-1)} Z(i1) is the output of the ( i − 1 ) (i-1) (i1)-th layer
    • and Z 0 = X . W ( i ) Z^{0}=X.W^{(i)} Z0=X.W(i) is the trainable parameters of the network
    • and σ ( ⋅ ) \sigma(\cdot) σ() denotes an activation function(e.g., ReLU, Sigmoid)
  • (2) he role of D ~ − 1 2 A ~ D ~ − 1 2 Z ( i − 1 ) \tilde{D}^{-\frac{1}{2}} \tilde{A} \tilde{D}^{-\frac{1}{2}} Z^{(i-1)} D~21A~D~21Z(i1) in Eq. (7) is to exactly conduct a 1-hop diffusion process in each layer. Namely, a node’s feature vector is enriched by linearly adding all the feature vectors of its neighbors. (精确地在每一层进行1跳的扩散过程。即通过将节点相邻节点的所有特征向量线性相加来丰富节点的特征向量)

开始讲故事了

  • (3) his discovery inspired the proposed concept. That is, this method can be further improved by reducing the exceptions to local consistency in semi-supervised learning: nearby points are likely to have the same label.(这个发现启发了这个提出的概念。也就是说,该方法可以进一步改进,通过减少半监督学习中局部一致性的异常:附近的点可能有相同的标签.)
  • (4) For example, from Figure 1a, as the directly connected data points x 8 x_8 x8 and x 30 x_{30} x30 have different labels, their convolved feature vectors should not be similar. However, Eq. (7) cannot deal with such an exception in an effective manner.
    (故事的开端,本文为什么要做这件事)
  • (5) From the visualized results in Figure 1d, as expected, x 8 x_8 x8 and x 30 x_{30} x30 are close together. However, they belong to different groups. To verify the proposed concept, we manually delete the edge between x 8 x_8 x8 and x 30 x_{30} x30, i.e., setting A [ 8 , 30 ] = A [ 30 , 8 ] = 0 A[8, 30] = A[30, 8] = 0 A[8,30]=A[30,8]=0. As a result, Figure 1e presents the new t-SNE distribution of all 34 data points, where x 8 x_8 x8 and x 30 x_{30} x30 are far apart. Hence, the attendant problem is how to automatically reduce the number of such exceptions.(作者猜想, x 8 x_8 x8 and x 30 x_{30} x30之间的边如果删掉,结果就会好。手动删除后,发现真的会好。那作者就要想办法,设计一种Model或者mechanism让程序或者method自动地删除这种情况)
  • (6) In the next subsection, we introduce a PPMI-based convolution method.(很自然的引出自己设计的method,开始讲创新点)
    • By encoding semantic information, this method allows different latent representations to be learnt for each data point.(通过对语义信息进行编码,该方法可以为每个数据点学习不同的潜在表示)

3.3 Global Consistency Convolution: C o n v P Conv_P ConvP

(本文的最大创新点)

  • (1) In addition to the graph structure information defined by the adjacency matrix A, we further apply PPMI to encode the semantic information, which is denoted as a matrix P ∈ R n × n P\in R^{n\times n} PRn×n.
  • (2) We first calculate a frequency matrix F F F using a random walk (自己设计的方法)
  • (3) Based on F, we then calculate P and explain why it leverages knowledge from the frequency to semantics.(本文的计算流程:先通过Random Walk得到 F F F, 再通过 F F F去计算 P P P)

3.3.1 Calculating frequency matrix F F F(本文创新的模型中,算法流程的起点–Random Walk)

  • (1) The Markov chain describing the sequence of nodes visited by a random walker is called a random walk.(描述随机漫步者访问的节点序列的马尔可夫链称为随机漫步。)阐述一下Random Walk是什么

    • If the random walker is on node x i x_i xi at time t t t, we define the state as s ( t ) = x i s(t) = x_i s(t)=xi.
    • The transition probability of jumping from the current node x i x_i xi to one of its neighbors x j x_j xj is denoted as p ( s ( t + 1 ) = x j ∣ s ( t ) = x i ) p(s(t + 1) = x_j |s(t) = x_i) p(s(t+1)=xjs(t)=xi).
  • (2) In our problem setting, given the adjacency matrix A A A, we assign:

  • (3) Algorithm 1 (阐述一些特点)

    • time complexity is O ( n γ q 2 ) O(n\gamma q^2) O(nγq2);
    • as the parameters γ \gamma γ and q q q are small integers, F F F can be calculated quickly.
    • Furthermore, the algorithm could be parallelized by conducting several random walks simultaneously on different parts of a graph(在不同的图上,可以并行)
  • (4) Random walks have been used (Random Walk 前人的一些应用)

    • as a similarity measure for a variety of problems in recommendation [11],
    • graph classification [1],
    • and semi-supervised learning [30].
    • In our method, we use a random walk to calculate the semantic similarity between nodes (本文用来计算相似性)

3.3.2 Calculating PPMI

  • (1) After calculating the frequency matrix F F F,

    • the i i i-th row in F is the row vector F i , : F_{i,:} Fi,:
    • and the j j j-th column in F F F is the column vector F : , j F_{:, j} F:,j
    • F i , : F_{i,:} Fi,:corresponds to a node x i x_i xi
    • and F : , j F_{:, j} F:,j corresponds to a context c j cj cj(本文定义的上下文).
  • (2) Based on Algorithm 1,

    • the contexts are defined as all nodes in X \mathcal{X} X.
    • The value of an entry F i , j F_{i, j} Fi,j is the number of times that x i x_i xi occurs in context c j c_j cj
  • (3) Based on F, we calculate the PPMI matrix P ∈ R n × n P\in R^{n\times n} PRn×n as:

  • (4) Applying Eq. (9) encodes the semantic information in P (通过公式9就得到了语义信息)

    • That is p i , j p_{i, j} pi,j is the estimated probability that node xi occurs in context c j c_j cj;
    • p i , ∗ p_{i,∗} pi, is the estimated probability of node x i x_i xi ;
    • and p ∗ , j p_{∗, j} p,j is the estimated probability of context c j c_j cj
    • Based on the definition of statistical independence, if x i x_i xi and c j c_j cj are independent (i.e., x i ) x_i) xi) occurs in c j c_j cj by pure random chance), then p i , j = p i , ∗ p ∗ , j p_{i, j}=p_{i,*} p_{*, j} pi,j=pi,p,j, and thus p m i i , j = 0 pmi_{i,j} = 0 pmii,j=0.
    • Accordingly, if there is a semantic relation between x i x_i xi and c j c_j cj , then p i , j p_{i,j} pi,j is expected to be greater than if KaTeX parse error: Expected group after '_' at position 3: xi_̲ and c j c_j cj are independent.(相比较于是独立的,有语义联系的话, p i , j 要 更 大 p_{i,j}要更大 pi,j)
      • Hence, when p i , j p_{i, j} pi,j > p i , ∗ p ∗ , j p_{i,∗} p_{∗, j} pi,p,j , p m i i , j pmi_{i, j} pmii,j should be positive.(如何区分positive)
      • If node xi is unrelated to context c j c_j cj , p m i i , j pmi_{i, j} pmii,j may be negative. (如何区分negative)
  • (5) As we are focusing on pairs ( x i , c j ) (x_i, c_j) (xi,cj) that have a semantic relation, our method uses a nonnegative p m i pmi pmi(本文使用非负的PMI).

  • (6) PPMI has been extensively investigated in terms of natural language processing (NLP) (说一下这个想法是从NLP里扒过来的)[4] [16] [28].

    • Indeed, the PPMI metric is known to perform well on semantic similarity tasks
  • (7) However, to the best of our knowledge, we are the first to introduce PPMI to the field of graph-based semi-supervised learning.

  • (8) Furthermore, using a novel PPMI-based convolution, our method applies the concept of global consistency: graph nodes that occur in similar contexts tend to have the same label.

    • (对比一下PPMI和adjacency的区别)
    • Figure 1c visualizes the normalized PPMI matrix P P P of the Karate club network.
    • Compared with the adjacency matrix of this network (shown in Figure 1b),
    • there are at least two obvious differences:
      • (1) P P P has reduced the effect of the hub nodes, e.g., x 0 x_0 x0 and x 33 x_{33} x33; and
      • (2) P P P has initiated more latent relations among different data points, which cannot be characterized by the adjacency matrix A A A.

3.3.3 PPMI-based convolution

  • (1) In addition to the convolution C o n v A Conv_A ConvA, which is based on the similarity defined by the adjacency matrix A A A, another feed-forward neural network C o n v P Conv_P ConvP is derived from the similarity defined by the PPMI matrix P P P.









































    本文Model







    Conv_A







    based on the similarity defined by the adjacency matrix A







    Conv_P







    derived from the similarity defined by the PPMI matrix P






    • This convolutional neural network is given by:
    • where P P P is the PPMI matrix and D i , i = ∑ j P i , j D_{i,i}=\sum_j P_{i,j} Di,i=jPi,j for normalization
  • (2) Obviously, applying diffusion based on such a node-contextual matrix P P P ensures global consistency(显然,基于这样的节点上下文矩阵P应用扩散可以确保全局一致性)(自圆其说,支持自己的观点)

  • (3) Additionally, by using the same neural network structure as C o n v A Conv_A ConvA, the two can be combined very concisely. (是一样的神经网络结构,区别是一个基于matrix A, 一个基于matrix P)

  • (4) 比较一下 Figure 1f, Figure 1d, Figure1e

3.4 Ensemble of Local and Global Consistencies (得想个办法,把2个部分缝合起来)

  • (1) To jointly consider the local consistency and global consistency for semi-supervised learning, we must overcome the challenge of

    • having very few labeled training data(有label的训练数据很少). That is, as the training data are limited, a general ensemble method (e.g., by concatenating the output of ConvA and ConvP ) cannot be utilized. (一般的集成方法不能用)
  • (2) supervised learning using training data, we further derive an unsupervised regularizer for the ensemble.(我们进一步推导出一个无监督正则化集成)
  • (3) Figure 2 presents the architecture of our dual graph convolutional networks method.
    • In addition to training C o n v A Conv_A ConvA using the labeled data (i.e., L 0 ( C o n v A ) \mathcal{L}_0 (Conv_A) L0(ConvA) in Eq. (3)
    • an unsupervised regularizer (i.e., L r e g ( C o n v A , C o n v P \mathcal{L}_{reg}(Conv_A, Conv_P Lreg(ConvA,ConvP in Eq. (3)) is introduced to train C o n v P Conv_P ConvP against the posterior probabilities of a previously trained model, i.e., the trained L 0 ( C o n v A ) \mathcal{L}_0(ConvA) L0(ConvA).(无监督正则化被引入,针对之前训练模型的后验概率,去训练 C o n v P Conv_P ConvP)

3.4.1 Calculating L 0 ( C o n v A ) \mathcal{L}_0(Conv_A) L0(ConvA)

  • (1) Assuming there are c c c different labels for prediction(预测 c c c个标签),

    • the softmax activation function is applied row-wise to the output Z A ∈ R n × c Z^A \in R^{n\times c} ZARn×c given by C o n v A Conv_A ConvA.
    • The output of the softmax layer is denoted as ˆZ A ∈ Rn×c. L0(ConvA), which evaluates the cross-entropy error over all labeled data points, is calculated as:

      • where Y L \mathcal{Y}_L YL is the set of data indices whose labels are observed for training
      • and Y ∈ R n × c Y \in R^{n×c} YRn×c is the ground truth.(真值)

3.4.2 Calculating L r e g ( C o n v A , C o n v P ) \mathcal{L}_{reg}(Conv_A, Conv_P) Lreg(ConvA,ConvP)

  • (1) The calculating of L r e g \mathcal{L}_{reg} Lreg is given by

    • Similar to C o n v A Conv_A ConvA, after applying the softmax activation function, the output of C o n v P Conv_P ConvP is denoted as Z ^ P ∈ R n × c \hat{Z}^P\in R^{n\times c} Z^PRn×c
    • Over all n n n data points, we introduce an unsupervised loss function that minimizes the mean squared differences between Z ^ P \hat{Z}^P Z^P and Z ^ A \hat{Z}^A Z^A.
  • (2) By looking at the formulation of Eq. (12), we could regard the unsupervised loss function as training C o n v P Conv_P ConvP against C o n v A Conv_A ConvA. (我们可以将无监督损失函数视为训练 C o n v P Conv_P ConvP去对抗 C o n v A Conv_A ConvA)

  • (3) That is, after the L 0 L_0 L0-based training (i.e., Eq. (11)), the softmaxed scores in Z ^ A ∈ R n × c \hat{Z}^A\in R^{n\times c} Z^ARn×c are then interpreted as a posterior distribution over the c c c labels. (被解释为c个标签之上的后验分布)

    • By minimizing the loss function in .Eq. (12),
    • despite different transformations by C o n v A , C o n v P Conv_A, Conv_P ConvA,ConvP, and random layer-wise dropout, the final predictions given by each model should then be the same.(每个模型给出的最终预测应该是相同的)
  • (4) As shown in Figure 2, the key to our model is to share the model parameters (i.e., neural network weights W W W in Eqs. (7) and (10)) in C o n v A Conv_A ConvA and C o n v P Conv_P ConvP .(本文Model的关键就是,2个部分的模型参数共享)

  • (5) Although sharing the same parameters W W W , the different diffusions (i.e., A A A and P P P)(其实matrix A A A 和 matrix P P P本质就是一个diffusion) and random dropout may cause the predictions of C o n v A Conv_A ConvA and C o n v P Conv_P ConvP (i.e., Z ^ A \hat{Z}^A Z^A and Z ^ P \hat{Z}^P Z^P ) to differ.

    • However, we know that each data point is assigned to only one class. (每个数据点只分配给一个类)
    • Therefore, the model (which is characterized by the parameters W W W ) is expected to give the same prediction from C o n v A Conv_A ConvA and C o n v P Conv_P ConvP , i.e., minimizing Eq. (12). (因此,通过公式(12),就给出一样的预测了。。。。。)
    • As a result, the trained parameters W W W have considered the opinions from both C o n v A Conv_A ConvA and C o n v P Conv_P ConvP .( W W W是同时考虑2个部分的得到的)
  • (6) By explicitly incorporating the prior knowledge (i.e., the diffusion matrices A and P in our method) during the data transformation stage, (通过在数据转换阶段显式合并先验知识)

    • Namely, by using multiple neural networks, different prior knowledge can be embedded during the data transformation stage.

3.4.3 the final model

  • (1) . Algorithm 2 describes the training process of our dual graph convolutional networks method

    • The loss function is defined as a weighted sum of L 0 ( C o n v A ) \mathcal{L}_0(Conv_A) L0(ConvA) and L r e g ( C o n v A , C o n v P ) \mathcal{L}_{reg}(ConvA, Conv_P) Lreg(ConvA,ConvP)
    • a dynamic weight function is devised to implement the idea described above
      • That is, at the beginning of the training process (i.e., small t t t), the loss function is mainly dominated by the supervised entry L 0 L_0 L0.
      • After obtaining a posterior distribution over the labels using C o n v A Conv_A ConvA, increasing λ ( t ) \lambda(t) λ(t) forces our model to simultaneously consider the knowledge encoded in C o n v P Conv_P ConvP .(同时考虑 C o n v P Conv_P ConvP)
  • (2) our implementation uses Batch Gradient Descent (BGD),

    • in which the full training dataset is used for every training iteration.
    • Although BGD is relatively slow,
    • it is guaranteed to converge to the global minimum for convex error surfaces (但对于凸误差面,它保证收敛到全局最小值)
    • and to a local minimum for non-convex surfaces (对于非凸误差面,它保证收敛到局部最小值)

4 Experiments

4.1 Dataset

  • 3 个citation network datasets

    • Citeseer
    • Cora
    • Pubmed
  • 1个knowledge graph dataset
    • NELL
  • Simplified NELL(作者自己根据NELL改的)

4.1.1 Citeseer

  • Only 3.6% of the nodes are labeled for training

4.1.2 Cora

  • Only 5.2% of the nodes are labeled for training

4.1.3 Pubmed

  • Only 0.3% of the nodes are labeled for training

4.1.4 NELL(Never Ending Language Learning(NELL) knowledge graph)

  • each relation in NELL is described as a triplet ( e h , r , e t ) (e_h, r, e_t) (eh,r,et)
  • By spliting each ( e h , r , e t ) (e_h, r, e_t) (eh,r,et) into two edges ( e h , r 1 ) (e_h, r_1) (eh,r1) and ( r 2 , e t ) (r_2, e_t) (r2,et), we obtain a graph …(author自己手动拆分的)
  • Only a single data point per class is labeled for training

4.1.5 Simplified NELL

  • In simplified NELL, the relation information (i.e., r r r) has been removed and edges among entities have been directly added.
  • By counting the co-occurrences of each (eh, et ) pair in all triplets, a weighted adjacency matrix A is constructed.

4.2 Methods for Comparison

4.2.1 DGCN

  • (1) This is the proposed method, as described in Algorithm 2. In our Dual Graph Convolutional Networks (DGCN) implementation, both C o n v A Conv_A ConvA and C o n v P Conv_P ConvP have two hidden layers. (有2层隐藏层)

    • Namely, there are two separate W W W vectors, W ( 1 ) W^{(1)} W(1) and W ( 2 ) W^{(2)} W(2), that need training in Algorithm 2.
  • (2) Table 2 presents detailed information about the implementation of our method for each dataset, including
    • (1) size of the hidden layer;
    • (2) layer-wise dropout rate;
    • (3) window size w w w in Algorithm 1;and
    • (4) learning rate η η η.

4.2.2 GCN(Graph Convolutional Networks)

4.2.3 PLANETOID

  • Inspired by the Skipgram model [18] from NLP, PLANETOID [30] embeds the graph information using positive and negative samplings.

    • During sampling, both the label information and graph structure are considered

4.2.4 DeepWalk

  • By taking random walks on a graph, different paths are generated.
  • By regarding the paths as “sentences, ” DeepWalk [21] generalizes language modeling techniques from sequences of words to paths in a graph. (通过将路径视为“句子”,DeepWalk[21]将语言建模技术从单词序列推广到图中的路径)(因为都是符合某个概率分布的得到的序列)

4.3 Results


4.4 Effect of Regularization Weight λ ( t ) \lambda (t) λ(t)


4.5 4.5 Effect of a Shifted PPMI Matrix P P P


  • (1) Eq. (13) presents the calculation of a shifted PPMI matrix, first introduced in [16] for word embedding
  • (2) On the basis of the derivation in [16], the value of k k k indicates the number of negative samplings required to calculate each entry of P P P(k的值表示计算P的每一项所需的负抽样个数)
  • (3) In research on semi-supervised learning, we are the first to verify whether such a shift can also be applied to understand a graph.(在半监督学习的研究中,我们首先验证了这种转移是否也可以应用于理解图)

5 Conclusions(大总结,高层次的干货)

  • (1) n this paper, we have proposed a Dual Graph Convolutional Net-work method for graph-based semi-supervised learning. (我们提出了一种基于图的半监督学习的对偶图卷积网络方法)
  • (2) In addition to the often-considered local consistency, our method uses the PPMI to encode the global consistency.(除了经常考虑的局部一致性之外,我们的方法还使用PPMI对全局一致性进行编码)
  • (3) Our work provides a solution for combining the prior knowledge learned from different views of raw data. (我们的工作为结合从原始数据的不同视图中获得的先验知识提供了一种解决方案。)

a Variants of our methods(其实就是消融实验)

  • DGCN: The original method used in our experiments.
  • DGCN-1: When the neural network parameters (i.e., W ( h ) s W ^{(h)}s W(h)s) are not shared between C o n v A Conv_A ConvA and C o n v P Conv_P ConvP
  • DGCN-2: Instead of Eq. (12), we use concatenation to form the ensemble.
    • Namely, the final predictions are derived from the latent representations: C o n v A ⊕ C o n v P Conv_A \oplus Conv_P ConvAConvP . Before softmax, a dense layer is added to ensure the same shape as the label matrix Y ∈ R n × c Y \in R^{n\times c} YRn×c
  • DGCN-3: When the parameters are not shared between C o n v A Conv_A ConvA and C o n v P Conv_P ConvP , and concatenation is used.
  • DGCN-4: Without ensemble, when only C o n v P Conv_P ConvP is used.

Acknowledgments

References

2018_WWW_Dual Graph Convolutional Networks for Graph-Based Semi-Supervised Classification相关推荐

  1. 21.Selective Attention Based Graph Convolutional Networks for Aspect-Level Sentiment Classification阅

    21.Selective Attention Based Graph Convolutional Networks for Aspect-Level Sentiment Classification阅 ...

  2. 知识图谱论文阅读(二十三)【SIGIR2020】Multi-behavior Recommendation with Graph Convolutional Networks

    题目: Multi-behavior Recommendation with Graph Convolutional Networks 论文地址: 论文代码: 想法 将相同行为的交互方式进行了聚合来计 ...

  3. 论文阅读课1-Attention Guided Graph Convolutional Networks for Relation Extraction(关系抽取,图卷积,ACL2019,n元)

    文章目录 abstract 1.introduction 1.1 dense connection+GCN 1.2 效果突出 1.3 contribution 2.Attention Guided G ...

  4. 【论文阅读】Spatio-Temporal Graph Convolutional Networks:...Traffic Forecasting[时空图卷积网络:用于交通预测的深度学习框架](1)

    [论文阅读]Spatio-Temporal Graph Convolutional Networks: A Deep Learning Framework for Traffic Forecastin ...

  5. 基于图卷积的价格推荐论文(Price-aware Recommendation with Graph Convolutional Networks)

    基于图卷积的价格推荐论文概述 ICDE2020非常好的文章 Price-aware Recommendation with Graph Convolutional Networks paper lin ...

  6. 2018 ESWC | Modeling Relational Data with Graph Convolutional Networks

    2018 ESWC | Modeling Relational Data with Graph Convolutional Networks Paper: https://arxiv.org/pdf/ ...

  7. Attentional Graph Convolutional Networks for Knowledge Concept Recommendation in MOOCs in a Heter---

    Attentional Graph Convolutional Networks for Knowledge Concept Recommendation in MOOCs in a Heteroge ...

  8. 图融合GCN(Graph Convolutional Networks)

    图融合GCN(Graph Convolutional Networks) 数据其实是图(graph),图在生活中无处不在,如社交网络,知识图谱,蛋白质结构等.本文介绍GNN(Graph Neural ...

  9. Semi-Supervised Classification with Graph Convolutional Networks

    Kipf, Thomas N., and Max Welling. "Semi-supervised classification with graph convolutional netw ...

最新文章

  1. Python 生成器 迭代器
  2. 【图像处理】——图像的灰度化处理(Python实现三种方法——最大值法、平均值法、加权均值法、gamma校正)
  3. stm32的语音识别_基于stm32循迹避障语音控制金属探测蓝牙小车设计(原理图+pcb+源码+参考文档)...
  4. qt 判断路径非英文符号与非英文字符_Qt中,软件多语言国际化翻译的方法与步骤...
  5. 问题 H: 判断闰年
  6. Twisted入门教程(10)
  7. C#之获取网卡IP地址
  8. ios13.7 iPhone/iPad Location-cleaned 驱动下载
  9. 高等数学-函数在线绘图工具推荐(精选好用)
  10. vCenter6.5+虚拟ISCSi
  11. C++ get()和put()读写文件详解
  12. 设计模式基础学习之单例模式
  13. Linux中find命令基本使用方法
  14. 聚焦传统网络,学习SDN基础和案例
  15. Linux vi vim 宏 批量操作 重复操作 宏录制
  16. k8s 使用HPA 进行弹性扩容pod节点,
  17. 优秀产品经理的18种能力
  18. 公众号如何运营?教你几招超实用的公众号运营方法
  19. 关于CMake的使用介绍
  20. C语言输出 1到20 的阶乘之和

热门文章

  1. 师者,传道授业解惑也――记侯捷在Yahoo中国的技术讲座
  2. 只需掌握这16点,学会seo站内优化
  3. 罗技压枪脚本教学(二)
  4. 如何查询物流信息并导出表格保存
  5. 股价暴涨59%后,美股二手车平台Carvana在短期内还会进一步上涨?
  6. JavaMail(三)——发送HTML格式带附件的邮件
  7. 两条直线的交点 c语言,计算两条直线的交点(C#)
  8. Educational Codeforces Round 90 (Rated for Div. 2) A. Donut Shops (1000)
  9. H264/AVC NAL单元和解码
  10. DIY智能家居设备:基于esp32和阿里云云智能APP完成智能灯泡的开发