Note
Go to the end to download the full example code
Graph Convolutional Network
Author: Qi Huang, Minjie Wang, Yu Gai, Quan Gan, Zheng Zhang
Warning
The tutorial aims at gaining insights into the paper, with code as a mean of explanation. The implementation thus is NOT optimized for running efficiency. For recommended implementation, please refer to the official examples.
This is a gentle introduction of using DGL to implement Graph Convolutional
Networks (Kipf & Welling et al., Semi-Supervised Classification with Graph
Convolutional Networks). We explain
what is under the hood of the GraphConv
module.
The reader is expected to learn how to define a new GNN layer using DGL’s
message passing APIs.
Model Overview
GCN from the perspective of message passing
We describe a layer of graph convolutional neural network from a message passing perspective; the math can be found here. It boils down to the following step, for each node \(u\):
1) Aggregate neighbors’ representations \(h_{v}\) to produce an intermediate representation \(\hat{h}_u\). 2) Transform the aggregated representation \(\hat{h}_{u}\) with a linear projection followed by a non-linearity: \(h_{u} = f(W_{u} \hat{h}_u)\).
We will implement step 1 with DGL message passing, and step 2 by
PyTorch nn.Module
.
GCN implementation with DGL
We first define the message and reduce function as usual. Since the aggregation on a node \(u\) only involves summing over the neighbors’ representations \(h_v\), we can simply use builtin functions:
import os
os.environ["DGLBACKEND"] = "pytorch"
import dgl
import dgl.function as fn
import torch as th
import torch.nn as nn
import torch.nn.functional as F
from dgl import DGLGraph
gcn_msg = fn.copy_u(u="h", out="m")
gcn_reduce = fn.sum(msg="m", out="h")
We then proceed to define the GCNLayer module. A GCNLayer essentially performs message passing on all the nodes then applies a fully-connected layer.
Note
This is showing how to implement a GCN from scratch. DGL provides a more
efficient builtin GCN layer module
.
class GCNLayer(nn.Module):
def __init__(self, in_feats, out_feats):
super(GCNLayer, self).__init__()
self.linear = nn.Linear(in_feats, out_feats)
def forward(self, g, feature):
# Creating a local scope so that all the stored ndata and edata
# (such as the `'h'` ndata below) are automatically popped out
# when the scope exits.
with g.local_scope():
g.ndata["h"] = feature
g.update_all(gcn_msg, gcn_reduce)
h = g.ndata["h"]
return self.linear(h)
The forward function is essentially the same as any other commonly seen NNs
model in PyTorch. We can initialize GCN like any nn.Module
. For example,
let’s define a simple neural network consisting of two GCN layers. Suppose we
are training the classifier for the cora dataset (the input feature size is
1433 and the number of classes is 7). The last GCN layer computes node embeddings,
so the last layer in general does not apply activation.
class Net(nn.Module):
def __init__(self):
super(Net, self).__init__()
self.layer1 = GCNLayer(1433, 16)
self.layer2 = GCNLayer(16, 7)
def forward(self, g, features):
x = F.relu(self.layer1(g, features))
x = self.layer2(g, x)
return x
net = Net()
print(net)
Net(
(layer1): GCNLayer(
(linear): Linear(in_features=1433, out_features=16, bias=True)
)
(layer2): GCNLayer(
(linear): Linear(in_features=16, out_features=7, bias=True)
)
)
We load the cora dataset using DGL’s built-in data module.
from dgl.data import CoraGraphDataset
def load_cora_data():
dataset = CoraGraphDataset()
g = dataset[0]
features = g.ndata["feat"]
labels = g.ndata["label"]
train_mask = g.ndata["train_mask"]
test_mask = g.ndata["test_mask"]
return g, features, labels, train_mask, test_mask
When a model is trained, we can use the following method to evaluate the performance of the model on the test dataset:
def evaluate(model, g, features, labels, mask):
model.eval()
with th.no_grad():
logits = model(g, features)
logits = logits[mask]
labels = labels[mask]
_, indices = th.max(logits, dim=1)
correct = th.sum(indices == labels)
return correct.item() * 1.0 / len(labels)
We then train the network as follows:
import time
import numpy as np
g, features, labels, train_mask, test_mask = load_cora_data()
# Add edges between each node and itself to preserve old node representations
g.add_edges(g.nodes(), g.nodes())
optimizer = th.optim.Adam(net.parameters(), lr=1e-2)
dur = []
for epoch in range(50):
if epoch >= 3:
t0 = time.time()
net.train()
logits = net(g, features)
logp = F.log_softmax(logits, 1)
loss = F.nll_loss(logp[train_mask], labels[train_mask])
optimizer.zero_grad()
loss.backward()
optimizer.step()
if epoch >= 3:
dur.append(time.time() - t0)
acc = evaluate(net, g, features, labels, test_mask)
print(
"Epoch {:05d} | Loss {:.4f} | Test Acc {:.4f} | Time(s) {:.4f}".format(
epoch, loss.item(), acc, np.mean(dur)
)
)
NumNodes: 2708
NumEdges: 10556
NumFeats: 1433
NumClasses: 7
NumTrainingSamples: 140
NumValidationSamples: 500
NumTestSamples: 1000
Done loading data from cached files.
/home/ubuntu/prod-doc/readthedocs.org/user_builds/dgl/envs/latest/lib/python3.8/site-packages/numpy/core/fromnumeric.py:3464: RuntimeWarning: Mean of empty slice.
return _methods._mean(a, axis=axis, dtype=dtype,
/home/ubuntu/prod-doc/readthedocs.org/user_builds/dgl/envs/latest/lib/python3.8/site-packages/numpy/core/_methods.py:192: RuntimeWarning: invalid value encountered in scalar divide
ret = ret.dtype.type(ret / rcount)
Epoch 00000 | Loss 1.9549 | Test Acc 0.3270 | Time(s) nan
Epoch 00001 | Loss 1.7860 | Test Acc 0.4900 | Time(s) nan
Epoch 00002 | Loss 1.6123 | Test Acc 0.5510 | Time(s) nan
Epoch 00003 | Loss 1.4706 | Test Acc 0.6190 | Time(s) 0.0175
Epoch 00004 | Loss 1.3491 | Test Acc 0.6570 | Time(s) 0.0175
Epoch 00005 | Loss 1.2461 | Test Acc 0.6610 | Time(s) 0.0175
Epoch 00006 | Loss 1.1572 | Test Acc 0.6640 | Time(s) 0.0175
Epoch 00007 | Loss 1.0793 | Test Acc 0.6710 | Time(s) 0.0175
Epoch 00008 | Loss 1.0089 | Test Acc 0.6790 | Time(s) 0.0174
Epoch 00009 | Loss 0.9431 | Test Acc 0.6930 | Time(s) 0.0174
Epoch 00010 | Loss 0.8801 | Test Acc 0.7050 | Time(s) 0.0174
Epoch 00011 | Loss 0.8196 | Test Acc 0.7110 | Time(s) 0.0174
Epoch 00012 | Loss 0.7623 | Test Acc 0.7230 | Time(s) 0.0174
Epoch 00013 | Loss 0.7081 | Test Acc 0.7170 | Time(s) 0.0174
Epoch 00014 | Loss 0.6570 | Test Acc 0.7220 | Time(s) 0.0174
Epoch 00015 | Loss 0.6089 | Test Acc 0.7280 | Time(s) 0.0174
Epoch 00016 | Loss 0.5638 | Test Acc 0.7190 | Time(s) 0.0174
Epoch 00017 | Loss 0.5217 | Test Acc 0.7130 | Time(s) 0.0174
Epoch 00018 | Loss 0.4827 | Test Acc 0.7160 | Time(s) 0.0174
Epoch 00019 | Loss 0.4464 | Test Acc 0.7100 | Time(s) 0.0174
Epoch 00020 | Loss 0.4121 | Test Acc 0.7130 | Time(s) 0.0174
Epoch 00021 | Loss 0.3793 | Test Acc 0.7140 | Time(s) 0.0174
Epoch 00022 | Loss 0.3484 | Test Acc 0.7150 | Time(s) 0.0174
Epoch 00023 | Loss 0.3200 | Test Acc 0.7200 | Time(s) 0.0174
Epoch 00024 | Loss 0.2939 | Test Acc 0.7250 | Time(s) 0.0174
Epoch 00025 | Loss 0.2697 | Test Acc 0.7320 | Time(s) 0.0174
Epoch 00026 | Loss 0.2473 | Test Acc 0.7360 | Time(s) 0.0174
Epoch 00027 | Loss 0.2264 | Test Acc 0.7370 | Time(s) 0.0174
Epoch 00028 | Loss 0.2070 | Test Acc 0.7380 | Time(s) 0.0174
Epoch 00029 | Loss 0.1891 | Test Acc 0.7360 | Time(s) 0.0174
Epoch 00030 | Loss 0.1726 | Test Acc 0.7360 | Time(s) 0.0174
Epoch 00031 | Loss 0.1575 | Test Acc 0.7370 | Time(s) 0.0174
Epoch 00032 | Loss 0.1437 | Test Acc 0.7390 | Time(s) 0.0174
Epoch 00033 | Loss 0.1311 | Test Acc 0.7390 | Time(s) 0.0174
Epoch 00034 | Loss 0.1196 | Test Acc 0.7410 | Time(s) 0.0174
Epoch 00035 | Loss 0.1090 | Test Acc 0.7400 | Time(s) 0.0174
Epoch 00036 | Loss 0.0994 | Test Acc 0.7370 | Time(s) 0.0174
Epoch 00037 | Loss 0.0906 | Test Acc 0.7380 | Time(s) 0.0174
Epoch 00038 | Loss 0.0826 | Test Acc 0.7370 | Time(s) 0.0174
Epoch 00039 | Loss 0.0754 | Test Acc 0.7350 | Time(s) 0.0174
Epoch 00040 | Loss 0.0688 | Test Acc 0.7360 | Time(s) 0.0174
Epoch 00041 | Loss 0.0629 | Test Acc 0.7360 | Time(s) 0.0174
Epoch 00042 | Loss 0.0576 | Test Acc 0.7340 | Time(s) 0.0174
Epoch 00043 | Loss 0.0527 | Test Acc 0.7340 | Time(s) 0.0174
Epoch 00044 | Loss 0.0483 | Test Acc 0.7340 | Time(s) 0.0174
Epoch 00045 | Loss 0.0443 | Test Acc 0.7330 | Time(s) 0.0174
Epoch 00046 | Loss 0.0407 | Test Acc 0.7330 | Time(s) 0.0174
Epoch 00047 | Loss 0.0374 | Test Acc 0.7320 | Time(s) 0.0174
Epoch 00048 | Loss 0.0344 | Test Acc 0.7330 | Time(s) 0.0174
Epoch 00049 | Loss 0.0318 | Test Acc 0.7280 | Time(s) 0.0174
GCN in one formula
Mathematically, the GCN model follows this formula:
\(H^{(l+1)} = \sigma(\tilde{D}^{-\frac{1}{2}}\tilde{A}\tilde{D}^{-\frac{1}{2}}H^{(l)}W^{(l)})\)
Here, \(H^{(l)}\) denotes the \(l^{th}\) layer in the network, \(\sigma\) is the non-linearity, and \(W\) is the weight matrix for this layer. \(\tilde{D}\) and \(\tilde{A}\) are separately the degree and adjacency matrices for the graph. With the superscript ~, we are referring to the variant where we add additional edges between each node and itself to preserve its old representation in graph convolutions. The shape of the input \(H^{(0)}\) is \(N \times D\), where \(N\) is the number of nodes and \(D\) is the number of input features. We can chain up multiple layers as such to produce a node-level representation output with shape \(N \times F\), where \(F\) is the dimension of the output node feature vector.
The equation can be efficiently implemented using sparse matrix multiplication kernels (such as Kipf’s pygcn code). The above DGL implementation in fact has already used this trick due to the use of builtin functions.
Note that the tutorial code implements a simplified version of GCN where we replace \(\tilde{D}^{-\frac{1}{2}}\tilde{A}\tilde{D}^{-\frac{1}{2}}\) with \(\tilde{A}\). For a full implementation, see our example here.
Total running time of the script: (0 minutes 1.605 seconds)