HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Extract the Knowledge of Graph Neural Networks and Go Beyond it: An Effective Knowledge Distillation Framework

Cheng Yang Jiawei Liu Chuan Shi

Extract the Knowledge of Graph Neural Networks and Go Beyond it: An Effective Knowledge Distillation Framework

Abstract

Semi-supervised learning on graphs is an important problem in the machine learning area. In recent years, state-of-the-art classification methods based on graph neural networks (GNNs) have shown their superiority over traditional ones such as label propagation. However, the sophisticated architectures of these neural models will lead to a complex prediction mechanism, which could not make full use of valuable prior knowledge lying in the data, e.g., structurally correlated nodes tend to have the same class. In this paper, we propose a framework based on knowledge distillation to address the above issues. Our framework extracts the knowledge of an arbitrary learned GNN model (teacher model), and injects it into a well-designed student model. The student model is built with two simple prediction mechanisms, i.e., label propagation and feature transformation, which naturally preserves structure-based and feature-based prior knowledge, respectively. In specific, we design the student model as a trainable combination of parameterized label propagation and feature transformation modules. As a result, the learned student can benefit from both prior knowledge and the knowledge in GNN teachers for more effective predictions. Moreover, the learned student model has a more interpretable prediction process than GNNs. We conduct experiments on five public benchmark datasets and employ seven GNN models including GCN, GAT, APPNP, SAGE, SGC, GCNII and GLP as the teacher models. Experimental results show that the learned student model can consistently outperform its corresponding teacher model by 1.4% - 4.7% on average. Code and data are available at https://github.com/BUPT-GAMMA/CPF

Code Repositories

BUPT-GAMMA/CPF
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
node-classification-on-amz-computersCPF-ind-GAT
Accuracy: 85.5%
node-classification-on-amz-photoCPF-ind-GAT
Accuracy: 94.10%
node-classification-on-citeseer-with-publicCPF-tra-APPNP
Accuracy: 74.6%
node-classification-on-cora-05CPF-ind_APPNP
Accuracy: 77.3%
node-classification-on-cora-1CPF-ind-APPNP
Accuracy: 80.24%
node-classification-on-cora-3CPF-tra-GCNII
Accuracy: 84.18%
node-classification-on-cora-fixed-10-node-perCPF-tra-GCNII
Accuracy: 84.1%
node-classification-on-cora-fixed-5-node-perCPF-tra-APPNP
Accuracy: 80.26%
node-classification-on-cora-with-public-splitCPF-ind-APPNP
Accuracy: 85.3%
node-classification-on-pubmed-with-publicCPF-tra-GCNII
Accuracy: 83.20%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Extract the Knowledge of Graph Neural Networks and Go Beyond it: An Effective Knowledge Distillation Framework | Papers | HyperAI