HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Converting Transformers into DGNNs Form

Jie Zhang Mao-Hsuan Mao Bo-Wei Chiu Min-Te Sun

Converting Transformers into DGNNs Form

Abstract

Recent advances in deep learning have established Transformer architectures as the predominant modeling paradigm. Central to the success of Transformers is the self-attention mechanism, which scores the similarity between query and key matrices to modulate a value matrix. This operation bears striking similarities to digraph convolution, prompting an investigation into whether digraph convolution could serve as an alternative to self-attention. In this study, we formalize this concept by introducing a synthetic unitary digraph convolution based on the digraph Fourier transform. The resulting model, which we term Converter, effectively converts a Transformer into a Directed Graph Neural Network (DGNN) form. We have tested Converter on Long-Range Arena benchmark, long document classification, and DNA sequence-based taxonomy classification. Our experimental results demonstrate that Converter achieves superior performance while maintaining computational efficiency and architectural simplicity, which establishes it as a lightweight yet powerful Transformer variant.

Code Repositories

hazdzz/Converter
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
long-range-modeling-on-lraConverter
Avg: 75.94
Image: 61.02
ListOps: 60.38
Pathfinder: 88.43
Retrieval: 83.41
Text: 86.44

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Converting Transformers into DGNNs Form | Papers | HyperAI