HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

SparseSwin: Swin Transformer with Sparse Transformer Block

Krisna Pinasthika Blessius Sheldo Putra Laksono Riyandi Banovbi Putera Irsal Syifa Hukma Shabiyya Novanto Yudistira

SparseSwin: Swin Transformer with Sparse Transformer Block

Abstract

Advancements in computer vision research have put transformer architecture as the state of the art in computer vision tasks. One of the known drawbacks of the transformer architecture is the high number of parameters, this can lead to a more complex and inefficient algorithm. This paper aims to reduce the number of parameters and in turn, made the transformer more efficient. We present Sparse Transformer (SparTa) Block, a modified transformer block with an addition of a sparse token converter that reduces the number of tokens used. We use the SparTa Block inside the Swin T architecture (SparseSwin) to leverage Swin capability to downsample its input and reduce the number of initial tokens to be calculated. The proposed SparseSwin model outperforms other state of the art models in image classification with an accuracy of 86.96%, 97.43%, and 85.35% on the ImageNet100, CIFAR10, and CIFAR100 datasets respectively. Despite its fewer parameters, the result highlights the potential of a transformer architecture using a sparse token converter with a limited number of tokens to optimize the use of the transformer and improve its performance.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
image-classification-on-cifar-10SparseSwin
Percentage correct: 97.43
image-classification-on-cifar-100SparseSwin
PARAMS: 17.58M
Percentage correct: 85.35
image-classification-on-imagenet-100SparseSwin with L2
Params: 17.58M
Percentage correct: 86.96

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
SparseSwin: Swin Transformer with Sparse Transformer Block | Papers | HyperAI