HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Local Multi-Head Channel Self-Attention for Facial Expression Recognition

Roberto Pecoraro Valerio Basile Viviana Bono Sara Gallo

Local Multi-Head Channel Self-Attention for Facial Expression Recognition

Abstract

Since the Transformer architecture was introduced in 2017 there has been many attempts to bring the self-attention paradigm in the field of computer vision. In this paper we propose a novel self-attention module that can be easily integrated in virtually every convolutional neural network and that is specifically designed for computer vision, the LHC: Local (multi) Head Channel (self-attention). LHC is based on two main ideas: first, we think that in computer vision the best way to leverage the self-attention paradigm is the channel-wise application instead of the more explored spatial attention and that convolution will not be replaced by attention modules like recurrent networks were in NLP; second, a local approach has the potential to better overcome the limitations of convolution than global attention. With LHC-Net we managed to achieve a new state of the art in the famous FER2013 dataset with a significantly lower complexity and impact on the "host" architecture in terms of computational cost when compared with the previous SOTA.

Code Repositories

bodhis4ttva/lhc_net
Official
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
facial-expression-recognition-on-fer2013LHC-Net
Accuracy: 74.42

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp