HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

COGMEN: COntextualized GNN based Multimodal Emotion recognitioN

Abhinav Joshi Ashwani Bhat Ayush Jain Atin Vikram Singh Ashutosh Modi

COGMEN: COntextualized GNN based Multimodal Emotion recognitioN

Abstract

Emotions are an inherent part of human interactions, and consequently, it is imperative to develop AI systems that understand and recognize human emotions. During a conversation involving various people, a person's emotions are influenced by the other speaker's utterances and their own emotional state over the utterances. In this paper, we propose COntextualized Graph Neural Network based Multimodal Emotion recognitioN (COGMEN) system that leverages local information (i.e., inter/intra dependency between speakers) and global information (context). The proposed model uses Graph Neural Network (GNN) based architecture to model the complex dependencies (local and global information) in a conversation. Our model gives state-of-the-art (SOTA) results on IEMOCAP and MOSEI datasets, and detailed ablation experiments show the importance of modeling information at both levels.

Code Repositories

m-muaz/Cogmen_SLT
pytorch
Mentioned in GitHub
exploration-lab/cogmen
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
emotion-recognition-in-conversation-on-7COGMEN
Weighted F1: 84.50
emotion-recognition-in-conversation-on-cmu-2COGMEN
Weighted F1: 43.90
multimodal-emotion-recognition-on-iemocap-4COGMEN
Weighted F1: 84.50

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
COGMEN: COntextualized GNN based Multimodal Emotion recognitioN | Papers | HyperAI