Poster
in
Workshop: UniReps: Unifying Representations in Neural Models
Deep Multimodal Emotion Recognition using Modality Aware Attention Network for Unifying Representations in Neural Models
Sungpil Woo · MUHAMMAD ZUBAIR · Sunhwan Lim · Daeyoung Kim
Abstract:
This paper introduces a multi-modal emotion recognition system aimed at enhancing emotion recognition by integrating representations from physiological signals. To accomplish this goal, we introduce a modality aware attention network to extract emotion-specific features by influencing and aligning the representation spaces of various modalities into a unified entity. Through a series of experiments and visualizations conducted on the AMIGO dataset, we demonstrate the efficacy of our proposed methodology for emotion classification, highlighting its capability to provide comprehensive representations of physiological signals.
Chat is not available.