TransFER: Learning Relation-aware Facial Expression Representations with Transformers
TransFER: Learning Relation-aware Facial Expression Representations with Transformers
Facial expression recognition (FER) has received increasing interest in computer vision. We propose the Trans-FER model which can learn rich relation-aware local representations. It mainly consists of three components: Multi-Attention Dropping (MAD), ViT-FER, and Multi-head Self-Attention Dropping (MSAD). First, local patches play an important role in distinguishing various expressions, however, …