A systematic review on affective computing: Emotion models, databases, and recent advances

Y Wang, W Song, W Tao, A Liotta, D Yang, X Li, S Gao… - Information …, 2022 - Elsevier
Affective computing conjoins the research topics of emotion recognition and sentiment
analysis, and can be realized with unimodal or multimodal data, consisting primarily of …

Deep facial expression recognition: A survey

S Li, W Deng - IEEE transactions on affective computing, 2020 - ieeexplore.ieee.org
With the transition of facial expression recognition (FER) from laboratory-controlled to
challenging in-the-wild conditions and the recent success of deep learning techniques in …

Progressive modality reinforcement for human multimodal emotion recognition from unaligned multimodal sequences

F Lv, X Chen, Y Huang, L Duan… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Human multimodal emotion recognition involves time-series data of different modalities,
such as natural language, visual motions, and acoustic behaviors. Due to the variable …

A review on methods and applications in multimodal deep learning

S Jabeen, X Li, MS Amin, O Bourahla, S Li… - ACM Transactions on …, 2023 - dl.acm.org
Deep Learning has implemented a wide range of applications and has become increasingly
popular in recent years. The goal of multimodal deep learning (MMDL) is to create models …

Deep learning-based late fusion of multimodal information for emotion classification of music video

YR Pandeya, J Lee - Multimedia Tools and Applications, 2021 - Springer
Affective computing is an emerging area of research that aims to enable intelligent systems
to recognize, feel, infer and interpret human emotions. The widely spread online and off-line …

In search of a robust facial expressions recognition model: A large-scale visual cross-corpus study

E Ryumina, D Dresvyanskiy, A Karpov - Neurocomputing, 2022 - Elsevier
Many researchers have been seeking robust emotion recognition system for already last two
decades. It would advance computer systems to a new level of interaction, providing much …

Learning modality-specific and-agnostic representations for asynchronous multimodal language sequences

D Yang, H Kuang, S Huang, L Zhang - Proceedings of the 30th ACM …, 2022 - dl.acm.org
Understanding human behaviors and intents from videos is a challenging task. Video flows
usually involve time-series data from different modalities, such as natural language, facial …

Target and source modality co-reinforcement for emotion understanding from asynchronous multimodal sequences

D Yang, Y Liu, C Huang, M Li, X Zhao, Y Wang… - Knowledge-Based …, 2023 - Elsevier
Perceiving human emotions from a multimodal perspective has received significant attention
in knowledge engineering communities. Due to the variable receiving frequency for …

Recent advances and trends in multimodal deep learning: A review

J Summaira, X Li, AM Shoib, S Li, J Abdul - arxiv preprint arxiv …, 2021 - arxiv.org
Deep Learning has implemented a wide range of applications and has become increasingly
popular in recent years. The goal of multimodal deep learning is to create models that can …

Attention is not enough: Mitigating the distribution discrepancy in asynchronous multimodal sequence fusion

T Liang, G Lin, L Feng, Y Zhang… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Videos flow as the mixture of language, acoustic, and vision modalities. A thorough video
understanding needs to fuse time-series data of different modalities for prediction. Due to the …