Understanding vision-based continuous sign language recognition

N Aloysius, M Geetha - Multimedia Tools and Applications, 2020 - Springer
Real-time sign language translation systems, that convert continuous sign sequences to
text/speech, will facilitate communication between the deaf-mute community and the normal …

A comprehensive survey of rgb-based and skeleton-based human action recognition

C Wang, J Yan - IEEE Access, 2023 - ieeexplore.ieee.org
With the advancement of computer vision, human action recognition (HAR) has shown its
broad research worth and application prospects in a wide range of fields such as intelligent …

Human action recognition and prediction: A survey

Y Kong, Y Fu - International Journal of Computer Vision, 2022 - Springer
Derived from rapid advances in computer vision and machine learning, video analysis tasks
have been moving from inferring the present state to predicting the future state. Vision-based …

Memory fusion network for multi-view sequential learning

A Zadeh, PP Liang, N Mazumder, S Poria… - Proceedings of the …, 2018 - ojs.aaai.org
Multi-view sequential learning is a fundamental problem in machine learning dealing with
multi-view sequences. In a multi-view sequence, there exists two forms of interactions …

Found in translation: Learning robust joint representations by cyclic translations between modalities

H Pham, PP Liang, T Manzini, LP Morency… - Proceedings of the …, 2019 - ojs.aaai.org
Multimodal sentiment analysis is a core research area that studies speaker sentiment
expressed from the language, visual, and acoustic modalities. The central challenge in …

Learning individual styles of conversational gesture

S Ginosar, A Bar, G Kohavi, C Chan… - Proceedings of the …, 2019 - openaccess.thecvf.com
Human speech is often accompanied by hand and arm gestures. We present a method for
cross-modal translation from" in-the-wild" monologue speech of a single speaker to their …

Learning factorized multimodal representations

YHH Tsai, PP Liang, A Zadeh, LP Morency… - arxiv preprint arxiv …, 2018 - arxiv.org
Learning multimodal representations is a fundamentally complex research problem due to
the presence of multiple heterogeneous sources of information. Although the presence of …

Multi-attention recurrent network for human communication comprehension

A Zadeh, PP Liang, S Poria, P Vij, E Cambria… - Proceedings of the …, 2018 - ojs.aaai.org
Human face-to-face communication is a complex multimodal signal. We use words
(language modality), gestures (vision modality) and changes in tone (acoustic modality) to …

Video-based sign language recognition without temporal segmentation

J Huang, W Zhou, Q Zhang, H Li, W Li - Proceedings of the AAAI …, 2018 - ojs.aaai.org
Millions of hearing impaired people around the world routinely use some variants of sign
languages to communicate, thus the automatic translation of a sign language is meaningful …

Multimodal language analysis with recurrent multistage fusion

PP Liang, Z Liu, A Zadeh, LP Morency - arxiv preprint arxiv:1808.03920, 2018 - arxiv.org
Computational modeling of human multimodal language is an emerging research area in
natural language processing spanning the language, visual and acoustic modalities …