[HTML][HTML] A survey of transformers

T Lin, Y Wang, X Liu, X Qiu - AI open, 2022 - Elsevier
Transformers have achieved great success in many artificial intelligence fields, such as
natural language processing, computer vision, and audio processing. Therefore, it is natural …

Fastformer: Additive attention can be all you need

C Wu, F Wu, T Qi, Y Huang, X **e - arxiv preprint arxiv:2108.09084, 2021 - arxiv.org
Transformer is a powerful model for text understanding. However, it is inefficient due to its
quadratic complexity to input sequence length. Although there are many methods on …

Parade: Passage representation aggregation fordocument reranking

C Li, A Yates, S MacAvaney, B He, Y Sun - ACM Transactions on …, 2023 - dl.acm.org
Pre-trained transformer models, such as BERT and T5, have shown to be highly effective at
ad hoc passage and document ranking. Due to the inherent sequence length limits of these …

Speechformer++: A hierarchical efficient framework for paralinguistic speech processing

W Chen, X **ng, X Xu, J Pang… - IEEE/ACM Transactions …, 2023 - ieeexplore.ieee.org
Paralinguistic speech processing is important in addressing many issues, such as sentiment
and neurocognitive disorder analyses. Recently, Transformer has achieved remarkable …

HiGNN: A hierarchical informative graph neural network for molecular property prediction equipped with feature-wise attention

W Zhu, Y Zhang, D Zhao, J Xu… - Journal of Chemical …, 2022 - ACS Publications
Elucidating and accurately predicting the druggability and bioactivities of molecules plays a
pivotal role in drug design and discovery and remains an open challenge. Recently, graph …

Neural natural language processing for long texts: A survey on classification and summarization

D Tsirmpas, I Gkionis, GT Papadopoulos… - … Applications of Artificial …, 2024 - Elsevier
Abstract The adoption of Deep Neural Networks (DNNs) has greatly benefited Natural
Language Processing (NLP) during the past decade. However, the demands of long …

SPT: Spatial pyramid transformer for image captioning

H Zhang, P Zeng, L Gao, X Lyu, J Song… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
The existing approaches to image captioning tend to adopt Transformer-based architectures
with grid features, which represent the state-of-the-art. However, the strategies are prone to …

Collaborative-Enhanced Prediction of Spending on Newly Downloaded Mobile Games under Consumption Uncertainty

P Sun, Y Wang, M Zhang, C Wu, Y Fang… - … Proceedings of the …, 2024 - dl.acm.org
With the surge in mobile gaming, accurately predicting user spending on newly downloaded
games has become paramount for maximizing revenue. However, the inherently …

Hierarchical multi-modal prompting transformer for multi-modal long document classification

T Liu, Y Hu, J Gao, Y Sun, B Yin - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
In the context of long document classification (LDC), effectively utilizing multi-modal
information encompassing texts and images within these documents has not received …

Integrating convolution and self-attention improves language model of human genome for interpreting non-coding regions at base-resolution

M Yang, L Huang, H Huang, H Tang… - Nucleic acids …, 2022 - academic.oup.com
Abstract Interpretation of non-coding genome remains an unsolved challenge in human
genetics due to impracticality of exhaustively annotating biochemically active elements in all …