[HTML][HTML] A review on dropout regularization approaches for deep neural networks within the scholarly domain

I Salehin, DK Kang - Electronics, 2023 - mdpi.com
Dropout is one of the most popular regularization methods in the scholarly domain for
preventing a neural network model from overfitting in the training phase. Develo** an …

A comprehensive survey of abstractive text summarization based on deep learning

M Zhang, G Zhou, W Yu, N Huang… - Computational …, 2022 - Wiley Online Library
With the rapid development of the Internet, the massive amount of web textual data has
grown exponentially, which has brought considerable challenges to downstream tasks, such …

St++: Make self-training work better for semi-supervised semantic segmentation

L Yang, W Zhuo, L Qi, Y Shi… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Self-training via pseudo labeling is a conventional, simple, and popular pipeline to leverage
unlabeled data. In this work, we first construct a strong baseline of self-training (namely ST) …

Domaindrop: Suppressing domain-sensitive channels for domain generalization

J Guo, L Qi, Y Shi - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Abstract Deep Neural Networks have exhibited considerable success in various visual tasks.
However, when applied to unseen test datasets, state-of-the-art models often suffer …

Galaxy: A generative pre-trained model for task-oriented dialog with semi-supervised learning and explicit policy injection

W He, Y Dai, Y Zheng, Y Wu, Z Cao, D Liu… - Proceedings of the …, 2022 - ojs.aaai.org
Pre-trained models have proved to be powerful in enhancing task-oriented dialog systems.
However, current pre-training methods mainly focus on enhancing dialog understanding …

Frequency enhanced hybrid attention network for sequential recommendation

X Du, H Yuan, P Zhao, J Qu, F Zhuang, G Liu… - Proceedings of the 46th …, 2023 - dl.acm.org
The self-attention mechanism, which equips with a strong capability of modeling long-range
dependencies, is one of the extensively used techniques in the sequential recommendation …

A survey on non-autoregressive generation for neural machine translation and beyond

Y **ao, L Wu, J Guo, J Li, M Zhang… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation
(NMT) to speed up inference, has attracted much attention in both machine learning and …

On the use of bert for automated essay scoring: Joint learning of multi-scale essay representation

Y Wang, C Wang, R Li, H Lin - arxiv preprint arxiv:2205.03835, 2022 - arxiv.org
In recent years, pre-trained models have become dominant in most natural language
processing (NLP) tasks. However, in the area of Automated Essay Scoring (AES), pre …

Findings of the IWSLT 2022 Evaluation Campaign.

A Anastasopoulos, L Barrault, L Bentivogli… - Proceedings of the 19th …, 2022 - cris.fbk.eu
The evaluation campaign of the 19th International Conference on Spoken Language
Translation featured eight shared tasks:(i) Simultaneous speech translation,(ii) Offline …

Diff-AMP: tailored designed antimicrobial peptide framework with all-in-one generation, identification, prediction and optimization

R Wang, T Wang, L Zhuo, J Wei, X Fu… - Briefings in …, 2024 - academic.oup.com
Antimicrobial peptides (AMPs), short peptides with diverse functions, effectively target and
combat various organisms. The widespread misuse of chemical antibiotics has led to …