A survey on curriculum learning

X Wang, Y Chen, W Zhu - IEEE transactions on pattern analysis …, 2021 - ieeexplore.ieee.org
Curriculum learning (CL) is a training strategy that trains a machine learning model from
easier data to harder data, which imitates the meaningful learning order in human curricula …

Continual lifelong learning in natural language processing: A survey

M Biesialska, K Biesialska, MR Costa-Jussa - arxiv preprint arxiv …, 2020 - arxiv.org
Continual learning (CL) aims to enable information systems to learn from a continuous data
stream across time. However, it is difficult for existing deep learning architectures to learn a …

Self-play fine-tuning converts weak language models to strong language models

Z Chen, Y Deng, H Yuan, K Ji, Q Gu - arxiv preprint arxiv:2401.01335, 2024 - arxiv.org
Harnessing the power of human-annotated data through Supervised Fine-Tuning (SFT) is
pivotal for advancing Large Language Models (LLMs). In this paper, we delve into the …

Curriculum learning: A survey

P Soviany, RT Ionescu, P Rota, N Sebe - International Journal of …, 2022 - Springer
Training machine learning models in a meaningful order, from the easy samples to the hard
ones, using curriculum learning can provide performance improvements over the standard …

Competence-based multimodal curriculum learning for medical report generation

F Liu, S Ge, Y Zou, X Wu - arxiv preprint arxiv:2206.14579, 2022 - arxiv.org
Medical report generation task, which targets to produce long and coherent descriptions of
medical images, has attracted growing research interests recently. Different from the general …

Competence-based curriculum learning for neural machine translation

EA Platanios, O Stretcu, G Neubig, B Poczos… - arxiv preprint arxiv …, 2019 - arxiv.org
Current state-of-the-art NMT systems use large neural networks that are not only slow to
train, but also often require many heuristics and optimization tricks, such as specialized …

Bridging pre-trained models and downstream tasks for source code understanding

D Wang, Z Jia, S Li, Y Yu, Y **ong, W Dong… - Proceedings of the 44th …, 2022 - dl.acm.org
With the great success of pre-trained models, the pretrain-then-finetune paradigm has been
widely adopted on downstream tasks for source code understanding. However, compared to …

Domain adaptation and multi-domain adaptation for neural machine translation: A survey

D Saunders - Journal of Artificial Intelligence Research, 2022 - jair.org
The development of deep learning techniques has allowed Neural Machine Translation
(NMT) models to become extremely powerful, given sufficient training data and training time …

When do curricula work?

X Wu, E Dyer, B Neyshabur - arxiv preprint arxiv:2012.03107, 2020 - arxiv.org
Inspired by human learning, researchers have proposed ordering examples during training
based on their difficulty. Both curriculum learning, exposing a network to easier examples …

Data-centric green artificial intelligence: A survey

S Salehi, A Schmeink - IEEE Transactions on Artificial …, 2023 - ieeexplore.ieee.org
With the exponential growth of computational power and the availability of large-scale
datasets in recent years, remarkable advancements have been made in the field of artificial …