A review on large Language Models: Architectures, applications, taxonomies, open issues and challenges

MAK Raiaan, MSH Mukta, K Fatema, NM Fahad… - IEEE …, 2024 - ieeexplore.ieee.org
Large Language Models (LLMs) recently demonstrated extraordinary capability in various
natural language processing (NLP) tasks including language translation, text generation …

Text mining in education

R Ferreira‐Mello, M André, A Pinheiro… - … : Data Mining and …, 2019 - Wiley Online Library
The explosive growth of online education environments is generating a massive volume of
data, specially in text format from forums, chats, social networks, assessments, essays …

Curriculum learning: A survey

P Soviany, RT Ionescu, P Rota, N Sebe - International Journal of …, 2022 - Springer
Training machine learning models in a meaningful order, from the easy samples to the hard
ones, using curriculum learning can provide performance improvements over the standard …

Lifelong pretraining: Continually adapting language models to emerging corpora

X **, D Zhang, H Zhu, W **ao, SW Li, X Wei… - arxiv preprint arxiv …, 2021 - arxiv.org
Pretrained language models (PTLMs) are typically learned over a large, static corpus and
further fine-tuned for various downstream tasks. However, when deployed in the real world …

Transfer learning

SJ Pan - Learning, 2020 - api.taylorfrancis.com
Supervised machine learning techniques have already been widely studied and applied to
various real-world applications. However, most existing supervised algorithms work well …

Adaptation algorithms for neural network-based speech recognition: An overview

P Bell, J Fainberg, O Klejch, J Li… - IEEE Open Journal …, 2020 - ieeexplore.ieee.org
We present a structured overview of adaptation algorithms for neural network-based speech
recognition, considering both hybrid hidden Markov model/neural network systems and end …

Multi-domain neural network language generation for spoken dialogue systems

TH Wen, M Gasic, N Mrksic… - arxiv preprint arxiv …, 2016 - arxiv.org
Moving from limited-domain natural language generation (NLG) to open domain is difficult
because the number of semantic input combinations grows exponentially with the number of …

Self-paced prioritized curriculum learning with coverage penalty in deep reinforcement learning

Z Ren, D Dong, H Li, C Chen - IEEE transactions on neural …, 2018 - ieeexplore.ieee.org
In this paper, a new training paradigm is proposed for deep reinforcement learning using
self-paced prioritized curriculum learning with coverage penalty. The proposed deep …

Recrecnet: Rectangling rectified wide-angle images by thin-plate spline model and dof-based curriculum learning

K Liao, L Nie, C Lin, Z Zheng… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
The wide-angle lens shows appealing applications in VR technologies, but it introduces
severe radial distortion into its captured image. To recover the realistic scene, previous …

Single image super-resolution for whole slide image using convolutional neural networks and self-supervised color normalization

B Li, A Keikhosravi, AG Loeffler, KW Eliceiri - Medical Image Analysis, 2021 - Elsevier
High-quality whole slide scanners used for animal and human pathology scanning are
expensive and can produce massive datasets, which limits the access to and adoption of …