Bringing order into the realm of Transformer-based language models for artificial intelligence and law

CM Greco, A Tagarelli - Artificial Intelligence and Law, 2024 - Springer
Transformer-based language models (TLMs) have widely been recognized to be a cutting-
edge technology for the successful development of deep-learning-based solutions to …

Crossfit: A few-shot learning challenge for cross-task generalization in nlp

Q Ye, BY Lin, X Ren - arxiv preprint arxiv:2104.08835, 2021 - arxiv.org
Humans can learn a new language task efficiently with only few examples, by leveraging
their knowledge obtained when learning prior tasks. In this paper, we explore whether and …

AUGER: automatically generating review comments with pre-training models

L Li, L Yang, H Jiang, J Yan, T Luo, Z Hua… - Proceedings of the 30th …, 2022 - dl.acm.org
Code review is one of the best practices as a powerful safeguard for software quality. In
practice, senior or highly skilled reviewers inspect source code and provide constructive …

Math-LLMs: AI cyberinfrastructure with pre-trained transformers for math education

F Zhang, C Li, O Henkel, W **ng, S Baral… - International Journal of …, 2024 - Springer
In recent years, the pre-training of Large Language Models (LLMs) in the educational
domain has garnered significant attention. However, a discernible gap exists in the …

[HTML][HTML] Challenges and opportunities of using transformer-based multi-task learning in NLP through ML lifecycle: A position paper

L Torbarina, T Ferkovic, L Roguski, V Mihelcic… - Natural Language …, 2024 - Elsevier
The increasing adoption of natural language processing (NLP) models across industries has
led to practitioners' need for machine learning (ML) systems to handle these models …

Enhancing molecular property prediction through task-oriented transfer learning: integrating universal structural insights and domain-specific knowledge

Y Duan, X Yang, X Zeng, W Wang… - Journal of Medicinal …, 2024 - ACS Publications
Precisely predicting molecular properties is crucial in drug discovery, but the scarcity of
labeled data poses a challenge for applying deep learning methods. While large-scale self …

Cluster & tune: Boost cold start performance in text classification

E Shnarch, A Gera, A Halfon, L Dankin… - arxiv preprint arxiv …, 2022 - arxiv.org
In real-world scenarios, a text classification task often begins with a cold start, when labeled
data is scarce. In such cases, the common practice of fine-tuning pre-trained models, such …

Finding the missing data: A bert-inspired approach against package loss in wireless sensing

Z Zhao, T Chen, F Meng, H Li, X Li… - IEEE INFOCOM 2024 …, 2024 - ieeexplore.ieee.org
Despite the development of various deep learning methods for Wi-Fi sensing, package loss
often results in noncontinuous estimation of the Channel State Information (CSI), which …

Deep representation learning: Fundamentals, technologies, applications, and open challenges

A Payandeh, KT Baghaei, P Fayyazsanavi… - IEEE …, 2023 - ieeexplore.ieee.org
Machine learning algorithms have had a profound impact on the field of computer science
over the past few decades. The performance of these algorithms heavily depends on the …

Deep representation learning: Fundamentals, perspectives, applications, and open challenges

KT Baghaei, A Payandeh, P Fayyazsanavi… - arxiv preprint arxiv …, 2022 - arxiv.org
Machine Learning algorithms have had a profound impact on the field of computer science
over the past few decades. These algorithms performance is greatly influenced by the …