A brief overview of ChatGPT: The history, status quo and potential future development

T Wu, S He, J Liu, S Sun, K Liu… - IEEE/CAA Journal of …, 2023 - ieeexplore.ieee.org
ChatGPT, an artificial intelligence generated content (AIGC) model developed by OpenAI,
has attracted world-wide attention for its capability of dealing with challenging language …

[HTML][HTML] A survey of GPT-3 family large language models including ChatGPT and GPT-4

KS Kalyan - Natural Language Processing Journal, 2024 - Elsevier
Large language models (LLMs) are a special class of pretrained language models (PLMs)
obtained by scaling model size, pretraining corpus and computation. LLMs, because of their …

Sam-clip: Merging vision foundation models towards semantic and spatial understanding

H Wang, PKA Vasu, F Faghri… - Proceedings of the …, 2024 - openaccess.thecvf.com
The landscape of publicly available vision foundation models (VFMs) such as CLIP and
SAM is expanding rapidly. VFMs are endowed with distinct capabilities stemming from their …

Forecasting of crop yield using remote sensing data, agrarian factors and machine learning approaches

JP Bharadiya, NT Tzenios… - Journal of Engineering …, 2023 - classical.goforpromo.com
The art of predicting crop production is done before the crop is harvested. Crop output
forecasts will help people make timely judgments concerning food policy, prices in markets …

Ties-merging: Resolving interference when merging models

P Yadav, D Tam, L Choshen… - Advances in Neural …, 2024 - proceedings.neurips.cc
Transfer learning–ie, further fine-tuning a pre-trained model on a downstream task–can
confer significant advantages, including improved downstream performance, faster …

[HTML][HTML] Large language models in law: A survey

J Lai, W Gan, J Wu, Z Qi, SY Philip - AI Open, 2024 - Elsevier
The advent of artificial intelligence (AI) has significantly impacted the traditional judicial
industry. Moreover, recently, with the development of the concept of AI-generated content …

Pengi: An audio language model for audio tasks

S Deshmukh, B Elizalde, R Singh… - Advances in Neural …, 2023 - proceedings.neurips.cc
In the domain of audio processing, Transfer Learning has facilitated the rise of Self-
Supervised Learning and Zero-Shot Learning techniques. These approaches have led to …

Conflict-averse gradient descent for multi-task learning

B Liu, X Liu, X **, P Stone… - Advances in Neural …, 2021 - proceedings.neurips.cc
The goal of multi-task learning is to enable more efficient learning than single task learning
by sharing model structures for a diverse set of tasks. A standard multi-task learning …

Efficiently identifying task grou**s for multi-task learning

C Fifty, E Amid, Z Zhao, T Yu… - Advances in Neural …, 2021 - proceedings.neurips.cc
Multi-task learning can leverage information learned by one task to benefit the training of
other tasks. Despite this capacity, naively training all tasks together in one model often …

Language models are super mario: Absorbing abilities from homologous models as a free lunch

L Yu, B Yu, H Yu, F Huang, Y Li - Forty-first International Conference …, 2024 - openreview.net
In this paper, we unveil that Language Models (LMs) can acquire new capabilities by
assimilating parameters from homologous models without retraining or GPUs. We first …