A brief overview of ChatGPT: The history, status quo and potential future development
ChatGPT, an artificial intelligence generated content (AIGC) model developed by OpenAI,
has attracted world-wide attention for its capability of dealing with challenging language …
has attracted world-wide attention for its capability of dealing with challenging language …
[HTML][HTML] A survey of GPT-3 family large language models including ChatGPT and GPT-4
KS Kalyan - Natural Language Processing Journal, 2024 - Elsevier
Large language models (LLMs) are a special class of pretrained language models (PLMs)
obtained by scaling model size, pretraining corpus and computation. LLMs, because of their …
obtained by scaling model size, pretraining corpus and computation. LLMs, because of their …
Sam-clip: Merging vision foundation models towards semantic and spatial understanding
The landscape of publicly available vision foundation models (VFMs) such as CLIP and
SAM is expanding rapidly. VFMs are endowed with distinct capabilities stemming from their …
SAM is expanding rapidly. VFMs are endowed with distinct capabilities stemming from their …
Forecasting of crop yield using remote sensing data, agrarian factors and machine learning approaches
The art of predicting crop production is done before the crop is harvested. Crop output
forecasts will help people make timely judgments concerning food policy, prices in markets …
forecasts will help people make timely judgments concerning food policy, prices in markets …
Ties-merging: Resolving interference when merging models
Transfer learning–ie, further fine-tuning a pre-trained model on a downstream task–can
confer significant advantages, including improved downstream performance, faster …
confer significant advantages, including improved downstream performance, faster …
[HTML][HTML] Large language models in law: A survey
The advent of artificial intelligence (AI) has significantly impacted the traditional judicial
industry. Moreover, recently, with the development of the concept of AI-generated content …
industry. Moreover, recently, with the development of the concept of AI-generated content …
Pengi: An audio language model for audio tasks
In the domain of audio processing, Transfer Learning has facilitated the rise of Self-
Supervised Learning and Zero-Shot Learning techniques. These approaches have led to …
Supervised Learning and Zero-Shot Learning techniques. These approaches have led to …
Conflict-averse gradient descent for multi-task learning
The goal of multi-task learning is to enable more efficient learning than single task learning
by sharing model structures for a diverse set of tasks. A standard multi-task learning …
by sharing model structures for a diverse set of tasks. A standard multi-task learning …
Efficiently identifying task grou**s for multi-task learning
Multi-task learning can leverage information learned by one task to benefit the training of
other tasks. Despite this capacity, naively training all tasks together in one model often …
other tasks. Despite this capacity, naively training all tasks together in one model often …
Language models are super mario: Absorbing abilities from homologous models as a free lunch
In this paper, we unveil that Language Models (LMs) can acquire new capabilities by
assimilating parameters from homologous models without retraining or GPUs. We first …
assimilating parameters from homologous models without retraining or GPUs. We first …