A survey on evaluation of large language models
Large language models (LLMs) are gaining increasing popularity in both academia and
industry, owing to their unprecedented performance in various applications. As LLMs …
industry, owing to their unprecedented performance in various applications. As LLMs …
Machine learning methods for small data challenges in molecular science
Small data are often used in scientific and engineering research due to the presence of
various constraints, such as time, cost, ethics, privacy, security, and technical limitations in …
various constraints, such as time, cost, ethics, privacy, security, and technical limitations in …
Domain-specific language model pretraining for biomedical natural language processing
Pretraining large neural language models, such as BERT, has led to impressive gains on
many natural language processing (NLP) tasks. However, most pretraining efforts focus on …
many natural language processing (NLP) tasks. However, most pretraining efforts focus on …
[PDF][PDF] Bert: Pre-training of deep bidirectional transformers for language understanding
J Devlin - arxiv preprint arxiv:1810.04805, 2018 - bibbase.org
We introduce a new language representation model called BERT, which stands for
Bidirectional Encoder Representations from Transformers. Unlike recent language …
Bidirectional Encoder Representations from Transformers. Unlike recent language …
Parameter-efficient transfer learning for NLP
Fine-tuning large pretrained models is an effective transfer mechanism in NLP. However, in
the presence of many downstream tasks, fine-tuning is parameter inefficient: an entire new …
the presence of many downstream tasks, fine-tuning is parameter inefficient: an entire new …
[PDF][PDF] Deep learning
I Goodfellow - 2016 - synapse.koreamed.org
An introduction to a broad range of topics in deep learning, covering mathematical and
conceptual background, deep learning techniques used in industry, and research …
conceptual background, deep learning techniques used in industry, and research …
Text and code embeddings by contrastive pre-training
Text embeddings are useful features in many applications such as semantic search and
computing text similarity. Previous work typically trains models customized for different use …
computing text similarity. Previous work typically trains models customized for different use …
[BOOK][B] Neural network methods in natural language processing
Y Goldberg - 2017 - books.google.com
Neural networks are a family of powerful machine learning models and this book focuses on
their application to natural language data. The first half of the book (Parts I and II) covers the …
their application to natural language data. The first half of the book (Parts I and II) covers the …
A survey on hate speech detection using natural language processing
A Schmidt, M Wiegand - … of the fifth international workshop on …, 2017 - aclanthology.org
This paper presents a survey on hate speech detection. Given the steadily growing body of
social media content, the amount of online hate speech is also increasing. Due to the …
social media content, the amount of online hate speech is also increasing. Due to the …
[PDF][PDF] Natural language processing (almost) from scratch
We propose a unified neural network architecture and learning algorithm that can be applied
to various natural language processing tasks including part-of-speech tagging, chunking …
to various natural language processing tasks including part-of-speech tagging, chunking …