Evaluating large language models for radiology natural language processing
The rise of large language models (LLMs) has marked a pivotal shift in the field of natural
language processing (NLP). LLMs have revolutionized a multitude of domains, and they …
language processing (NLP). LLMs have revolutionized a multitude of domains, and they …
Cif-bench: A chinese instruction-following benchmark for evaluating the generalizability of large language models
The advancement of large language models (LLMs) has enhanced the ability to generalize
across a wide range of unseen natural language processing (NLP) tasks through instruction …
across a wide range of unseen natural language processing (NLP) tasks through instruction …
Nlebench+ norglm: A comprehensive empirical analysis and benchmark dataset for generative language models in norwegian
Norwegian, spoken by only 5 million population, is under-representative within the most
impressive breakthroughs in NLP tasks. To the best of our knowledge, there has not yet …
impressive breakthroughs in NLP tasks. To the best of our knowledge, there has not yet …
Persianllama: Towards building first persian large language model
Despite the widespread use of the Persian language by millions globally, limited efforts have
been made in natural language processing for this language. The use of large language …
been made in natural language processing for this language. The use of large language …
LMTuner: An user-friendly and highly-integrable Training Framework for fine-tuning Large Language Models
With the burgeoning development in the realm of large language models (LLMs), the
demand for efficient incremental training tailored to specific industries and domains …
demand for efficient incremental training tailored to specific industries and domains …
Create and find flatness: Building flat training spaces in advance for continual learning
W Shi, Y Chen, Z Zhao, W Lu, K Yan, X Du - ECAI 2023, 2023 - ebooks.iospress.nl
Catastrophic forgetting remains a critical challenge in the field of continual learning, where
neural networks struggle to retain prior knowledge while assimilating new information. Most …
neural networks struggle to retain prior knowledge while assimilating new information. Most …
FLIP-80M: 80 Million Visual-Linguistic Pairs for Facial Language-Image Pre-Training
While significant progress has been made in multi-modal learning driven by large-scale
image-text datasets, there is still a noticeable gap in the availability of such datasets within …
image-text datasets, there is still a noticeable gap in the availability of such datasets within …
Weight-inherited distillation for task-agnostic bert compression
Knowledge Distillation (KD) is a predominant approach for BERT compression. Previous KD-
based methods focus on designing extra alignment losses for the student model to mimic the …
based methods focus on designing extra alignment losses for the student model to mimic the …
Design as Desired: Utilizing Visual Question Answering for Multimodal Pre-training
Multimodal pre-training demonstrates its potential in the medical domain, which learns
medical visual representations from paired medical reports. However, many pre-training …
medical visual representations from paired medical reports. However, many pre-training …
Enhanced ICD-10 code assignment of clinical texts: A summarization-based approach
Y Sun, L Sang, D Wu, S He, Y Chen, H Duan… - Artificial Intelligence in …, 2024 - Elsevier
Abstract Background Assigning International Classification of Diseases (ICD) codes to
clinical texts is a common and crucial practice in patient classification, hospital management …
clinical texts is a common and crucial practice in patient classification, hospital management …