Pruning as a Domain-specific LLM Extractor
Large Language Models (LLMs) have exhibited remarkable proficiency across a wide array
of NLP tasks. However, the escalation in model size also engenders substantial deployment …
of NLP tasks. However, the escalation in model size also engenders substantial deployment …
MEDVOC: Vocabulary Adaptation for Fine-tuning Pre-trained Language Models on Medical Text Summarization
This work presents a dynamic vocabulary adaptation strategy, MEDVOC, for fine-tuning pre-
trained language models (PLMs) like BertSumAbs, BART, and PEGASUS for improved …
trained language models (PLMs) like BertSumAbs, BART, and PEGASUS for improved …
Adaptive BPE Tokenization for Enhanced Vocabulary Adaptation in Finetuning Pretrained Language Models
In this work, we show a fundamental limitation in vocabulary adaptation approaches that use
Byte-Pair Encoding (BPE) tokenization scheme for fine-tuning pretrained language models …
Byte-Pair Encoding (BPE) tokenization scheme for fine-tuning pretrained language models …
Self-supervised Segment Contrastive Learning for Medical Document Representation
Learning high-quality text embedding is vital for biomedical topic classification and many
other NLP tasks. Contrastive learning has shown remarkable performance in generating …
other NLP tasks. Contrastive learning has shown remarkable performance in generating …