Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
On efficient training of large-scale deep learning models: A literature review
The field of deep learning has witnessed significant progress, particularly in computer vision
(CV), natural language processing (NLP), and speech. The use of large-scale models …
(CV), natural language processing (NLP), and speech. The use of large-scale models …
Towards making the most of chatgpt for machine translation
ChatGPT shows remarkable capabilities for machine translation (MT). Several prior studies
have shown that it achieves comparable results to commercial systems for high-resource …
have shown that it achieves comparable results to commercial systems for high-resource …
A survey on non-autoregressive generation for neural machine translation and beyond
Non-autoregressive (NAR) generation, which is first proposed in neural machine translation
(NMT) to speed up inference, has attracted much attention in both machine learning and …
(NMT) to speed up inference, has attracted much attention in both machine learning and …
On Efficient Training of Large-Scale Deep Learning Models
The field of deep learning has witnessed significant progress in recent times, particularly in
areas such as computer vision (CV), natural language processing (NLP), and speech. The …
areas such as computer vision (CV), natural language processing (NLP), and speech. The …
Improving sharpness-aware minimization with fisher mask for better generalization on language models
Fine-tuning large pretrained language models on a limited training corpus usually suffers
from poor generalization. Prior works show that the recently-proposed sharpness-aware …
from poor generalization. Prior works show that the recently-proposed sharpness-aware …
Token-level self-evolution training for sequence-to-sequence learning
Adaptive training approaches, widely used in sequence-to-sequence models, commonly
reweigh the losses of different target tokens based on priors, eg word frequency. However …
reweigh the losses of different target tokens based on priors, eg word frequency. However …
Where Does the Performance Improvement Come From? -A Reproducibility Concern about Image-Text Retrieval
This article aims to provide the information retrieval community with some reflections on
recent advances in retrieval learning by analyzing the reproducibility of image-text retrieval …
recent advances in retrieval learning by analyzing the reproducibility of image-text retrieval …
Revisiting catastrophic forgetting in large language model tuning
Catastrophic Forgetting (CF) means models forgetting previously acquired knowledge when
learning new data. It compromises the effectiveness of large language models (LLMs) …
learning new data. It compromises the effectiveness of large language models (LLMs) …
Selectit: Selective instruction tuning for large language models via uncertainty-aware self-reflection
Instruction tuning (IT) is crucial to tailoring large language models (LLMs) towards human-
centric interactions. Recent advancements have shown that the careful selection of a small …
centric interactions. Recent advancements have shown that the careful selection of a small …
Redistributing low-frequency words: Making the most of monolingual data in non-autoregressive translation
Abstract Knowledge distillation (KD) is the preliminary step for training non-autoregressive
translation (NAT) models, which eases the training of NAT models at the cost of losing …
translation (NAT) models, which eases the training of NAT models at the cost of losing …