Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Dart-math: Difficulty-aware rejection tuning for mathematical problem-solving
Solving mathematical problems requires advanced reasoning abilities and presents notable
challenges for large language models. Previous works usually synthesize data from …
challenges for large language models. Previous works usually synthesize data from …
Mammoth2: Scaling instructions from the web
Instruction tuning improves the reasoning abilities of large language models (LLMs), with
data quality and scalability being the crucial factors. Most instruction tuning data come from …
data quality and scalability being the crucial factors. Most instruction tuning data come from …
Internal consistency and self-feedback in large language models: A survey
Large language models (LLMs) often exhibit deficient reasoning or generate hallucinations.
To address these, studies prefixed with" Self-" such as Self-Consistency, Self-Improve, and …
To address these, studies prefixed with" Self-" such as Self-Consistency, Self-Improve, and …
A survey on data synthesis and augmentation for large language models
K Wang, J Zhu, M Ren, Z Liu, S Li, Z Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
The success of Large Language Models (LLMs) is inherently linked to the availability of vast,
diverse, and high-quality data for training and evaluation. However, the growth rate of high …
diverse, and high-quality data for training and evaluation. However, the growth rate of high …
A theoretical understanding of self-correction through in-context alignment
Going beyond mimicking limited human experiences, recent studies show initial evidence
that, like humans, large language models (LLMs) are capable of improving their abilities …
that, like humans, large language models (LLMs) are capable of improving their abilities …
Exploring automated energy optimization with unstructured building data: A multi-agent based framework leveraging large language models
The building sector is a significant energy consumer, making building energy optimization
crucial for reducing energy demand. Automating energy optimization tasks eases the …
crucial for reducing energy demand. Automating energy optimization tasks eases the …
Self-generated critiques boost reward modeling for language models
Reward modeling is crucial for aligning large language models (LLMs) with human
preferences, especially in reinforcement learning from human feedback (RLHF). However …
preferences, especially in reinforcement learning from human feedback (RLHF). However …
Do not think that much for 2+ 3=? on the overthinking of o1-like llms
The remarkable performance of models like the OpenAI o1 can be attributed to their ability to
emulate human-like long-time thinking during inference. These models employ extended …
emulate human-like long-time thinking during inference. These models employ extended …
Ptd-sql: Partitioning and targeted drilling with llms in text-to-sql
Large Language Models (LLMs) have emerged as powerful tools for Text-to-SQL tasks,
exhibiting remarkable reasoning capabilities. Different from tasks such as math word …
exhibiting remarkable reasoning capabilities. Different from tasks such as math word …
Evaluating the Evaluator: Measuring LLMs' Adherence to Task Evaluation Instructions
LLMs-as-a-judge is a recently popularized method which replaces human judgements in
task evaluation (Zheng et al. 2024) with automatic evaluation using LLMs. Due to …
task evaluation (Zheng et al. 2024) with automatic evaluation using LLMs. Due to …