Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Adaptable logical control for large language models
Despite the success of Large Language Models (LLMs) on various tasks following human
instructions, controlling model generation to follow strict constraints at inference time poses …
instructions, controlling model generation to follow strict constraints at inference time poses …
Optimizing instructions and demonstrations for multi-stage language model programs
Language Model Programs, ie sophisticated pipelines of modular language model (LM)
calls, are increasingly advancing NLP tasks, but they require crafting prompts that are jointly …
calls, are increasingly advancing NLP tasks, but they require crafting prompts that are jointly …
Kimi k1. 5: Scaling reinforcement learning with llms
K Team, A Du, B Gao, B **ng, C Jiang, C Chen… - arxiv preprint arxiv …, 2025 - arxiv.org
Language model pretraining with next token prediction has proved effective for scaling
compute but is limited to the amount of available training data. Scaling reinforcement …
compute but is limited to the amount of available training data. Scaling reinforcement …
Stateful large language model serving with pensieve
L Yu, J Lin, J Li - arxiv preprint arxiv:2312.05516, 2023 - arxiv.org
Large Language Models (LLMs) are wildly popular today and it is important to serve them
efficiently. Existing LLM serving systems are stateless across requests. Consequently, when …
efficiently. Existing LLM serving systems are stateless across requests. Consequently, when …
vattention: Dynamic memory management for serving llms without pagedattention
Efficient management of GPU memory is essential for high throughput LLM inference. Prior
systems used to reserve KV-cache memory ahead-of-time that resulted in wasted capacity …
systems used to reserve KV-cache memory ahead-of-time that resulted in wasted capacity …
Neo: Saving gpu memory crisis with cpu offloading for online llm inference
Online LLM inference powers many exciting applications such as intelligent chatbots and
autonomous agents. Modern LLM inference engines widely rely on request batching to …
autonomous agents. Modern LLM inference engines widely rely on request batching to …
Structuredrag: Json response formatting with large language models
The ability of Large Language Models (LLMs) to generate structured outputs, such as JSON,
is crucial for their use in Compound AI Systems. However, evaluating and improving this …
is crucial for their use in Compound AI Systems. However, evaluating and improving this …
User Behavior Simulation with Large Language Model-based Agents
Simulating high quality user behavior data has always been a fundamental yet challenging
problem in human-centered applications such as recommendation systems, social networks …
problem in human-centered applications such as recommendation systems, social networks …
TokenSelect: Efficient Long-Context Inference and Length Extrapolation for LLMs via Dynamic Token-Level KV Cache Selection
With the development of large language models (LLMs), the ability to handle longer contexts
has become a key capability for Web applications such as cross-document understanding …
has become a key capability for Web applications such as cross-document understanding …
UBER: Uncertainty-Based Evolution with Large Language Models for Automatic Heuristic Design
NP-hard problem-solving traditionally relies on heuristics, but manually crafting effective
heuristics for complex problems remains challenging. While recent work like FunSearch has …
heuristics for complex problems remains challenging. While recent work like FunSearch has …