Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
A survey on mixture of experts
Large language models (LLMs) have garnered unprecedented advancements across
diverse fields, ranging from natural language processing to computer vision and beyond …
diverse fields, ranging from natural language processing to computer vision and beyond …
Llama-moe: Building mixture-of-experts from llama with continual pre-training
Abstract Mixture-of-Experts (MoE) has gained increasing popularity as a promising
framework for scaling up large language models (LLMs). However, training MoE from …
framework for scaling up large language models (LLMs). However, training MoE from …
Maplm: A real-world large-scale vision-language benchmark for map and traffic scene understanding
Vision-language generative AI has demonstrated remarkable promise for empowering cross-
modal scene understanding of autonomous driving and high-definition (HD) map systems …
modal scene understanding of autonomous driving and high-definition (HD) map systems …
Model compression and efficient inference for large language models: A survey
Transformer based large language models have achieved tremendous success. However,
the significant memory and computational costs incurred during the inference process make …
the significant memory and computational costs incurred during the inference process make …
Mvmoe: Multi-task vehicle routing solver with mixture-of-experts
Learning to solve vehicle routing problems (VRPs) has garnered much attention. However,
most neural solvers are only structured and trained independently on a specific problem …
most neural solvers are only structured and trained independently on a specific problem …
Lory: Fully differentiable mixture-of-experts for autoregressive language model pre-training
Mixture-of-experts (MoE) models facilitate efficient scaling; however, training the router
network introduces the challenge of optimizing a non-differentiable, discrete objective …
network introduces the challenge of optimizing a non-differentiable, discrete objective …
Combining fine-tuning and llm-based agents for intuitive smart contract auditing with justifications
Smart contracts are decentralized applications built atop blockchains like Ethereum. Recent
research has shown that large language models (LLMs) have potential in auditing smart …
research has shown that large language models (LLMs) have potential in auditing smart …
Branch-train-mix: Mixing expert llms into a mixture-of-experts llm
We investigate efficient methods for training Large Language Models (LLMs) to possess
capabilities in multiple specialized domains, such as coding, math reasoning and world …
capabilities in multiple specialized domains, such as coding, math reasoning and world …
Demystifying the compression of mixture-of-experts through a unified framework
Scaling large language models has revolutionized the performance across diverse domains,
yet the continual growth in model size poses significant challenges for real-world …
yet the continual growth in model size poses significant challenges for real-world …
Shortcut-connected expert parallelism for accelerating mixture-of-experts
Expert parallelism has been introduced as a strategy to distribute the computational
workload of sparsely-gated mixture-of-experts (MoE) models across multiple computing …
workload of sparsely-gated mixture-of-experts (MoE) models across multiple computing …