Multi-layer transformers gradient can be approximated in almost linear time
The computational complexity of the self-attention mechanism in popular transformer
architectures poses significant challenges for training and inference, and becomes the …
architectures poses significant challenges for training and inference, and becomes the …
Retrieval augmented generation or long-context llms? a comprehensive study and hybrid approach
Abstract Retrieval Augmented Generation (RAG) has been a powerful tool for Large
Language Models (LLMs) to efficiently process overly lengthy contexts. However, recent …
Language Models (LLMs) to efficiently process overly lengthy contexts. However, recent …
Flooding spread of manipulated knowledge in llm-based multi-agent communities
The rapid adoption of large language models (LLMs) in multi-agent systems has highlighted
their impressive capabilities in various applications, such as collaborative problem-solving …
their impressive capabilities in various applications, such as collaborative problem-solving …
Longmemeval: Benchmarking chat assistants on long-term interactive memory
Recent large language model (LLM)-driven chat assistant systems have integrated memory
components to track user-assistant chat histories, enabling more accurate and personalized …
components to track user-assistant chat histories, enabling more accurate and personalized …
Circuit Complexity Bounds for RoPE-based Transformer Architecture
Characterizing the express power of the Transformer architecture is critical to understanding
its capacity limits and scaling law. Recent works provide the circuit complexity bounds to …
its capacity limits and scaling law. Recent works provide the circuit complexity bounds to …
[HTML][HTML] Leveraging Large Language Models for Enhancing Safety in Maritime Operations
Maritime operations play a critical role in global trade but face persistent safety challenges
due to human error, environmental factors, and operational complexities. This review …
due to human error, environmental factors, and operational complexities. This review …
What is Wrong with Perplexity for Long-context Language Modeling?
Handling long-context inputs is crucial for large language models (LLMs) in tasks such as
extended conversations, document summarization, and many-shot in-context learning …
extended conversations, document summarization, and many-shot in-context learning …
Moba: A two-level agent system for efficient mobile task automation
Current mobile assistants are limited by dependence on system APIs or struggle with
complex user instructions and diverse interfaces due to restricted comprehension and …
complex user instructions and diverse interfaces due to restricted comprehension and …
Stark: Social Long-Term Multi-Modal Conversation with Persona Commonsense Knowledge
Humans share a wide variety of images related to their personal experiences within
conversations via instant messaging tools. However, existing works focus on (1) image …
conversations via instant messaging tools. However, existing works focus on (1) image …
Memsim: A bayesian simulator for evaluating memory of llm-based personal assistants
LLM-based agents have been widely applied as personal assistants, capable of memorizing
information from user messages and responding to personal queries. However, there still …
information from user messages and responding to personal queries. However, there still …