A comprehensive survey on applications of transformers for deep learning tasks
Abstract Transformers are Deep Neural Networks (DNN) that utilize a self-attention
mechanism to capture contextual relationships within sequential data. Unlike traditional …
mechanism to capture contextual relationships within sequential data. Unlike traditional …
Autonomous driving system: A comprehensive survey
Automation is increasingly at the forefront of transportation research, with the potential to
bring fully autonomous vehicles to our roads in the coming years. This comprehensive …
bring fully autonomous vehicles to our roads in the coming years. This comprehensive …
Decision transformer: Reinforcement learning via sequence modeling
We introduce a framework that abstracts Reinforcement Learning (RL) as a sequence
modeling problem. This allows us to draw upon the simplicity and scalability of the …
modeling problem. This allows us to draw upon the simplicity and scalability of the …
Offline reinforcement learning as one big sequence modeling problem
Reinforcement learning (RL) is typically viewed as the problem of estimating single-step
policies (for model-free RL) or single-step models (for model-based RL), leveraging the …
policies (for model-free RL) or single-step models (for model-based RL), leveraging the …
Megalodon: Efficient llm pretraining and inference with unlimited context length
The quadratic complexity and weak length extrapolation of Transformers limits their ability to
scale to long sequences, and while sub-quadratic solutions like linear attention and state …
scale to long sequences, and while sub-quadratic solutions like linear attention and state …
Structured state space models for in-context reinforcement learning
Structured state space sequence (S4) models have recently achieved state-of-the-art
performance on long-range sequence modeling tasks. These models also have fast …
performance on long-range sequence modeling tasks. These models also have fast …
History aware multimodal transformer for vision-and-language navigation
Vision-and-language navigation (VLN) aims to build autonomous visual agents that follow
instructions and navigate in real scenes. To remember previously visited locations and …
instructions and navigate in real scenes. To remember previously visited locations and …
A survey of meta-reinforcement learning
While deep reinforcement learning (RL) has fueled multiple high-profile successes in
machine learning, it is held back from more widespread adoption by its often poor data …
machine learning, it is held back from more widespread adoption by its often poor data …
Frozen pretrained transformers as universal computation engines
We investigate the capability of a transformer pretrained on natural language to generalize
to other modalities with minimal finetuning--in particular, without finetuning of the self …
to other modalities with minimal finetuning--in particular, without finetuning of the self …
Mega: moving average equipped gated attention
The design choices in the Transformer attention mechanism, including weak inductive bias
and quadratic computational complexity, have limited its application for modeling long …
and quadratic computational complexity, have limited its application for modeling long …