Large sequence models for sequential decision-making: a survey
Transformer architectures have facilitated the development of large-scale and general-
purpose sequence models for prediction tasks in natural language processing and computer …
purpose sequence models for prediction tasks in natural language processing and computer …
Rt-1: Robotics transformer for real-world control at scale
A Brohan, N Brown, J Carbajal, Y Chebotar… - arxiv preprint arxiv …, 2022 - arxiv.org
By transferring knowledge from large, diverse, task-agnostic datasets, modern machine
learning models can solve specific downstream tasks either zero-shot or with small task …
learning models can solve specific downstream tasks either zero-shot or with small task …
Towards generalist biomedical AI
Background Medicine is inherently multimodal, requiring the simultaneous interpretation
and integration of insights between many data modalities spanning text, imaging, genomics …
and integration of insights between many data modalities spanning text, imaging, genomics …
Learning universal policies via text-guided video generation
A goal of artificial intelligence is to construct an agent that can solve a wide variety of tasks.
Recent progress in text-guided image synthesis has yielded models with an impressive …
Recent progress in text-guided image synthesis has yielded models with an impressive …
Perceiver-actor: A multi-task transformer for robotic manipulation
Transformers have revolutionized vision and natural language processing with their ability to
scale with large datasets. But in robotic manipulation, data is both limited and expensive …
scale with large datasets. But in robotic manipulation, data is both limited and expensive …
Q-transformer: Scalable offline reinforcement learning via autoregressive q-functions
In this work, we present a scalable reinforcement learning method for training multi-task
policies from large offline datasets that can leverage both human demonstrations and …
policies from large offline datasets that can leverage both human demonstrations and …
Diffusion model is an effective planner and data synthesizer for multi-task reinforcement learning
Diffusion models have demonstrated highly-expressive generative capabilities in vision and
NLP. Recent studies in reinforcement learning (RL) have shown that diffusion models are …
NLP. Recent studies in reinforcement learning (RL) have shown that diffusion models are …
Supervised pretraining can learn in-context reinforcement learning
Large transformer models trained on diverse datasets have shown a remarkable ability to
learn in-context, achieving high few-shot performance on tasks they were not explicitly …
learn in-context, achieving high few-shot performance on tasks they were not explicitly …
Foundation models for decision making: Problems, methods, and opportunities
Foundation models pretrained on diverse data at scale have demonstrated extraordinary
capabilities in a wide range of vision and language tasks. When such models are deployed …
capabilities in a wide range of vision and language tasks. When such models are deployed …
Stealing part of a production language model
We introduce the first model-stealing attack that extracts precise, nontrivial information from
black-box production language models like OpenAI's ChatGPT or Google's PaLM-2 …
black-box production language models like OpenAI's ChatGPT or Google's PaLM-2 …