Transformers in reinforcement learning: a survey

P Agarwal, AA Rahman, PL St-Charles… - arxiv preprint arxiv …, 2023 - arxiv.org
Transformers have significantly impacted domains like natural language processing,
computer vision, and robotics, where they improve performance compared to other neural …

Object-centric slot diffusion

J Jiang, F Deng, G Singh, S Ahn - arxiv preprint arxiv:2303.10834, 2023 - arxiv.org
The recent success of transformer-based image generative models in object-centric learning
highlights the importance of powerful image generators for handling complex scenes …

Rotating features for object discovery

S Löwe, P Lippe, F Locatello… - Advances in Neural …, 2024 - proceedings.neurips.cc
The binding problem in human cognition, concerning how the brain represents and
connects objects within a fixed network of neural connections, remains a subject of intense …

Unsupervised Musical Object Discovery from Audio

J Gha, V Herrmann, B Grewe, J Schmidhuber… - arxiv preprint arxiv …, 2023 - arxiv.org
Current object-centric learning models such as the popular SlotAttention architecture allow
for unsupervised visual scene decomposition. Our novel MusicSlots method adapts …

Learning rational subgoals from demonstrations and instructions

Z Luo, J Mao, J Wu, T Lozano-Pérez… - Proceedings of the …, 2023 - ojs.aaai.org
We present a framework for learning useful subgoals that support efficient long-term
planning to achieve novel goals. At the core of our framework is a collection of rational …

Parallelized Spatiotemporal Slot Binding for Videos

G Singh, Y Wang, J Yang, B Ivanovic, S Ahn… - … on Machine Learning, 2024 - openreview.net
While modern best practices advocate for scalable architectures that support long-range
interactions, object-centric models are yet to fully embrace these architectures. In particular …

Interaction Asymmetry: A General Principle for Learning Composable Abstractions

J Brady, J von Kügelgen, S Lachapelle… - arxiv preprint arxiv …, 2024 - arxiv.org
Learning disentangled representations of concepts and re-composing them in unseen ways
is crucial for generalizing to out-of-domain situations. However, the underlying properties of …

Inverted-attention transformers can learn object representations: Insights from slot attention

YF Wu, K Greff, GF Elsayed, MC Mozer… - … Learning Workshop at …, 2023 - openreview.net
Visual reasoning is supported by a causal understanding of the physical world, and theories
of human cognition suppose that a necessary step to causal understanding is the discovery …

Recurrent Complex-Weighted Autoencoders for Unsupervised Object Discovery

A Gopalakrishnan, A Stanić, J Schmidhuber… - arxiv preprint arxiv …, 2024 - arxiv.org
Current state-of-the-art synchrony-based models encode object bindings with complex-
valued activations and compute with real-valued weights in feedforward architectures. We …

Unsupervised tokenization learning

A Kolonin, V Ramesh - arxiv preprint arxiv:2205.11443, 2022 - arxiv.org
In the presented study, we discover that the so-called" transition freedom" metric appears
superior for unsupervised tokenization purposes in comparison to statistical metrics such as …