Towards nonlinear disentanglement in natural data with temporal sparse coding

D Klindt, L Schott, Y Sharma, I Ustyuzhaninov… - arxiv preprint arxiv …, 2020 - arxiv.org
We construct an unsupervised learning model that achieves nonlinear disentanglement of
underlying factors of variation in naturalistic videos. Previous work suggests that …

[HTML][HTML] Symmetry-based representations for artificial and biological general intelligence

I Higgins, S Racanière, D Rezende - Frontiers in Computational …, 2022 - frontiersin.org
Biological intelligence is remarkable in its ability to produce complex behaviour in many
diverse situations through data efficient, generalisable and transferable skill acquisition. It is …

The challenges of exploration for offline reinforcement learning

N Lambert, M Wulfmeier, W Whitney, A Byravan… - arxiv preprint arxiv …, 2022 - arxiv.org
Offline Reinforcement Learning (ORL) enablesus to separately study the two interlinked
processes of reinforcement learning: collecting informative experience and inferring optimal …

Pact: Perception-action causal transformer for autoregressive robotics pre-training

R Bonatti, S Vemprala, S Ma, F Frujeri… - 2023 IEEE/RSJ …, 2023 - ieeexplore.ieee.org
Robotics has long been a field riddled with complex systems architectures whose modules
and connections, whether traditional or learning-based, require significant human expertise …

Disentangled representations for causal cognition

F Torresan, M Baltieri - Physics of Life Reviews, 2024 - Elsevier
Complex adaptive agents consistently achieve their goals by solving problems that seem to
require an understanding of causal information, information pertaining to the causal …

Solving continuous control via q-learning

T Seyde, P Werner, W Schwarting… - arxiv preprint arxiv …, 2022 - arxiv.org
While there has been substantial success for solving continuous control with actor-critic
methods, simpler critic-only methods such as Q-learning find limited application in the …

The treachery of images: Bayesian scene keypoints for deep policy learning in robotic manipulation

JO von Hartz, E Chisari, T Welschehold… - IEEE Robotics and …, 2023 - ieeexplore.ieee.org
In policy learning for robotic manipulation, sample efficiency is of paramount importance.
Thus, learning and extracting more compact representations from camera observations is a …

APEX: Unsupervised, object-centric scene segmentation and tracking for robot manipulation

Y Wu, OP Jones, M Engelcke… - 2021 IEEE/RSJ …, 2021 - ieeexplore.ieee.org
Recent advances in unsupervised learning for object detection, segmentation, and tracking
hold significant promise for applications in robotics. A common approach is to frame these …

ASIMO: Agent-centric scene representation in multi-object manipulation

CH Min, YM Kim - The International Journal of Robotics …, 2025 - journals.sagepub.com
Vision-based reinforcement learning (RL) is a generalizable way to control an agent
because it is agnostic of specific hardware configurations. As visual observations are highly …

Zero-shot sim-to-real transfer using Siamese-Q-Based reinforcement learning

Z Zhang, S **e, H Zhang, X Luo, H Yu - Information Fusion, 2025 - Elsevier
To address real world decision problems in reinforcement learning, it is common to train a
policy in a simulator first for safety. Unfortunately, the sim-real gap hinders effective …