Don't flatten, tokenize! Unlocking the key to SoftMoE's efficacy in deep RL

G Sokar, J Obando-Ceron, A Courville… - arxiv preprint arxiv …, 2024 - arxiv.org
The use of deep neural networks in reinforcement learning (RL) often suffers from
performance degradation as model size increases. While soft mixtures of experts (SoftMoEs) …

Mixtures of Experts for Scaling up Neural Networks in Order Execution

K Li, M Cucuringu, L Sánchez-Betancourt… - Proceedings of the 5th …, 2024 - dl.acm.org
We develop a methodology that employs mixture of experts to scale up the parameters of
reinforcement learning (RL) models in optimal execution tasks. The innovation of our …

Neo-FREE: Policy Composition Through Thousand Brains And Free Energy Optimization

F Rossi, É Garrabé, G Russo - arxiv preprint arxiv:2412.06636, 2024 - arxiv.org
We consider the problem of optimally composing a set of primitives to tackle control tasks. To
address this problem, we introduce Neo-FREE: a control architecture inspired by the …

[CITATION][C] Learning Robust Representations for Transfer in Reinforcement Learning

F Mohamed, RC Castanyer, H Tang, Z Sheikhbahaee… - … 2024 Workshop on Fine-Tuning in …