Don't flatten, tokenize! Unlocking the key to SoftMoE's efficacy in deep RL
The use of deep neural networks in reinforcement learning (RL) often suffers from
performance degradation as model size increases. While soft mixtures of experts (SoftMoEs) …
performance degradation as model size increases. While soft mixtures of experts (SoftMoEs) …
Mixtures of Experts for Scaling up Neural Networks in Order Execution
We develop a methodology that employs mixture of experts to scale up the parameters of
reinforcement learning (RL) models in optimal execution tasks. The innovation of our …
reinforcement learning (RL) models in optimal execution tasks. The innovation of our …
Neo-FREE: Policy Composition Through Thousand Brains And Free Energy Optimization
We consider the problem of optimally composing a set of primitives to tackle control tasks. To
address this problem, we introduce Neo-FREE: a control architecture inspired by the …
address this problem, we introduce Neo-FREE: a control architecture inspired by the …