Flow to better: Offline preference-based reinforcement learning via preferred trajectory generation Z Zhang, Y Sun, J Ye, TS Liu, J Zhang, Y Yu The Twelfth International Conference on Learning Representations, 2023 | 13 | 2023 |
Any-step Dynamics Model Improves Future Predictions for Online and Offline Reinforcement Learning H Lin, YY Xu, Y Sun, Z Zhang, YC Li, C Jia, J Ye, J Zhang, Y Yu arXiv preprint arXiv:2405.17031, 2024 | 1 | 2024 |
UDCA may promote COVID-19 recovery: a cohort study with AI-aided analysis Y Yu, G Yu, LY Han, J Li, ZL Zhang, TS Liu, MF Li, DC Zhan, SQ Tang, ... MedRxiv, 2023.05. 02.23289410, 2023 | 1 | 2023 |
Limited Preference Aided Imitation Learning from Imperfect Demonstrations X Cao, FM Luo, J Ye, T Xu, Z Zhang, Y Yu Forty-first International Conference on Machine Learning, 0 | 1 | |
WHALE: Towards Generalizable and Scalable World Models for Embodied Decision-making Z Zhang, R Chen, J Ye, Y Sun, P Wang, J Pang, K Li, T Liu, H Lin, Y Yu, ... arXiv preprint arXiv:2411.05619, 2024 | | 2024 |
Provably and Practically Efficient Adversarial Imitation Learning with General Function Approximation T Xu, Z Zhang, R Chen, Y Sun, Y Yu arXiv preprint arXiv:2411.00610, 2024 | | 2024 |
Energy-guided diffusion sampling for offline-to-online reinforcement learning XH Liu, TS Liu, S Jiang, R Chen, Z Zhang, X Chen, Y Yu arXiv preprint arXiv:2407.12448, 2024 | | 2024 |
BWArea Model: Learning World Model, Inverse Dynamics, and Policy for Controllable Language Generation C Jia, P Wang, Z Li, YC Li, Z Zhang, N Tang, Y Yu arXiv preprint arXiv:2405.17039, 2024 | | 2024 |