フォロー
Haoran Xu
Haoran Xu
確認したメール アドレス: utexas.edu - ホームページ
タイトル
引用先
引用先
Constraints penalized q-learning for safe offline reinforcement learning
H Xu, X Zhan, X Zhu
AAAI 2022, 2022
902022
Offline RL with No OOD Actions: In-Sample Learning via Implicit Value Regularization
H Xu, L Jiang, J Li, Z Yang, Z Wang, VWK Chan, X Zhan
ICLR 2023, 2023
812023
Discriminator-weighted offline imitation learning from suboptimal demonstrations
H Xu, X Zhan, H Yin, H Qin
ICML 2022, 2022
792022
Deepthermal: Combustion optimization for thermal power generating units using offline reinforcement learning
X Zhan, H Xu, Y Zhang, X Zhu, H Yin, Y Zheng
AAAI 2022, 2022
792022
A Policy-Guided Imitation Approach for Offline Reinforcement Learning
H Xu, L Jiang, J Li, X Zhan
NeurIPS 2022, 2022
672022
When data geometry meets deep function: Generalizing offline reinforcement learning
J Li, X Zhan, H Xu, X Zhu, J Liu, YQ Zhang
ICLR 2023, 2023
38*2023
Offline reinforcement learning with soft behavior regularization
H Xu, X Zhan, J Li, H Yin
NeurIPS 2021 Offline Reinforcement Learning Workshop, 2021
312021
Model-based offline planning with trajectory pruning
X Zhan, X Zhu, H Xu
IJCAI 2022, 2021
312021
Mind the Gap: Offline Policy Optimization for Imperfect Rewards
J Li, X Hu, H Xu, J Liu, X Zhan, QS Jia, YQ Zhang
ICLR 2023, 2023
242023
Offline multi-agent reinforcement learning with implicit global-to-local value regularization
X Wang, H Xu, Y Zheng, X Zhan
NeurIPS 2023, 2023
202023
Saformer: A conditional sequence modeling approach to offline safe reinforcement learning
Q Zhang, L Zhang, H Xu, L Shen, B Wang, Y Chang, X Wang, B Yuan, ...
arXiv preprint arXiv:2301.12203, 2023
202023
Proto: Iterative policy regularized offline-to-online reinforcement learning
J Li, X Hu, H Xu, J Liu, X Zhan, YQ Zhang
arXiv preprint arXiv:2305.15669, 2023
182023
Discriminator-Guided Model-Based Offline Imitation Learning
W Zhang, H Xu, H Niu, P Cheng, M Li, H Zhang, G Zhou, X Zhan
CoRL 2022, 2022
172022
Robust spatio-temporal purchase prediction via deep meta learning
H Qin, S Ke, X Yang, H Xu, X Zhan, Y Zheng
AAAI 2021, 2021
162021
ECoalVis: visual analysis of control strategies in coal-fired power plants
S Liu, D Weng, Y Tian, Z Deng, H Xu, X Zhu, H Yin, X Zhan, Y Wu
IEEE transactions on visualization and computer graphics 29 (1), 1091-1101, 2022
132022
ODICE: Revealing the Mystery of Distribution Correction Estimation via Orthogonal-gradient Update
L Mao, H Xu, W Zhang, X Zhan
ICLR 2024, 2024
112024
Curriculum goal-conditioned imitation for offline reinforcement learning
X Feng, L Jiang, X Yu, H Xu, X Sun, J Wang, X Zhan, WK Chan
IEEE Transactions on Games 16 (1), 102-112, 2022
52022
Offline reinforcement learning with imbalanced datasets
L Jiang, S Cheng, J Qiu, H Xu, WK Chan, Z Ding
arXiv preprint arXiv:2307.02752, 2023
42023
Diffusion-DICE: In-Sample Diffusion Guidance for Offline Reinforcement Learning
L Mao, H Xu, W Zhang, X Zhan, A Zhang
NeurIPS 2024, 2024
32024
Learning to Achieve Goals with Belief State Transformers
ES Hu, K Ahn, Q Liu, H Xu, M Tomar, A Langford, D Jayaraman, A Lamb, ...
arXiv preprint arXiv:2410.23506, 2024
12024
現在システムで処理を実行できません。しばらくしてからもう一度お試しください。
論文 1–20