The rise and potential of large language model based agents: A survey Z Xi, W Chen, X Guo, W He, Y Ding, B Hong, M Zhang, J Wang, S Jin, ... Science China Information Sciences 68 (2), 121101, 2025 | 773 | 2025 |
Secrets of rlhf in large language models part ii: Reward modeling B Wang, R Zheng, L Chen, Y Liu, S Dou, C Huang, W Shen, S Jin, E Zhou, ... arXiv preprint arXiv:2401.06080, 2024 | 70 | 2024 |
The rise and potential of large language model based agents: A survey, 2023 Z Xi, W Chen, X Guo, W He, Y Ding, B Hong, M Zhang, J Wang, S Jin, ... arXiv preprint arXiv:2309.07864, 2023 | 32 | 2023 |
Zhiheng Xi B Wang, R Zheng, L Chen, Y Liu, S Dou, C Huang, W Shen, S Jin, E Zhou, ... Jun Zhao, Xiao Wang, Tao Ji, Hang Yan, Lixing Shen, Zhan Chen, Tao Gui, Qi …, 2024 | 22 | 2024 |
The rise and potential of large language model based agents: A survey. arXiv 2023 Z Xi, W Chen, X Guo, W He, Y Ding, B Hong, M Zhang, J Wang, S Jin, ... arXiv preprint arXiv:2309.07864, 2023 | 22 | 2023 |
Loramoe: Revolutionizing mixture of experts for maintaining world knowledge in language model alignment S Dou, E Zhou, Y Liu, S Gao, J Zhao, W Shen, Y Zhou, Z Xi, X Wang, ... arXiv preprint arXiv:2312.09979 4 (7), 2023 | 19 | 2023 |
ToolEyes: fine-grained evaluation for tool learning capabilities of large language models in real-world scenarios J Ye, G Li, S Gao, C Huang, Y Wu, S Li, X Fan, S Dou, Q Zhang, T Gui, ... arXiv preprint arXiv:2401.00741, 2024 | 18 | 2024 |
LoRAMoE: Alleviate world knowledge forgetting in large language models via MoE-style plugin S Dou, E Zhou, Y Liu, S Gao, J Zhao, W Shen, Y Zhou, Z Xi, X Wang, ... arXiv preprint arXiv:2312.09979, 2023 | 17 | 2023 |
Training large language models for reasoning through reverse curriculum reinforcement learning Z Xi, W Chen, B Hong, S Jin, R Zheng, W He, Y Ding, S Liu, X Guo, ... arXiv preprint arXiv:2402.05808, 2024 | 16 | 2024 |
Mousi: Poly-visual-expert vision-language models X Fan, T Ji, C Jiang, S Li, S Jin, S Song, J Wang, B Hong, L Chen, ... arXiv preprint arXiv:2401.17221, 2024 | 14 | 2024 |
Stepcoder: Improve code generation with reinforcement learning from compiler feedback S Dou, Y Liu, H Jia, L Xiong, E Zhou, W Shen, J Shan, C Huang, X Wang, ... arXiv preprint arXiv:2402.01391, 2024 | 12 | 2024 |
Zhiheng Xi, et al. 2024. StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback S Dou, Y Liu, H Jia, L Xiong, E Zhou, J Shan, C Huang, W Shen, X Fan arXiv preprint arXiv:2402.01391, 2024 | 7 | 2024 |
The rise and potential of large language model based agents: a survey. CoRR abs/2309.07864 (2023) Z Xi, W Chen, X Guo, W He, Y Ding, B Hong, M Zhang, J Wang, S Jin, ... | 7 | 2023 |
The art of balancing: revolutionizing mixture of experts for maintaining world knowledge in language model alignment S Dou, E Zhou, Y Liu, S Gao, J Zhao, W Shen, Y Zhou, Z Xi, X Wang, ... arXiv preprint arXiv:2312.09979 5, 2023 | 4 | 2023 |
abcbpc at SemEval-2021 Task 7: ERNIE-based multi-task model for detecting and rating humor and offense C Pang, X Fan, W Su, X Chen, S Wang, J Liu, X Ouyang, S Feng, Y Sun Proceedings of the 15th International Workshop on Semantic Evaluation …, 2021 | 3 | 2021 |
RealBehavior: A Framework for Faithfully Characterizing Foundation Models' Human-like Behavior Mechanisms E Zhou, R Zheng, Z Xi, S Gao, X Fan, Z Fei, J Ye, T Gui, Q Zhang, ... arXiv preprint arXiv:2310.11227, 2023 | 1 | 2023 |
Predicting Large Language Model Capabilities on Closed-Book QA Tasks Using Only Information Available Prior to Training C Jiang, M Zhang, J Ye, X Fan, Y Cao, J Sun, Z Xi, S Dou, Y Dong, ... arXiv preprint arXiv:2502.04066, 2025 | | 2025 |
Distill Visual Chart Reasoning Ability from LLMs to MLLMs W He, Z Xi, W Zhao, X Fan, Y Ding, Z Shan, T Gui, Q Zhang, X Huang arXiv preprint arXiv:2410.18798, 2024 | | 2024 |
Have the VLMs Lost Confidence? A Study of Sycophancy in VLMs S Li, T Ji, X Fan, L Lu, L Yang, Y Yang, Z Xi, R Zheng, Y Wang, X Zhao, ... arXiv preprint arXiv:2410.11302, 2024 | | 2024 |
ERNIE-SAT: Speech and Text Joint Pretraining for Cross-Lingual Multi-Speaker Text-to-Speech X Fan, C Pang, T Yuan, H Bai, R Zheng, P Zhu, S Wang, J Chen, Z Chen, ... arXiv preprint arXiv:2211.03545, 2022 | | 2022 |