Deepseek-v2: A strong, economical, and efficient mixture-of-experts language model A Liu, B Feng, B Wang, B Wang, B Liu, C Zhao, C Dengr, C Ruan, D Dai, ... arXiv preprint arXiv:2405.04434, 2024 | 104 | 2024 |
MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback X Wang*, Z Wang*, J Liu, Y Chen, L Yuan, H Peng, H Ji ICLR 2024, 2023 | 99 | 2023 |
NOVO: learnable and interpretable document identifiers for model-based IR Z Wang, Y Zhou, Y Tu, Z Dou CIKM 2023, 2023 | 25 | 2023 |
RETA-LLM: A Retrieval-Augmented Large Language Model Toolkit J Liu, J Jin, Z Wang, J Cheng, Z Dou, J Wen arXiv preprint arXiv:2306.05212, 2023 | 2* | 2023 |
RAGEN: A General-Purpose Reasoning Agent Training Framework Z Wang, K Wang, Q Wang, P Zhang, M Li Github: https://github.com/ZihanWang314/ragen, 2025 | | 2025 |
Let the Expert Stick to His Last: Expert-Specialized Fine-Tuning for Sparse Architectural Large Language Models Z Wang, D Chen, D Dai, R Xu, Z Li, Y Wu EMNLP 2024, 2024 | | 2024 |
Learning on Structured Documents for Conditional Question Answering W Zihan, Q Hongjin, D Zhicheng CCL 2023, 2023 | | 2023 |