mplug-owl: Modularization empowers large language models with multimodality Q Ye, H Xu, G Xu, J Ye, M Yan, Y Zhou, J Wang, A Hu, P Shi, Y Shi, C Li, ... arXiv preprint arXiv:2304.14178, 2023 | 828 | 2023 |
mplug: Effective and efficient vision-language learning by cross-modal skip-connections C Li, H Xu, J Tian, W Wang, M Yan, B Bi, J Ye, H Chen, G Xu, Z Cao, ... arXiv preprint arXiv:2205.12005, 2022 | 178* | 2022 |
Guiding generation for abstractive text summarization based on key information guide network C Li, W Xu, S Li, S Gao Proceedings of the 2018 Conference of the North American Chapter of the …, 2018 | 166 | 2018 |
StructuralLM: Structural pre-training for form understanding C Li, B Bi, M Yan, W Wang, S Huang, F Huang, L Si arXiv preprint arXiv:2105.11210, 2021 | 130 | 2021 |
mplug-2: A modularized multi-modal foundation model across text, image and video H Xu, Q Ye, M Yan, Y Shi, J Ye, Y Xu, C Li arXiv preprint arXiv:2302.00402 3, 2023 | 129 | 2023 |
E2E-VLP: End-to-end vision-language pre-training enhanced by visual learning H Xu, M Yan, C Li, B Bi, S Huang, W Xiao, F Huang arXiv preprint arXiv:2106.01804, 2021 | 119 | 2021 |
Ureader: Universal ocr-free visually-situated language understanding with multimodal large language model J Ye, A Hu, H Xu, Q Ye, M Yan, G Xu, C Li, J Tian, Q Qian, J Zhang, Q Jin, ... arXiv preprint arXiv:2310.05126, 2023 | 113 | 2023 |
mplug-docowl: Modularized multimodal large language model for document understanding J Ye, A Hu, H Xu, Q Ye, M Yan, Y Dan, C Zhao, G Xu, C Li, J Tian, Q Qi, ... arXiv preprint arXiv:2307.02499, 2023 | 109 | 2023 |
Palm: Pre-training an autoencoding&autoregressive language model for context-conditioned generation B Bi, C Li, C Wu, M Yan, W Wang, S Huang, F Huang, L Si arXiv preprint arXiv:2004.07159, 2020 | 76 | 2020 |
Bin Bi, Jiabo Ye, Hehong Chen, Guohai Xu, Zheng Cao, et al. mplug: Effective and efficient vision-language learning by cross-modal skip-connections C Li, H Xu, J Tian, W Wang, M Yan arXiv preprint arXiv:2205.12005 1 (2), 2022 | 57 | 2022 |
Incorporating external knowledge into machine reading for generative question answering B Bi, C Wu, M Yan, W Wang, J Xia, C Li arXiv preprint arXiv:1909.02745, 2019 | 51 | 2019 |
Small llms are weak tool learners: A multi-llm agent W Shen, C Li, H Chen, M Yan, X Quan, H Chen, J Zhang, F Huang arXiv preprint arXiv:2401.07324, 2024 | 33 | 2024 |
mplug-paperowl: Scientific diagram analysis with the multimodal large language model A Hu, Y Shi, H Xu, J Ye, Q Ye, M Yan, C Li, Q Qian, J Zhang, F Huang Proceedings of the 32nd ACM International Conference on Multimedia, 6929-6938, 2024 | 30 | 2024 |
Multi-task learning for abstractive text summarization with key information guide network W Xu, C Li, M Lee, C Zhang EURASIP Journal on Advances in Signal Processing 2020, 1-11, 2020 | 30 | 2020 |
Modelscope-agent: Building your customizable agent system with open-source large language models C Li, H Chen, M Yan, W Shen, H Xu, Z Wu, Z Zhang, W Zhou, Y Chen, ... arXiv preprint arXiv:2309.00986, 2023 | 25 | 2023 |
Addressing semantic drift in generative question answering with auxiliary extraction C Li, B Bi, M Yan, W Wang, S Huang Proceedings of the 59th Annual Meeting of the Association for Computational …, 2021 | 25 | 2021 |
Bin Bi, Wei Wang, Jiangnan Xia, and Luo Si. 2019. IDST at TREC 2019 deep learning track: Deep cascade ranking with generation-based document expansion and pre-trained language … M Yan, C Li, C Wu Proceedings of the Twenty-Eighth Text REtrieval Conference, TREC, 13-15, 2019 | 25 | 2019 |
SemVLP: Vision-language pre-training by aligning semantics at multiple levels C Li, M Yan, H Xu, F Luo, W Wang, B Bi, S Huang arXiv preprint arXiv:2103.07829, 2021 | 23 | 2021 |
IDST at TREC 2019 Deep Learning Track: Deep Cascade Ranking with Generation-based Document Expansion and Pre-trained Language Modeling. M Yan, C Li, C Wu, B Bi, W Wang, J Xia, L Si TREC, 2019 | 22 | 2019 |
Youku-mplug: A 10 million large-scale chinese video-language dataset for pre-training and benchmarks H Xu, Q Ye, X Wu, M Yan, Y Miao, J Ye, G Xu, A Hu, Y Shi, G Xu, C Li, ... arXiv preprint arXiv:2306.04362, 2023 | 21 | 2023 |