Transformer-patcher: One mistake worth one neuron Z Huang, Y Shen, X Zhang, J Zhou, W Rong, Z Xiong arXiv preprint arXiv:2301.09785, 2023 | 144 | 2023 |
Prompt-based meta-learning for few-shot text classification H Zhang, X Zhang, H Huang, L Yu Proceedings of the 2022 conference on empirical methods in natural language …, 2022 | 44 | 2022 |
Mixture of attention heads: Selecting attention heads per token X Zhang, Y Shen, Z Huang, J Zhou, W Rong, Z Xiong arXiv preprint arXiv:2210.05144, 2022 | 37 | 2022 |
How to determine the most powerful pre-trained language model without brute force fine-tuning? an empirical survey J Bai, X Zhang, C Li, H Hong, X Xu, C Lin, W Rong arXiv preprint arXiv:2312.04775, 2023 | 9 | 2023 |
Token Relation Aware Chinese Named Entity Recognition Z Huang, W Rong, X Zhang, Y Ouyang, C Lin, Z Xiong ACM Transactions on Asian and Low-Resource Language Information Processing …, 2022 | 8 | 2022 |
Solving Math Word Problems Following Logically Consistent Template Z Huang, X Zhang, J Bai, W Rong, Y Ouyang, Z Xiong 2023 International Joint Conference on Neural Networks (IJCNN), 01-08, 2023 | 2 | 2023 |