Segui
Hongkun Yu
Hongkun Yu
Email verificata su google.com
Titolo
Citata da
Citata da
Anno
Scaling instruction-finetuned language models
HW Chung, L Hou, S Longpre, B Zoph, Y Tay, W Fedus, Y Li, X Wang, ...
Journal of Machine Learning Research 25 (70), 1-53, 2024
32752024
Gemini: a family of highly capable multimodal models
G Team, R Anil, S Borgeaud, JB Alayrac, J Yu, R Soricut, J Schalkwyk, ...
arXiv preprint arXiv:2312.11805, 2023
24942023
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
G Team, P Georgiev, VI Lei, R Burnell, L Bai, A Gulati, G Tanzer, ...
arXiv preprint arXiv:2403.05530, 2024
9662024
Mobilebert: a compact task-agnostic bert for resource-limited devices
Z Sun, H Yu, X Song, R Liu, Y Yang, D Zhou
arXiv preprint arXiv:2004.02984, 2020
8432020
Large language models can self-improve
J Huang, SS Gu, L Hou, Y Wu, X Wang, H Yu, J Han
arXiv preprint arXiv:2210.11610, 2022
4902022
TensorFlow model garden
H Yu, C Chen, X Du, Y Li, A Rashwan, L Hou, P Jin, F Yang, F Liu, J Kim, ...
Model Garden for TensorFlow., 2020
1092020
Huai hsin Chi, Jeff Dean, Jacob Devlin, Adam Roberts, Denny Zhou, Quoc V
HW Chung, L Hou, S Longpre, B Zoph, Y Tay, W Fedus, E Li, X Wang, ...
Le, and Jason Wei, 2022
1032022
Latent factor transition for dynamic collaborative filtering
C Zhang, K Wang, H Yu, J Sun, EP Lim
Proceedings of the 2014 SIAM international conference on data mining, 452-460, 2014
972014
Generating representative headlines for news stories
X Gu, Y Mao, J Han, J Liu, Y Wu, C Yu, D Finnie, H Yu, J Zhai, N Zukoski
Proceedings of The Web Conference 2020, 1773-1784, 2020
802020
Mixture-of-experts meets instruction tuning: A winning combination for large language models
S Shen, L Hou, Y Zhou, N Du, S Longpre, J Wei, HW Chung, B Zoph, ...
arXiv preprint arXiv:2305.14705, 2023
672023
Scaling instruction-finetuned language models (2022)
HW Chung, L Hou, S Longpre, B Zoph, Y Tay, W Fedus, Y Li, X Wang, ...
arXiv preprint arXiv:2210.11416 3, 2022
632022
On the transformer growth for progressive bert training
X Gu, L Liu, H Yu, J Li, C Chen, J Han
arXiv preprint arXiv:2010.12562, 2020
592020
Mining multi-aspect reflection of news events in twitter: Discovery, linking and presentation
J Wang, W Tong, H Yu, M Li, X Ma, H Cai, T Hanratty, J Han
2015 IEEE International Conference on Data Mining, 429-438, 2015
412015
Flan-moe: Scaling instruction-finetuned language models with sparse mixture of experts
S Shen, L Hou, Y Zhou, N Du, S Longpre, J Wei, HW Chung, B Zoph, ...
arXiv preprint arXiv:2305.14705 2, 2023
302023
Are features equally representative? A feature-centric recommendation
C Zhang, K Wang, E Lim, Q Xu, J Sun, H Yu
Proceedings of the AAAI Conference on Artificial Intelligence 29 (1), 2015
242015
Data-driven contextual valence shifter quantification for multi-theme sentiment analysis
H Yu, J Shang, M Hsu, M Castellanos, J Han
Proceedings of the 25th ACM international on conference on information and …, 2016
212016
Enct5: Fine-tuning t5 encoder for non-autoregressive tasks
F Liu, S Shakeri, H Yu, J Li
arXiv preprint arXiv:2110.08426 2, 2021
192021
Multi-step problem solving through a verifier: An empirical analysis on model-induced process supervision
Z Wang, Y Li, Y Wu, L Luo, L Hou, H Yu, J Shang
arXiv preprint arXiv:2402.02658, 2024
182024
TensorFlow model garden. 2020
H Yu, C Chen, X Du, Y Li, A Rashwan, L Hou, P Jin, F Yang, F Liu, J Kim, ...
URL https://github. com/tensorflow/models, 2020
182020
Mobilebert: Task-agnostic compression of bert by progressive knowledge transfer
Z Sun, H Yu, X Song, R Liu, Y Yang, D Zhou
172019
Il sistema al momento non può eseguire l'operazione. Riprova più tardi.
Articoli 1–20