Suivre
Yaru Hao
Yaru Hao
Microsoft Research Asia
Adresse e-mail validée de microsoft.com
Titre
Citée par
Citée par
Année
Knowledge neurons in pretrained transformers
D Dai, L Dong, Y Hao, Z Sui, B Chang, F Wei
arXiv preprint arXiv:2104.08696, 2021
2802021
Language is not all you need: Aligning perception with language models
S Huang, L Dong, W Wang, Y Hao, S Singhal, S Ma, T Lv, L Cui, ...
Advances in Neural Information Processing Systems 36, 2024
2632024
Kosmos-2: Grounding multimodal large language models to the world
Z Peng, W Wang, L Dong, Y Hao, S Huang, S Ma, F Wei
arXiv preprint arXiv:2306.14824, 2023
2362023
Why can gpt learn in-context? language models implicitly perform gradient descent as meta-optimizers
D Dai, Y Sun, L Dong, Y Hao, S Ma, Z Sui, F Wei
arXiv preprint arXiv:2212.10559, 2022
2052022
Visualizing and understanding the effectiveness of BERT
Y Hao, L Dong, F Wei, K Xu
arXiv preprint arXiv:1908.05620, 2019
1902019
Self-attention attribution: Interpreting information interactions inside transformer
Y Hao, L Dong, F Wei, K Xu
Proceedings of the AAAI Conference on Artificial Intelligence 35 (14), 12963 …, 2021
1592021
Language models are general-purpose interfaces
Y Hao, H Song, L Dong, S Huang, Z Chi, W Wang, S Ma, F Wei
arXiv preprint arXiv:2206.06336, 2022
832022
Optimizing prompts for text-to-image generation
Y Hao, Z Chi, L Dong, F Wei
Advances in Neural Information Processing Systems 36, 2024
752024
Subhojit Som, Xia Song, and Furu Wei
S Huang, L Dong, W Wang, Y Hao, S Singhal, S Ma, T Lv, L Cui, ...
Language is not all you need: Aligning perception with language models …, 2023
572023
Investigating learning dynamics of BERT fine-tuning
Y Hao, L Dong, F Wei, K Xu
Proceedings of the 1st Conference of the Asia-Pacific Chapter of the …, 2020
392020
Prototypical calibration for few-shot learning of language models
Z Han, Y Hao, L Dong, Y Sun, F Wei
The Eleventh International Conference on Learning Representations, 2022
242022
Structured prompting: Scaling in-context learning to 1,000 examples
Y Hao, Y Sun, L Dong, Z Han, Y Gu, F Wei
arXiv preprint arXiv:2212.06713, 2022
232022
Large language model for science: A study on p vs. np
Q Dong, L Dong, K Xu, G Zhou, Y Hao, Z Sui, F Wei
arXiv preprint arXiv:2309.05689, 2023
102023
Prototypical fine-tuning: Towards robust performance under varying data sizes
Y Jin, X Wang, Y Hao, Y Sun, X Xie
Proceedings of the AAAI Conference on Artificial Intelligence 37 (11), 12968 …, 2023
82023
Learning to sample replacements for electra pre-training
Y Hao, L Dong, H Bao, K Xu, F Wei
arXiv preprint arXiv:2106.13715, 2021
82021
Grounding multimodal large language models to the world
Z Peng, W Wang, L Dong, Y Hao, S Huang, S Ma, Q Ye, F Wei
The Twelfth International Conference on Learning Representations, 2023
42023
Towards Optimal Learning of Language Models
Y Gu, L Dong, Y Hao, Q Dong, M Huang, F Wei
arXiv preprint arXiv:2402.17759, 2024
2024
Le système ne peut pas réaliser cette opération maintenant. Veuillez réessayer plus tard.
Articles 1–17