关注
Yaru Hao
Yaru Hao
Microsoft Research Asia
在 microsoft.com 的电子邮件经过验证
标题
引用次数
引用次数
年份
Knowledge neurons in pretrained transformers
D Dai, L Dong, Y Hao, Z Sui, B Chang, F Wei
arXiv preprint arXiv:2104.08696, 2021
3482021
Kosmos-2: Grounding multimodal large language models to the world
Z Peng, W Wang, L Dong, Y Hao, S Huang, S Ma, F Wei
arXiv preprint arXiv:2306.14824, 2023
3232023
Language is not all you need: Aligning perception with language models
S Huang, L Dong, W Wang, Y Hao, S Singhal, S Ma, T Lv, L Cui, ...
Advances in Neural Information Processing Systems 36, 2024
3142024
Why can gpt learn in-context? language models implicitly perform gradient descent as meta-optimizers
D Dai, Y Sun, L Dong, Y Hao, S Ma, Z Sui, F Wei
arXiv preprint arXiv:2212.10559, 2022
2472022
Visualizing and understanding the effectiveness of BERT
Y Hao, L Dong, F Wei, K Xu
arXiv preprint arXiv:1908.05620, 2019
2152019
Self-attention attribution: Interpreting information interactions inside transformer
Y Hao, L Dong, F Wei, K Xu
Proceedings of the AAAI Conference on Artificial Intelligence 35 (14), 12963 …, 2021
1832021
Optimizing prompts for text-to-image generation
Y Hao, Z Chi, L Dong, F Wei
Advances in Neural Information Processing Systems 36, 2024
932024
Language models are general-purpose interfaces
Y Hao, H Song, L Dong, S Huang, Z Chi, W Wang, S Ma, F Wei
arXiv preprint arXiv:2206.06336, 2022
882022
Investigating learning dynamics of BERT fine-tuning
Y Hao, L Dong, F Wei, K Xu
Proceedings of the 1st Conference of the Asia-Pacific Chapter of the …, 2020
442020
Subhojit Som, Xia Song, and Furu Wei. Language is not all you need: Aligning perception with language models
S Huang, L Dong, W Wang, Y Hao, S Singhal, S Ma, T Lv, L Cui, ...
arXiv preprint arXiv:2302.14045 1 (2), 3, 2023
402023
Subhojit Som, Xia Song, and Furu Wei
S Huang, L Dong, W Wang, Y Hao, S Singhal, S Ma, T Lv, L Cui, ...
Language is not all you need: Aligning perception with language models …, 2023
372023
Prototypical calibration for few-shot learning of language models
Z Han, Y Hao, L Dong, Y Sun, F Wei
The Eleventh International Conference on Learning Representations, 2023
292023
Structured prompting: Scaling in-context learning to 1,000 examples
Y Hao, Y Sun, L Dong, Z Han, Y Gu, F Wei
arXiv preprint arXiv:2212.06713, 2022
292022
Large language model for science: A study on P vs. NP
Q Dong, L Dong, K Xu, G Zhou, Y Hao, Z Sui, F Wei
arXiv preprint arXiv:2309.05689, 2023
132023
Grounding multimodal large language models to the world
Z Peng, W Wang, L Dong, Y Hao, S Huang, S Ma, Q Ye, F Wei
The Twelfth International Conference on Learning Representations, 2024
102024
Prototypical fine-tuning: Towards robust performance under varying data sizes
Y Jin, X Wang, Y Hao, Y Sun, X Xie
Proceedings of the AAAI Conference on Artificial Intelligence 37 (11), 12968 …, 2023
92023
Learning to sample replacements for electra pre-training
Y Hao, L Dong, H Bao, K Xu, F Wei
arXiv preprint arXiv:2106.13715, 2021
82021
Towards Optimal Learning of Language Models
Y Gu, L Dong, Y Hao, Q Dong, M Huang, F Wei
arXiv preprint arXiv:2402.17759, 2024
2024
系统目前无法执行此操作,请稍后再试。
文章 1–18