关注
Ofir Press
Ofir Press
在 princeton.edu 的电子邮件经过验证 - 首页
标题
引用次数
引用次数
年份
Bloom: A 176b-parameter open-access multilingual language model
T Le Scao, A Fan, C Akiki, E Pavlick, S Ilić, D Hesslow, R Castagné, ...
12802023
Using the Output Embedding to Improve Language Models
O Press, L Wolf
EACL 2017, 2017
7522017
Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation
O Press, NA Smith, M Lewis
ICLR 2022, 2021
4072021
Measuring and narrowing the compositionality gap in language models
O Press, M Zhang, S Min, L Schmidt, NA Smith, M Lewis
Findings of EMNLP 2023, 2022
277*2022
How language model hallucinations can snowball
M Zhang, O Press, W Merrill, A Liu, NA Smith
arXiv preprint arXiv:2305.13534, 2023
1532023
Language Generation with Recurrent Generative Adversarial Networks without Pre-training
O Press, A Bar, B Bogin, J Berant, L Wolf
1st Workshop on Learning to Generate Natural Language at ICML 2017, 2017
1372017
What Language Model to Train if You Have One Million GPU Hours?
T Le Scao, T Wang, D Hesslow, L Saulnier, S Bekman, MS Bari, ...
Findings of EMNLP 2022, 2022
802022
SWE-bench: Can Language Models Resolve Real-World GitHub Issues?
CE Jimenez, J Yang, A Wettig, S Yao, K Pei, O Press, K Narasimhan
ICLR 2024, 2023
772023
Improving Transformer Models by Reordering their Sublayers
O Press, NA Smith, O Levy
ACL 2020, 2019
732019
Shortformer: Better Language Modeling using Shorter Inputs
O Press, NA Smith, M Lewis
ACL 2021, 2020
692020
Transformer Language Models without Positional Encodings Still Learn Positional Information
A Haviv, O Ram, O Press, P Izsak, O Levy
Findings of EMNLP 2022, 2022
622022
You may not need attention
O Press, NA Smith
arXiv preprint arXiv:1810.13409, 2018
282018
SWE-agent: Agent-Computer Interfaces Enable Automated Software Engineering
J Yang, CE Jimenez, A Wettig, K Lieret, S Yao, K Narasimhan, O Press
52024
Partially shuffling the training data to improve language models
O Press
arXiv preprint arXiv:1903.04167, 2019
42019
Complementing Scale: Novel Guidance Methods for Improving Language Models
O Press
University of Washington, 2023
2023
系统目前无法执行此操作,请稍后再试。
文章 1–15