关注
Jack Rae
Jack Rae
Principal Research Scientist @ Google DeepMind
在 google.com 的电子邮件经过验证
标题
引用次数
引用次数
年份
Gpt-4 technical report
J Achiam, S Adler, S Agarwal, L Ahmad, I Akkaya, FL Aleman, D Almeida, ...
arXiv preprint arXiv:2303.08774, 2023
21852023
Training compute-optimal large language models
J Hoffmann, S Borgeaud, A Mensch, E Buchatskaya, T Cai, E Rutherford, ...
arXiv preprint arXiv:2203.15556, 2022
11432022
Gemini: a family of highly capable multimodal models
G Team, R Anil, S Borgeaud, Y Wu, JB Alayrac, J Yu, R Soricut, ...
arXiv preprint arXiv:2312.11805, 2023
9242023
Scaling Language Models: Methods, Analysis & Insights from Training Gopher
JW Rae, S Borgeaud, T Cai, K Millican, J Hoffmann, F Song, J Aslanides, ...
9192021
A clinically applicable approach to continuous prediction of future acute kidney injury
N Tomašev, X Glorot, JW Rae, M Zielinski, H Askham, A Saraiva, ...
Nature 572 (7767), 116-119, 2019
8622019
Improving language models by retrieving from trillions of tokens
S Borgeaud, A Mensch, J Hoffmann, T Cai, E Rutherford, K Millican, ...
International conference on machine learning, 2206-2240, 2022
7642022
Compressive transformers for long-range sequence modelling
JW Rae, A Potapenko, SM Jayakumar, TP Lillicrap
arXiv preprint arXiv:1911.05507, 2019
4942019
Stabilizing transformers for reinforcement learning
E Parisotto, F Song, J Rae, R Pascanu, C Gulcehre, S Jayakumar, ...
International conference on machine learning, 7487-7498, 2020
3562020
Model-free episodic control
C Blundell, B Uria, A Pritzel, Y Li, A Ruderman, JZ Leibo, J Rae, ...
arXiv preprint arXiv:1606.04460, 2016
2952016
Relational recurrent neural networks
A Santoro, R Faulkner, D Raposo, J Rae, M Chrzanowski, T Weber, ...
Advances in neural information processing systems 31, 2018
2642018
Neural arithmetic logic units
A Trask, F Hill, SE Reed, J Rae, C Dyer, P Blunsom
Advances in neural information processing systems 31, 2018
2422018
Unsupervised predictive memory in a goal-directed agent
G Wayne, CC Hung, D Amos, M Mirza, A Ahuja, A Grabska-Barwinska, ...
arXiv preprint arXiv:1803.10760, 2018
1962018
Scaling memory-augmented neural networks with sparse reads and writes
J Rae, JJ Hunt, I Danihelka, T Harley, AW Senior, G Wayne, A Graves, ...
Advances in Neural Information Processing Systems 29, 2016
1822016
Reducing sentiment bias in language models via counterfactual evaluation
PS Huang, H Zhang, R Jiang, R Stanforth, J Welbl, J Rae, V Maini, ...
arXiv preprint arXiv:1911.03064, 2019
1772019
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
M Reid, N Savinov, D Teplyashin, D Lepikhin, T Lillicrap, J Alayrac, ...
arXiv preprint arXiv:2403.05530, 2024
1562024
Multiplicative interactions and where to find them
SM Jayakumar, WM Czarnecki, J Menick, J Schwarz, J Rae, S Osindero, ...
International conference on learning representations, 2020
1272020
V-mpo: On-policy maximum a posteriori policy optimization for discrete and continuous control
HF Song, A Abdolmaleki, JT Springenberg, A Clark, H Soyer, JW Rae, ...
arXiv preprint arXiv:1909.12238, 2019
1112019
Memory-based parameter adaptation
P Sprechmann, SM Jayakumar, JW Rae, A Pritzel, AP Badia, B Uria, ...
International Conference on Learning Representations, 2018
1072018
An empirical analysis of compute-optimal large language model training
J Hoffmann, S Borgeaud, A Mensch, E Buchatskaya, T Cai, E Rutherford, ...
Advances in Neural Information Processing Systems 35, 30016-30030, 2022
1042022
Top-kast: Top-k always sparse training
S Jayakumar, R Pascanu, J Rae, S Osindero, E Elsen
Advances in Neural Information Processing Systems 33, 20744-20754, 2020
902020
系统目前无法执行此操作,请稍后再试。
文章 1–20