Grounded language acquisition through the eyes and ears of a single child

WK Vong, W Wang, AE Orhan, BM Lake - Science, 2024 - science.org
Starting around 6 to 9 months of age, children begin acquiring their first words, linking
spoken words to their visual counterparts. How much of this knowledge is learnable from …

Beyond learnability: understanding human visual development with DNNs

L Yuan - Trends in cognitive sciences, 2024 - cell.com
Abstract Recently, Orhan and Lake demonstrated the computational plausibility that children
can acquire sophisticated visual representations from natural input data without inherent …

The limitations of large language models for understanding human language and cognition

C Cuskley, R Woods, M Flaherty - Open Mind, 2024 - direct.mit.edu
Researchers have recently argued that the capabilities of Large Language Models (LLMs)
can provide new insights into longstanding debates about the role of learning and/or …

Parallel development of object recognition in newborn chicks and deep neural networks

L Pandey, D Lee, SMW Wood… - PLOS Computational …, 2024 - journals.plos.org
How do newborns learn to see? We propose that visual systems are space-time fitters,
meaning visual development can be understood as a blind fitting process (akin to evolution) …

Parallel development of social behavior in biological and artificial fish

JD McGraw, D Lee, JN Wood - Nature Communications, 2024 - nature.com
Our algorithmic understanding of vision has been revolutionized by a reverse engineering
paradigm that involves building artificial systems that perform the same tasks as biological …

Rethinking category-selectivity in human visual cortex

JB Ritchie, SG Wardle, M Vaziri-Pashkam… - arXiv preprint arXiv …, 2024 - arxiv.org
A wealth of studies report evidence that occipitotemporal cortex tessellates into" category-
selective" brain regions that are apparently specialized for representing ecologically …

Neural networks that overcome classic challenges through practice

K Irie, BM Lake - arXiv preprint arXiv:2410.10596, 2024 - arxiv.org
Since the earliest proposals for neural network models of the mind and brain, critics have
pointed out key weaknesses in these models compared to human cognitive abilities. Here …

Human Gaze Boosts Object-Centered Representation Learning

T Schaumlöffel, A Aubret, G Roig, J Triesch - arXiv preprint arXiv …, 2025 - arxiv.org
Recent self-supervised learning (SSL) models trained on human-like egocentric visual
inputs substantially underperform on image recognition tasks compared to humans. These …

The BabyView dataset: High-resolution egocentric videos of infants' and young children's everyday experiences

B Long, V Xiang, S Stojanov, RZ Sparks, Z Yin… - arXiv preprint arXiv …, 2024 - arxiv.org
Human children far exceed modern machine learning algorithms in their sample efficiency,
achieving high performance in key domains with much less data than current models …

Self-supervised learning of video representations from a child's perspective

AE Orhan, W Wang, AN Wang, M Ren… - arXiv preprint arXiv …, 2024 - arxiv.org
Children learn powerful internal models of the world around them from a few years of
egocentric visual experience. Can such internal models be learned from a child's visual …