Are deep neural networks adequate behavioral models of human visual perception?

FA Wichmann, R Geirhos - Annual Review of Vision Science, 2023 - annualreviews.org
Deep neural networks (DNNs) are machine learning algorithms that have revolutionized
computer vision due to their remarkable successes in tasks like object classification and …

Beyond neural scaling laws: beating power law scaling via data pruning

B Sorscher, R Geirhos, S Shekhar… - Advances in …, 2022 - proceedings.neurips.cc
Widely observed neural scaling laws, in which error falls off as a power of the training set
size, model size, or both, have driven substantial performance improvements in deep …

D4: Improving llm pretraining via document de-duplication and diversification

K Tirumala, D Simig, A Aghajanyan… - Advances in Neural …, 2023 - proceedings.neurips.cc
Over recent years, an increasing amount of compute and data has been poured into training
large language models (LLMs), usually by doing one-pass learning on as many tokens as …

Partial success in closing the gap between human and machine vision

R Geirhos, K Narayanappa, B Mitzkus… - Advances in …, 2021 - proceedings.neurips.cc
A few years ago, the first CNN surpassed human performance on ImageNet. However, it
soon became clear that machines lack robustness on more challenging test cases, a major …

Getting aligned on representational alignment

I Sucholutsky, L Muttenthaler, A Weller, A Peng… - arXiv preprint arXiv …, 2023 - arxiv.org
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …

Data pruning via moving-one-sample-out

H Tan, S Wu, F Du, Y Chen, Z Wang… - Advances in Neural …, 2024 - proceedings.neurips.cc
In this paper, we propose a novel data-pruning approach called moving-one-sample-out
(MoSo), which aims to identify and remove the least informative samples from the training …

On the diversity and realism of distilled dataset: An efficient dataset distillation paradigm

P Sun, B Shi, D Yu, T Lin - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
Contemporary machine learning which involves training large neural networks on massive
datasets faces significant computational challenges. Dataset distillation as a recent …

Less: Selecting influential data for targeted instruction tuning

M Xia, S Malladi, S Gururangan, S Arora… - arXiv preprint arXiv …, 2024 - arxiv.org
Instruction tuning has unlocked powerful capabilities in large language models (LLMs),
effectively using combined datasets to develop generalpurpose chatbots. However, real …

You only condense once: Two rules for pruning condensed datasets

Y He, L Xiao, JT Zhou - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Dataset condensation is a crucial tool for enhancing training efficiency by reducing the size
of the training dataset, particularly in on-device scenarios. However, these scenarios have …

Instructiongpt-4: A 200-instruction paradigm for fine-tuning minigpt-4

L Wei, Z Jiang, W Huang, L Sun - arXiv preprint arXiv:2308.12067, 2023 - arxiv.org
Multimodal large language models acquire their instruction-following capabilities through a
two-stage training process: pre-training on image-text pairs and fine-tuning on supervised …