Are deep neural networks adequate behavioral models of human visual perception?
FA Wichmann, R Geirhos - Annual Review of Vision Science, 2023 - annualreviews.org
Deep neural networks (DNNs) are machine learning algorithms that have revolutionized
computer vision due to their remarkable successes in tasks like object classification and …
computer vision due to their remarkable successes in tasks like object classification and …
Beyond neural scaling laws: beating power law scaling via data pruning
Widely observed neural scaling laws, in which error falls off as a power of the training set
size, model size, or both, have driven substantial performance improvements in deep …
size, model size, or both, have driven substantial performance improvements in deep …
D4: Improving llm pretraining via document de-duplication and diversification
Over recent years, an increasing amount of compute and data has been poured into training
large language models (LLMs), usually by doing one-pass learning on as many tokens as …
large language models (LLMs), usually by doing one-pass learning on as many tokens as …
Partial success in closing the gap between human and machine vision
A few years ago, the first CNN surpassed human performance on ImageNet. However, it
soon became clear that machines lack robustness on more challenging test cases, a major …
soon became clear that machines lack robustness on more challenging test cases, a major …
Getting aligned on representational alignment
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …
Data pruning via moving-one-sample-out
In this paper, we propose a novel data-pruning approach called moving-one-sample-out
(MoSo), which aims to identify and remove the least informative samples from the training …
(MoSo), which aims to identify and remove the least informative samples from the training …
On the diversity and realism of distilled dataset: An efficient dataset distillation paradigm
Contemporary machine learning which involves training large neural networks on massive
datasets faces significant computational challenges. Dataset distillation as a recent …
datasets faces significant computational challenges. Dataset distillation as a recent …
Less: Selecting influential data for targeted instruction tuning
Instruction tuning has unlocked powerful capabilities in large language models (LLMs),
effectively using combined datasets to develop generalpurpose chatbots. However, real …
effectively using combined datasets to develop generalpurpose chatbots. However, real …
You only condense once: Two rules for pruning condensed datasets
Dataset condensation is a crucial tool for enhancing training efficiency by reducing the size
of the training dataset, particularly in on-device scenarios. However, these scenarios have …
of the training dataset, particularly in on-device scenarios. However, these scenarios have …
Instructiongpt-4: A 200-instruction paradigm for fine-tuning minigpt-4
Multimodal large language models acquire their instruction-following capabilities through a
two-stage training process: pre-training on image-text pairs and fine-tuning on supervised …
two-stage training process: pre-training on image-text pairs and fine-tuning on supervised …