A survey on curriculum learning

X Wang, Y Chen, W Zhu - IEEE transactions on pattern analysis …, 2021 - ieeexplore.ieee.org
Curriculum learning (CL) is a training strategy that trains a machine learning model from
easier data to harder data, which imitates the meaningful learning order in human curricula …

Curriculum learning: A survey

P Soviany, RT Ionescu, P Rota, N Sebe - International Journal of …, 2022 - Springer
Training machine learning models in a meaningful order, from the easy samples to the hard
ones, using curriculum learning can provide performance improvements over the standard …

Skill-it! a data-driven skills framework for understanding and training language models

M Chen, N Roberts, K Bhatia, J Wang… - Advances in …, 2024 - proceedings.neurips.cc
The quality of training data impacts the performance of pre-trained large language models
(LMs). Given a fixed budget of tokens, we study how to best select data that leads to good …

Identifying mislabeled data using the area under the margin ranking

G Pleiss, T Zhang, E Elenberg… - Advances in Neural …, 2020 - proceedings.neurips.cc
Not all data in a typical training set help with generalization; some samples can be overly
ambiguous or outrightly mislabeled. This paper introduces a new method to identify such …

Learning fast sample re-weighting without reward data

Z Zhang, T Pfister - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Training sample re-weighting is an effective approach for tackling data biases such as
imbalanced and corrupted labels. Recent methods develop learning-based algorithms to …

Beyond synthetic noise: Deep learning on controlled noisy labels

L Jiang, D Huang, M Liu… - … conference on machine …, 2020 - proceedings.mlr.press
Performing controlled experiments on noisy data is essential in understanding deep
learning across noise levels. Due to the lack of suitable datasets, previous research has only …

When do curricula work?

X Wu, E Dyer, B Neyshabur - arXiv preprint arXiv:2012.03107, 2020 - arxiv.org
Inspired by human learning, researchers have proposed ordering examples during training
based on their difficulty. Both curriculum learning, exposing a network to easier examples …

Superloss: A generic loss for robust curriculum learning

T Castells, P Weinzaepfel… - Advances in Neural …, 2020 - proceedings.neurips.cc
Curriculum learning is a technique to improve a model performance and generalization
based on the idea that easy samples should be presented before difficult ones during …

Correlated input-dependent label noise in large-scale image classification

M Collier, B Mustafa, E Kokiopoulou… - Proceedings of the …, 2021 - openaccess.thecvf.com
Large scale image classification datasets often contain noisy labels. We take a principled
probabilistic approach to modelling input-dependent, also known as heteroscedastic, label …

Provable advantage of curriculum learning on parity targets with mixed inputs

E Abbe, E Cornacchia, A Lotfi - Advances in Neural …, 2023 - proceedings.neurips.cc
Experimental results have shown that curriculum learning, ie, presenting simpler examples
before more complex ones, can improve the efficiency of learning. Some recent theoretical …