The boosting approach to machine learning: An overview

RE Schapire - Nonlinear estimation and classification, 2003 - Springer
Boosting is a general method for improving the accuracy of any given learning algorithm.
Focusing primarily on the AdaBoost algorithm, this chapter overviews some of the recent …

Divergence measures for statistical data processing—An annotated bibliography

M Basseville - Signal Processing, 2013 - Elsevier
Divergence measures for statistical data processing—An annotated bibliography -
ScienceDirect Skip to main contentSkip to article Elsevier logo Journals & Books Search …

Boosting: Foundations and algorithms

RE Schapire, Y Freund - Kybernetes, 2013 - emerald.com
The term “boosting” denotes a powerful means of facilitating machine learning that was
invented by the book's authors 20 years ago and intensively developed since. Despite this …

[PDF][PDF] Foundations of machine learning

M Mohri - 2018 - dlib.hust.edu.vn
A new edition of a graduate-level machine learning textbook that focuses on the analysis
and theory of algorithms. This book is a general introduction to machine learning that can …

[图书][B] Information geometry and its applications

S Amari - 2016 - books.google.com
This is the first comprehensive book on information geometry, written by the founder of the
field. It begins with an elementary introduction to dualistic geometry and proceeds to a wide …

Convexity, classification, and risk bounds

PL Bartlett, MI Jordan, JD McAuliffe - Journal of the American …, 2006 - Taylor & Francis
Many of the classification algorithms developed in the machine learning literature, including
the support vector machine and boosting, can be viewed as minimum contrast methods that …

Learning from the wisdom of crowds by minimax entropy

D Zhou, S Basu, Y Mao, J Platt - Advances in neural …, 2012 - proceedings.neurips.cc
An important way to make large training sets is to gather noisy labels from crowds of
nonexperts. We propose a minimax entropy principle to improve the quality of these labels …

Dual coordinate descent methods for logistic regression and maximum entropy models

HF Yu, FL Huang, CJ Lin - Machine Learning, 2011 - Springer
Most optimization methods for logistic regression or maximum entropy solve the primal
problem. They range from iterative scaling, coordinate descent, quasi-Newton, and …

Herding dynamical weights to learn

M Welling - Proceedings of the 26th annual international …, 2009 - dl.acm.org
A new" herding" algorithm is proposed which directly converts observed moments into a
sequence of pseudo-samples. The pseudo-samples respect the moment constraints and …

Discriminative reranking for natural language parsing

M Collins, T Koo - Computational Linguistics, 2005 - direct.mit.edu
This article considers approaches which rerank the output of an existing probabilistic parser.
The base parser produces a set of candidate parses for each input sentence, with …