Multi Layers Perceptron, PMC (Perceptron Multi-Couches)
Rétropropagation du Gradient
Recuit Simulé
Converges faster than gradient descent
Second order
Radial Basis Functions Neural Networks
You have to chose k
k-means then gradient descent
incremental addition of neurons then exact method
Support Vectors Machine
arbre de décision
based on entropy
k plus proches voisins
[Freund,Schapire, 1995]
Consists in combining a lot of weak classifiers to get a strong one.
ADAptive BOOSTing, [Freund,Schapire, 1996]
The first and standard version is refered as Discrete AdaBoost.
Greedy approach
[Freund,Schapire, 1996]
[Friedman,Hastie,Tibshirani, 1998]
[Friedman,Hastie,Tibshirani, 1998]
[Friedman,Hastie,Tibshirani, 1998]
[Friedman,Hastie,Tibshirani, 1998]
AdaBoost is a sequential forward search procedure using the greedy selection strategy to minimize a certain margin on the training set. A crucial heuristic assumption used in such a sequential forward search procedure is the monotonicity (i.e. that addition of a new weak classifier to the current set does not decrease the value of the performance criterion). The premise offered by the sequential procedure in AdaBoost breaks down when this assumption is violated. Floating Search is a sequential feature selection procedure with backtracking, aimed to deal with nonmonotonic criterion functions for feature selection
[Freund,Schapire, 1997]
An extension of AdaBoost to regression problems
[Freund,Schapire, 1997]
Basic extension of AdaBoost to multiclass problems. A weak classifier needs to have an error rate less than 1/2, which is stronger than random guessing (1/k) and is often too difficult to obtain.
A weak classifier associates to an example a label in {0,…,k}
.
[Freund,Schapire, 1997]
Tries to overcome the difficulty of AdaBoost.M1 by extending the communication between the boosting algorithm and the weak learner. The algorithm not only focuses on hard instances, but also on classes which are hard to distinguish.
A weak classifier associates to an example a vector in [0,1]^k
, and the pseudo-loss takes also into account weights according to the performance of the weak classifier over the different classes for the same example.
[Schapire,Singer, 1998]
[Schapire,Singer, 1998]
[Schapire,Singer, 1998]
[Schapire, 1997]
[Guruswami,Sahai, 1999]
[Eibl,Pfeiffer, 2002]
[Eibl,Pfeiffer, 2003]
[Eibl,Pfeiffer, 2003]
Stagewise Additive Modeling using a Multi-class Exponential loss function, [Zhu,Rosset,Zou, 2006]
Gentle Adaptive Multiclass Boosting Learning, [Huang,Ertekin,Song, 2005]
Uneven loss function + greedy
Not greedy, exact.
TOTALly corrective BOOSTing, [Warmuth,Liao,Ratsch, 2006]
[Li,Abu-Mostafa,Pratap, 2003]
[Li,Abu-Mostafa,Pratap, 2003]
(Multiple Instance Learning BOOSting), [Viola,Platt, 2005]
Conjugate Gradient BOOSTing, [Li,Abu-Mostafa,Pratap, 2003]
A cascade of classifiers is a degenerated decision tree, where at each stage a classifier is trained to detect almost all objects of interest, while rejecting a certain fraction of the non-object patterns (eg eliminates 50% of non-object patterns and falsely eliminates 0.1%, then after 20 stages it can be expected a false alarm rate of 0.5^20 and a hit rate of 0.999^20). It enables to focus attention on certain regions and dramatically increases speed.