Become an expert in R — Interactive courses, Cheat Sheets, certificates and more!
Get Started for Free

adaboost

Adaboost.M1 algorithm


Description

Implements Freund and Schapire's Adaboost.M1 algorithm

Usage

adaboost(formula, data, nIter, ...)

Arguments

formula

Formula for models

data

Input dataframe

nIter

no. of classifiers

...

other optional arguments, not implemented now

Details

This implements the Adaboost.M1 algorithm for a binary classification task. The target variable must be a factor with exactly two levels. The final classifier is a linear combination of weak decision tree classifiers.

Value

object of class adaboost

References

Freund, Y. and Schapire, R.E. (1996):“Experiments with a new boosting algorithm” . In Proceedings of the Thirteenth International Conference on Machine Learning, pp. 148–156, Morgan Kaufmann.

See Also

Examples

fakedata <- data.frame( X=c(rnorm(100,0,1),rnorm(100,1,1)), Y=c(rep(0,100),rep(1,100) ) )
fakedata$Y <- factor(fakedata$Y)
test_adaboost <- adaboost(Y~X, data=fakedata,10)

fastAdaboost

a Fast Implementation of Adaboost

v1.0.0
MIT + file LICENSE
Authors
Sourav Chatterjee [aut, cre]
Initial release
2016-02-23

We don't support your browser anymore

Please choose more modern alternatives, such as Google Chrome or Mozilla Firefox.