GAMBoostModel {MachineShop} | R Documentation |
Gradient Boosting with Additive Models
Description
Gradient boosting for optimizing arbitrary loss functions, where component-wise arbitrary base-learners, e.g., smoothing procedures, are utilized as additive base-learners.
Usage
GAMBoostModel(
family = NULL,
baselearner = c("bbs", "bols", "btree", "bss", "bns"),
dfbase = 4,
mstop = 100,
nu = 0.1,
risk = c("inbag", "oobag", "none"),
stopintern = FALSE,
trace = FALSE
)
Arguments
family |
optional |
baselearner |
character specifying the component-wise
|
dfbase |
gobal degrees of freedom for P-spline base learners
( |
mstop |
number of initial boosting iterations. |
nu |
step size or shrinkage parameter between 0 and 1. |
risk |
method to use in computing the empirical risk for each boosting iteration. |
stopintern |
logical inidicating whether the boosting algorithm stops internally when the out-of-bag risk increases at a subsequent iteration. |
trace |
logical indicating whether status information is printed during the fitting process. |
Details
- Response types:
binary factor
,BinomialVariate
,NegBinomialVariate
,numeric
,PoissonVariate
,Surv
- Automatic tuning of grid parameter:
-
mstop
Default argument values and further model details can be found in the source See Also links below.
Value
MLModel
class object.
See Also
gamboost
, Family
,
baselearners
, fit
,
resample
Examples
## Requires prior installation of suggested package mboost to run
data(Pima.tr, package = "MASS")
fit(type ~ ., data = Pima.tr, model = GAMBoostModel)