Gradient boosting is a machine learning technique that produces a prediction model in the form of an ensemble of weak prediction models. A gradient boosting algorithm tries to solve the minimization error problem on learning samples in a functional space where each function is a model. Each model in this composition tries to predict a gradient of error for points in a feature space and these predictions will be summed with some weight to model an answer. This algorithm may be used for regression and classification problems. For more information please see Wikipedia.
In Ignite ML there is an implementation of a general GDB algorithm and GDB-on-trees algorithm. General GDB (
GDBBinaryClassifierTrainer) allows any trainer for training each model in composition. GDB on trees uses some optimizations specific for trees, such as indexes, for avoiding sorting during the decision tree build phase.
Apache Ignite ML purposes all implementations of the GDB algorithm to use GDBModel, wrapping ModelsComposition for representing the composition of a few models. ModelsComposition implements a common Model interface and can be used as follows:
GDBModel model = ...; double prediction = model.apply(featureVector);
WeightedPredictionsAggregator as the model answer reducer. This aggregator computes an answer of a meta-model, since "result = bias + p1w1 + p2w2 + …" where:
pi- answer of i-th model.
wi- weight of model in composition.
GDB uses the mean value of labels for the bias-parameter in the aggregator.
Training of GDB is represented by
GDBBinaryClassificationOnTreesTrainer for general GDB and GDB on trees respectively. All trainers have the following parameters:
gradStepSize- sets the constant weight of each model in composition; in future versions of Ignite ML this parameter may be computed dynamically.
cntOfIterations- sets the maximum of models in the composition after training.
checkConvergenceFactory- sets factory for construction of convergence checker used for preventing overfitting and learning of many useless models while training.
For classifier trainers there is addition parameter:
loss- sets loss computer on some learning example from a training dataset.
There are several factories for convergence checkers:
ConvergenceCheckerStubFactorycreates a checker that always returns false for a convergence check. So in this case, model composition size will have cntOfIterations models.
MeanAbsValueConvergenceCheckerFactorycreates a checker that compute a mean value of the absolute gradient values on each example from a dataset and returns true if this it is less than the used-defined threshold.
MedianOfMedianConvergenceCheckerFactorycreates a checker that computes the median of median absolute gradient values on each data partition. This method is less sensitive for anomalies in the learning dataset, but GDB may converge longer.
Example of training:
// Set up trainer GDBTrainer trainer = new GDBBinaryClassifierOnTreesTrainer( learningRate, countOfIterations, new LogLoss() ).withCheckConvergenceStgyFactory(new MedianOfMedianConvergenceCheckFactory(precision)); // Build the model GDBModel mdl = trainer.fit( datasetBuilder, featureExtractor, labelExtractor );
To see how GDB Classifier can be used in practice, try this example, available on GitHub and delivered with every Apache Ignite distribution.
Apache, Apache Ignite, the Apache feather and the Apache Ignite logo are either registered trademarks or trademarks of The Apache Software Foundation.