Classification Ensembles
A classification ensemble is a predictive model composed of a weighted combination of multiple classification models. In general, combining multiple classification models increases predictive performance.
To explore classification ensembles interactively, use the Classification Learner app. For greater flexibility, use fitcensemble
in the command-line interface to boost or bag classification trees, or to grow a random forest [12]. For details on all supported ensembles, see Ensemble Algorithms. To reduce a multiclass problem into an ensemble of binary classification problems, train an error-correcting output codes (ECOC) model. For details, see fitcecoc
.
To boost regression trees using LSBoost, or to grow a random forest of regression trees[12], see Regression Ensembles.
Apps
Classification Learner | Train models to classify data using supervised machine learning |
Blocks
ClassificationEnsemble Predict | Classify observations using ensemble of decision trees (Since R2021a) |
ClassificationECOC Predict | Classify observations using error-correcting output codes (ECOC) classification model (Since R2023a) |
Functions
Classes
Topics
- Train Ensemble Classifiers Using Classification Learner App
Create and compare ensemble classifiers, and export trained models to make predictions for new data.
- Framework for Ensemble Learning
Obtain highly accurate predictions by using many weak learners.
- Ensemble Algorithms
Learn about different algorithms for ensemble learning.
- Train Classification Ensemble
Train a simple classification ensemble.
- Test Ensemble Quality
Learn methods to evaluate the predictive quality of an ensemble.
- Handle Imbalanced Data or Unequal Misclassification Costs in Classification Ensembles
Learn how to set prior class probabilities and misclassification costs.
- Classification with Imbalanced Data
Use the RUSBoost algorithm for classification when one or more classes are over-represented in your data.
- LPBoost and TotalBoost for Small Ensembles
Create small ensembles by using the LPBoost and TotalBoost algorithms.
- Tune RobustBoost
Tune RobustBoost parameters for better predictive accuracy.
- Surrogate Splits
Gain better predictions when you have missing data by using surrogate splits.
- Train Classification Ensemble in Parallel
Train a bagged ensemble in parallel reproducibly.
- Bootstrap Aggregation (Bagging) of Classification Trees Using TreeBagger
Create a
TreeBagger
ensemble for classification. - Credit Rating by Bagging Decision Trees
This example shows how to build an automated credit rating tool.
- Random Subspace Classification
Increase the accuracy of classification by using a random subspace ensemble.
- Predict Class Labels Using ClassificationEnsemble Predict Block
Train a classification ensemble model with optimal hyperparameters, and then use the ClassificationEnsemble Predict block for label prediction.
- Predict Class Labels Using ClassificationECOC Predict Block
Train an ECOC classification model, and then use the ClassificationECOC Predict block for label prediction. (Since R2023a)