General-purpose and introductory examples for the imbalanced-learn toolbox.

Examples showing API imbalanced-learn usage#

Examples that show some details regarding the API of imbalanced-learn.

Examples based on real world datasets#

Examples which use real-word dataset.

Examples using combine class methods#

Combine methods mixed over- and under-sampling methods. Generally SMOTE is used for over-sampling while some cleaning methods (i.e., ENN and Tomek links) are used to under-sample.

Dataset examples#

Examples concerning the imblearn.datasets module.

Example using ensemble class methods#

Under-sampling methods implies that samples of the majority class are lost during the balancing procedure. Ensemble methods offer an alternative to use most of the samples. In fact, an ensemble of balanced sets is created and used to later train any classifier.

Evaluation examples#

Examples illustrating how classification using imbalanced dataset can be done.

Model Selection#

Examples related to the selection of balancing methods.

Example using over-sampling class methods#

Data balancing can be performed by over-sampling such that new samples are generated in the minority class to reach a given balancing ratio.

Pipeline examples#

Example of how to use the a pipeline to include under-sampling with scikit-learn estimators.

Example using under-sampling class methods#

Under-sampling refers to the process of reducing the number of samples in the majority classes. The implemented methods can be categorized into 2 groups: (i) fixed under-sampling and (ii) cleaning under-sampling.

Gallery generated by Sphinx-Gallery