Awesome
Cost-sensitive-Boosting-Tutorial
The tutorial 'CalibratedAdaMEC_ExtendedVersion.ipynb' introduces the concepts of asymmetric (cost-sensitive and/or imbalanced class) learning, decision theory and boosting. It briefly describes the results of the paper:
Nikolaou, N., Edakunni, N. U., Kull, M., Flach, P. A., and Brown, G., 'Cost-sensitive boosting algorithms: Do we really need them?', Machine Learning, 104(2), pages 359-384, 2016. [http://link.springer.com/article/10.1007/s10994-016-5572-x]
It presents the Calibrated AdaMEC method (AdaBoost with calibrated probability estimates and a shifted decision threshold) found to be the most flexible, empirically successful and theoretically valid way to handle asymmetric classification with adaboost ensembles.
The code provided allows the user to reproduce the papers experiments, but also to extend them by choosing different calibration techniques, weak learners, ensemble sizes, AdaBoost variants, train\calibration splits, etc. We provide the tutorial along with standalone code and all the datasets used in the paper.
For a straightforward, ready-to-use but less flexible implementation of Calibrated AdaMEC (following the same syntax of AdaBoostClassifier() in scikit-learn), please visit: http://www.cs.man.ac.uk/~gbrown/costsensitiveboosting/
If you make use of the code found here, please cite the paper above.