Extended Methods to Handle Classification Biases

Emma Beauxis-Aussalet, Lynda Hardman

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

1 Citation (Scopus)
1 Downloads (Pure)


Classifiers can provide counts of items per class, but systematic classification errors yield biases (e.g., if a class is often misclassified as another, its size may be under-estimated). To handle classification biases, the statistics and epidemiology domains devised methods for estimating unbiased class sizes (or class probabilities) without identifying which individual items are misclassified. These bias correction methods are applicable to machine learning classifiers, but in some cases yield high result variance and increased biases. We present the applicability and drawbacks of existing methods and extend them with three novel methods. Our Sample-to-Sample method provides accurate confidence intervals for the bias correction results. Our Maximum Determinant method predicts which classifier yields the least result variance. Our Ratio-to-TP method details the error decomposition in classifier outputs (i.e., how many items classified as class C_y truly belong to C_x, for all possible classes) and has properties of interest for applying the Maximum Determinant method. Our methods are demonstrated empirically, and we discuss the need for establishing theory and guidelines for choosing the methods and classifier to apply.
Original languageEnglish
Title of host publication2017 IEEE International Conference on Data Science and Advanced Analytics (DSAA)
PublisherIEEE Press
ISBN (Electronic)9781509050048
Publication statusPublished - 2017
Externally publishedYes
EventThe 4th IEEE International Conference on Data Science and Advanced Analytics 2017 - Tokyo, Japan
Duration: 19 Oct 201721 Oct 2017


ConferenceThe 4th IEEE International Conference on Data Science and Advanced Analytics 2017
Abbreviated titleDSAA 2017


Dive into the research topics of 'Extended Methods to Handle Classification Biases'. Together they form a unique fingerprint.

Cite this