Next Article in Journal
Development and Validation of a Novel Surface Defect Index (SDI) Method for the Effective Quality Evaluation of Concrete Surfaces
Previous Article in Journal
BIMO: Bootstrap Inter–Intra Modality at Once Unsupervised Learning for Multivariate Time Series
Previous Article in Special Issue
TER-CA-WGNN: Trimodel Emotion Recognition Using Cumulative Attribute-Weighted Graph Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comprehensive Validation on Reweighting Samples for Bias Mitigation via AIF360

1
Department of Electrical and Computer Engineering, Texas A&M University System, Prairie View A&M University, Prairie View, TX 77446, USA
2
Texas Juvenile Crime Prevention Center, College of Juvenile Justice, Texas A&M University System, Prairie View A&M University, Prairie View, TX 77446, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2024, 14(9), 3826; https://doi.org/10.3390/app14093826
Submission received: 15 February 2024 / Revised: 26 April 2024 / Accepted: 27 April 2024 / Published: 30 April 2024
(This article belongs to the Special Issue Trustworthy Artificial Intelligence (AI) and Robotics)

Abstract

:
Fairness Artificial Intelligence (AI) aims to identify and mitigate bias throughout the AI development process, spanning data collection, modeling, assessment, and deployment—a critical facet of establishing trustworthy AI systems. Tackling data bias through techniques like reweighting samples proves effective for promoting fairness. This paper undertakes a systematic exploration of reweighting samples for conventional Machine-Learning (ML) models, utilizing five models for binary classification on datasets such as Adult Income and COMPAS, incorporating various protected attributes. In particular, AI Fairness 360 (AIF360) from IBM, a versatile open-source library aimed at identifying and mitigating bias in machine-learning models throughout the entire AI application lifecycle, is employed as the foundation for conducting this systematic exploration. The evaluation of prediction outcomes employs five fairness metrics from AIF360, elucidating the nuanced and model-specific efficacy of reweighting samples in fostering fairness within traditional ML frameworks. Experimental results illustrate that reweighting samples effectively reduces bias in traditional ML methods for classification tasks. For instance, after reweighting samples, the balanced accuracy of Decision Tree (DT) improves to 100%, and its bias, as measured by fairness metrics such as Average Odds Difference (AOD), Equal Opportunity Difference (EOD), and Theil Index (TI), is mitigated to 0. However, reweighting samples does not effectively enhance the fairness performance of K Nearest Neighbor (KNN). This sheds light on the intricate dynamics of bias, underscoring the complexity involved in achieving fairness across different models and scenarios.

1. Introduction

Fairness Artificial Intelligence (AI) endeavors to identify and mitigate bias throughout the entire life cycle of AI technique development, spanning data curation and preparation, modeling, evaluation, and deployment, which is crucial for the successful implementation of trustworthy AI [1]. Bias can manifest in various forms, potentially leading to unfairness in different downstream learning tasks. These biases typically originate from different stages of the Machine-Learning (ML) pipeline, including data curation, algorithm design, and user interactions. Data bias may arise when the data are collected from skewed sources, are incomplete, lack crucial information, or contain errors. Such biases result in unrepresentative or incomplete data, leading to biased outputs. Algorithmic bias stems from biased assumptions or criteria in algorithm design, resulting in biased outputs for downstream tasks. Bias introduced through user interactions occurs when individuals using AI systems inject their own biases or prejudices, whether consciously or unconsciously. To address these sources of bias, various approaches have been proposed. Dataset augmentation involves adding more diverse data to training datasets, enhancing representativeness and reducing bias [2]. Bias-aware algorithms are designed to consider different types of bias, working to minimize their impact on system outputs [3]. User feedback mechanisms, such as human-in-the-loop systems [4], involve soliciting feedback from users to identify and rectify biases in the system.
Addressing data bias can be efficiently achieved through sample reweighting, contributing to the advancement of fairness in AI [5]. This method involves assigning weights to each sample based on the ratio of its population proportion to its sampling proportion [6]. The process ensures the dataset becomes discrimination-free through two key steps. Firstly, specific attributes, such as gender and race, are identified within the datasets. Subsequently, higher weights are assigned to samples from underrepresented groups, while lower weights are assigned to those from overrepresented groups with respect to these specific attributes. These steps collectively contribute to achieving balance across all groups, thereby fostering fairness in the final outputs of AI algorithms trained on the reweighed data. Nevertheless, existing efforts appear to lack a thorough and comprehensive evaluation of the effectiveness of reweighting samples in mitigating bias associated with traditional machine-learning models.
This paper undertakes a comprehensive evaluation of the effectiveness of sample reweighting in mitigating bias linked with traditional machine-learning models, employing AI Fairness 360 (AIF360) [7]. AIF360 is a versatile open-source library designed to identify and alleviate bias in machine-learning models across the entire AI application lifecycle. It encompasses a comprehensive array of metrics for scrutinizing biases in datasets and models, along with detailed explanations for these metrics and algorithms for bias mitigation. The evaluation in this paper focuses on the reweighting samples methods available in AIF360, applied to classification tasks performed by five traditional machine-learning models: Decision Tree (DT), K Nearest Neighbor (KNN), Gaussian Naive Bayes (GNB), Logistic Regression (LR), and Random Forest (RF). The reweighting process is carried out with respect to privileged attributes such as sex and race. Subsequently, each traditional machine-learning model undergoes classification tasks on both the original datasets and the new datasets resulting from reweighting samples. The comparative analysis involves assessing the performance of these models on the original and new datasets, respectively. This evaluation is based on metrics such as balanced accuracy and fairness metrics. Experimental results highlight the model-specific nature of reweighting samples effectiveness in achieving fairness in traditional ML models and reveal the complexity of bias dynamics.
The contributions in this paper can be summarized as follows:
  • In contrast to the previous work [8], our research involves a systematic comparison of reweighting samples for mitigating bias on five traditional ML models through the AIF360 platform.
  • We systematically examine the fairness of experimental results with five fairness metrics and provide insights of effectiveness of reweighting samples for bias mitigation.

2. Methodology

This paper aims to examine the effectiveness of reweighting samples to enhance the fairness of traditional machine-learning algorithms on classification tasks. It covers three AI techniques, namely, reweighting samples, traditional machine learning for implementing classification, and fairness metrics to performance evaluation.

2.1. Reweighting Samples

Fairness in AI can be conceptualized as a multi-objective optimization challenge, aiming to optimize learning objectives while mitigating discrimination with respect to sensitive attributes [5]. In essence, achieving fairness may involve a trade-off in learning performance to minimize bias. Data preprocessing emerges as an effective technique for molding training data to foster fairness in AI. Techniques such as suppression, dataset massaging, and reweighting samples have proven effective [5].
Reweighting samples, a specific preprocessing technique, involves adjusting the significance or contribution of individual samples within the training dataset. By strategically assigning weights, it becomes possible to render the training dataset free from discrimination concerning sensitive attributes, all without altering the existing labels. One approach to determining these weights involves measuring them based on the frequency counts associated with the sensitive attribute [9].
This paper leveraged the reweighting samples technique from AIF360 during data preprocessing to enhance fairness in binary classification. The input for the reweighting process comprises a training dataset with samples containing attributes (including a sensitive attribute) and labels, along with the specification of the sensitive attribute. The output is a transformed dataset where sample weights are adjusted concerning the sensitive attribute, mitigating potential classification bias. Throughout the reweighting process, an analysis of the distribution of the sensitive attribute within different groups is conducted. This analysis informs the calculation of reweighting coefficients, which, in turn, adjust sample weights to promote a more uniform distribution across groups.
Given a sensitive (protected) attribute, the privileged group of samples includes the samples with the positive sensitive attribute while the unprivileged group of samples includes the samples with the negative sensitive attribute. For a binary classification task, reweighting samples adjusts the weights of four categories of samples, namely w p p (the weight of the positive privileged sample (pp)), w p u p (the weight of positive unprivileged samples (pup)), w n p (the weight of negative privileged samples (np)), and w n u p (the weight of negative unprivileged samples (nup)), as below.
w p p = N p N t o t a l × N p o s N p p
w p u p = N u p N t o t a l × N p o s N p u p
w n p = N p N t o t a l × N n e g N n p
w n u p = N u p N t o t a l × N n e g N u p ,
where
  • N p : the number of samples in the privileged group.
  • N p p : the number of samples with the positive class in the privileged group.
  • N n p : the number of samples with the negative class in the privileged group.
  • N u p : the number of samples in the unprivileged group.
  • N p u p : the number of samples with the positive class in the unprivileged group.
  • N n u p : the number of samples with the negative class in the unprivileged group.
  • N p o s : the number of samples with the positive class.
  • N n e g : the number of samples with the negative class.
  • N t o t a l : the number of samples.

2.2. Traditional Machine Learning

This paper applied five traditional machine-learning models to implement binary classification tasks, including LR, DT, KNN, GNB, and RF [10].
LR is a statistical technique employed to model the probability of a binary outcome. It finds widespread application in machine learning for scenarios involving binary classification, aiming to predict whether an instance belongs to one of two classes. The posterior probability of class c 1 is expressed through a logistic sigmoid applied to a linear function of the feature vector ϕ ,
p ( c 1 | ϕ ) = y ( ϕ ) = σ ( w T ϕ ) ,
where p ( c 2 | ϕ ) = 1 p ( c 1 | ϕ ) and σ ( · ) is the logistic sigmoid function.
DT is a data mining technique employed to establish classification systems using multiple covariates or to create prediction algorithms for a target variable. This method organizes a population into branch-like segments, forming an inverted tree structure with a root node, internal nodes, and leaf nodes. Notably, the algorithm is non-parametric, allowing it to effectively handle large and complex datasets without imposing a rigid parametric structure [11]. However, one limitation of DT lies in its reliance on hard splits in the input space, where a single model is responsible for predictions for any given value of the input variables.
KNN is a supervised machine-learning algorithm used for classification and regression tasks. It is a type of instance-based learning, where the model makes predictions based on the similarity of new data points to existing labeled data points in the training set [12,13].
GNB is a straightforward classification algorithm [10]. Its primary principle involves assigning labels to classes by maximizing the posterior probability for each sample. This method operates under the assumption that voxel contributions are conditionally independent and follow a Gaussian (normal) distribution. The discriminant function is formulated as the sum of squared distances to the centroid of each class across all voxels in the searchlight. This sum is then weighted by the variance and the logarithm of the a priori probability, computed in the training set using Bayes rule. In essence, GNB provides a probabilistic approach to classification, leveraging assumptions about the distribution of features to make predictions.
RF has proven to be remarkably successful as a classification and regression method. This approach involves the combination of multiple randomized decision trees, and it aggregates their predictions through averaging. Notably, it has demonstrated exceptional performance in scenarios where the number of variables is significantly greater than the number of observations. Its versatility extends to large-scale problems, making it adaptable to various ad hoc learning tasks. Additionally, it provides measures of variable importance, adding to its utility and interpretability [14].

2.3. Fairness Metrics

This paper employed five fairness metrics to evaluate the effectiveness of reweighting samples for mitigating bias.
Given sensitive (protected) attributes, Disparate Impact (DI) [7] denotes inadvertent bias that may arise when predictions exhibit varying error rates or outcomes across demographic groups, where the sensitive attributes, such as race, sex, disability, and age, are deemed protected. This bias can emerge either from training models on biased data or from the predictive model itself being discriminatory. In the context of this study, Disparate Impact pertains to divergent impacts on prediction results as defined by
D I = p p u p p p p ,
where p p u p refers to the prediction probability for the unprivileged samples with positive predictions while p p p denotes the prediction probability for the privileged samples with positive predictions. If the disparate impact of the predictions approaches 0, it signifies bias in favor of the privileged group. Conversely, if it exceeds 1 it indicates a bias in favor of the unprivileged group. A value of 1 implies perfect fairness in the predictions [15].
Average Odds Difference (AOD) [7] is the average of difference in the False Positive Rates (FPRs) and True Positive Rates (TPRs) between the unprivileged and privileged groups. It is defined by
A O D = ( F P R u p F P R p ) + ( T P R u p T P R p ) 2 ,
where F P R u p and F P R p denote the False Positive Rate for unprivileged and privileged samples, respectively, within predictions, while T P R u p and T P R p refer to the True Positive Rate for unprivileged and privileged samples, respectively, within predictions. A result of 0 signifies perfect fairness. A positive value indicates bias in favor of the unprivileged group, while a negative value indicates bias in favor of the privileged group.
Statistical Parity Difference (SPD) [7] is used to calculate the difference between the ratio of favorable outcomes in unprivileged and privileged groups. It is defined by
S P D = p p u p p p p .
A score below 0 suggests benefits for the unprivileged group, while a score above 0 implies benefits for the privileged group. A score of 0 indicates that both groups receive equal benefits.
Equal Opportunity Difference (EOD) [7] involves evaluating the equal opportunity for benefiting all groups. EOD specifically centers on the True Positive Rate (TPR), representing the accurate identification of positives in both the unprivileged and privileged groups. The measure is defined by
E O D = T P R u p T P R p .
A value of 0 signifies perfect fairness. A positive value indicates bias in favor of the unprivileged group, while a negative value indicates bias in favor of the privileged group.
Theil Index (TI) [7] is also called the entropy index and measures both the group and individual fairness. It is defined by
T I = 1 n i = 1 n b i μ l n b i μ ,
where b i = y i ^ y i + 1 and μ is the average of b i . A lower absolute value of TI value in this context would indicate a more equitable distribution of classification outcomes, while a higher absolute value suggests greater disparity.

3. Results and Discussions

In order to comprehensively validate the reweighting of samples for mitigating classification bias, we utilized five classifiers to conduct classification tasks on two datasets. Subsequently, we assessed the fairness of classification using five evaluation metrics. This enables us to systematically investigate whether reweighting samples effectively mitigates classification bias.

3.1. Dataset

This paper employed two datasets, including the Adult Income dataset and the Correctional Offender Management Profiling for Alternative Sanctions (COMPAS) dataset, to evaluate the effectiveness of reweighting samples for mitigating fairness.
Adult Income dataset: it includes 48,842 samples with 14 attributes, which can be used for predicting whether income exceeds $50K/yr based on census data [16].
COMPAS dataset: the COMPAS system is a case management system and decision support tool initially developed and owned by Northpointe (now Equivant). It was designed for the purpose of assessing the likelihood of an individual committing a future crime. The dataset associated with COMPAS comprises more than 20 attributes and includes a substantial sample size of 11,000 instances [17].

3.2. Experimental Metrics

Balance Accuracy (BA) is applicable to both binary and multi-class classification scenarios by computing the mean of sensitivity and specificity. Sensitivity gauges the correct prediction of true positives, representing accurately identified positive instances, while specificity measures the true negatives over the total negatives predicted by the model. A result nearing 0 signifies poor model performance, whereas a result approaching 1 indicates effective performance across both sensitivity and specificity [18].

3.3. Experimental Results and Discussion

This paper implemented comprehensive validation on reweighting samples for mitigating bias for traditional supervised machine-learning models with the binary classification tasks. It involved two components of experiments on two datasets: the Adult Income dataset and the COMPAS dataset.

3.3.1. Adult Income

Figure 1 and Figure 2 illustrate the performance comparison before and after reweighting samples with respect to the protected attribute R a c e . Prior to reweighting, various ML models exhibit distinct biases in AOD and DI curves. Following reweighting, biases in ML models are mitigated to varying degrees. Notably, DT models demonstrate bias-free behavior in terms of AOD and retain bias in DI values. Conversely, biases in LR and KNN appear unchanged, and GNB produced changed with a high optimal classification threshold, while the bias in RF becomes more unstable. This suggests that the effectiveness of reweighting samples depends on the specific ML model and may not be universally applied.
Table 1 provides a systematic comparison using additional fairness metrics. DT’s bias is eliminated in terms of AOD, EOD, and TI values. However, regarding SPD and DI values, bias persists towards the unprivileged group, emphasizing the need for a comprehensive examination of bias. Furthermore, although other ML models experience slight reductions in BA, their fairness is marginally improved, emphasizing the inherent trade-off between BA and fairness.
Additionally, Figure 3 and Figure 4 depict comparison results for the protected attribute S e x . Similar observations are noted, where reweighting samples are effective mainly for DT, less so for KNN, and have varying impacts on LR, GNB, and RF. Table 2 echoes these trends, revealing slight reductions in BA and modest improvements in fairness for other ML models across various fairness metrics (SPD, AOD, DI, EOD, and TI). This underscores the nuanced effectiveness of the same debiasing technique across different ML models.

3.3.2. COMPAS

To thoroughly assess the effectiveness of reweighting samples in mitigating bias, Figure 5 and Figure 6 provide a performance comparison before and after reweighting samples for the protected attribute R a c e in the COMPAS dataset. Similarly, prior to reweighting, various ML models exhibit diverse biases in the AOD and DI curves. Post-reweighting, DT models demonstrate a lack of bias in AOD while showing significant bias against unprivileged groups in DI values. Conversely, biases in LR, GNB, and RF show significant change over the classification threshold cycle as seen in the curves of AOD and DI. KNN became more bias after reweighing was applied.
Table 3 offers a systematic comparison using additional fairness metrics on the COMPAS dataset. Similar to the Adult Income dataset, DT’s bias is eradicated concerning AOD, EOD, and TI values. However, biases persist in terms of SPD and DI values, showing a continued bias toward the unprivileged group. The LR and RF models displayed significant bias in favor of the unprivileged group. While GNB and KNN models experience slight reductions in BA, their fairness is marginally improved.
Moreover, Figure 7 and Figure 8 present comparison results for the protected attribute S e x . Similar observations are made, where reweighting samples are effective primarily for DT but less so for KNN. Notably, LR, GNB, and RF exhibit more significant effects after reweighting. Table 4 reveals similar trends, with DT showing bias in SPD and DI but less bias overall. All other models bias was reduced while retaining bias against the unpriveleged group. LR showed bias in favor of the unprivileged group for EOD.
In summary, the figures and tables illustrate the impact of reweighting samples on bias mitigation in traditional machine-learning models for both the Adult Income and COMPAS datasets. Notably, DT models showcase effective bias reduction, while KNN exhibits more resistance to debiasing for the Adult Income dataset but work well with the COMPAS dataset. The trade-off between BA and fairness is evident. Results differ across protected attributes, emphasizing the nuanced effectiveness of reweighting. Comprehensive assessments, including additional fairness metrics, reveal the complexity of bias dynamics. This study highlights the model-specific nature of reweighting sample effectiveness in achieving fairness in traditional machine-learning models.

4. Related Work

ML fairness has become one of the most pivotal challenges of the decade [19]. Intended to intelligently prevent errors and biases in decision-making, ML models sometimes unintentionally become sources of bias and discrimination in society. Concerns have been raised about various forms of unfairness in ML, including racial biases in criminal justice systems, disparities in employment, and biases in loan approval processes [20]. The entire life cycle of an ML model, covering input data, modeling, evaluation, and feedback, is susceptible to both external and inherent biases, potentially resulting in unjust outcomes.
Techniques for mitigating bias in ML models can be categorized into pre-processing, in-processing, and post-processing methods [21]. Pre-processing acknowledges that data itself often introduces bias, with distributions of sensitive or protected variables being discriminatory or imbalanced. This approach modifies sample distributions or transforms data to eliminate discrimination during training [5]. It is considered the most flexible part of the data-science pipeline, making no assumptions about subsequent modeling techniques [22]. In-processing adjusts modeling techniques to counter biases and incorporates fairness metrics into model optimization [23]. Post-processing addresses unfair model outputs, applying transformations to enhance prediction fairness [24].
Pre-processing assumes that the disparate impact of the trained classifier mirrors that of the training data. Techniques include massaging the dataset by adjusting mislabeled class labels due to bias [15,25] and reweighting training samples to assign greater importance to sensitive ones [5,26].

5. Conclusions

Understanding the impact of the biases of machine-learning techniques becomes crucial to prevent unintended behaviors towards specific groups. Fairness AI techniques, such as reweighting samples, have proven effective across various fields This study presents a comprehensive validation of reweighting samples’ application to address bias in binary classification using traditional ML models. The comparative analysis involves assessing model performance on original and new datasets, respectively, based on metrics like balanced accuracy and fairness metrics. It offers systematic insights into the effectiveness of different traditional classification algorithms concerning various protected attributes, contributing to the advancement of fairness-enhanced AI. Experimental results underscore the model-specific nature of reweighting sample effectiveness in achieving fairness in traditional ML models, while also revealing the complexity of bias dynamics.
Future research will expand upon this exploration by integrating more advanced machine-learning algorithms, such as generative AI models. For example, generative AI can be utilized for data augmentation with original data to bolster the fairness of AI techniques. Furthermore, the study will integrate new datasets and investigate a wider array of protected attributes to deepen the validation process.

Author Contributions

Conceptualization, P.O., L.Q. and X.D.; methodology, C.H.B., P.O., L.Q. and X.D.; validation, C.H.B., P.O. and X.D.; formal analysis, C.H.B., C.G., P.O., L.Q. and X.D.; data curation, C.H.B. and X.D.; writing—original draft preparation, C.H.B. and X.D.; writing—review and editing, C.H.B., C.G., P.O., L.Q. and X.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study involved two datasets, including Adult Income (https://archive.ics.uci.edu/dataset/2/adult, accessed on 10 January 2023) and COMPAS (https://www.kaggle.com/datasets/danofer/compass, accessed on 10 January 2023).

Acknowledgments

This research work is supported by NSF under award number 2323419. Any opinions, findings, and conclusions or recommendations expressed in this work are those of the author(s) and do not necessarily reflect the views of NSF.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIF360AI Fairness 360
SPDStatistical parity difference
AODAverage odds difference
DIDisparate Impact
EODEqual opportunity difference
TITheil index
LRLogistic Regression
DTDecision Tree
KNNK Nearest Neighbor
GNBGaussian Naive Bayes
RFRandom Forest

References

  1. Li, B.; Qi, P.; Liu, B.; Di, S.; Liu, J.; Pei, J.; Yi, J.; Zhou, B. Trustworthy AI: From principles to practices. ACM Comput. Surv. 2023, 55, 1–46. [Google Scholar] [CrossRef]
  2. Sharma, S.; Zhang, Y.; Ríos Aliaga, J.M.; Bouneffouf, D.; Muthusamy, V.; Varshney, K.R. Data augmentation for discrimination prevention and bias disambiguation. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, New York, NY, USA, 7–9 February 2020; pp. 358–364. [Google Scholar]
  3. Ahsen, M.E.; Ayvaci, M.U.S.; Raghunathan, S. When algorithmic predictions use human-generated data: A bias-aware classification algorithm for breast cancer diagnosis. Inf. Syst. Res. 2019, 30, 97–116. [Google Scholar] [CrossRef]
  4. Ghai, B.; Mueller, K. D-BIAS: A causality-based human-in-the-loop system for tackling algorithmic bias. IEEE Trans. Vis. Comput. Graph. 2022, 29, 473–482. [Google Scholar] [CrossRef] [PubMed]
  5. Kamiran, F.; Calders, T. Data preprocessing techniques for classification without discrimination. Knowl. Inf. Syst. 2012, 33, 1–33. [Google Scholar] [CrossRef]
  6. An, J.; Ying, L.; Zhu, Y. Why resampling outperforms reweighting for correcting sampling bias with stochastic gradients. arXiv 2020, arXiv:2009.13447. [Google Scholar]
  7. Bellamy, R.K.; Dey, K.; Hind, M.; Hoffman, S.C.; Houde, S.; Kannan, K.; Lohia, P.; Martino, J.; Mehta, S.; Mojsilovic, A.; et al. AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. arXiv 2018, arXiv:1810.01943. [Google Scholar]
  8. Hufthammer, K.T.; Aasheim, T.H.; Ånneland, S.; Brynjulfsen, H.; Slavkovik, M. Bias Mitigation with AIF360: A Comparative Study. 2020. Available online: https://bora.uib.no/bora-xmlui/handle/11250/2764230 (accessed on 10 January 2023).
  9. Calders, T.; Kamiran, F.; Pechenizkiy, M. Building classifiers with independency constraints. In Proceedings of the 2009 IEEE International Conference on Data Mining Workshops, Miami, FL, USA, 6 December 2009; IEEE: Piscataway Township, NJ, USA, 2009; pp. 13–18. [Google Scholar]
  10. Bishop, C.M.; Nasrabadi, N.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006; Volume 4. [Google Scholar]
  11. Song, Y.Y.; Ying, L. Decision tree methods: Applications for classification and prediction. Shanghai Arch. Psychiatry 2015, 27, 130. [Google Scholar] [PubMed]
  12. Peterson, L.E. K-nearest neighbor. Scholarpedia 2009, 4, 1883. [Google Scholar] [CrossRef]
  13. Kramer, O.; Kramer, O. K-nearest neighbors. In Dimensionality Reduction with Unsupervised Nearest Neighbors; Springer: Berlin, Germany, 2013; pp. 13–23. [Google Scholar]
  14. Biau, G.; Scornet, E. A random forest guided tour. Test 2016, 25, 197–227. [Google Scholar] [CrossRef]
  15. Feldman, M.; Friedler, S.A.; Moeller, J.; Scheidegger, C.; Venkatasubramanian, S. Certifying and removing disparate impact. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 259–268. [Google Scholar]
  16. Adult Income Dataset. UCI Irvine Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/2/adult (accessed on 10 January 2023).
  17. Larson, J.; Surya Mattu, L.K.; Angwin, J. How We Analyzed the COMPAS Recidivism Algorithm. 2016. Available online: https://www.propublica.org/article/how-we-analyzed-the-compas-recidivism-algorithm (accessed on 10 January 2023).
  18. Brodersen, K.H.; Ong, C.S.; Stephan, K.E.; Buhmann, J.M. The balanced accuracy and its posterior distribution. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; IEEE: Piscataway Township, NJ, USA, 2010; pp. 3121–3124. [Google Scholar]
  19. Shaham, S.; Hajisafi, A.; Quan, M.K.; Nguyen, D.C.; Krishnamachari, B.; Peris, C.; Ghinita, G.; Shahabi, C.; Pathirana, P.N. Holistic Survey of Privacy and Fairness in Machine Learning. arXiv 2023, arXiv:2307.15838. [Google Scholar]
  20. Angwin, J.; Larson, J.; Mattu, S.; Kirchner, L. Machine bias. In Ethics of Data and Analytics; Auerbach Publications: Boca Raton, FL, USA, 2022; pp. 254–264. [Google Scholar]
  21. Caton, S.; Haas, C. Fairness in machine learning: A survey. ACM Comput. Surv. 2020, 56, 1–38. [Google Scholar] [CrossRef]
  22. du Pin Calmon, F.; Wei, D.; Vinzamuri, B.; Ramamurthy, K.N.; Varshney, K.R. Data pre-processing for discrimination prevention: Information-theoretic optimization and analysis. IEEE J. Sel. Top. Signal Process. 2018, 12, 1106–1119. [Google Scholar] [CrossRef]
  23. Jiang, H.; Nachum, O. Identifying and correcting label bias in machine learning. In Proceedings of the International Conference on Artificial Intelligence and Statistics, PMLR, Online, 26–28 August 2020; pp. 702–712. [Google Scholar]
  24. Kim, M.P.; Ghorbani, A.; Zou, J. Multiaccuracy: Black-box post-processing for fairness in classification. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, Honolulu, HI, USA, 27–28 January 2019; pp. 247–254. [Google Scholar]
  25. Kamiran, F.; Calders, T. Classifying without discriminating. In Proceedings of the 2009 2nd International Conference on Computer, Control and Communication, Karachi, Pakistan, 17–18 February 2009; IEEE: Piscataway Township, NJ, USA, 2009; pp. 1–6. [Google Scholar]
  26. Doherty, N.A.; Kartasheva, A.V.; Phillips, R.D. Information effect of entry into credit ratings market: The case of insurers’ ratings. J. Financ. Econ. 2012, 106, 308–330. [Google Scholar] [CrossRef]
Figure 1. Performance comparison via BA vs. AOD before and after reweighting samples on Adult Income dataset with respect to the protected attribute R a c e .
Figure 1. Performance comparison via BA vs. AOD before and after reweighting samples on Adult Income dataset with respect to the protected attribute R a c e .
Applsci 14 03826 g001
Figure 2. Performance comparison via BA vs. DI before and after reweighting samples on Adult Income dataset with respect to the protected attribute R a c e .
Figure 2. Performance comparison via BA vs. DI before and after reweighting samples on Adult Income dataset with respect to the protected attribute R a c e .
Applsci 14 03826 g002
Figure 3. Performance comparison via BA vs. AOD before and after reweighting samples on Adult Income dataset with respect to the protected attribute S e x .
Figure 3. Performance comparison via BA vs. AOD before and after reweighting samples on Adult Income dataset with respect to the protected attribute S e x .
Applsci 14 03826 g003
Figure 4. Performance comparison via BA vs. DI before and after reweighting samples on Adult Income dataset with respect to the protected attribute S e x .
Figure 4. Performance comparison via BA vs. DI before and after reweighting samples on Adult Income dataset with respect to the protected attribute S e x .
Applsci 14 03826 g004
Figure 5. Performance comparison via BA vs. AOD before and after reweighting samples on COMPAS dataset with respect to the protected attribute R a c e .
Figure 5. Performance comparison via BA vs. AOD before and after reweighting samples on COMPAS dataset with respect to the protected attribute R a c e .
Applsci 14 03826 g005
Figure 6. Performance comparison via BA vs. DI before and after reweighting samples on COMPAS dataset with respect to the protected attribute R a c e .
Figure 6. Performance comparison via BA vs. DI before and after reweighting samples on COMPAS dataset with respect to the protected attribute R a c e .
Applsci 14 03826 g006
Figure 7. Performance comparison via BA vs. AOD before and after reweighting samples on COMPAS dataset with respect to the protected attribute S e x .
Figure 7. Performance comparison via BA vs. AOD before and after reweighting samples on COMPAS dataset with respect to the protected attribute S e x .
Applsci 14 03826 g007
Figure 8. Performance comparison via BA vs. DI before and after reweighting samples on COMPAS dataset with respect to the protected attribute S e x .
Figure 8. Performance comparison via BA vs. DI before and after reweighting samples on COMPAS dataset with respect to the protected attribute S e x .
Applsci 14 03826 g008
Table 1. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on Adult Income dataset regarding the protected attribute R a c e .
Table 1. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on Adult Income dataset regarding the protected attribute R a c e .
Performance before reweighting samples
ModelBASPDAODDIEODTI
DT0.7426−0.2416−0.19590.4196−0.20260.1130
GNB0.7416−0.2952−0.26230.3252−0.28720.1111
KNN0.7390−0.1904−0.14090.4882−0.14160.1207
LR0.7437−0.2435−0.19660.4122−0.20200.1129
RF0.7471−0.2014−0.13360.5380−0.10970.1066
Performance after reweighting samples
DT1.0−0.10660.00.58630.00.0
GNB0.7432−0.1147−0.02520.73790.03100.1058
KNN0.7390−0.1904−0.14090.4882−0.14160.1207
LR0.7311−0.05230.04190.85080.10830.1247
RF0.7447−0.1072−0.02010.74490.03210.1081
Table 2. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on Adult Income dataset regarding the protected attribute S e x .
Table 2. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on Adult Income dataset regarding the protected attribute S e x .
Performance before reweighting samples
ModelBASPDAODDIEODTI
DT0.7426−0.3608−0.32040.2785−0.37750.1130
GNB0.7416−0.3353−0.28050.3369−0.31840.1111
KNN0.7390−0.3983−0.40750.1616−0.53110.1207
LR0.7437−0.3580−0.31810.2794−0.37690.1129
RF0.7471−0.3777−0.32920.2884−0.37630.1066
Performance after reweighting samples
DT1.0−0.19100.00.37400.00.0
GNB0.7209−0.08610.00730.79970.02030.1192
KNN0.7390−0.3983−0.40750.1616−0.53110.1207
LR0.7134−0.07050.01880.77850.02930.1401
RF0.7271−0.1386−0.06380.7220−0.07740.1065
Table 3. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on COMPAS dataset regarding the protected attribute R a c e .
Table 3. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on COMPAS dataset regarding the protected attribute R a c e .
Performance before reweighting samples
ModelBASPDAODDIEODTI
DT0.6586−0.1516−0.09700.7791−0.12120.1835
GNB0.6553−0.2483−0.19940.6155−0.19800.2382
KNN0.6414−0.2139−0.17270.7282−0.11310.1607
LR0.6774−0.2494−0.19270.6600−0.18770.1774
RF0.6432−0.1539−0.10570.6873−0.11660.3003
Performance after reweighting samples
DT1.0−0.17690.00.70600.00.0
GNN0.6437−0.1782−0.13180.6926−0.12510.2594
KNN0.6311−0.3432−0.31050.5945−0.23910.1762
LR0.63420.05460.10421.10620.12150.2257
RF0.62340.14590.19531.40120.19770.2867
Table 4. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on COMPAS dataset regarding the protected attribute S e x .
Table 4. Performance comparison between before and after reweighting samples through one classification metric, BA, and fairness metrics including SPD, AOD, DI, EOD, and TI on COMPAS dataset regarding the protected attribute S e x .
Performance before reweighting samples
ModelBASPDAODDIEODTI
DT0.6586−0.1637−0.13400.7759−0.05970.1835
GNB0.6553−0.4129−0.38770.5066−0.30900.2382
KNN0.6414−0.2336−0.20950.7256−0.13500.1607
LR0.6774−0.2724−0.24390.6631−0.13920.1774
RF0.6432−0.3759−0.34840.4700−0.30020.3003
Performance after reweighting samples
DT1.0−0.13830.00.77320.00.0
GNB0.6581−0.1998−0.17200.7154−0.08990.2146
KNN0.6311−0.2551−0.23180.7003−0.17080.1762
LR0.6562−0.1188−0.09460.83420.01110.1730
RF0.6585−0.1615−0.12790.7081−0.07600.2776
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Blow, C.H.; Qian, L.; Gibson, C.; Obiomon, P.; Dong, X. Comprehensive Validation on Reweighting Samples for Bias Mitigation via AIF360. Appl. Sci. 2024, 14, 3826. https://doi.org/10.3390/app14093826

AMA Style

Blow CH, Qian L, Gibson C, Obiomon P, Dong X. Comprehensive Validation on Reweighting Samples for Bias Mitigation via AIF360. Applied Sciences. 2024; 14(9):3826. https://doi.org/10.3390/app14093826

Chicago/Turabian Style

Blow, Christina Hastings, Lijun Qian, Camille Gibson, Pamela Obiomon, and Xishuang Dong. 2024. "Comprehensive Validation on Reweighting Samples for Bias Mitigation via AIF360" Applied Sciences 14, no. 9: 3826. https://doi.org/10.3390/app14093826

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop