TY - JOUR
T1 - Metrics and Algorithms for Identifying and Mitigating Bias in AI Design
T2 - A Counterfactual Fairness Approach
AU - Moon, Dongsoo
AU - Ahn, Seongjin
N1 - Publisher Copyright:
© 2013 IEEE.
PY - 2025
Y1 - 2025
N2 - The rapid advancements in artificial intelligence (AI) have revolutionized industries such as healthcare, finance, and education. However, these advancements have also intensified ethical concerns regarding bias, fairness, and accountability in AI systems. Traditional fairness evaluation methods primarily focus on dataset-level biases, overlooking biases arising from model decision-making processes. This study introduces a novel framework for identifying, evaluating, and mitigating biases in AI models using counterfactual fairness, a robust approach that simulates alternative outcomes to minimize discriminatory effects. The proposed methodology integrates fairness-aware data preprocessing, feature selection, and model optimization strategies, ensuring equitable treatment across demographic groups. To validate the framework, we conducted empirical experiments using random forest and eXtreme Gradient Boosting models on the xAPI-Edu-Data dataset. Our results demonstrate significant improvements in demographic parity and equal opportunity fairness metrics while maintaining high predictive performance. Furthermore, comparative analysis with existing bias mitigation techniques confirms that our approach effectively reduces bias propagation in AI decision-making processes. By incorporating counterfactual fairness into AI design, this study provides a scalable and adaptable solution for ensuring ethical AI deployments aligned with regulatory standards.
AB - The rapid advancements in artificial intelligence (AI) have revolutionized industries such as healthcare, finance, and education. However, these advancements have also intensified ethical concerns regarding bias, fairness, and accountability in AI systems. Traditional fairness evaluation methods primarily focus on dataset-level biases, overlooking biases arising from model decision-making processes. This study introduces a novel framework for identifying, evaluating, and mitigating biases in AI models using counterfactual fairness, a robust approach that simulates alternative outcomes to minimize discriminatory effects. The proposed methodology integrates fairness-aware data preprocessing, feature selection, and model optimization strategies, ensuring equitable treatment across demographic groups. To validate the framework, we conducted empirical experiments using random forest and eXtreme Gradient Boosting models on the xAPI-Edu-Data dataset. Our results demonstrate significant improvements in demographic parity and equal opportunity fairness metrics while maintaining high predictive performance. Furthermore, comparative analysis with existing bias mitigation techniques confirms that our approach effectively reduces bias propagation in AI decision-making processes. By incorporating counterfactual fairness into AI design, this study provides a scalable and adaptable solution for ensuring ethical AI deployments aligned with regulatory standards.
KW - AI bias
KW - bias mitigation
KW - counterfactual fairness
KW - ethical AI
KW - fairness evaluation
UR - https://www.scopus.com/pages/publications/105003088043
U2 - 10.1109/ACCESS.2025.3556082
DO - 10.1109/ACCESS.2025.3556082
M3 - Article
AN - SCOPUS:105003088043
SN - 2169-3536
VL - 13
SP - 59118
EP - 59129
JO - IEEE Access
JF - IEEE Access
ER -