An Analytical Framework for Bias Mitigation in Credit Scoring Systems through Fairness-Constrained Neural Optimization
DOI:
https://doi.org/10.63282/3050-9262.IJAIDSML-V6I1P120Keywords:
Fairness in Machine Learning, Credit Scoring, Bias Mitigation, Neural Networks, Fairness Constraints, Disparate Impact, Lagrangian Optimization, Algorithmic Fairness, Ethical AI, Group FairnessAbstract
Machine learning has significantly enhanced predictive accuracy in credit scoring systems; however, it has also intensified concerns regarding algorithmic bias and fairness. This paper introduces an analytical framework that integrates fairness constraints into neural network optimization to mitigate such biases. We propose a constrained optimization methodology based on Lagrangian relaxation and fairness-aware loss functions to align predictive performance with equity objectives. Using a real-world credit dataset, we demonstrate that the proposed framework effectively reduces disparate impact across sensitive attributes such as race and gender while maintaining predictive performance. Additionally, the model incorporates group fairness constraintssuch as demographic parity and equal opportunitydirectly into the neural network’s loss function. Empirical evaluations show that our method consistently outperforms baseline models in terms of both fairness metrics and classification accuracy. This study offers a systematic approach to ethically aligning financial decision-making algorithms with broader societal fairness imperatives
References
[1] Hardt, M., Price, E., & Srebro, N. (2016). Equality of opportunity in supervised learning. NeurIPS.
[2] Barocas, S., Hardt, M., & Narayanan, A. (2019). Fairness and Machine Learning: Limitations and Opportunities.
[3] Mehrabi, N. et al. (2021). A Survey on Bias and Fairness in Machine Learning. ACM Computing Surveys.
[4] Chouldechova, A. (2017). Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big Data.
[5] Zemel, R. et al. (2013). Learning fair representations. ICML.
[6] Binns, R. (2018). Fairness in Machine Learning: Lessons from Political Philosophy. Proceedings of FAT*.
[7] Berk, R. et al. (2018). Fairness in criminal justice risk assessments. Sociological Methods & Research.
[8] Donini, M. et al. (2018). Empirical risk minimization under fairness constraints. NeurIPS.
[9] Dwork, C. et al. (2012). Fairness through awareness. ITCS.
[10] Kamiran, F., & Calders, T. (2012). Data preprocessing techniques for classification without discrimination. Knowledge and Information Systems.
[11] Berk R. A. 2016b. “A Primer on Fairness in Criminal Justice Risk Assessments.” The Criminologist 41(6):6–9. Google Scholar