Quick Search:

A Fairness-Aware Machine Learning Framework for Sexual and Reproductive Health: Evaluating Algorithmic Bias Across Models

Osagie, Efosa ORCID logoORCID: https://orcid.org/0009-0004-3462-7175, Shemi, Ayo-Ogbor and Balasundaram, Rebecca (2026) A Fairness-Aware Machine Learning Framework for Sexual and Reproductive Health: Evaluating Algorithmic Bias Across Models. Journal of Data Science and Intelligent Systems.

[thumbnail of Accepted Manuscript - A Fairness-Aware Machine Learning Framework for Sexual and Reproductive Health_ Evaluating Algorithmic Bias Across Models.pdf]
Preview
Text
Accepted Manuscript - A Fairness-Aware Machine Learning Framework for Sexual and Reproductive Health_ Evaluating Algorithmic Bias Across Models.pdf - Accepted Version

| Preview

Abstract

Advances in computational infrastructure and the widespread adoption of Electronic Health Record (EHR) systems have accelerated the integration of Artificial Intelligence (AI) and Machine Learning (ML) into Sexual and Reproductive Health (SRH) services. These technologies enhance diagnostic accuracy, support clinical decision‑making, and enable predictive analytics using diverse healthcare data. However, biases within training datasets can produce unfair outcomes, particularly for underrepresented groups. This study proposes a fairness‑aware ML framework designed to detect and mitigate algorithmic bias in SRH services.The framework is evaluated using two open‑source datasets: a large SRH dataset from England (2014–2015) containing 2,126,413 records, and the PCOS dataset covering the top 75 countries, enabling assessment of generalisability and intersectional fairness. It integrates pre‑processing, in‑processing, and post‑processing techniques, including model‑specific and group‑specific thresholding. Results show that on the SRH England dataset, Logistic Regression (LR) achieved near‑optimal parity fairness with minimal performance loss, improving Disparate Impact from 0.99 to 1.00 while maintaining 0.66 accuracy. Random Forest (RF) and Gradient Boosting (GB) exhibited larger fairness shifts, with Disparate Impact decreasing from 0.94 to 0.66 (RF) and 0.93 to 0.77 (GB), though accuracy remained stable. On the PCOS dataset, LR reduced bias with only a 1.96% accuracy drop, while GB improved performance but saw fairness decline, with Disparate Impact falling from 1.08 to 0.57. RF improved fairness but experienced a 28% accuracy reduction. Overall, the findings show that fairness‑aware ML can substantially reduce bias, though equity–performance trade‑offs vary across models and datasets.

Item Type: Article
Status: Published
DOI: 10.47852/bonviewJDSIS62027678
Subjects: Q Science > QA Mathematics
Q Science > QA Mathematics > QA75 Electronic computers. Computer science
R Medicine > RA Public aspects of medicine
R Medicine > RA Public aspects of medicine > RA0421 Public health. Hygiene. Preventive Medicine
School/Department: London Campus
URI: https://ray.yorksj.ac.uk/id/eprint/14755

University Staff: Request a correction | RaY Editors: Update this record