Algorithmic Bias and Justice in AI Systems: A Data-Driven NLP Analysis of Decision-Making Frameworks in Global Governance

  • Jhennys Becerra Ossa Mississippi Statate University
  • Leidy Perez Coronell Universidad de la Costa
Keywords: Algorithmic Bias, AI Governance, Natural Language Processing, Legal Decision-Making, Algorithmic Fairness, Governance Systems, Responsible AI

Abstract

The integration of artificial intelligence (AI) into decision-making systems has intensified concerns about algorithmic bias and its implications for fairness in global governance. Despite growing interest in AI ethics, empirical evidence based on real-world legal data remains limited. This study develops a data-driven framework to analyze bias in AI-assisted legal and administrative decisions using Natural Language Processing (NLP). A dataset of 420 legal cases across five continents (2018–2025) was examined using lexical and semantic bias indicators, fairness scores, and governance variables such as transparency and accountability. The methodology combines computational text analysis (SpaCy, NLTK) with statistical modeling, including correlation and regression, supported by expert validation. Results reveal a significant inverse relationship between bias and fairness, as well as a moderating effect of transparency, which reduces the impact of bias. AI involvement is found to amplify existing structural biases under low-transparency conditions. The findings demonstrate that algorithmic bias is a governance-dependent phenomenon and provide a scalable framework for improving fairness in AI-driven decision systems.

Downloads

Download data is not yet available.
Published
2026-03-15
How to Cite
Becerra Ossa, J., & Perez Coronell, L. (2026). Algorithmic Bias and Justice in AI Systems: A Data-Driven NLP Analysis of Decision-Making Frameworks in Global Governance. International Journal of Management Science and Operations Research, 11(1). Retrieved from https://ijmsoridi.com/index.php/ijmsor/article/view/159