Abstract

Abstract: There has been an exponential increase in discussions about bias in Artificial Intelligence (AI) systems. Bias in AI has typically been defined as a divergence from standard statistical patterns in the output of an AI model, which could be due to a biased dataset or biased assumptions. While the bias in artificially taught models is attributed able to bias in the dataset provided by humans, there is still room for advancement in terms of bias mitigation in AI models. The failure to detect bias in datasets or models stems from the "black box" problem or a lack of understanding of algorithmic outcomes. This paper provides a comprehensive review of the analysis of the approaches provided by researchers and scholars to mitigate AI bias and investigate the several methods of employing a responsible AI model for decision-making processes. We clarify what bias means to different people, as well as provide the actual definition of bias in AI systems. In addition, the paper discussed the causes of bias in AI systems thereby permitting researchers to focus their efforts on minimising the causes and mitigating bias. Finally, we recommend the best direction for future research to ensure the discovery of the most accurate method for reducing bias in algorithms. We hope that this study will help researchers to think from different perspectives while developing unbiased systems.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call