As the line between governance, technology, and decision-making becomes ever-blurred, advocates and disparagers of Artificial Intelligence (AI) debate the impacts of its societal integration. Despite efforts to promote Fairness, Accountability, Transparency, and Ethics (FATE), vulnerable populations continue to be systematically marginalized and made “invisible” by the racialised, classed, and colonial institutions buttressing Algorithmic Decision-Making Systems (ADMS). To address these sociotechnical risks and acknowledge our privileged, Western “standpoint epistemology,” we employ a “metaparadigm perspective” to engage the literature and case studies through a critical theory lens. The cross-analysis of three case studies: Systems Theory, the United States’ “Blueprint for an AI Bill of Rights,” and predictive policing demonstrate that our current risk mitigation frameworks are incapable of addressing transhistorical injustices. Therefore, we call for operationalizing intersectional risk theory via the decolonial turn to address ADMS risks. On that note, we propose three algorithmic accountability recommendations. We theorize that through a reformulation of FATE principles to that of the algorithmic (decolonial) self, understanding power imbalances through marginalized risk discourses, and advancing intersectional ML decision-making practices, policymakers may be better positioned to enact equitable and just ADMS.
Read full abstract