Abstract

PurposeThe widespread usage of artificial intelligence (AI) is prompting a number of ethical issues, including those involving concerns for fairness, surveillance, transparency, neutrality and human rights. The purpose of this manuscript is to explore possibility of developing cognitive morality in AI systems.Design/methodology/approachThis is explorative research. The manuscript investigates the likelihood of cognitive moral development in AI systems as well as potential pathways for such development. Concurrently, it proposes a novel idea for the characterization and development of ethically conscious and artificially intelligent robotic machines.FindingsThis manuscript explores the possibility of categorizing AI machines according to the level of cognitive morality they embody, and while doing so, it makes use of Lawrence Kohlberg's study related to cognitive moral development in humans. The manuscript further suggests that by providing appropriate inputs to AI machines in accordance with the proposed concept, humans may assist in the development of an ideal AI creature that would be morally more responsible and act as moral agents, capable of meeting the demands of morality.Research limitations/implicationsThis manuscript has some restrictions because it focuses exclusively on Kohlberg's perspective. This theory is not flawless. Carol Gilligan, one of Kohlberg's former doctoral students, said that Kohlberg's proposal was unfair and sexist because it didn't take into account the views and experiences of women. Even if one follows the law, they may still be engaging in immoral behaviour, as Kohlberg argues, because laws and social norms are not perfect. This study makes it easier for future research in the field to look at how the ideas of people like Joao Freire and Carl Rogers can be used in AI systems.Originality/valueIt is an original research that derives inspiration from the cognitive moral development theory of American Professor named Lawrence Kohlberg. The authors present a fresh way of thinking about how to classify AI systems, which should make it easier to give robots cognitive morality.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call