Abstract
Since the release of ChatGPT, popular demand has driven the use of social chatbots as pseudo-AI psychotherapists. With time, it is inevitable that these technologies will be deployed in some form as dedicated psychotherapy interventions. Here, we attempt to forecast the implications for psychotherapy including the unique benefits to distributive justice as well as concerns for the quality of the therapy and its societal impact. An AI psychotherapist (APT) has the potential to provide engaging clinical interactions given its capacity for highly realistic interaction as well as its high level cognitive and emotional capabilities. Moreover, it can potentially address financial and workforce limitations on access to therapy. However, an APT may cause significant iatrogenic harm if released without adequate quality control and oversight by trained psychotherapists. If not appropriately designed and regulated, APTs have potential to mislead and reinforce maladaptive coping behaviours. Given societal drivers and possible benefits, these technologies will inevitably be deployed; thus, it is incumbent upon us as a professional body to consider their regulation.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Australasian psychiatry : bulletin of Royal Australian and New Zealand College of Psychiatrists
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.