Abstract
AbstractThis study analyzes how ChatGPT portrays and describes itself, revealing misleading myths about AI technologies, specifically conversational agents based on large language models. This analysis allows for critical reflection on the potential harm these misconceptions may pose for public understanding of AI and related technologies. While previous research has explored AI discourses and representations more generally, few studies focus specifically on AI chatbots. To narrow this research gap, an experimental-qualitative investigation into auto-generated AI representations based on prompting was conducted. Over the course of a month, ChatGPT (both in its GPT-4 and GPT-4o models) was prompted to “Draw an image of yourself,” “Represent yourself visually,” and “Envision yourself visually.” The resulting data (n = 50 images and 58 texts) was subjected to a critical exploratory visual semiotic analysis to identify recurring themes and tendencies in how ChatGPT is represented and characterized. Three themes emerged from the analysis: anthropomorphism, futuristic/futurism and (social)intelligence. Importantly, compared to broader AI imaginations, the findings emphasize ChatGPT as a friendly AI assistant. These results raise critical questions about trust in these systems, not only in terms of their capability to produce reliable information and handle personal data, but also in terms of human–computer relations.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.