Abstract
This article considers the rhetorical risks of using generative AI to compose organizational communication during crises or in the aftermath of tragedies. It focuses on a case study in which representatives of Vanderbilt University’s Peabody College of Education and Human Development disclosed their use of ChatGPT to write a response to a school shooting at another university. The author argues that although generative AI can often be useful in technical and professional communication, it can also undermine perceptions of “rhetorical humanity” if its use is disclosed or discovered, making it rhetorically risky in certain contexts. Thus, knowing when not to utilize AI is an important aspect of AI literacy for practitioners.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Similar Papers
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.