Machine learning and natural language processing have led to the development of powerful language models such as ChatGPT, which can generate consistent and human-like responses to a wide range of queries. In many domains, ChatGPT provides appropriate responses to given commands. One of the aims of this study is to investigate the use of these association lists, such as the Deese-Roediger-McDermott (DRM) lists popular in cognitive psychology studies, by ChatGPT by giving the necessary instructions. The same method was then used to create association lists around a specific topic (climate change). The results of the first study showed that participants gave more false answers when discriminating whether critical words were presented during the test phase than when related and unrelated words were presented. This finding shows that DRM lists generated by ChatGPT can be used to search for memory errors. In line with the literature, false answers for critical words were predominantly rated as ‘remember’. The results of the second study, which was applied to the lists created on the topic of climate change and compared the responses of the groups with the climate denial scores, show that there is no significant difference in the emergence of false memories between the two groups. The level of climate change denial did not significantly affect the participants’ responses to the critical words in the climate- related lists. The low level of climate denial in the sample is a limitation of this study. It is recommended that future studies compare memory performance across an appropriate sample.