Cookies on this website

We use cookies to ensure that we give you the best experience on our website. If you click 'Accept all cookies' we'll assume that you are happy to receive all cookies and you won't see this message again. If you click 'Reject all non-essential cookies' only necessary cookies providing core functionality such as security, network management, and accessibility will be enabled. Click 'Find out more' for information on how to change your cookie settings.

Humans may have evolved to be "hyperactive agency detectors". Upon hearing a rustle in a pile of leaves, it would be safer to assume that an agent, like a lion, hides beneath (even if there may ultimately be nothing there). Can this evolutionary cognitive mechanism-and related mechanisms of anthropomorphism-explain some of people's contemporary experience with using chatbots (e.g., ChatGPT, Gemini)? In this paper, we sketch how such mechanisms may engender the seemingly irresistible anthropomorphism of large language-based chatbots. We then explore the implications of this within the educational context. Specifically, we argue that people's tendency to perceive a "mind in the machine" is a double-edged sword for educational progress: Though anthropomorphism can facilitate motivation and learning, it may also lead students to trust-and potentially over-trust-content generated by chatbots. To be sure, students do seem to recognize that LLM-generated content may, at times, be inaccurate. We argue, however, that the rise of anthropomorphism towards chatbots will only serve to further camouflage these inaccuracies. We close by considering how research can turn towards aiding students in becoming digitally literate-avoiding the pitfalls caused by perceiving agency and humanlike mental states in chatbots.

Original publication

DOI

10.3390/proceedings2025114004

Type

Conference paper

Publication Date

26/02/2025

Volume

114

Keywords

anthropomorphism, artificial intelligence, education, hyperactive agency detection (HAAD)