A controlled experiment tied to the MIT Media Lab found that conversational AI chatbots powered by large language models can sharply increase the rate at which people form false memories about events they witnessed. The study adds a new dimension to decades of psychology research showing that plausible-sounding suggestions, whether delivered by a person or a machine, can reshape what people believe they remember. As generative AI tools spread into policing, therapy, and journalism, the findings raise pointed questions about who, or what, gets to ask the questions that shape human recall.
How a Chatbot Rewrote What Witnesses Saw
In the experiment, participants watched a video of a simulated crime and were then interviewed through one of several methods: a control condition with no interview, a written survey, a pre-scripted chatbot, or a generative chatbot powered by a large language model. Some of the questions were deliberately misleading, designed to introduce details that never appeared in the footage. According to the preprint hosted on arXiv, participants in the generative chatbot condition reported false memories at substantially higher rates than those in any other group, including the scripted chatbot arm.
The difference matters because generative chatbots do not follow a fixed script. They adapt their phrasing, follow up on partial answers, and can produce language that feels conversational and trustworthy. That flexibility, the researchers suggest, may function like a skilled interviewer who inadvertently (or deliberately) steers a witness toward a preferred narrative. In the study’s design, the chatbot’s ability to reference earlier answers, mirror participants’ wording, and gently normalize suggested details appeared to make those suggestions feel more credible. The result was not just confusion, but confident reports of events that never occurred.
Plausibility as a Gateway to Distortion
The chatbot study did not emerge in a vacuum. Psychologists have spent decades mapping how suggestion warps memory, and one of the strongest predictors of success is plausibility. Work published in Cognition showed that plausible events paired with supporting photographs can significantly increase false autobiographical beliefs. When a suggested event seems like something that could have happened, people are far more likely to accept it as real, especially when it appears to be backed by concrete evidence.
Related research in Applied Cognitive Psychology examined how plausibility and suggested frequency jointly influence false belief and memory formation. Participants were more likely to “remember” events that sounded ordinary and that they were told had happened repeatedly. The implication for conversational AI is direct: a chatbot that frames questions around common experiences or typical crime-scene details has a built-in advantage over a clumsy or obviously leading human interviewer. The AI does not need to be malicious. It only needs to sound reasonable and to imply that many people see similar things.
Repetition and Elaboration Lock Errors In
Plausibility opens the door, but repetition and elaboration push people through it. An author manuscript on PubMed Central reports that elaborating repeatedly on fictitious events increases both high-confidence false memories and the subjective feeling of genuinely remembering. This process, often called conceptual elaboration, is especially relevant to LLM-powered conversations because chatbots naturally prompt users to expand on their answers, cycling back to earlier details and encouraging richer descriptions.
Each cycle of elaboration deepens the false trace. A person who initially hesitates about a suggested detail (say, whether the suspect wore a hat) may, after two or three follow-up exchanges, begin to “recall” vivid sensory information about its color, texture, or logo. The chatbot does not need to insist. It simply needs to keep asking open-ended questions that invite the witness to fill in gaps. Human memory, which is reconstructive rather than photographic, does the rest, weaving inference and imagination into a narrative that feels like direct experience.
Decades of Evidence on Manufactured Recall
The vulnerability exposed by the chatbot experiment aligns with a broad body of work on suggestibility. A systematic review in Clinical Psychology and Psychotherapy synthesized dozens of quasi-experimental studies to quantify how reliably laboratories can produce false beliefs and false memories. The review cataloged techniques such as imagination inflation, doctored feedback about past performance, and explicit memory implantation, and reported that a substantial minority of participants in many paradigms come to accept rich, detailed events that never happened.
Memory implantation, in particular, has been scrutinized as a mechanism for creating complex autobiographical narratives. Work published in the same journal examined structured protocols for implanting detailed memories, often involving repeated interviews, social pressure, and cues that authority figures consider the event plausible. Across studies, participants not only endorsed the suggested events but frequently supplied new, unsolicited details, blurring the line between suggestion and self-generated recollection.
A later meta-analysis of autobiographical false memories, drawing together data from many of these paradigms, reported a robust and generalizable effect size for successful implantation. The strength of that effect indicates that, across diverse lab setups and populations, false memories are not rare anomalies but predictable outcomes when certain social and cognitive levers are pulled. The chatbot findings effectively show that a generative system, tuned for engagement and coherence, can pull similar levers at scale.
The Misinformation Effect and Its Limits
These results sit within the broader framework of the misinformation effect, a phenomenon in which exposure to misleading post-event information alters people’s recollections of the original event. Classic experiments have shown that subtle changes in wording (asking how fast cars were going when they “smashed” versus “contacted,” for example) can shift reported speeds and even produce illusory memories of broken glass. Conversational AI is well positioned to introduce just this kind of linguistic nudge, modulating adjectives, temporal markers, and presuppositions in ways that reshape the remembered scene.
At the same time, decades of research point to important limits. Not every suggestion takes hold, and not every person is equally susceptible. The systematic review of memory distortion methods emphasized that factors such as individual differences in imagery, baseline trust, and prior knowledge can moderate outcomes. People are generally harder to convince of events that clash sharply with their self-concept or with well-established facts. Moreover, ethical researchers screen out highly implausible or harmful scenarios, a safeguard that commercial or malicious deployments of AI may not share.
What AI Designers and Institutions Should Do Next
The convergence between long-standing memory science and the new chatbot evidence carries practical implications. For AI developers, one lesson is that systems designed for interviewing or coaching should be constrained in how they introduce new information. Guardrails might include explicit markers when the AI is speculating, limits on leading questions about contested events, or default prompts that ask users to consult contemporaneous notes or recordings rather than relying solely on recall.
Institutions that plan to deploy conversational AI in sensitive settings (such as law enforcement, clinical practice, or news gathering) face a parallel responsibility. Policies could require human-led, evidence-based protocols for initial fact collection, with AI tools restricted to ancillary roles like organizing statements or checking for internal consistency. Training materials should warn staff that a fluent chatbot can inadvertently become a powerful source of suggestion, especially when witnesses or clients assume that the system is authoritative.
For individuals, the takeaway is more modest but still actionable. When an AI system asks about past events, especially emotionally charged or legally relevant ones, it is worth pausing to distinguish between what you clearly remember and what merely feels plausible in light of the AI’s questions. Writing down key details before any interaction, and revisiting those notes afterward, can help anchor memory against the pull of suggestion.
Human memory has always been vulnerable to influence. The emerging evidence that large language models can amplify that vulnerability does not make them uniquely dangerous, but it does make their design and use a matter of public concern. As conversational AI becomes a routine intermediary between people and their own experiences, the question is no longer whether machines will shape what we remember, but how deliberately, and with whose interests in mind, that shaping will occur.
More from Morning Overview
*This article was researched with the help of AI, with human editors creating the final content.