Your brain works differently when you’re using generative AI to complete a task than it does when you use your brain alone. Namely, you’re less likely to remember what you did. That’s the somewhat obvious-sounding conclusion of an MIT study that looked at how people think when they write an essay — one of the earliest scientific studies of how using gen AI affects us.
The study, a preprint that has not yet been peer-reviewed, is pretty small (54 participants) and preliminary, but it points toward the need for more research into how using tools like OpenAI’s ChatGPT is affecting how our brains function. OpenAI did not immediately respond to a request for comment on the research (Disclosure: Ziff Davis, ‘s parent company, in April filed a lawsuit against OpenAI, alleging it infringed Ziff Davis copyrights in training and operating its AI systems.)
The findings show a significant difference in what happens in your brain and with your memory when you complete a task using an AI tool rather than when you do it with just your brain. But don’t read too much into those differences — this is just a glimpse at brain activity in the moment, not long-term evidence of changes in how your brain operates all the time, researchers said.
“We want to try to give some first steps in this direction and also encourage others to ask the question,” Nataliya Kosmyna, a research scientist at MIT and the lead author of the study, told me.
The growth of AI tools like chatbots is quickly changing how we work, search for information and write. All of this has happened so fast that it’s easy to forget that ChatGPT first emerged as a popular tool just a few years ago, at the end of 2022. That means we’re just now beginning to see research on how AI use is affecting us.
Here’s a look at what the MIT study found about what happened in the brains of ChatGPT users, and what future studies might tell us.
This is your brain on ChatGPT
The MIT researchers split their 54 research participants into three groups and asked them to write essays during separate sessions over several weeks. One group was given access to ChatGPT, another was allowed to use a standard search engine (Google), and the third had none of those tools, just their own brains. The researchers analyzed the texts they produced, interviewed the subjects immediately after they wrote the essays, and recorded the participants’ brain activity using electroencephalography (EEG).
An analysis of the language used in the essays found that those in the “brain-only” group wrote in more distinct ways, while those who used large language models produced fairly similar essays. More interesting findings came from the interviews after the essays were written. Those who used their brains alone showed better recall and were better able to quote from their writing than those who used search engines or LLMs.
Read more: AI Essentials: 29 Ways to Make Gen AI Work for You, According to Our Experts
It might be unsurprising that those who relied more heavily on LLMs, who may have copied and pasted from the chatbot’s responses, would be less able to quote what they had “written.” Kosmyna said these interviews were done immediately after the writing happened, and the lack of recall is notable. “You wrote it, didn’t you?” she said. “Aren’t you supposed to know what it was?”
The EEG results also showed significant differences between the three groups. There was more neural connectivity — interaction between the components of the brain — among the brain-only participants than in the search engine group, and the LLM group had the least activity. Again, not an entirely surprising conclusion. Using tools means you use less of your brain to complete a task. But Kosmyna said the research helped show what the differences were: “The idea was to look closer to understand that it’s different, but how is it different?” she said.
The LLM group showed “weaker memory traces, reduced self-monitoring and fragmented authorship,” the study authors wrote. That can be a concern in a learning environment: “If users rely heavily on AI tools, they may achieve superficial fluency but fail to internalize the knowledge or feel a sense of ownership over it.”
After the first three essays, the researchers invited participants back for a fourth session in which they were assigned to a different group. The findings there, from a significantly smaller group of subjects (just 18), found that those who were in the brain-only group at first showed more activity even when using an LLM, while those in the LLM-only group showed less neural connectivity without the LLM than the initial brain-only group had.
This isn’t ‘brainrot’
When the MIT study was released, many headlines claimed it showed ChatGPT use was “rotting” brains or causing significant long-term problems. That’s not exactly what the researchers found, Kosmyna said. The study focused on the brain activity that happened while the participants were working — their brain’s internal circuitry in the moment. It also examined their memory of their work in that moment.
Understanding the long-term effects of AI use would require a longer-term study and different methods. Kosmyna said future research could look at other gen AI use cases, like coding, or use technology that examines different parts of the brain, like functional magnetic resonance imaging, or fMRI. “The whole idea is to encourage more experiments, more scientific data collection,” she said.
While the use of LLMs is still being researched, it’s also likely that the effect on our brains isn’t as significant as you might think, said Genevieve Stein-O’Brien, assistant professor of neuroscience at Johns Hopkins University, who was not involved in the MIT study. She studies how genetics and biology help develop and build the brain — which occurs early in life. Those critical periods tend to close during childhood or adolescence, she said.
“All of this happens way before you ever interact with ChatGPT or anything like that,” Stein-O’Brien told me. “There is a lot of infrastructure that is set up, and that is very robust.”
The situation might be different in children, who are increasingly coming into contact with AI technology, although the study of children raises ethical concerns for scientists wanting to research human behavior, Stein-O’Brien said.
You can have a chatbot help you write an essay, but will you remember what you write?
Why care about essay writing anyway?
The idea of studying the effect of AI use on essay writing might sound pointless to some. After all, wasn’t the point of writing an essay in school to get a grade? Why not outsource that work to a machine that can do it, if not better, then more easily?
The MIT study gets to the point of the task: Writing an essay is about developing your thinking, about understanding the world around you.
“We start out with what we know when we begin writing, but in the act of writing, we end up framing the next questions and thinking about new ideas or new content to explore,” said Robert Cummings, a professor of writing and rhetoric at the University of Mississippi.
Cummings has done similar research on the way computer technologies affect how we write. One study involved sentence completion technology — what you might know informally as autocomplete. He took 119 writers and tasked them with writing an essay. Roughly half had computers with Google Smart Compose enabled, while the rest didn’t. Did it make writers faster, or did they spend more time and write less because they had to navigate the choices proposed? The result was that they wrote about the same amount in the same time period. “They weren’t writing in different sentence lengths, with different levels of complexity of ideas,” he told me. “It was straight-up equal.”
ChatGPT and its ilk are a different beast. With a sentence completion technology, you still have control over the words, you still have to make writing choices. In the MIT study, some participants just copied and pasted what ChatGPT said. They might not have even read the work they turned in as their own.
“My personal opinion is that when students are using generative AI to replace their writing, they’re kind of surrendering, they’re not actively engaged in their project any longer,” Cummings said.
The MIT researchers found something interesting in that fourth session, when they noticed that the group who had written three essays without tools had higher levels of engagement when finally given tools.
“Taken together, these findings support an educational model that delays AI integration until learners have engaged in sufficient self-driven cognitive effort,” they wrote. “Such an approach may promote both immediate tool efficacy and lasting cognitive autonomy.”
Cummings said he has started teaching his composition class with no devices. Students write by hand in class, generally on topics that are more personal and would be harder to feed into an LLM. He said he doesn’t feel like he’s grading papers written by AI, that his students are getting a chance to engage with their own ideas before seeking help from a tool. “I’m not going back,” he said.