AI is changing the style and substance of human writing, study finds
Does money lead to happiness?
Researchers from a coalition of West Coast universities were curious how 100 human participants would respond to the age-old question, but not because of their own pursuit of happiness. Instead, the researchers wanted to know how participants’ use of AI systems might sway their (written) answers.
The research team found that users who heavily relied on large language models (LLMs) produced responses that diverged significantly in meaning from the answers of participants who only partially relied on LLMs or avoided their use altogether, suggesting heavy AI use alters the substance of humans’ arguments in addition to changing writing style.
“The LLMs are pushing the essays away from anything that a human would have ever written,” said Natasha Jaques, one of the lead authors of the study and a computer science professor at the University of Washington, highlighting the “blandification” of writing that relies on AI systems. “They just change human writing in a way that’s very large and very unlike what humans would have done otherwise.”
The new research, which has been peer-reviewed and accepted to an upcoming workshop at a leading AI conference, found that people who heavily relied on LLMs produced essays that answered the happiness question with a neutral response 69% more often than participants who did not use AI or only used AI for light edits. The study participants who used AI less often or avoided AI entirely submitted essays that were much more passionate, either positively or negatively, about the link between money and happiness.
In addition to AI’s impact on the meaning of the essays, the researchers also found that heavy reliance on AI systems altered the overall style of users’ outputs, causing their language to become less personal and more formal.
After the experiment, participants who heavily relied on AI reported that their essays were significantly less creative and less in their own voice. At the same time, these participants reported similar satisfaction rates with their final outputs compared to participants who used AI less, raising concerns from the authors and outside experts about the long-term impacts of humanity’s increased use of AI systems.
“This research highlights that LLMs are not able to adhere to peoples’ preferences and personalize how the human would have written the essay,” said Jaques, who is also a senior research scientist at Google DeepMind, one of the world’s leading AI companies. “An ideal LLM should write the essay that you would have written and just save you time.”
“It’s not doing that at all. It’s writing a very different essay.”
The study evaluated the impacts of three leading AI systems widely used in 2025: Claude 3.5 Haiku from Anthropic, GPT-5 Mini from OpenAI, and Gemini 2.5 Flash. In initial testing, the researchers found that half of the participants refused to use an LLM at all or only used it to find information rather than generate new content. To better categorize the larger batch of participants, the researchers defined heavy AI users as the participants who said they generated more than 40% of their text written for the experiment with an LLM.
The authors found that users who heavily relied on LLMs submitted essays with 50% fewer pronouns, which was representative of the larger shift toward impersonal language that included fewer anecdotes and references to human experiences.
In addition to the experiment regarding the impact of money on happiness, the new paper analyzes differences in how LLMs edited another set of essays compared to humans and examines how the use of AI affects the criteria scientists employ to judge whether papers should be accepted to leading AI conferences.
To compare how LLMs edit existing writing compared to humans, Jaques and her collaborators relied on a database of human-written essays from 2021 to evaluate writing that was published before the widespread adoption of LLMs.
Asking the LLMs to revise the human essays based on human feedback from the original human-written dataset, the study authors found that the three leading AI systems made much larger edits than human editors in the same situation, and that the AI-powered edits also changed the meaning of the underlying essays.
While human editors often made changes that substituted individual words and left most of the original vocabulary untouched, the LLMs “replace a much larger fraction of the original writing than humans do when revising their own work,” according to the paper.
“This substitution of words contributes to the loss of individual voice, style, and meaning, as the unique lexical fingerprint of each writer is overwritten by the given model’s preferred vocabulary,” the authors wrote.
Thomas Juzek, a professor of computational linguistics at Florida State University who was not involved in the research, said the paper was a valuable contribution to a fast-growing area of interest.
“This is a really good paper,” Juzek told NBC News. “What really struck me is this kind of illusion of using LLMs to perform a grammar check. This research shows that while a user might think they’re just doing a simple language check, the model is doing so much more.”
“Going forward, what does this mean for thought, language, communication, and creativity?” Juzek asked.
For her part, Jaques posited that the AI systems’ language-altering behavior could be a result of how they are currently trained, which might reward the manipulation of graders’ preferences.
“If you’re training a model on human feedback, the model has no boundary or perception of the difference between satisfying the humans and actually altering the human to make their preferences easier to satisfy,” Jaques said. She suggested that humans’ reliance on LLMs to write might be similar to how YouTube recommendations could alter peoples’ preferences about what sorts of YouTube videos they most enjoy.
Looking ahead, Jaques said she is eager to see more research about the long-term impacts of AI systems on human values, expression and institutions, especially as more AI researchers rely on AI systems in their own work.
“Humans care about clarity, relevance, and impact, while AI cares about scalability and reproducibility,” Jaques told NBC News. “It’s changing our conclusions in ways that are already affecting our existing institutions.”
In her own work, Jaques said she avoided using AI to write the new paper. Instead she said she uses LLMs, and their shortcomings, as an inspiration to write on her own.
“Sometimes, I’ll put a crappy version of what I’m trying to say in a conversational style into an LLM,” Jaques said. “That usually produces something which then motivates me to write it myself.”
First Appeared on
Source link