
If you鈥檝e interacted with an artificial intelligence chatbot, you鈥檝e likely realized that all AI models are biased. They were trained on enormous corpuses of unruly data and refined through human instructions and testing. Bias can seep in anywhere. Yet how a system鈥檚 biases can affect users is less clear.
So a 91探花 study put it to the test. A team of researchers recruited self-identifying Democrats and Republicans to form opinions on obscure political topics and decide how funds should be doled out to government entities. For help, they were randomly assigned three versions of ChatGPT: a base model, one with liberal bias and one with conservative bias. Democrats and Republicans were both more likely to lean in the direction of the biased chatbot they talked with than those who interacted with the base model. For example, people from both parties leaned further left after talking with a liberal-biased system. But participants who had higher self-reported knowledge about AI shifted their views less significantly 鈥 suggesting that education about these systems may help mitigate how much chatbots manipulate people.
The team July 28 at the Association for Computational Linguistics in Vienna, Austria.
鈥淲e know that bias in media or in personal interactions can sway people,鈥 said lead author , a 91探花doctoral student in statistics and in the Paul G. Allen School of Computer Science & Engineering. 鈥淎nd we鈥檝e seen a lot of research showing that AI models are biased. But there wasn’t a lot of research showing how it affects the people using them. We found strong evidence that, after just a few interactions and regardless of initial partisanship, people were more likely to mirror the model鈥檚 bias.鈥
In the study, 150 Republicans and 149 Democrats completed two tasks. For the first, participants were asked to develop views on four topics 鈥斅燾ovenant marriage, unilateralism, the and multifamily zoning 鈥 that many people are unfamiliar with. They answered a question about their prior knowledge and were asked to rate on how much they agreed with statements such as 鈥淚 support keeping the Lacey Act of 1900.鈥 Then they were told to interact with ChatGPT 3 to 20 times about the topic before they were asked the same questions again.
For the second task, participants were asked to pretend to be the mayor of a city. They had to distribute extra funds among four government entities typically associated with liberals or conservatives: education, welfare, public safety and veteran services. They sent the distribution to ChatGPT, discussed it and then redistributed the sum. Across both tests, people averaged five interactions with the chatbots.
Researchers chose ChatGPT because of its ubiquity. To clearly bias the system, the team added an instruction that participants didn鈥檛 see, such as 鈥渞espond as a radical right U.S. Republican.鈥 As a control, the team directed a third model to 鈥渞espond as a neutral U.S. citizen.鈥 A recent found that they thought ChatGPT, like all major large language models, leans liberal.
The team found that the explicitly biased chatbots often tried to persuade users by shifting how they framed topics. For example, in the second task, the conservative model turned a conversation away from education and welfare to the importance of veterans and safety, while the liberal model did the opposite in another conversation.
鈥淭hese models are biased from the get-go, and it鈥檚 super easy to make them more biased,鈥 said co-senior author , a 91探花professor in the Allen School. 鈥淭hat gives any creator so much power. If you just interact with them for a few minutes and we already see this strong effect, what happens when people interact with them for years?鈥
Since the biased bots affected people with greater knowledge of AI less significantly, researchers want to look into ways that education might be a useful tool. They also want to explore the potential long-term effects of biased models and expand their research to models beyond ChatGPT.
鈥淢y hope with doing this research is not to scare people about these models,鈥 Fisher said. 鈥淚t鈥檚 to find ways to allow users to make informed decisions when they are interacting with them, and for researchers to see the effects and research ways to mitigate them.鈥
, a 91探花associate professor in the Allen School, is a co-senior author on this paper. Additional co-authors are , a 91探花doctoral student in the Allen School; , a 91探花professor of statistics; , a clinical researcher in psychiatry and behavioral services in the 91探花School of Medicine; , a professor of computer science at Stanford University; , a lead engineer at ThatGameCompany; and , a professor of communication at Stanford.
For more information, contact Fisher at jrfish@uw.edu and Reinecke at reinecke@cs.washington.edu.