91探花

Skip to content

Computer engineers at the world鈥檚 largest companies and universities are using machines to scan through tomes of written material. The goal? Teach these machines the gift of language. Do that, some even claim, and computers will be able to mimic the human brain.

But this impressive compute capability comes with real costs, including perpetuating racism and causing significant environmental damage, according to a new paper, 鈥 🦜鈥 The paper is being presented Wednesday, March 10 at the .听

This is the first exhaustive review of the literature surrounding the risks that come with rapid growth of language-learning technologies, said , a 91探花 professor of linguistics and a lead author of the paper along with , a well-known AI researcher.

Emily Bender Photo: Corrine Thrash/91探花

鈥淭he question we’re asking is what are the possible dangers of this approach and the answers that we’re giving involve surveying literature across a broad range of fields and pulling them together,鈥 said Bender, who is the 91探花Howard and Frances Nostrand Endowed Professor.

What the researchers surfaced was that there are downsides to the ever-growing computing power put into natural language models. They discuss how the ever-increasing size of training data for language modeling exacerbates social and environmental issues. Alarmingly, such language models perpetuate hegemonic language and can deceive people into thinking they are having a “real” conversation with a person rather than a machine. The increased computational needs of these models further contributes to environmental degradation.

The authors were motivated to write the paper because of a trend within the field towards ever-larger language models and their growing spheres of influence.

The paper already has generated wide-spread attention due, in part, to the fact that two of the paper鈥檚 co-authors say they were fired recently from Google for reasons that remain unsettled. Margaret Mitchell and Gebru, the two now-former Google researchers, said they stand by the paper鈥檚 scholarship and point to its conclusions as a clarion call to industry to take heed.听

鈥淚t’s very clear that putting in the concerns has to happen right now, because it’s already becoming too late,鈥 said Mitchell, a researcher in AI.

It takes an enormous amount of computing power to fuel the model language programs, Bender said. That takes up energy at tremendous scale, and that, the authors argue, causes environmental degradation. And those costs aren鈥檛 borne by the computer engineers, but rather by marginalized people who cannot afford the environmental costs.

鈥淚t’s not just that there’s big energy impacts here, but also that the carbon impacts of that will bring costs first to people who are not benefiting from this technology,鈥 Bender said. 鈥淲hen we do the cost-benefit analysis, it鈥檚 important to think of who鈥檚 getting the benefit and who’s paying the cost because they’re not the same people.鈥

The large scale of this compute power also can restrict access to only the most well-resourced companies and research groups, leaving out smaller developers outside of the U.S., Canada, Europe and China. That鈥檚 because it takes huge machines to run the software necessary to make computers mimic human thought and speech.听

Another risk comes from the training data itself, the authors say. Because the computers read language from the Web and from other sources, they can pick up and perpetuate racist, sexist, ableist, extremist and other harmful ideologies.

鈥淥ne of the fallacies that people fall into is well, the internet is big, the internet is everything. If I just scrape the whole internet then clearly I’ve incorporated diverse viewpoints,鈥 Bender said. 鈥淏ut when we did a step-by-step review of the literature, it says that’s not the case right now because not everybody’s on the internet, and of the people who are on the internet, not everybody is socially comfortable participating in the same way.鈥

And, people can confuse the language models for real human interaction, believing that they鈥檙e actually talking with a person or reading something that a person has spoken or written, when, in fact, the language comes from a machine. Thus, the stochastic parrots.听

鈥淚t produces this seemingly coherent text, but it has no communicative intent. It has no idea what it’s saying. There’s no there there,鈥 Bender said.

, a doctoral student in linguistics at UW, also co-authored the paper.

For more information, contact Bender .听