FutureBriefing
Back to Readings
Beginner

On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?

Emily M. Bender, Timnit Gebru, Angelina McMillan-Major, Shmargaret Shmitchell · FAccT · 2021

Read the original paper

Plain-English Summary

A landmark paper arguing that ever-larger language models carry significant risks including environmental costs, biased training data, and the illusion of understanding. The paper's publication and the firing of co-author Timnit Gebru from Google became a defining moment in AI ethics discourse.

EthicsBiasLLMsClimateGovernance

published: false

Why This Paper Matters

Published in 2021, this paper was one of the first high-profile critiques of the "bigger is better" approach to language models. Its arguments about environmental cost, data bias, and the gap between statistical pattern-matching and genuine understanding remain central to AI discourse today.

Key Concepts

  • Environmental costs: Training large language models requires enormous computational resources with significant carbon emissions.

  • Training data bias: Models trained on internet text absorb and amplify existing societal biases, including racism, sexism, and other forms of discrimination.

  • The illusion of understanding: Language models produce fluent text that can appear meaningful without any underlying comprehension, creating risks when people treat model outputs as authoritative.

Discussion Questions

  • Have the concerns raised in this paper been addressed by the AI industry since 2021?
  • How should the environmental costs of AI training be weighed against potential benefits?
  • What responsibility do AI companies have for biases in their training data?

Further Reading