← Back to Glossary
Developer Glossary
Whisper logo

Whisper

Speech Recognition

Whisper is OpenAI's open-source automatic speech recognition model that turns audio into text with startling accuracy across dozens of languages. When a client needs transcription, meeting notes, call recordings, voice commands, podcast processing, Whisper is the model I build on. You can run it locally for privacy-sensitive applications or call it through the OpenAI API for convenience. It handles accents, background noise, technical jargon, and multilingual conversations in a way that legacy speech-to-text services from Google and AWS simply could not match when it launched. The fact that it's open-source and MIT-licensed means you can self-host it, fine-tune it, and embed it directly into a product without per-minute API costs eating into margins.

How It Started

Whisper was released by OpenAI in September 2022, developed by Alec Radford and a small team that included Jong Wook Kim, Tao Xu, Greg Brockman, and Ilya Sutskever. Radford is the same researcher who led the development of GPT-2 and was a core contributor to GPT-3. The model was trained on 680,000 hours of multilingual audio data scraped from the web, an enormous dataset that dwarfed anything used by prior speech recognition models. OpenAI's approach was deliberately brute-force: instead of engineering clever acoustic features or language-specific pipelines, they threw a standard Transformer architecture at a massive pile of weakly supervised data and let scale do the work. The research paper was titled "Robust Speech Recognition via Large-Scale Weak Supervision," and the results validated the approach. Whisper immediately outperformed commercial speech-to-text APIs on benchmarks, particularly for non-English languages and noisy environments.

The Unknown Fact

Whisper's training data was so large and diverse that it accidentally learned to perform tasks it was never explicitly trained for. Beyond basic transcription, Whisper can detect what language is being spoken, translate speech from one language to English in real time, and identify timestamps for individual words, all from the same model with zero additional training. This emergent capability came from the sheer diversity of its training data, which included subtitled YouTube videos, podcast transcripts, and audiobooks in 97 languages. Another little-known detail: the smallest Whisper model (tiny) has only 39 million parameters and can run on a Raspberry Pi. The largest model (large-v3) has 1.5 billion parameters and rivals human transcriptionists in accuracy. This range means the same architecture scales from edge devices to cloud servers, which is unusual for models in this class.

Visit: openai.com/research/whisper

Want speech recognition built into your application?

or hi@mikelatimer.ai