OpenAI logo

GPT-3: The 175 Billion Parameter Leap

What Happened

OpenAI released GPT-3 with 175 billion parameters — 100x larger than GPT-2. Without any fine-tuning, GPT-3 could write essays, code, poetry, translate languages, and answer questions through 'few-shot learning' (learning from just a few examples in the prompt). The API launched in beta, enabling thousands of applications.

Why It Mattered

A paradigm shift. GPT-3 showed that sheer scale could produce emergent capabilities no one predicted. It spawned an entire ecosystem of AI startups building on the API. The 'scaling hypothesis' — that bigger models would get smarter — gained enormous credibility.

Key People

Organizations

Tags

Related Milestones

OpenAI logo
Research

GPT-1: Generative Pre-training

OpenAI released GPT-1, demonstrating that a Transformer trained on vast amounts of text using unsupervised pre-training could then be fine-tuned for specific NLP tasks. With 117 million parameters, it showed the potential of scaling language models.

Alec RadfordOpenAI
GPT-2 language model generating text about itself
Research

GPT-2: 'Too Dangerous to Release'

OpenAI announced GPT-2 (1.5 billion parameters) but initially refused to release the full model, calling it 'too dangerous' due to its ability to generate convincing fake text. The decision was controversial — some praised the caution, others called it a publicity stunt. The full model was eventually released in November 2019.

Alec RadfordOpenAI
AI-generated image by DALL-E
Research

DALL-E: Text to Image Generation

OpenAI unveiled DALL-E, a model that could generate images from text descriptions — 'an armchair in the shape of an avocado' became iconic. Built on GPT-3's architecture adapted for images, it showed that language models could bridge the gap between text and visual creativity.

OpenAI
OpenAI logo
Research

GPT-4: Multimodal Intelligence

OpenAI released GPT-4, a multimodal model that could understand both text and images. It passed the bar exam (90th percentile), scored 1410 on the SAT, and demonstrated remarkably nuanced reasoning. It was a massive leap from GPT-3.5 in accuracy, safety, and capability.

OpenAI
GitHub Copilot AI coding assistant logo
Product

GitHub Copilot: AI Writes Code

GitHub launched Copilot as a technical preview — an AI pair programmer powered by OpenAI Codex that could autocomplete entire functions, write boilerplate, and suggest code from natural language comments. It was trained on billions of lines of public code.

GitHubOpenAI

Get the latest AI milestones as they happen

Join the newsletter. No spam, just signal.