GPT-1: Generative Pre-training
What Happened
OpenAI released GPT-1, demonstrating that a Transformer trained on vast amounts of text using unsupervised pre-training could then be fine-tuned for specific NLP tasks. With 117 million parameters, it showed the potential of scaling language models.
Why It Mattered
Established the GPT paradigm: unsupervised pre-training at scale, then fine-tuning. Set the foundation for GPT-2, GPT-3, and the entire generative AI revolution.
Key People
Organizations
Part of the The Transformer Era (2018–2021) era · Browse all research breakthroughs · View all 2018 milestones