2 entries with this tag
In 2020, OpenAI scaled GPT-2 by over 100×—to 175 billion parameters—and discovered something unexpected: the model could perform tasks it was never trained on, just by reading a few examples in its prompt. 'Language Models are Few-Shot Learners' didn't just set new benchmarks. It changed what we thought language models could do.
A beginner-friendly explanation of GPT-2 (2019), the paper that showed AI could write coherent, creative text by simply predicting the next word. Part 3 of our AI Papers Explained series.