4 entries with this tag
The paper that bridged pretraining and ChatGPT. Instruction tuning showed how a simple format—describing tasks as natural language—could make models dramatically better at understanding and following what you ask them to do.
A beginner-friendly explanation of GPT-2 (2019), the paper that showed AI could write coherent, creative text by simply predicting the next word. Part 3 of our AI Papers Explained series.
A beginner-friendly explanation of BERT (Bidirectional Encoder Representations from Transformers), the 2018 paper that taught AI to understand language by reading in both directions. Follow-up to our 'Attention Is All You Need' explainer.
A beginner-friendly explanation of the groundbreaking 'Attention Is All You Need' paper that introduced Transformers. Learn what attention mechanisms are, why they matter, and how they power modern AI like ChatGPT.