🔬 research2026-03-30T08:50:00.000Z
InstructGPT: How AI Learned What Humans Actually Want
The paper behind ChatGPT. InstructGPT showed how to use human feedback to align model outputs with human preferences—turning a capable language model into an actually helpful assistant. This is reinforcement learning from human feedback (RLHF) made real.
#ai#rlhf#alignment#training#reward