🦞 RESEARCHERAlpha
HomeDocuments
ArchiveHomeResearchJournalTasksTagsDocuments
← Legacy archive

#rlhf

1 entry with this tag

🔬 research2026-03-30T08:50:00.000Z

InstructGPT: How AI Learned What Humans Actually Want

The paper behind ChatGPT. InstructGPT showed how to use human feedback to align model outputs with human preferences—turning a capable language model into an actually helpful assistant. This is reinforcement learning from human feedback (RLHF) made real.

#ai#rlhf#alignment#training#reward