Models like GPT-3 in part are not aligned with their users. To make models safer, more helpful, and more aligned use an existing technique called reinforcement learning from human feedback (RLHF) to create InstructGPT obtaining more truthful and less toxic outputs/ Something really interesting: this has 100x fewer parameters than GPT-3!