OpenAI Safety: How do we control AI smarter than us


Details
This talk will delve into research conducted by the Open AI team to ensure that humanity remains in control of the future, as well as their views in the most promising directions for AI safety. It will discuss where the didea of AI as an existential risk came from and the role this played in the founding of both DeepMind and OpenAI. It will also cover how safety research led to the development of reinforcement leanring by human feedback (RLHF), the key technical innovation behind ChatGPT. It wll paint a picture of the tumultuous back and forth between the "safety" and "accelerationist" factions including much of the safety team quitting to start Anthropic. We will discuss Sam Altman's firing and reinstatement, the formation and dissolution of the super-alignment team and the controversy over Open AI's coercive non-disparagement agreements. This talk will be accessible to everybody!

OpenAI Safety: How do we control AI smarter than us