Skip to content

Details

This talk will delve into research conducted by the Open AI team to ensure that humanity remains in control of the future, as well as their views in the most promising directions for AI safety. It will discuss where the didea of AI as an existential risk came from and the role this played in the founding of both DeepMind and OpenAI. It will also cover how safety research led to the development of reinforcement leanring by human feedback (RLHF), the key technical innovation behind ChatGPT. It wll paint a picture of the tumultuous back and forth between the "safety" and "accelerationist" factions including much of the safety team quitting to start Anthropic. We will discuss Sam Altman's firing and reinstatement, the formation and dissolution of the super-alignment team and the controversy over Open AI's coercive non-disparagement agreements. This talk will be accessible to everybody!

Events in Ultimo, AU
Artificial Intelligence
Artificial Intelligence Applications
Machine Learning
Neural Networks
New Technology

Members are also interested in