In this 58-minute talk from the Simons Institute, UC Berkeley professor Dawn Song explores safety-guaranteed Large Language Models (LLMs) and outlines critical future directions in AI safety research. Discover key challenges and emerging approaches to ensuring AI systems remain safe, reliable, and aligned with human values as capabilities continue to advance. Learn about cutting-edge methodologies for implementing safety guarantees in LLM development and deployment.
Overview
Syllabus
Future Directions In AI Safety Research
Taught by
Simons Institute