Overview
Syllabus
- The importance of interpretability in AI alignment
- Alignment as a system, not a single model
- The importance of testing intelligence in complex systems
- The Right Research
- Self-stabilizing systems
- Alignment experiment 2
- The second agent model
- Alignment research with existing technology
- Alignment research on superintelligence
- The dangers of nanotechnology and genetic alteration
- The dangers of an AI with no hard goals
- The instability of a simple for loop
- The process of creating a machine that can write novels
- The stability of instruct models
- The stability of agent models
Taught by
David Shapiro ~ AI