A Metauni talk about AI Safety by Elliot Catt, a worker at deepmind.
Notes
- AI systems are highly capable, they might lead to x-risk fates.
- Ideally, we want to run experiments on AGI, though AGI has not been developed yet
- Next best thing is to use Universal Artificial Intelligence, which is a mathematical model that tells us how optimal ai systems will act
- Notations will be:
- are the finite sets of actions, observations and rewards repectively
- to denote set of percepts (additionally )
- denotes set of finite interaction histories
- We use to denote some environment
- The Value Function