A Metauni talk about AI Safety by Elliot Catt, a worker at deepmind.

Notes

  • AI systems are highly capable, they might lead to x-risk fates.
  • Ideally, we want to run experiments on AGI, though AGI has not been developed yet
  • Next best thing is to use Universal Artificial Intelligence, which is a mathematical model that tells us how optimal ai systems will act
  • Notations will be:
    • are the finite sets of actions, observations and rewards repectively
    • to denote set of percepts (additionally )
    • denotes set of finite interaction histories
    • We use to denote some environment
  • The Value Function