More dates

Payment plans

How does it work?

  • Reserve your order today and pay over time in regular, automatic payments.
  • You’ll receive your tickets and items once the final payment is complete.
  • No credit checks or third-party accounts - just simple, secure, automatic payments using your saved card.

OpenAI's Richard Ngo on AI Safety

Share
To be announced
Add to calendar
 

Event description

Within the coming decades, artificial general intelligence (AGI) may surpass human capabilities at a wide range of important tasks.

In this talk Richard outlines a case for expecting that AGIs could learn to pursue goals which are undesirable (i.e. misaligned) from a human perspective. He argues that if AGIs are trained in ways similar to today's most capable models, they could learn to act deceptively to receive higher reward, learn internally-represented goals which generalize beyond their training distributions, and pursue those goals using power-seeking strategies.

Richard is an AI safety researcher, currently working on the Governance team at OpenAI. He previously worked as a research engineer on the AGI safety team at DeepMind. Originally from Auckland, he studied at Oxford and Cambridge; while at the latter he designed and ran the AGI Safety Fundamentals course

Powered by

Tickets for good, not greed Humanitix dedicates 100% of profits from booking fees to charity

This event has passed
This event has passed
To be announced
Hosted by EA UoA