DSI Seminar Series | How Could We Design Aligned and Provably Safe AI?

  Рет қаралды 176

Inside Livermore Lab

Inside Livermore Lab

Күн бұрын

On April 19, 2024, Dr. Yoshua Bengio presented “How Could We Design Aligned and Provably Safe AI?” His talk was co-sponsored by LLNL’s Data Science Institute and the Center for Advanced Signal and Image Sciences. A Turning Award winner, Bengio is recognized as one of the world’s leading AI experts, known for his pioneering work in deep learning. He is a full professor at the University of Montreal, and the founder and scientific director of the Mila - Quebec AI Institute. In 2022, Bengio became the most-cited computer scientist in the world.
Evaluating the risks with a learned AI system statically seems hopeless because the number of contexts in which it could behave is infinite or exponentially large and static checks can only verify a finite and relatively small set of such contexts. However, if we had a run-time evaluation of risk, we could potentially prevent actions with an unacceptable level of risk. The probability of harm produced by an action or a plan in a given context and past data under the true explanation for how the world works is unknown. However, under reasonable hypotheses related to Occam's Razor and having a non-parametric Bayesian prior (that thus includes the true explanation) it can be shown to be bounded by quantities that can in principle be numerically approximated or estimated by large neural networks, all based on a Bayesian view that captures epistemic uncertainty about what is harm and how the world works. Capturing this uncertainty is essential: The AI could otherwise be confidently wrong about what is “good” and produce catastrophic existential risks, for example through instrumental goals or taking control of the reward mechanism (wrongly thinking that the rewards recorded in the computer are what it should maximize). The bound relies on a kind of paranoid theory, the one that has maximal probability given that it predicts harm and given the past data. The talk will discuss the research program based on these ideas and how amortized inference with large neural networks could be made to estimate the required quantities.
LLNL-VIDEO-865371

Пікірлер: 1
@Senecamarcus
@Senecamarcus 4 ай бұрын
Thanks for sharing this video. Could u please upload the q/a
DSI | GeoAI: Past, Present, and Future
1:13:30
Inside Livermore Lab
Рет қаралды 183
It's Not About Scale, It's About Abstraction
46:22
Machine Learning Street Talk
Рет қаралды 51 М.
小路飞嫁祸姐姐搞破坏 #路飞#海贼王
00:45
路飞与唐舞桐
Рет қаралды 29 МЛН
VAMPIRE DESTROYED GIRL???? 😱
00:56
INO
Рет қаралды 7 МЛН
龟兔赛跑:好可爱的小乌龟#short #angel #clown
01:00
Super Beauty team
Рет қаралды 20 МЛН
What makes learning to control easy or hard?
1:18:31
Control Seminar
Рет қаралды 373
CNN: Convolutional Neural Networks Explained - Computerphile
14:17
Computerphile
Рет қаралды 859 М.
Think Fast, Talk Smart: Communication Techniques
58:20
Stanford Graduate School of Business
Рет қаралды 41 МЛН
Stuart Russell, "AI: What If We Succeed?" April 25, 2024
1:29:57
Neubauer Collegium
Рет қаралды 25 М.
小路飞嫁祸姐姐搞破坏 #路飞#海贼王
00:45
路飞与唐舞桐
Рет қаралды 29 МЛН