Dan Hendrycks on Catastrophic AI Risks

  Рет қаралды 2,618

Future of Life Institute

Future of Life Institute

Күн бұрын

Dan Hendrycks joins the podcast again to discuss X.ai, how AI risk thinking has evolved, malicious use of AI, AI race dynamics between companies and between militaries, making AI organizations safer, and how representation engineering could help us understand AI traits like deception. You can learn more about Dan's work at www.safe.ai
Timestamps:
00:00 X.ai - Elon Musk's new AI venture
02:41 How AI risk thinking has evolved
12:58 AI bioengeneering
19:16 AI agents
24:55 Preventing autocracy
34:11 AI race - corporations and militaries
48:04 Bulletproofing AI organizations
1:07:51 Open-source models
1:15:35 Dan's textbook on AI safety
1:22:58 Rogue AI
1:28:09 LLMs and value specification
1:33:14 AI goal drift
1:41:10 Power-seeking AI
1:52:07 AI deception
1:57:53 Representation engineering

Пікірлер: 9
@kimholder
@kimholder 6 ай бұрын
I got a lot out of this and am reading the associated paper. I have some questions. Why isn't criminal liability also included?
@PauseAI
@PauseAI 7 ай бұрын
Is there a source for Elon Musk's p(doom)?
@mrpicky1868
@mrpicky1868 6 ай бұрын
he is much better in making a serious risk taken seriously then Eliezer. hope he does more interviews
@geaca3222
@geaca3222 5 ай бұрын
I hope so too, he also recently published a very informative safety book online
@mrpicky1868
@mrpicky1868 5 ай бұрын
books have no power .sadly so more interviews and broader public understanding is what will make difference @@geaca3222
@geaca3222
@geaca3222 5 ай бұрын
@@mrpicky1868 Agree, but as an addition I think the online book is very helpful as a good source of information. It gives a concise overview of the CAIS research findings that is readily accessible for international AI safety agents and the general public. the website also offers courses on the subject.
@michaelsbeverly
@michaelsbeverly 7 ай бұрын
_Knock, knock!"_ "Who's there?" _"Hello Amazon, I'm agent of the court with service..."_ "This is about that destroying humanity thing?" _"That's right."_ "Yeah, um, about that..."
@Dan-dy8zp
@Dan-dy8zp 3 ай бұрын
He doesn't provide any justification for why we should be more concerned about these problems than about the alignment of true super-intelligence, nor any for why he thinks we are in a 'medium take-off' situation, or why we should be replaced with a 'species' instead of a singleton. *(These programs don't mate. They are not related to each other. They don't age and die and replace themselves. One would probably triumph in the end, I think, however long that takes)*. I'm left with the impression he just likes to tackle easier problems. Though if the former problem, super-alignment, is totally intractable you could argue that it makes sense to focus on what is doable and just hope we get lucky about the alignment. He doesn't really make that argument though.
Roman Yampolskiy on Shoggoth, Scaling Laws, and Evidence for AI being Uncontrollable
1:31:14
Liron Shapira on Superintelligence Goals
1:26:30
Future of Life Institute
Рет қаралды 2,6 М.
100❤️
00:20
Nonomen ノノメン
Рет қаралды 61 МЛН
1 класс vs 11 класс  (игрушка)
00:30
БЕРТ
Рет қаралды 1,4 МЛН
ДЕНЬ РОЖДЕНИЯ БАБУШКИ #shorts
00:19
Паша Осадчий
Рет қаралды 6 МЛН
Dan Hendryks Says AI Could Lead to 'Extinction'
1:34
NowThis Impact
Рет қаралды 1,4 М.
Carl Robichaud on Preventing Nuclear War
1:39:04
Future of Life Institute
Рет қаралды 1,1 М.
Why Does AI Lie, and What Can We Do About It?
9:24
Robert Miles AI Safety
Рет қаралды 251 М.
Dan Hendrycks on Why Evolution Favors AIs over Humans
2:26:38
Future of Life Institute
Рет қаралды 6 М.
3 Most chilling nuclear war aftermath movies of ALL TIMES
2:26
Minute Before Midnight Videos
Рет қаралды 30 М.
Tom Davidson on How Quickly AI Could Automate the Economy
1:56:23
Future of Life Institute
Рет қаралды 2,5 М.
How might AI be weaponized? | Al, Social Media and Nukes at SXSW 2024
57:53
Future of Life Institute
Рет қаралды 1,5 М.
The AI Safety Summit at Bletchley Park
2:19
Bletchley Park
Рет қаралды 3,1 М.
Dan Faggella on the Race to AGI
1:45:21
Future of Life Institute
Рет қаралды 7 М.
i love you subscriber ♥️ #iphone #iphonefold #shortvideo
0:14
#miniphone
0:16
Miniphone
Рет қаралды 1 МЛН
ЭТОТ ЗАБЫТЫЙ ФЛАГМАН СИЛЬНО ПОДЕШЕВЕЛ! Стоит купить...
12:54
Thebox - о технике и гаджетах
Рет қаралды 151 М.