meta_pixel
Tapesearch Logo
Log in
Robert Wright's Nonzero

Brave New AI World (Robert Wright & Jeffrey Ladish)

Robert Wright's Nonzero

Nonzero

News & Politics, Society & Culture, Philosophy

4.7618 Ratings

🗓️ 27 June 2024

⏱️ 55 minutes

🧾️ Download transcript

Summary

This is a free preview of a paid episode. To hear more, visit www.nonzero.org

1:23 Why Jeffrey helped turn Meta’s Llama into Bad Llama 9:19 Bioweapons, hacking, and other AI dangers 20:03 How America unwittingly boosts China’s AI sector 29:01 The dark side of AI “alignment” 40:15 Is superintelligence the super-weapon of the future? 49:37 The perils of AI acceleration

Robert Wright (Nonzero, The Evolution of God, Why Buddhism Is True) and Jeffrey Ladish (Palisade Research). Recorded June 19, 2024.

Twitter: https://twitter.com/NonzeroPods

Transcript

Click on a timestamp to play from that location

0:00.0

You're listening to Robert Wright's Non-Zero podcast. You're listening to Robert Wright's Non-Zero Podcast.

0:28.6

Hi, Jeffrey.

0:30.5

Hey, Bob.

0:31.4

Thanks for having me on.

0:32.9

Well, thanks for coming on.

0:34.5

When we introduce us, I'm Robert Wright, publisher,

0:36.2

The Non-Zero Newsletter.

0:37.1

This is the Non-Zero podcast.

0:39.1

You're Jeffrey Laddish, Executive Director of Palisade Research, which studies AI risk.

0:47.1

And in particular, ways that AI could be used offensively by bad actors.

0:54.5

And before that, you were that, you were on the security team at Anthropic, which is one of the big,

1:02.7

I think of it as one of the big three large language model makers along with Google and

1:07.7

Open AI.

1:08.7

And of those three, the one whose identity is kind of most closely associated with concern

1:13.1

about AI risk, I would say.

1:16.6

Now, I'm really excited about this because there's a lot I want to talk about.

1:20.8

For starters, researchers at Anthropic actually published some important findings a few weeks ago about what's

1:30.9

going on inside the mind of Claude, their large language model, which I'm a fan of. And that has

1:38.4

implications for both risk and, you know, alignment that is keeping the AI in sync with human values and interests.

1:47.2

And then there was a paper that got a ton of attention in AI circles called situational awareness by Leopold Ashinbrenner, who was fired from OpenAI purportedly for just closing, you know, important

2:04.3

information. I get the sense that that was more of a pretext for firing a guy that Sam

2:09.9

Altman found annoying, but maybe I'm reading too much into it. But anyway, that paper, a lot of people found alarming.

...

Please login to see the full transcript.

Disclaimer: The podcast and artwork embedded on this page are from Nonzero, and are the property of its owner and not affiliated with or endorsed by Tapesearch.

Generated transcripts are the property of Nonzero and are distributed freely under the Fair Use doctrine. Transcripts generated by Tapesearch are not guaranteed to be accurate.

Copyright © Tapesearch 2025.