Inside the Biden admin’s AI policy approach | Jake Sullivan, Biden’s NSA | via The Cognitive Revolution

Inside the Biden admin’s AI policy approach | Jake Sullivan, Biden’s NSA | via The Cognitive Revolution

Jake Sullivan was the US National Security Advisor from 2021-2025. He joined our friends on The Cognitive Revolution podcast in August to discuss AI as a critical national security issue. We thought it was such a good interview and we wanted more people to see it, so we’re cross-posting it here on The 80,000 Hours Podcast.

Jake and host Nathan Labenz discuss:

  • Jake’s four-category framework to think about AI risks and opportunities: security, economics, society, and existential.
  • Why Jake advocates for "managed competition" with China — where the US and China "compete like hell" while maintaining sufficient guardrails to prevent conflict.
  • Why Jake thinks competition is a "chronic condition" of the US-China relationship that cannot be solved with “grand bargains.”
  • How current conflicts are providing "glimpses of the future" with lessons about scale, attritability, and the potential for autonomous weapons as AI gets integrated into modern warfare.
  • Why Jake worries that Pentagon bureaucracy prevents rapid AI adoption while China's People’s Liberation Army may be better positioned to integrate AI capabilities.
  • And why we desperately need private sector leadership: AI is "the first technology with such profound national security applications that the government really had very little to do with."

Check out more of Nathan’s interviews on The Cognitive Revolution YouTube channel: https://www.youtube.com/@CognitiveRevolutionPodcast

What did you think of the episode? https://forms.gle/g7cj6TkR9xmxZtCZ9

Originally produced by: https://aipodcast.ing

This edit by: Simon Monsour, Dominic Armstrong, and Milo McGuire | 80,000 Hours

Chapters:

  • Cold open (00:00:00)
  • Luisa's intro (00:01:06)
  • Jake’s AI worldview (00:02:08)
  • What Washington gets — and doesn’t — about AI (00:04:43)
  • Concrete AI opportunities (00:10:53)
  • Trump’s AI Action Plan (00:19:36)
  • Middle East AI deals (00:23:26)
  • Is China really a threat? (00:28:52)
  • Export controls strategy (00:35:55)
  • Managing great power competition (00:54:51)
  • AI in modern warfare (01:01:47)
  • Economic impacts in people’s daily lives (01:04:13)

Jaksot(326)

#226 – Holden Karnofsky on unexploited opportunities to make AI safer — and all his AGI takes

#226 – Holden Karnofsky on unexploited opportunities to make AI safer — and all his AGI takes

For years, working on AI safety usually meant theorising about the ‘alignment problem’ or trying to convince other people to give a damn. If you could find any way to help, the work was frustrating an...

30 Loka 20254h 30min

#225 – Daniel Kokotajlo on what a hyperspeed robot economy might look like

#225 – Daniel Kokotajlo on what a hyperspeed robot economy might look like

When Daniel Kokotajlo talks to security experts at major AI labs, they tell him something chilling: “Of course we’re probably penetrated by the CCP already, and if they really wanted something, they c...

27 Loka 20252h 12min

#224 – There's a cheap and low-tech way to save humanity from any engineered disease | Andrew Snyder-Beattie

#224 – There's a cheap and low-tech way to save humanity from any engineered disease | Andrew Snyder-Beattie

Conventional wisdom is that safeguarding humanity from the worst biological risks — microbes optimised to kill as many as possible — is difficult bordering on impossible, making bioweapons humanity’s ...

2 Loka 20252h 31min

#223 – Neel Nanda on leading a Google DeepMind team at 26 – and advice if you want to work at an AI company (part 2)

#223 – Neel Nanda on leading a Google DeepMind team at 26 – and advice if you want to work at an AI company (part 2)

At 26, Neel Nanda leads an AI safety team at Google DeepMind, has published dozens of influential papers, and mentored 50 junior researchers — seven of whom now work at major AI companies. His secret?...

15 Syys 20251h 46min

#222 – Can we tell if an AI is loyal by reading its mind? DeepMind's Neel Nanda (part 1)

#222 – Can we tell if an AI is loyal by reading its mind? DeepMind's Neel Nanda (part 1)

We don’t know how AIs think or why they do what they do. Or at least, we don’t know much. That fact is only becoming more troubling as AIs grow more capable and appear on track to wield enormous cultu...

8 Syys 20253h 1min

#221 – Kyle Fish on the most bizarre findings from 5 AI welfare experiments

#221 – Kyle Fish on the most bizarre findings from 5 AI welfare experiments

What happens when you lock two AI systems in a room together and tell them they can discuss anything they want?According to experiments run by Kyle Fish — Anthropic’s first AI welfare researcher — som...

28 Elo 20252h 28min

How not to lose your job to AI (article by Benjamin Todd)

How not to lose your job to AI (article by Benjamin Todd)

About half of people are worried they’ll lose their job to AI. They’re right to be concerned: AI can now complete real-world coding tasks on GitHub, generate photorealistic video, drive a taxi more sa...

31 Heinä 202551min

Suosittua kategoriassa Koulutus

rss-murhan-anatomia
voi-hyvin-meditaatiot-2
psykopodiaa-podcast
rss-narsisti
adhd-podi
rss-rahamania
kesken
rss-liian-kuuma-peruna
psykologia
rss-duodecim-lehti
rss-eron-alkemiaa
rss-arkea-ja-aurinkoa-podcast-espanjasta
rss-niinku-asia-on
rss-tietoinen-yhteys-podcast-2
rss-vapaudu-voimaasi
rss-finnish-daily-dialogues
rahapuhetta
rss-valo-minussa-2
rss-luonnollinen-synnytys-podcast
rss-helppoa-suomea-learn-finnish-through-comprehensible-input