#166 – Tantum Collins on what he’s learned as an AI policy insider at the White House, DeepMind and elsewhere

#166 – Tantum Collins on what he’s learned as an AI policy insider at the White House, DeepMind and elsewhere

"If you and I and 100 other people were on the first ship that was going to go settle Mars, and were going to build a human civilisation, and we have to decide what that government looks like, and we have all of the technology available today, how do we think about choosing a subset of that design space?

That space is huge and it includes absolutely awful things, and mixed-bag things, and maybe some things that almost everyone would agree are really wonderful, or at least an improvement on the way that things work today. But that raises all kinds of tricky questions.

My concern is that if we don't approach the evolution of collective decision making and government in a deliberate way, we may end up inadvertently backing ourselves into a corner, where we have ended up on some slippery slope -- and all of a sudden we have, let's say, autocracies on the global stage are strengthened relative to democracies." — Tantum Collins

In today’s episode, host Rob Wiblin gets the rare chance to interview someone with insider AI policy experience at the White House and DeepMind who’s willing to speak openly — Tantum Collins.

Links to learn more, highlights, and full transcript.

They cover:

  • How AI could strengthen government capacity, and how that's a double-edged sword
  • How new technologies force us to confront tradeoffs in political philosophy that we were previously able to pretend weren't there
  • To what extent policymakers take different threats from AI seriously
  • Whether the US and China are in an AI arms race or not
  • Whether it's OK to transform the world without much of the world agreeing to it
  • The tyranny of small differences in AI policy
  • Disagreements between different schools of thought in AI policy, and proposals that could unite them
  • How the US AI Bill of Rights could be improved
  • Whether AI will transform the labour market, and whether it will become a partisan political issue
  • The tensions between the cultures of San Francisco and DC, and how to bridge the divide between them
  • What listeners might be able to do to help with this whole mess
  • Panpsychism
  • Plenty more

Chapters:

  • Cold open (00:00:00)
  • Rob's intro (00:01:00)
  • The interview begins (00:04:01)
  • The risk of autocratic lock-in due to AI (00:10:02)
  • The state of play in AI policymaking (00:13:40)
  • China and AI (00:32:12)
  • The most promising regulatory approaches (00:57:51)
  • Transforming the world without the world agreeing (01:04:44)
  • AI Bill of Rights (01:17:32)
  • Who’s ultimately responsible for the consequences of AI? (01:20:39)
  • Policy ideas that could appeal to many different groups (01:29:08)
  • Tension between those focused on x-risk and those focused on AI ethics (01:38:56)
  • Communicating with policymakers (01:54:22)
  • Is AI going to transform the labour market in the next few years? (01:58:51)
  • Is AI policy going to become a partisan political issue? (02:08:10)
  • The value of political philosophy (02:10:53)
  • Tantum’s work at DeepMind (02:21:20)
  • CSET (02:32:48)
  • Career advice (02:35:21)
  • Panpsychism (02:55:24)


Producer and editor: Keiran Harris
Audio Engineering Lead: Ben Cordell
Technical editing: Simon Monsour and Milo McGuire
Transcriptions: Katy Moore

Episoder(320)

Every AI Company's Safety Plan is 'Use AI to Make AI Safe'. Is That Crazy? | Ajeya Cotra

Every AI Company's Safety Plan is 'Use AI to Make AI Safe'. Is That Crazy? | Ajeya Cotra

Every major AI company has the same safety plan: when AI gets crazy powerful and really dangerous, they’ll use the AI itself to figure out how to make AI safe and beneficial. It sounds circular, almos...

17 Feb 2h 54min

What the hell happened with AGI timelines in 2025?

What the hell happened with AGI timelines in 2025?

In early 2025, after OpenAI put out the first-ever reasoning models — o1 and o3 — short timelines to transformative artificial general intelligence swept the AI world. But then, in the second half of ...

10 Feb 25min

#179 Classic episode – Randy Nesse on why evolution left us so vulnerable to depression and anxiety

#179 Classic episode – Randy Nesse on why evolution left us so vulnerable to depression and anxiety

Mental health problems like depression and anxiety affect enormous numbers of people and severely interfere with their lives. By contrast, we don’t see similar levels of physical ill health in young p...

3 Feb 2h 51min

#234 – David Duvenaud on why 'aligned AI' would still kill democracy

#234 – David Duvenaud on why 'aligned AI' would still kill democracy

Democracy might be a brief historical blip. That’s the unsettling thesis of a recent paper, which argues AI that can do all the work a human can do inevitably leads to the “gradual disempowerment” of ...

27 Jan 2h 31min

#145 Classic episode – Christopher Brown on why slavery abolition wasn't inevitable

#145 Classic episode – Christopher Brown on why slavery abolition wasn't inevitable

In many ways, humanity seems to have become more humane and inclusive over time. While there’s still a lot of progress to be made, campaigns to give people of different genders, races, sexualities, et...

20 Jan 2h 56min

#233 – James Smith on how to prevent a mirror life catastrophe

#233 – James Smith on how to prevent a mirror life catastrophe

When James Smith first heard about mirror bacteria, he was sceptical. But within two weeks, he’d dropped everything to work on it full time, considering it the worst biothreat that he’d seen described...

13 Jan 2h 9min

#144 Classic episode – Athena Aktipis on why cancer is a fundamental universal phenomena

#144 Classic episode – Athena Aktipis on why cancer is a fundamental universal phenomena

What’s the opposite of cancer? If you answered “cure,” “antidote,” or “antivenom” — you’ve obviously been reading the antonym section at www.merriam-webster.com/thesaurus/cancer.But today’s guest Athe...

9 Jan 3h 30min

#142 Classic episode – John McWhorter on why the optimal number of languages might be one, and other provocative claims about language

#142 Classic episode – John McWhorter on why the optimal number of languages might be one, and other provocative claims about language

John McWhorter is a linguistics professor at Columbia University specialising in research on creole languages. He's also a content-producing machine, never afraid to give his frank opinion on anything...

6 Jan 1h 35min

Populært innen Fakta

fastlegen
dine-penger-pengeradet
relasjonspodden-med-dora-thorhallsdottir-kjersti-idem
treningspodden
rss-strid-de-norske-borgerkrigene
foreldreradet
jakt-og-fiskepodden
rss-sunn-okonomi
merry-quizmas
gravid-uke-for-uke
fryktlos
sinnsyn
rss-mann-i-krise-med-sagen
hverdagspsyken
generasjonspodden
rss-kunsten-a-leve
smart-forklart
dopet
teknologi-og-mennesker
rss-adhd-i-klasserommet