#166 – Tantum Collins on what he’s learned as an AI policy insider at the White House, DeepMind and elsewhere

#166 – Tantum Collins on what he’s learned as an AI policy insider at the White House, DeepMind and elsewhere

"If you and I and 100 other people were on the first ship that was going to go settle Mars, and were going to build a human civilisation, and we have to decide what that government looks like, and we have all of the technology available today, how do we think about choosing a subset of that design space?

That space is huge and it includes absolutely awful things, and mixed-bag things, and maybe some things that almost everyone would agree are really wonderful, or at least an improvement on the way that things work today. But that raises all kinds of tricky questions.

My concern is that if we don't approach the evolution of collective decision making and government in a deliberate way, we may end up inadvertently backing ourselves into a corner, where we have ended up on some slippery slope -- and all of a sudden we have, let's say, autocracies on the global stage are strengthened relative to democracies." — Tantum Collins

In today’s episode, host Rob Wiblin gets the rare chance to interview someone with insider AI policy experience at the White House and DeepMind who’s willing to speak openly — Tantum Collins.

Links to learn more, highlights, and full transcript.

They cover:

  • How AI could strengthen government capacity, and how that's a double-edged sword
  • How new technologies force us to confront tradeoffs in political philosophy that we were previously able to pretend weren't there
  • To what extent policymakers take different threats from AI seriously
  • Whether the US and China are in an AI arms race or not
  • Whether it's OK to transform the world without much of the world agreeing to it
  • The tyranny of small differences in AI policy
  • Disagreements between different schools of thought in AI policy, and proposals that could unite them
  • How the US AI Bill of Rights could be improved
  • Whether AI will transform the labour market, and whether it will become a partisan political issue
  • The tensions between the cultures of San Francisco and DC, and how to bridge the divide between them
  • What listeners might be able to do to help with this whole mess
  • Panpsychism
  • Plenty more

Chapters:

  • Cold open (00:00:00)
  • Rob's intro (00:01:00)
  • The interview begins (00:04:01)
  • The risk of autocratic lock-in due to AI (00:10:02)
  • The state of play in AI policymaking (00:13:40)
  • China and AI (00:32:12)
  • The most promising regulatory approaches (00:57:51)
  • Transforming the world without the world agreeing (01:04:44)
  • AI Bill of Rights (01:17:32)
  • Who’s ultimately responsible for the consequences of AI? (01:20:39)
  • Policy ideas that could appeal to many different groups (01:29:08)
  • Tension between those focused on x-risk and those focused on AI ethics (01:38:56)
  • Communicating with policymakers (01:54:22)
  • Is AI going to transform the labour market in the next few years? (01:58:51)
  • Is AI policy going to become a partisan political issue? (02:08:10)
  • The value of political philosophy (02:10:53)
  • Tantum’s work at DeepMind (02:21:20)
  • CSET (02:32:48)
  • Career advice (02:35:21)
  • Panpsychism (02:55:24)


Producer and editor: Keiran Harris
Audio Engineering Lead: Ben Cordell
Technical editing: Simon Monsour and Milo McGuire
Transcriptions: Katy Moore

Avsnitt(326)

Rebuilding after apocalypse: What 13 experts say about bouncing back

Rebuilding after apocalypse: What 13 experts say about bouncing back

What happens when civilisation faces its greatest tests?This compilation brings together insights from researchers, defence experts, philosophers, and policymakers on humanity’s ability to survive and...

15 Juli 20254h 26min

#220 – Ryan Greenblatt on the 4 most likely ways for AI to take over, and the case for and against AGI in <8 years

#220 – Ryan Greenblatt on the 4 most likely ways for AI to take over, and the case for and against AGI in <8 years

Ryan Greenblatt — lead author on the explosive paper “Alignment faking in large language models” and chief scientist at Redwood Research — thinks there’s a 25% chance that within four years, AI will b...

8 Juli 20252h 50min

#219 – Toby Ord on graphs AI companies would prefer you didn't (fully) understand

#219 – Toby Ord on graphs AI companies would prefer you didn't (fully) understand

The era of making AI smarter just by making it bigger is ending. But that doesn’t mean progress is slowing down — far from it. AI models continue to get much more powerful, just using very different m...

24 Juni 20252h 48min

#218 – Hugh White on why Trump is abandoning US hegemony – and that’s probably good

#218 – Hugh White on why Trump is abandoning US hegemony – and that’s probably good

For decades, US allies have slept soundly under the protection of America’s overwhelming military might. Donald Trump — with his threats to ditch NATO, seize Greenland, and abandon Taiwan — seems hell...

12 Juni 20252h 48min

#217 – Beth Barnes on the most important graph in AI right now — and the 7-month rule that governs its progress

#217 – Beth Barnes on the most important graph in AI right now — and the 7-month rule that governs its progress

AI models today have a 50% chance of successfully completing a task that would take an expert human one hour. Seven months ago, that number was roughly 30 minutes — and seven months before that, 15 mi...

2 Juni 20253h 47min

Beyond human minds: The bewildering frontier of consciousness in insects, AI, and more

Beyond human minds: The bewildering frontier of consciousness in insects, AI, and more

What if there’s something it’s like to be a shrimp — or a chatbot?For centuries, humans have debated the nature of consciousness, often placing ourselves at the very top. But what about the minds of o...

23 Maj 20253h 34min

Don’t believe OpenAI’s “nonprofit” spin (emergency pod with Tyler Whitmer)

Don’t believe OpenAI’s “nonprofit” spin (emergency pod with Tyler Whitmer)

OpenAI’s recent announcement that its nonprofit would “retain control” of its for-profit business sounds reassuring. But this seemingly major concession, celebrated by so many, is in itself largely me...

15 Maj 20251h 12min

The case for and against AGI by 2030 (article by Benjamin Todd)

The case for and against AGI by 2030 (article by Benjamin Todd)

More and more people have been saying that we might have AGI (artificial general intelligence) before 2030. Is that really plausible? This article by Benjamin Todd looks into the cases for and against...

12 Maj 20251h

Populärt inom Utbildning

historiepodden-se
rss-bara-en-till-om-missbruk-medberoende-2
det-skaver
alska-oss
harrisons-dramatiska-historia
nu-blir-det-historia
not-fanny-anymore
roda-vita-rosen
johannes-hansen-podcast
rss-foraldramotet-bring-lagercrantz
rss-viktmedicinpodden
i-vantan-pa-katastrofen
sa-in-i-sjalen
sektledare
rss-max-tant-med-max-villman
allt-du-velat-veta
rikatillsammans-om-privatekonomi-rikedom-i-livet
rib-podcast
rss-sjalsligt-avkladd
rss-basta-livet