#87 – Russ Roberts on whether it's more effective to help strangers, or people you know

#87 – Russ Roberts on whether it's more effective to help strangers, or people you know

If you want to make the world a better place, would it be better to help your niece with her SATs, or try to join the State Department to lower the risk that the US and China go to war?

People involved in 80,000 Hours or the effective altruism community would be comfortable recommending the latter. This week's guest — Russ Roberts, host of the long-running podcast EconTalk, and author of a forthcoming book on decision-making under uncertainty and the limited ability of data to help — worries that might be a mistake.

Links to learn more, summary and full transcript.

I've been a big fan of Russ' show EconTalk for 12 years — in fact I have a list of my top 100 recommended episodes — so I invited him to talk about his concerns with how the effective altruism community tries to improve the world.

These include:

• Being too focused on the measurable
• Being too confident we've figured out 'the best thing'
• Being too credulous about the results of social science or medical experiments
• Undermining people's altruism by encouraging them to focus on strangers, who it's naturally harder to care for
• Thinking it's possible to predictably help strangers, who you don't understand well enough to know what will truly help
• Adding levels of wellbeing across people when this is inappropriate
• Encouraging people to pursue careers they won't enjoy

These worries are partly informed by Russ' 'classical liberal' worldview, which involves a preference for free market solutions to problems, and nervousness about the big plans that sometimes come out of consequentialist thinking.

While we do disagree on a range of things — such as whether it's possible to add up wellbeing across different people, and whether it's more effective to help strangers than people you know — I make the case that some of these worries are founded on common misunderstandings about effective altruism, or at least misunderstandings of what we believe here at 80,000 Hours.

We primarily care about making the world a better place over thousands or even millions of years — and we wouldn’t dream of claiming that we could accurately measure the effects of our actions on that timescale.

I'm more skeptical of medicine and empirical social science than most people, though not quite as skeptical as Russ (check out this quiz I made where you can guess which academic findings will replicate, and which won't).

And while I do think that people should occasionally take jobs they dislike in order to have a social impact, those situations seem pretty few and far between.

But Russ and I disagree about how much we really disagree. In addition to all the above we also discuss:

• How to decide whether to have kids
• Was the case for deworming children oversold?
• Whether it would be better for countries around the world to be better coordinated

Chapters:

  • Rob’s intro (00:00:00)
  • The interview begins (00:01:48)
  • RCTs and donations (00:05:15)
  • The 80,000 Hours project (00:12:35)
  • Expanding the moral circle (00:28:37)
  • Global coordination (00:39:48)
  • How to act if you're pessimistic about improving the long-term future (00:55:49)
  • Communicating uncertainty (01:03:31)
  • How much to trust empirical research (01:09:19)
  • How to decide whether to have kids (01:24:13)
  • Utilitarianism (01:34:01)


Producer: Keiran Harris.
Audio mastering: Ben Cordell.
Transcriptions: Zakee Ulhaq.

Avsnitt(321)

Why Teaching AI Right from Wrong Could Get Everyone Killed | Max Harms, MIRI

Why Teaching AI Right from Wrong Could Get Everyone Killed | Max Harms, MIRI

Most people in AI are trying to give AIs ‘good’ values. Max Harms wants us to give them no values at all. According to Max, the only safe design is an AGI that defers entirely to its human operators, ...

24 Feb 2h 41min

Every AI Company's Safety Plan is 'Use AI to Make AI Safe'. Is That Crazy? | Ajeya Cotra

Every AI Company's Safety Plan is 'Use AI to Make AI Safe'. Is That Crazy? | Ajeya Cotra

Every major AI company has the same safety plan: when AI gets crazy powerful and really dangerous, they’ll use the AI itself to figure out how to make AI safe and beneficial. It sounds circular, almos...

17 Feb 2h 54min

What the hell happened with AGI timelines in 2025?

What the hell happened with AGI timelines in 2025?

In early 2025, after OpenAI put out the first-ever reasoning models — o1 and o3 — short timelines to transformative artificial general intelligence swept the AI world. But then, in the second half of ...

10 Feb 25min

#179 Classic episode – Randy Nesse on why evolution left us so vulnerable to depression and anxiety

#179 Classic episode – Randy Nesse on why evolution left us so vulnerable to depression and anxiety

Mental health problems like depression and anxiety affect enormous numbers of people and severely interfere with their lives. By contrast, we don’t see similar levels of physical ill health in young p...

3 Feb 2h 51min

#234 – David Duvenaud on why 'aligned AI' would still kill democracy

#234 – David Duvenaud on why 'aligned AI' would still kill democracy

Democracy might be a brief historical blip. That’s the unsettling thesis of a recent paper, which argues AI that can do all the work a human can do inevitably leads to the “gradual disempowerment” of ...

27 Jan 2h 31min

#145 Classic episode – Christopher Brown on why slavery abolition wasn't inevitable

#145 Classic episode – Christopher Brown on why slavery abolition wasn't inevitable

In many ways, humanity seems to have become more humane and inclusive over time. While there’s still a lot of progress to be made, campaigns to give people of different genders, races, sexualities, et...

20 Jan 2h 56min

#233 – James Smith on how to prevent a mirror life catastrophe

#233 – James Smith on how to prevent a mirror life catastrophe

When James Smith first heard about mirror bacteria, he was sceptical. But within two weeks, he’d dropped everything to work on it full time, considering it the worst biothreat that he’d seen described...

13 Jan 2h 9min

#144 Classic episode – Athena Aktipis on why cancer is a fundamental universal phenomena

#144 Classic episode – Athena Aktipis on why cancer is a fundamental universal phenomena

What’s the opposite of cancer? If you answered “cure,” “antidote,” or “antivenom” — you’ve obviously been reading the antonym section at www.merriam-webster.com/thesaurus/cancer.But today’s guest Athe...

9 Jan 3h 30min

Populärt inom Utbildning

historiepodden-se
rss-bara-en-till-om-missbruk-medberoende-2
det-skaver
nu-blir-det-historia
sektledare
harrisons-dramatiska-historia
not-fanny-anymore
johannes-hansen-podcast
rss-viktmedicinpodden
roda-vita-rosen
rss-sjalsligt-avkladd
allt-du-velat-veta
sa-in-i-sjalen
rss-beratta-alltid-det-har
rss-max-tant-med-max-villman
i-vantan-pa-katastrofen
alska-oss
rikatillsammans-om-privatekonomi-rikedom-i-livet
sektpodden
polisutbildningspodden