#25 – Why we have to lie to ourselves about why we do what we do, according to economist Robin Hanson

On February 2, 1685, England’s King Charles II was struck by a sudden illness. Fortunately his physicians were the best of the best. To reassure the public they were kept abreast of the King’s treatment regimen. King Charles was made to swallow a toxic metal; had blistering agents applied to his scalp; had pigeon droppings attached to his feet; was prodded with a red-hot poker; given forty drops of ooze from “the skull of a man that was never buried”; and, finally, had crushed stones from the intestines of an East Indian goat forced down his throat. Sadly, despite these heroic efforts, he passed away the following week.

Why did the doctors go this far?

Prof Robin Hanson – Associate Professor of Economics at George Mason University – suspects that on top of any medical beliefs the doctors had a hidden motive: it needed to be clear, to the King and the public, that the physicians cared enormously about saving His Royal Majesty. Only extreme measures could make it undeniable that they had done everything they could.

If you believe Hanson, the same desire to prove we care about our family and friends explains much of what’s perverse about our medical system today.

And not only what’s perverse about medicine – Robin thinks we’re mostly kidding ourselves when we say our charities exist to help others, our schools exist to educate students, and our political expression is about choosing wise policies.

So important are hidden motives for navigating our social world that we have to deny them to ourselves, lest we accidentally reveal them to others.

Robin is a polymath economist, and a font of surprising and novel ideas in a range of fields including psychology, politics and futurology. In this extensive episode we discuss his latest book with Kevin Simler, The Elephant in the Brain: Hidden Motives in Everyday Life. We also dive into:

  • What was it like being part of a competitor group to the ‘World Wide Web’, but being beaten to the post?
  • If people aren’t going to school to learn, what’s education for?
  • What split brain patients show about our capacity for self-justification
  • Why we choose the friends we do
  • What’s puzzling about our attitude to medicine?
  • How would it look if people were focused on doing as much good as possible?
  • Are we better off donating now, when we’re older, or even after our deaths?
  • How much of the behavior of ‘effective altruists’ can we assume is genuinely motivated by wanting to do as much good as possible?
  • What does Robin mean when he refers to effective altruism as a youth movement? Is that a good or bad thing?
  • Should people make peace with their hidden motives, or remain ignorant of them?
  • How might we change policy if we fully understood these hidden motivations?
  • Is this view of human nature depressing?
  • Could we let betting markets run much of the government?
  • Why don’t big ideas for institutional reform get adopted?
  • Does history show we’re capable of predicting when new technologies will arise, or what their social impact will be?
  • What are the problems with thinking about the future in an abstract way?
  • Why has Robin shifted from mainly writing papers, to writing blog posts, to writing books?
  • Why are people working in policy reluctant to accept conclusions from psychology?
  • How did being publicly denounced by senators help Robin’s career?
  • Is contrarianism good or bad?
  • The relationship between the quality of an argument and its popularity
  • What would Robin like to see effective altruism do differently?
  • What has Robin changed his mind about over the last 5 years?

The 80,000 Hours podcast is produced by Keiran Harris.

Continue reading →

#24 – Stefan Schubert on why it’s a bad idea to break the rules, even if it’s for a good cause

How honest should we be? How helpful? How friendly? If our society claims to value honesty, for instance, but in reality accepts an awful lot of lying – should we go along with those lax standards? Or, should we attempt to set a new norm for ourselves?

Dr Stefan Schubert, a researcher at the Social Behaviour and Ethics Lab at Oxford University, has been modelling this in the context of the effective altruism community. He thinks people trying to improve the world should hold themselves to very high standards of integrity, because their minor sins can impose major costs on the thousands of others who share their goals.

In addition, when a norm is uniquely important to our situation, we should be willing to question society and come up with something different and hopefully better.

But in other cases, we can be better off sticking with whatever our culture expects, both to save time, avoid making mistakes, and ensure others can predict our behaviour.

In this interview Stefan offers a range of views on the projects and culture that make up ‘effective altruism’ – including where it’s going right and where it’s going wrong.

Stefan did his PhD in formal epistemology, before moving on to a postdoc in political rationality at the London School of Economics, while working on advocacy projects to improve truthfulness among politicians. At the time the interview was recorded Stefan was a researcher at the Centre for Effective Altruism in Oxford.

We also discuss:

  • Should we trust our own judgement more than others’?
  • How hard is it to improve political discourse?
  • What should we make of well-respected academics writing articles that seem to be completely misinformed?
  • How is effective altruism (EA) changing? What might it be doing wrong?
  • How has Stefan’s view of EA changed?
  • Should EA get more involved in politics, or steer clear of it? Would it be a bad idea for a talented graduate to get involved in party politics?
  • How much should we cooperate with those with whom we have disagreements?
  • What good reasons are there to be inconsiderate?
  • Should effective altruism potentially focused on a more narrow range of problems?

The 80,000 Hours podcast is produced by Keiran Harris.

Continue reading →

#23 – Jan Leike on how to become a machine learning alignment researcher

Want to help steer the 21st century’s most transformative technology? First complete an undergrad degree in computer science and mathematics. Prioritize harder courses over easier ones. Publish at least one paper before you apply for a PhD. Find a supervisor who’ll have a lot of time for you. Go to the top conferences and meet your future colleagues. And finally, get yourself hired.

That’s Dr Jan Leike’s advice on how to join him as a Research Scientist at DeepMind, the world’s leading AI team.

Jan is also a Research Associate at the Future of Humanity Institute at the University of Oxford, and his research aims to make machine learning robustly beneficial. His current focus is getting AI systems to learn good objective functions in cases where we can’t easily specify the outcome we actually want.

How might you know you’re a good fit for this kind of research?

Jan says to check whether you get obsessed with puzzles and problems, and find yourself mulling over questions that nobody knows the answer to. To do research in a team you also have to be good at clearly and concisely explaining your new ideas to other people.

We also discuss:

  • Where do Jan’s views differ from those expressed by Dario Amodei in episode 3?
  • Why is AGI alignment one of the world’s most pressing problems?
  • Common misconceptions about artificial intelligence
  • What are some of the specific things DeepMind is researching?
  • The ways in which today’s AI systems can fail
  • What are the best techniques available today for teaching an AI the right objective function?
  • What’s it like to have some of the world’s greatest minds as coworkers?
  • Who should do empirical research and who should do theoretical research
  • What’s the DeepMind application process like?
  • The importance of researchers being comfortable with the unknown.

The 80,000 Hours podcast is produced by Keiran Harris.

Continue reading →

#22 – Leah Utyasheva on how to massively cut suicide rates in Sri Lanka, and her non-profit’s plan to do the same around the world

How people kill themselves varies enormously depending on which means are most easily available. In the United States, suicide by firearm stands out. In Hong Kong, where most people live in high rise buildings, jumping from a height is more common. And in some countries in Asia and Africa with many poor agricultural communities, the leading means is drinking pesticide.

There’s a good chance you’ve never heard of this issue before. And yet, of the 800,000 people who kill themselves globally each year 20% die from pesticide self-poisoning.

Research suggests most people who try to kill themselves with pesticides reflect on the decision for less than 30 minutes, and that less than 10% of those who don’t die the first time around will try again.

Unfortunately, the fatality rate from pesticide ingestion is 40% to 70%.

Having such dangerous chemicals near people’s homes is therefore an enormous public health issue not only for the direct victims, but also the partners and children they leave behind.

Fortunately researchers like Dr Leah Utyasheva have figured out a very cheap way to massively reduce pesticide suicide rates.

In 2016, Leah co-founded the first organisation focused on this problem – The Centre for Pesticide Suicide Prevention – which recently received an incubation grant from GiveWell. She’s a human rights expert and law reform specialist, and has participated in drafting legal aid, human rights, gender equality, and anti-discrimination legislation in various countries across Europe and Canada.

In this episode, Leah and I discuss:

  • How do you prevent pesticide suicide and what’s the evidence it works?
  • How do you know that most people attempting suicide don’t want to die?
  • What types of events are causing people to have the crises that lead to attempted suicide?
  • How much money does it cost to save a life in this way?
  • How do you estimate the probability of getting law reform passed in a particular country?
  • Have you generally found politicians to be sympathetic to the idea of banning these pesticides? What are their greatest reservations?
  • The comparison of getting policy change rather than helping person-by-person
  • The importance of working with locals in places like India and Nepal, rather than coming in exclusively as outsiders
  • What are the benefits of starting your own nonprofit versus joining an existing org and persuading them of the merits of the cause?
  • Would Leah in general recommend starting a new charity? Is it more exciting than it is scary?
  • Is it important to have an academic leading this kind of work?
  • How did The Centre for Pesticide Suicide Prevention get seed funding?
  • How does the value of saving a life from suicide compare to savings someone from malaria
  • Leah’s political campaigning for the rights of vulnerable groups in Eastern Europe
  • What are the biggest downsides of human rights work?

Keiran Harris helped produce today’s episode.

Continue reading →

#21 – The world’s most intellectual foundation is hiring. Holden Karnofsky, founder of GiveWell, on how philanthropy can have maximum impact by taking big risks.

The Green Revolution averted mass famine during the 20th century. The contraceptive pill gave women unprecedented freedom in planning their own lives. Both are widely recognised as scientific breakthroughs that transformed the world. But few know that those breakthroughs only happened when they did because of two donors willing to take risky bets on new ideas.

Today’s guest, Holden Karnofsky, has been looking for philanthropy’s biggest success stories because he’s Executive Director of Open Philanthropy, which gives away over $100 million per year – and he’s hungry for big wins.

As he learned, in the 1940s, poverty reduction overseas was not a big priority for many. But the Rockefeller Foundation decided to fund agricultural scientists to breed much better crops for the developing world – thereby massively increasing their food production.

Similarly in the 1950s, society was a long way from demanding effective birth control. Activist Margaret Sanger had the idea for the pill, and endocrinologist Gregory Pincus the research team – but they couldn’t proceed without a $40,000 research check from biologist and women’s rights activist Katherine McCormick.

In both cases, it was philanthropists rather than governments that led the way.

The reason, according to Holden, is that while governments have enormous resources, they’re constrained by only being able to fund reasonably sure bets. Philanthropists can transform the world by filling the gaps government leaves – but to seize that opportunity they have to hire the best researchers, think long-term and be willing to fail most of the time.

Holden knows more about this type of giving than almost anyone. As founder of GiveWell and then Open Philanthropy, he has been working feverishly since 2007 to find outstanding giving opportunities. This practical experience has made him one of the most influential figures in the development of the school of thought that has come to be known as effective altruism.

We’ve recorded this episode now because Open Philanthropy is hiring for a large number of positions, which we think would allow the right person to have a very large positive influence on the world. They’re looking for a large number of entry lever researchers to train up, 3 specialist researchers into potential risks from advanced artificial intelligence, as well as a Director of Operations, Operations Associate and General Counsel.

But the conversation goes well beyond specifics about these jobs. We also discuss:

  • How did they pick the problems they focus on, and how will they change over time?
  • What would Holden do differently if he were starting Open Phil again today?
  • What can we learn from the history of philanthropy?
  • What makes a good Program Officer.
  • The importance of not letting hype get ahead of the science in an emerging field.
  • The importance of honest feedback for philanthropists, and the difficulty getting it.
  • How do they decide what’s above the bar to fund, and when it’s better to hold onto the money?
  • How philanthropic funding can most influence politics.
  • What Holden would say to a new billionaire who wanted to give away most of their wealth.
  • Why Open Phil is building a research field around the safe development of artificial intelligence
  • Why they invested in OpenAI.
  • Academia’s faulty approach to answering practical questions.
  • What kind of people do and don’t thrive in Open Phil’s culture.
  • What potential utopias do people most want, according to opinion polls?

Keiran Harris helped produce today’s episode.

Continue reading →

#20 – Bruce Friedrich makes the case that inventing outstanding meat replacements is the most effective way to help animals

Before the US Civil War, it was easier for the North to morally oppose slavery. Why? Because unlike the South they weren’t profiting much from its existence. The fight for abolition was partly won because many no longer saw themselves as having a selfish stake in its continuation.

Bruce Friedrich, executive director of The Good Food Institute (GFI), thinks the same may be true in the fight against speciesism. 98% of people currently eat meat. But if eating meat stops being part of most people’s daily lives — it should be a lot easier to convince them that farming practices are just as cruel as they look, and that the suffering of these animals really matters.

That’s why GFI is “working with scientists, investors, and entrepreneurs” to create plant-based meat, dairy and eggs as well as clean meat alternatives to animal products. In 2016, Animal Charity Evaluators named GFI one of its recommended charities.

In this interview I’m joined by my colleague Natalie Cargill, and we ask Bruce about:

  • What’s the best meat replacement product out there right now?
  • How effective is meat substitute research for people who want to reduce animal suffering as much as possible?
  • When will we get our hands on clean meat? And why does Bruce call it clean meat, rather than in vitro meat or cultured meat?
  • What are the challenges of producing something structurally identical to meat?
  • Can clean meat be healthier than conventional meat?
  • Do plant-based alternatives have a better shot at success than clean meat?
  • Is there a concern that, even if the product is perfect, people still won’t eat it? Why might that happen?
  • What’s it like being a vegan in a family made up largely of hunters and meat-eaters?
  • What kind of pushback should be expected from the meat industry?

Keiran Harris helped produce today’s episode.

Continue reading →

#19 – Samantha Pitts-Kiefer on her job worrying about any way nukes could get used

Rogue elements within a state’s security forces enrich dozens of kilograms of uranium. It’s then assembled into a crude nuclear bomb. The bomb is transported on a civilian aircraft to Washington D.C, and loaded onto a delivery truck. The truck is driven by an American citizen midway between the White House and the Capitol Building. The driver casually steps out of the vehicle, and detonates the weapon. There are more than 80,000 instant deaths. There are also at least 100,000 seriously wounded, with nowhere left to treat them.

It’s likely that one of those immediately killed would be Samantha Pitts-Kiefer, who works only one block away from the White House.

Samantha serves as Senior Director of The Global Nuclear Policy Program at the Nuclear Threat Initiative, and warns that the chances of a nuclear terrorist attack are alarmingly high. Terrorist groups have expressed a desire for nuclear weapons, and the material required to build those weapons is scattered throughout the world at a diverse range of sites – some of which lack the necessary security.

When you combine the massive death toll with the accompanying social panic and economic disruption – a nuclear 9/11 would be unthinkably bad. And yet, Samantha reminds us, we must confront the possibility.

Clearly, this is far from the only nuclear nightmare. We also discuss:

  • In the case of nuclear war, what fraction of the world’s population would die?
  • What is the biggest nuclear threat?
  • How concerned should we be about North Korea?
  • How often has the world experienced nuclear near misses?
  • How might a conflict between India and Pakistan escalate to the nuclear level?
  • How quickly must a president make a decision in the result of a suspected first strike?
  • Are global sources of nuclear material safely secured?
  • What role does cyber security have in preventing nuclear disasters?
  • How can we improve relations between nuclear armed states?
  • What do you think about the campaign for complete nuclear disarmament?
  • If you could tell the US government to do three things, what are the key priorities today?
  • Is it practical to get members of congress to pay attention to nuclear risks?
  • Could modernisation of nuclear weapons actually make the world safer?

Keiran Harris helped produce today’s episode.

Continue reading →

#18 – Ofir Reich on using data science to end poverty and the spurious action/inaction distinction

Ofir Reich spent 6 years doing math in the military, before spending another 2 in tech startups – but then made a sharp turn to become a data scientist focussed on helping the global poor.

At UC Berkeley’s Center for Effective Global Action he helps prevent tax evasion by identifying fake companies in India, enable Afghanistan to pay its teachers electronically, and raise yields for Ethiopian farmers by messaging them when local conditions make it ideal to apply fertiliser. Or at least that’s the hope – he’s also working on ways to test whether those interventions actually work.

Why dedicate his life to helping the global poor?

Ofir sees little moral difference between harming people and failing to help them. After all, if you had to press a button to keep all of your money from going to charity, and you pressed that button, would that be an action, or an inaction? Is there even an answer?

After reflecting on cases like this, he decided that to not engage with a problem is an active choice, one whose consequences he is just as morally responsible for as if he were directly involved. On top of his life philosophy we also discuss:

  • The benefits of working in a top academic environment
  • How best to start a career in global development
  • Are RCTs worth the money? Should we focus on big picture policy change instead? Or more economic theory?
  • How the delivery standards of nonprofits compare to top universities
  • Why he doesn’t enjoy living in the San Francisco bay area
  • How can we fix the problem of most published research being false?
  • How good a career path is data science?
  • How important is experience in development versus technical skills?
  • How he learned much of what he needed to know in the army
  • How concerned should effective altruists be about burnout?

Keiran Harris helped produce today’s episode.

Continue reading →

#17 – Will MacAskill fears our descendants will probably see us as moral monsters. What should we do about that?

Immanuel Kant is a profoundly influential figure in modern philosophy, and was one of the earliest proponents for universal democracy and international cooperation. He also thought that women have no place in civil society, that illegitimate children should receive fewer legal protections, and that there was a ranking in the moral worth of different races.

Throughout history we’ve consistently believed, as common sense, truly horrifying things by today’s standards. According to philosopher Will MacAskill, it’s extremely likely that we’re in the same boat today. If we accept that we’re probably making major moral errors, how should we proceed?

If our morality is tied to common sense intuitions, we’re probably just preserving these biases and moral errors. Instead we need to develop a moral view that criticises common sense intuitions, and gives us a chance to move beyond them. And if humanity is going to spread to the stars it could be worth dedicating hundreds or thousands of years to moral reflection, lest we spread our errors far and wide.

Will is an Associate Professor in Philosophy at Oxford University, author of Doing Good Better, and one of the co-founders of the effective altruism community. In this interview we discuss a wide range of topics:

  • How would we go about a ‘long reflection’ to fix our moral errors?
  • Will’s forthcoming book on how one should reason and act if you don’t know which moral theory is correct. What are the practical implications of so-called ‘moral uncertainty’?
  • If we basically solve existential risks, what does humanity do next?
  • What are some of Will’s most unusual philosophical positions?
  • What are the best arguments for and against utilitarianism?
  • Given disagreements among philosophers, how much should we believe the findings of philosophy as a field?
  • What are some the biases we should be aware of within academia?
  • What are some of the downsides of becoming a professor?
  • What are the merits of becoming a philosopher?
  • How does the media image of EA differ to the actual goals of the community?
  • What kinds of things would you like to see the EA community do differently?
  • How much should we explore potentially controversial ideas?
  • How focused should we be on diversity?
  • What are the best arguments against effective altruism?

Keiran Harris helped produce today’s episode.

Continue reading →

#16 – Michelle Hutchinson hopes to shape the world by shaping the ideas of intellectuals. Will global priorities research succeed?

In the 40s and 50s neoliberalism was a fringe movement within economics. But by the 80s it had become a dominant school of thought in public policy, and achieved major policy changes across the English speaking world. How did this happen?

In part because its leaders invested heavily in training academics to study and develop their ideas. Whether you think neoliberalism was good or bad, its history demonstrates the impact building a strong intellectual base within universities can have.

Dr Michelle Hutchinson is working to get a different set of ideas a hearing in academia by setting up the Global Priorities Institute (GPI) at Oxford University. The Institute, which is currently hiring for three roles, aims to bring together outstanding philosophers and economists to research how to most improve the world. The hope is that it will spark widespread academic engagement with effective altruist thinking, which will hone the ideas and help them gradually percolate into society more broadly.

Its research agenda includes questions like:

  • How do we compare the good done by focussing on really different types of causes?
  • How does saving lives actually affect the world relative to other things we could do?
  • What are the biggest wins governments should be focussed on getting?

Before moving to GPI, Michelle was the Executive Director of Giving What We Can and a founding figure of the effective altruism movement. She has a PhD in Applied Ethics from Oxford on prioritization and global health.

We discuss:

  • What is global priorities research and why does it matter?
  • How is effective altruism seen in academia? Is it important to convince academics of the value of your work, or is it OK to ignore them?
  • Operating inside a university is quite expensive, so is it even worth doing? Who can pay for this kind of thing?
  • How hard is it to do something innovative inside a university? How serious are the administrative and other barriers?
  • Is it harder to fundraise for a new institute, or hire the right people?
  • Have other social movements benefitted from having a prominent academic arm?
  • How can people prepare themselves to get research roles at a place like GPI?
  • Many people want to have roles doing this kind of research. How many are actually cut out for it? What should those who aren’t do instead?
  • What are the odds of the Institute’s work having an effect on the real world?

If you’re interesting in donating to or working at GPI, you can email Michelle at [email protected].

Continue reading →

#15 – Phil Tetlock on predicting catastrophes, why keep your politics secret, and when experts know more than you

Prof Philip Tetlock is a social science legend. Over forty years he has researched whose forecasts we can trust, whose we can’t and why – and developed methods that allow all of us to be better at predicting the future.

After the Iraq WMDs fiasco, the US intelligence services hired him to figure out how to ensure they’d never screw up that badly again. The result of that work – Superforecasting – was a media sensation in 2015.

It described Tetlock’s Good Judgement Project, which found forecasting methods so accurate they beat everyone else in open competition, including thousands of people in the intelligence services with access to classified information.

Today he’s working to develop the best forecasting process ever by combining the best of human and machine intelligence in the Hybrid Forecasting Competition, which you can start participating in now to sharpen your own judgement.

In this interview we describe his key findings and then push to the edge of what’s known about how to foresee the unforeseeable:

  • Should people who want to be right just adopt the views of experts rather than apply their own judgement?
  • Why are Berkeley undergrads worse forecasters than dart-throwing chimps?
  • Should I keep my political views secret, so it will be easier to change them later?
  • How can listeners contribute to his latest cutting-edge research?
  • What do we know about our accuracy at predicting low-probability high-impact disasters?
  • Does his research provide an intellectual basis for populist political movements?
  • Was the Iraq War caused by bad politics, or bad intelligence methods?
  • What can we learn about forecasting from the 2016 election?
  • Can experience help people avoid overconfidence and underconfidence?
  • When does an AI easily beat human judgement?
  • Could more accurate forecasting methods make the world more dangerous?
  • How much does demographic diversity line up with cognitive diversity?
  • What are the odds we’ll go to war with China?
  • Should we let prediction tournaments run most of the government?

Continue reading →

#14 – Sharon Nuñez & Jose Valle on going undercover to expose animal cruelty, get rabbit cages banned, and reduce meat consumption

What if you knew that ducks were being killed with pitchforks? Rabbits dumped alive into containers? Or pigs being strangled with forklifts? Would you be willing to go undercover to expose the crime?

That’s a real question that confronts volunteers at Animal Equality (AE). In this episode we speak to Sharon Nunez and Jose Valle, who founded AE in 2006 and then grew it into a multi-million dollar international animal rights organisation. They’ve been chosen as one of the most effective animal protection orgs in the world by Animal Charity Evaluators for the last 3 consecutive years.

In addition to undercover investigations AE has also designed a 3D virtual-reality farm experience called iAnimal360. People get to experience being trapped in a cage – in a room designed to kill then – and can’t just look away. How big an impact is this having on users?

In this interview I’m joined by my colleague Natalie Cargill – Sharon Nuñez and Jose Valle also tackle:

  • How do they track their goals and metrics week to week?
  • How much does an undercover investigation cost?
  • Why don’t people donate more to factory farmed animals, given that they’re the vast majority of animals harmed directly by humans?
  • How risky is it to attempt to build a career in animal advocacy?
  • What led to a change in their focus from bullfighting in Spain to animal farming?
  • How does working with governments or corporate campaigns compare with early strategies like creating new vegans/vegetarians?
  • Has their very rapid growth been difficult to handle?
  • What should our listeners study or do if they want to work in this area?
  • How can we get across the message that horrific cases are a feature – not a bug – of factory farming?
  • Do the owners or workers of factory farms ever express shame at what they do?

If you’re interested in this episode you’ll also want to hear our comprehensive review of ways to help animals with Lewis Bollard.

Continue reading →

#13 – Claire Walsh on testing which policies work & how to get governments to listen to the results

In both rich and poor countries, government policy is often based on no evidence at all and many programs don’t work. This has particularly harsh effects on the global poor – in some countries governments only spend $100 on each citizen a year so they can’t afford to waste a single dollar.

Enter MIT’s Poverty Action Lab (J-PAL). Since 2003 they’ve conducted experiments to figure out what policies actually help recipients, and then try to get them implemented by governments and nonprofits.

Claire Walsh leads J-PAL’s Government Partnership Initiative, which works to evaluate policies and programs in collaboration with developing world governments, scale policies that have been shown to work, and generally promote a culture of evidence-based policymaking.

We discussed (her views only, not J-PAL’s):

  • How can they get evidence backed policies adopted? Do politicians in the developing world even care whether their programs actually work? Is the norm evidence-based policy, or policy-based evidence?
  • Is evidence-based policy an evidence-based strategy itself?
  • Which policies does she think would have a particularly large impact on human welfare relative to their cost?
  • How did she come to lead one of J-PAL’s departments at 29?
  • How do you evaluate the effectiveness of energy and environment programs (Walsh’s area of expertise), and what are the standout approaches in that area?
  • 80,000 Hours has warned people about the downsides of starting your career in a nonprofit. Walsh started her career in a nonprofit and has thrived, so are we making a mistake?
  • Other than J-PAL, what are the best places to work in development? What are the best subjects to study? Where can you go network to break into the sector?
  • Is living in poverty as bad as we think?

And plenty of other things besides.

We haven’t run an RCT to test whether this episode will actually help your career, but I suggest you listen anyway. Trust my intuition on this one.

Continue reading →

#12 – Beth Cameron fought Ebola for the White House. Now she works to stop something even worse.

When you’re in the middle of a crisis and you have to ask for money, you’re already too late.

That’s Dr. Beth Cameron, and she’s someone who should know. Beth runs Global Biological Policy and Programs at the Nuclear Threat Initiative.

She has years of experience preparing for and fighting the diseases of our nightmares, on the White House Ebola Taskforce, in the National Security Council staff, and as the senior advisor to the Assistant Secretary of Defense for Nuclear, Chemical and Biological Defense Programs.

Unfortunately, the nations of the world aren’t prepared for a crisis – and like children crowded into daycare, there’s a real danger that something nasty will come along and make us all sick at once.

During previous pandemics, countries have dragged their feet over who will pay to contain them, or struggled to move people and supplies to where they needed to be. Unfortunately, there’s no reason to think that the same wouldn’t happen again today. And at the same time, advances in biotechnology may make it possible for terrorists to bring back smallpox – or create something even worse.

In this interview we look at the current state of play in disease control, what needs to change, and how you can work towards a job where you can help make those changes yourself. Topics covered include:

  • The best strategies for containing pandemics.
  • Why we lurch from panic, to neglect, to panic again when it comes to protecting ourselves from contagious diseases.
  • Current reform efforts within the World Health Organization, and attempts to prepare partial vaccines ahead of time.
  • How the Nuclear Threat Initiative, with just 50 people, collaborates with governments around the world to reduce the risk of nuclear or biological catastrophes (also, whether they might want to hire you).
  • Which global health security groups most impress Beth, and what they’re doing.
  • What new technologies could be invented to make us safer.
  • Whether it’s possible to help solve the problem through mass advocacy.
  • What and where to study, and how to begin a career in pandemic preparedness (below you’ll find a lengthy list of people and places mentioned in the interview, and others we’ve had recommended to us).
  • Much more besides.

Below you’ll find a coaching application form, three key points from the interview, extra resources to learn more, and dozens of people and places you can contact to begin a career in this field.

If you know nothing about this topic, it is recommended that you listen to the first hour or two of the episode with Howie Lempel first, as it lays out the problem more gradually.

Continue reading →

#11 – Spencer Greenberg on speeding up social science 10-fold & why plenty of startups cause harm

What is the best, state-of-the-art therapy for depression? Do most meat eaters think it’s wrong to hurt animals? How likely do Americans think climate change is to cause human extinction? How do we make academics more intellectually honest, so we can actually trust their findings? How can we speed up social science research 10-fold? Do most startups improve the world, or make it worse? Why is research in top journals less reliable?

If you’re interested in these questions, this interview is for you.

A scientist, entrepreneur, writer and mathematician, Spencer Greenberg is constantly working to create tools to speed up and improve research and critical thinking. These include:

  • Rapid public opinion surveys – which he has used to learn public opinion on animal consciousness, farm animal welfare, the impact of developing world charities and the likelihood of extinction by various different means;
  • Tools to enable social science research to be run en masse very cheaply by anyone;
  • ClearerThinking.org, a highly popular site for improving people’s judgement and decision-making;
  • Ways to transform data analysis methods to ensure that papers only show true findings;
  • Ways to decide which research projects are actually worth pursuing.

In this episode of the show, Spencer discusses all of these and more. If you don’t feel like listening, that just shows that you have poor judgement and need to benefit from his wisdom even more!

Continue reading →

#10 – Nick Beckstead on how to spend billions of dollars preventing human extinction

What if you were in a position to give away billions of dollars to improve the world? What would you do with it? This is the problem facing Program Officers at Open Philanthropy – people like Dr Nick Beckstead.

Following a PhD in philosophy, Nick works to figure out where money can do the most good. He’s been involved in major grants in a wide range of areas, including ending factory farming through technological innovation, safeguarding the world from advances in biotechnology and artificial intelligence, and spreading rational compassion.

This episode is a tour through some of the toughest questions ‘effective altruists’ face when figuring out how to best improve the world, including:

  • Should we mostly try to help people currently alive, or future generations? Nick studied this question for years in his PhD thesis, On the Overwhelming Importance of Shaping the Far Future. (The first 31 minutes is a snappier version of my conversation with Toby Ord.)
  • Is clean meat (aka in vitro meat) technologically feasible any time soon, or should we be looking for plant-based alternatives?
  • To stop malaria is it more cost-effective to use technology to eliminate mosquitos than to distribute bed nets?
  • What are the greatest risks to human civilisation continuing?
  • Should people who want to improve the future work for changes that will be very useful in a specific scenario, or just generally try to improve how well humanity makes decisions?
  • What specific jobs should our listeners take in order for Nick to be able to spend more money in useful ways to improve the world?
  • Should we expect the future to be better if the economy grows more quickly – or more slowly?

We also cover some more personal issues like:

  • Nick’s top book recommendations.
  • How he developed (what is in my view) exceptional judgement.
  • How he made his toughest career decisions.
  • Why he wants to see less dilettantism and more expertise in the effective altruism community.

Don’t miss it.

Continue reading →

#9 – Christine Peterson on the ’80s futurist movement and its lessons for today’s idealists

Take a trip to Silicon Valley in the 70s and 80s, when going to space sounded like a good way to get around environmental limits, people started cryogenically freezing themselves, and nanotechnology looked like it might revolutionise industry – or turn us all into grey goo.

In this episode of the 80,000 Hours Podcast Christine Peterson takes us back to her youth in the Bay Area, the ideas she encountered there, and what the dreamers she met did as they grew up. We also discuss how she came up with the term ‘open source software’ (and how she had to get someone else to propose it).

Today Christine helps runs the Foresight Institute, which fills a gap left by for-profit technology companies – predicting how new revolutionary technologies could go wrong, and ensuring we steer clear of the downsides.

We dive into:

  • Can technology ‘move fast and break things’ without eventually breaking the world? Would it be better for technology to advance more quickly, or more slowly?
  • Whether the poor security of computer systems poses a catastrophic risk for the world.
  • Could all our essential services be taken down at once? And if so, what can be done about it? Christine makes a radical proposal for solving the problem.
  • Will AIs designed for wide-scale automated hacking make computers more or less secure?
  • Would it be good to radically extend human lifespan? Is it sensible to cryogenically freeze yourself in the hope of being resurrected in the future?
  • Could atomically precise manufacturing (nanotechnology) really work? Why was it initially so controversial and why did people stop worrying about it?
  • Should people who try to do good in their careers work long hours and take low salaries? Or should they take care of themselves first of all?
  • How she thinks the the effective altruism community resembles the scene she was involved with when she was young, and where it might be going wrong.

Continue reading →

#8 – Lewis Bollard on ending factory farming as soon as possible

Every year tens of billions of animals are raised in terrible conditions in factory farms before being killed for human consumption. Despite the enormous scale of suffering this causes, the issue is largely neglected, with only about $50 million dollars spent each year tackling the problem globally.

Over the last two years Lewis Bollard – Project Officer for Farm Animal Welfare at Open Philanthropy – has conducted extensive research into the best ways to eliminate animal suffering in farms as soon as possible.

This has resulted in $30 million in grants, making Open Philanthropy one of the largest funders in the area.

Our conversation covers almost every approach being taken, which ones work, how individuals can best contribute through their careers, as well as:

  • How young people can set themselves up to contribute to scientific research into meat alternatives
  • How genetic manipulation of chickens has caused them to suffer much more than their ancestors, but could also be used to make them better off
  • Why Lewis is skeptical of vegan advocacy
  • Open Phil’s grants to improve animal welfare in China, India and South America
  • Why Lewis thinks insect farming would be worse than the status quo, and whether we should look for ‘humane’ insecticides
  • Why Lewis doubts that much can be done to tackle factory farming through legal advocacy or electoral politics
  • Which species of farm animals is best to focus on first
  • Whether fish and crustaceans are conscious, and if so what can be done for them
  • Many other issues

Listening to this episode is among the fastest ways to get up to speed on how animals are mistreated and the best ways to help them.

Continue reading →

#7 – Julia Galef on making humanity more rational, what EA does wrong, and why Twitter isn’t all bad

The scientific revolution in the 16th century was one of the biggest societal shifts in human history, driven by the discovery of new and better methods of figuring out who was right and who was wrong.

Julia Galef – a well-known writer and researcher focused on improving human judgment, especially about high stakes questions – believes that if we could develop new techniques to resolve disagreements, predict the future and make sound decisions together, we could again dramatically improve the world. We brought her in to talk about her ideas.

Julia has hosted the Rationally Speaking podcast since 2010, co-founded the Center for Applied Rationality in 2012, and is currently working for Open Philanthropy on an investigation of expert disagreements.

This interview complements a new detailed review of whether and how to follow Julia’s career path

We ended up speaking about a wide range of topics, including:

  • Her research on how people can have productive intellectual disagreements.
  • Why she once planned on becoming an urban designer.
  • Why she doubts people are more rational than 200 years ago.
  • What the effective altruism community is doing wrong.
  • What makes her a fan of Twitter (while I think it’s dystopian).
  • Whether more people should write books.
  • Whether it’s a good idea to run a podcast, and how she grew her audience.
  • Why saying you don’t believe X often won’t convince people you don’t.
  • Why she started a PhD in economics but then quit.
  • Whether she would recommend an unconventional ‘public intellectual’ career like her own.
  • Whether the incentives in the intelligence community actually support sound thinking.
  • Whether big institutions will actually pick up new tools for improving decision-making if they are developed.
  • How to start out pursuing a career in which you also try to enhance human judgement and foresight.

Continue reading →

#6 – Toby Ord on why the long-term future of humanity matters more than anything else, and what we should do about it

Of all the people whose well-being we should care about, only a small fraction are alive today. The rest are members of future generations who are yet to exist. Whether they’ll be born into a world that is flourishing or disintegrating – and indeed, whether they will ever be born at all – is in large part up to us. As such, the welfare of future generations should be our number one moral concern.

This conclusion holds true regardless of whether your moral framework is based on common sense, consequences, rules of ethical conduct, cooperating with others, virtuousness, keeping options open – or just a sense of wonder about the universe we find ourselves in.

That’s the view of Dr Toby Ord, a philosophy Fellow at the University of Oxford and co-founder of the effective altruism community. In this episode of the 80,000 Hours podcast Dr Ord makes the case that aiming for a positive long-term future is likely the best way to improve the world.

We then discuss common objections to long-termism, such as the idea that benefits to future generations are less valuable than those to people alive now, or that we can’t meaningfully benefit future generations beyond taking the usual steps to improve the present.

Later the conversation turns to how individuals can and have changed the course of history, what could go wrong and why, and whether plans to colonise Mars would actually put humanity in a safer position than it is today.

This episode goes deep into one of the most distinctive features of 80,000 Hours’ advice on doing good. It’s likely the most in-depth discussion of how we and the effective altruism community think about the long term future, and why we so often give it top priority.

book cover

If you prefer a book, Dr Toby Ord, an Oxford philosopher and advisor to 80,000 Hours, has recently published The Precipice: Existential Risk and the Future of Humanity which gives an overview of the moral importance of future generations, and what we can do to help them today.

We’ll mail you the book, for free

Join the 80,000 Hours newsletter and we’ll send you a free copy of the book.

We’ll also send you updates on our latest research, opportunities to work on existential risk, and news from the author.

If you’re already on our newsletter, email us at [email protected] to get a copy.

Continue reading →