#215 – Tom Davidson on how AI-enabled coups could allow a tiny group to seize power

Throughout history, technological revolutions have fundamentally shifted the balance of power in society. The Industrial Revolution created conditions where democracies could dominate for the first time — as nations needed educated, informed, and empowered citizens to deploy advanced technologies and remain competitive.

Unfortunately there’s every reason to think artificial general intelligence (AGI) will reverse that trend.

In a new paper published today, Tom Davidson — senior research fellow at the Forethought Centre for AI Strategy — argues that advanced AI systems will enable unprecedented power grabs by tiny groups of people, primarily by removing the need for other human beings to participate.

Come work with us on the 80,000 Hours podcast team! We’re accepting expressions of interest for the new host and chief of staff until May 6 in order to deliver as much incredibly insightful AGI-related content as we can. Learn more about our shift in strategic direction and apply soon!

When a country’s leaders no longer need citizens for economic production, or to serve in the military, there’s much less need to share power with them. “Over the broad span of history, democracy is more the exception than the rule,” Tom points out. “With AI, it will no longer be important to a country’s competitiveness to have an empowered and healthy citizenship.”

Citizens in established democracies are not typically that concerned about coups. We doubt anyone will try, and if they do, we expect human soldiers to refuse to join in. Unfortunately, the AI-controlled military systems of the future will lack those inhibitions. As Tom lays out, “Human armies today are very reluctant to fire on their civilians. If we get instruction-following AIs, then those military systems will just fire.”

Why would AI systems follow the instructions of a would-be tyrant? One answer is that, as militaries worldwide race to incorporate AI to remain competitive, they risk leaving the door open for exploitation by malicious actors in a few ways:

  1. AI systems could be programmed to simply follow orders from the top of the chain of command, without any checks on that power — potentially handing total power indefinitely to any leader willing to abuse that authority.
  2. Systems could contain “secret loyalties” inserted during development that activate at critical moments, as demonstrated in Anthropic’s recent paper on “sleeper agents”.
  3. Superior cyber capabilities could enable small groups to hack into and take full control of AI-operated military infrastructure.

It’s also possible that the companies with the most advanced AI, if it conveyed a significant enough advantage over competitors, could quickly develop armed forces sufficient to overthrow an incumbent regime. History suggests that as few as 10,000 obedient military drones could be sufficient to kill competitors, take control of key centres of power, and make your success fait accompli.

Without active effort spent mitigating risks like these, it’s reasonable to fear that AI systems will destabilise the current equilibrium that enables the broad distribution of power we see in democratic nations.

In this episode, host Rob Wiblin and Tom discuss new research on the question of whether AI-enabled coups are likely, and what we can do about it if they are, as well as:

  • Whether preventing coups and preventing ‘rogue AI’ require opposite interventions, leaving us in a bind
  • Whether open sourcing AI weights could be helpful, rather than harmful, for advancing AI safely
  • Why risks of AI-enabled coups have been relatively neglected in AI safety discussions
  • How persuasive AGI will really be
  • How many years we have before these risks become acute
  • The minimum number of military robots needed to stage a coup

This episode was originally recorded on January 20, 2025.

Video editing: Simon Monsour
Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
Camera operator: Jeremy Chevillotte
Transcriptions and web: Katy Moore

Continue reading →

Expression of interest: Shortform Video Editing Contractor

Video Cartoon

Help make spectacular videos that reach a huge audience.

We’re looking for someone to contract as a video editor, who can quickly learn our style and make our videos successful on shortform video platforms. We want these videos to start changing and informing the conversation about transformative AI and AGI.

Why this role?

In 2025, 80,000 Hours is planning to focus especially on helping explain why and how our audience can help society safely navigate a transition to a world with transformative AI. Right now not nearly enough people are talking about these ideas and their implications.

A great video program could change this. Time spent on the internet is increasingly spent watching video, and for many people in our target audience, video is the main way that they both find entertainment and learn about topics that matter to them.

To get our video program off the ground, we need great editors who understand our style and vision and can work quickly and to a high standard.

Responsibilities

  • Be able to work at least 10 hours a week
  • Be able to turn around drafts of edited shortform videos in 24-48 hours
  • Take feedback well
  • Learn our style and adapt to it quickly

About you

We’re looking for someone who ideally has:

  • Experience making shortform videos
  • Experience with Capcut, Descript or similar
  • Good taste in shortform video
  • Knowledge of the current trends and what succeeds in shortform video
  • The ability to work quickly and take feedback well
  • Familiarity with AI Safety

If you don’t have experience here but think you’d be a great fit,

Continue reading →

    Should you quit your job — and work on risks from advanced AI?

    Should you quit your job — and work on risks from advanced AI instead?

    There’s a real chance that by 2030 AI systems could cause explosive technological and economic change, leading to risks like war between US and China, concentration of power in a small minority, and even loss of control and human disempowerment.

    80,000 Hours has ranked AI as the most pressing global problem since 2016, but now we’re doubling down.

    We don’t think everyone should work on it, but we’d encourage everyone reading to seriously consider switching their career to one helping advanced AI go well.

    Here’s why.

    (If you already know you’d like to switch, apply to speak to us one-on-one.)

    1) World-changing AI systems could come much sooner than expected

    Last newsletter I explained why there’s a significant chance that AI could contribute to scientific research or automate many jobs by 2030. Forecasters and experts widely agree the probability is much higher than it was even a year ago.

    2) The impact on society would be explosive

    People say AI will be transformative, but few really get just how wild it could be.

    Here are three types of explosive impact we might see, which are now all supported by substantially more empirical research than before:

    Continue reading →

    Bonus: Guilt, imposter syndrome, and doing good: 16 past guests share their mental health journeys

    What happens when your desire to do good starts to undermine your own wellbeing?

    Over the years, we’ve heard from therapists, charity directors, researchers, psychologists, and career advisors — all wrestling with how to do good without falling apart. Today’s episode brings together insights from 16 past guests on the emotional and psychological costs of pursuing a high-impact career to improve the world — and how to best navigate the all-too-common guilt, burnout, perfectionism, and imposter syndrome along the way.

    You’ll hear from:

    • 80,000 Hours’ former CEO on managing anxiety, self-doubt, and a chronic sense of falling short (from episode #100)
    • Randy Nesse on why we evolved to be anxious and depressed (episode #179)
    • Hannah Boettcher on how ‘optimisation framing’ can quietly distort our sense of self-worth (from our 80k After Hours feed)
    • Luisa Rodriguez on grieving the gap between who you are and who you wish you were (from our 80k After Hours feed)
    • Cameron Meyer Shorb on how guilt and shame became his biggest source of suffering — and what helped (episode #210)
    • Tim LeBon on the trap of moral perfectionism, and why we should strive for excellence instead (episode #149)
    • Cal Newport on why we need to make time to be alone with our thoughts (episode #106)
    • Michelle Hutchinson and Habiba Islam on when to prioritise wellbeing over impact (episode #122)
    • Sarah Eustis-Guthrie on the emotional weight of founding a charity (episode #207)
    • Hannah Ritchie on feeling like an imposter, even after writing a book and giving a TED Talk (episode #160)
    • Will MacAskill on why he’s five to 10 times happier than he used to be after learning to work in a way that’s genuinely sustainable (episode #130)
    • Ajeya Cotra on handling the pressure of high-stakes research (episode #90)
    • Christian Ruhl on pursuing a high-impact career while managing a stutter (from our 80k After Hours feed)
    • Leah Garcés on insisting on self-care when witnessing trauma regularly (episode #99)
    • Kelsey Piper on recognising that you’re not alone in your struggles (episode #53)

    And if you’re dealing with your own mental health concerns, here are some resources that might help:

    Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
    Content editing: Katy Moore and Milo McGuire
    Transcriptions and web: Katy Moore

    Continue reading →

    #214 – Buck Shlegeris on controlling AI that wants to take over – so we can use it anyway

    Most AI safety conversations centre on alignment: ensuring AI systems share our values and goals. But despite progress, we’re unlikely to know we’ve solved the problem before the arrival of human-level and superhuman systems in as little as three years.

    So some are developing a backup plan to safely deploy models we fear are actively scheming to harm us — so-called “AI control.” While this may sound mad, given the reluctance of AI companies to delay deploying anything they train, not developing such techniques is probably even crazier.

    Today’s guest — Buck Shlegeris, CEO of Redwood Research — has spent the last few years developing control mechanisms, and for human-level systems they’re more plausible than you might think. He argues that given companies’ unwillingness to incur large costs for security, accepting the possibility of misalignment and designing robust safeguards might be one of our best remaining options.

    Buck asks us to picture a scenario where, in the relatively near future, AI companies are employing 100,000 AI systems running 16 times faster than humans to automate AI research itself. These systems would need dangerous permissions: the ability to run experiments, access model weights, and push code changes. As a result, a misaligned system could attempt to hack the data centre, exfiltrate weights, or sabotage research. In such a world, misalignment among these AIs could be very dangerous.

    But in the absence of a method for reliably aligning frontier AIs, Buck argues for implementing practical safeguards to prevent catastrophic outcomes. His team has been developing and testing a range of straightforward, cheap techniques to detect and prevent risky behaviour by AIs — such as auditing AI actions with dumber but trusted models, replacing suspicious actions, and asking questions repeatedly to catch randomised attempts at deception.

    Most importantly, these methods are designed to be cheap and shovel-ready. AI control focuses on harm reduction using practical techniques — techniques that don’t require new, fundamental breakthroughs before companies could reasonably implement them, and that don’t ask us to forgo the benefits of deploying AI.

    As Buck puts it:

    Five years ago I thought of misalignment risk from AIs as a really hard problem that you’d need some really galaxy-brained fundamental insights to resolve. Whereas now, to me the situation feels a lot more like we just really know a list of 40 things where, if you did them — none of which seem that hard — you’d probably be able to not have very much of your problem.

    Of course, even if Buck is right, we still need to do those 40 things — which he points out we’re not on track for. And AI control agendas have their limitations: they aren’t likely to work once AI systems are much more capable than humans, since greatly superhuman AIs can probably work around whatever limitations we impose.

    Still, AI control agendas seem to be gaining traction within AI safety. Buck and host Rob Wiblin discuss all of the above, plus:

    • Why he’s more worried about AI hacking its own data centre than escaping
    • What to do about “chronic harm,” where AI systems subtly underperform or sabotage important work like alignment research
    • Why he might want to use a model he thought could be conspiring against him
    • Why he would feel safer if he caught an AI attempting to escape
    • Why many control techniques would be relatively inexpensive
    • How to use an untrusted model to monitor another untrusted model
    • What the minimum viable intervention in a “lazy” AI company might look like
    • How even small teams of safety-focused staff within AI labs could matter
    • The moral considerations around controlling potentially conscious AI systems, and whether it’s justified

    This episode was originally recorded on February 21, 2025.

    Video: Simon Monsour and Luke Monsour
    Audio engineering: Ben Cordell, Milo McGuire, and Dominic Armstrong
    Transcriptions and web: Katy Moore

    Continue reading →

    To understand AI, you should use it. Here’s how to get started.

    To truly understand what AI can do — and what is coming soon — you should make regular use of the latest AI services.

    This article will help you learn how to do this.

    1. You have a team of experts

    Think of ChatGPT as a team of experts, ready to assist you.

    Whenever you have a question, task, or problem, consider who you’d want in the room if you could talk to — or hire — anyone in the world.

    Here are some of the experts I work with on a regular basis:

    • Programmer: write, debug, and explain code. I’m an experienced software developer, yet more than 95% of the code I ship is now written by AI.
    • Manager: plan and debug my days, weeks, and months. There’s an example transcript in section four below.
    • Product designer: ideate, create prototypes, design and critique user interfaces, and analyse user interview transcripts.
    • Writer and editor: give feedback on my writing, rewrite things, write entire drafts based on a rough dictation, proofread, and fix markdown.
    • Data analyst: write SQL queries, analyse spreadsheets, create graphs, and infographics.
    • Language tutor: do written exercises and have spoken conversations in French and Icelandic.
    • Intern: extract data from text, get key quotes from an interview, resize a folder of images, merge a spreadsheet, and write an SQL query.
    • Handyman: answer household questions,

    Continue reading →

    Gradual disempowerment

    The proliferation of advanced AI systems may lead to the gradual disempowerment of humanity, even if efforts to prevent them from becoming power-seeking or scheming are successful. Humanity may be incentivised to hand over increasing amounts of control to AIs, giving them power over the economy, politics, culture, and more. Over time, humanity’s interests may be sidelined and our control over the future undermined, potentially constituting an existential catastrophe.

    There’s disagreement over how serious a problem this is and how it relates to other concerns about AI alignment. It’s also unclear, if this is a genuine risk, what we could do about it. But we think it’s potentially very important, and more people should work on clarifying the issue and perhaps figuring out how to address it.

    Continue reading →

    We’re shifting our strategic approach to focus more on AGI

    Why we’re updating our strategic direction

    Since 2016, we’ve ranked ‘risks from artificial intelligence’ as our top pressing problem. Whilst we’ve provided research and support on how to work on reducing AI risks since that point (and before!), we’ve put in varying amounts of investment over time and between programmes.

    We think we should consolidate our effort and focus because:

    • We think that AGI by 2030 is plausible — and this is much sooner than most of us would have predicted five years ago. This is far from guaranteed, but we think the view is compelling based on analysis of the current flow of inputs into AI development and the speed of recent AI progress. You can read about this argument in far more detail in this new article.
    • We are in a window of opportunity to influence AGI, before laws and norms are set in place.
    • 80,000 Hours has an opportunity to help more people take advantage of this window. We want our strategy to be responsive to changing events in the world, and we think that prioritising reducing risks from AI is probably the best way to achieve our high-level, cause-impartial goal of doing the most good for others over the long term by helping people have high-impact careers. We expect the landscape to move faster in the coming years, so we’ll need a faster moving culture to keep up.

    Continue reading →

    Bonus: 15 expert takes on infosec in the age of AI

    What happens when a USB cable can secretly control your system? Are we hurtling toward a security nightmare as critical infrastructure connects to the internet? Is it possible to secure AI model weights from sophisticated attackers? And could AI might actually make computer security better rather than worse?

    With AI security concerns becoming increasingly urgent, we bring you insights from 15 top experts across information security, AI safety, and governance, examining the challenges of protecting our most powerful AI models and digital infrastructure — including a sneak peek from an episode that hasn’t yet been released with Tom Davidson, where he explains how we should be more worried about “secret loyalties” in AI agents.

    You’ll hear:

    • Holden Karnofsky on why every good future relies on strong infosec, and how hard it’s been to hire security experts (from episode #158)
    • Tantum Collins on why infosec might be the rare issue everyone agrees on (episode #166)
    • Nick Joseph on whether AI companies can develop frontier models safely with the current state of information security (episode #197)
    • Sella Nevo on why AI model weights are so valuable to steal, the weaknesses of air-gapped networks, and the risks of USBs (episode #195)
    • Kevin Esvelt on what cryptographers can teach biosecurity experts (episode #164)
    • Lennart Heim on on Rob’s computer security nightmares (episode #155)
    • Zvi Mowshowitz on the insane lack of security mindset at some AI companies (episode #184)
    • Nova DasSarma on the best current defences against well-funded adversaries, politically motivated cyberattacks, and exciting progress in infosecurity (episode #132)
    • Bruce Schneier on whether AI could eliminate software bugs for good, and why it’s bad to hook everything up to the internet (episode #64)
    • Nita Farahany on the dystopian risks of hacked neurotech (episode #174)
    • Vitalik Buterin on how cybersecurity is the key to defence-dominant futures (episode #194)
    • Nathan Labenz on how even internal teams at AI companies may not know what they’re building (episode #176)
    • Allan Dafoe on backdooring your own AI to prevent theft (episode #212)
    • Tom Davidson on how dangerous “secret loyalties” in AI models could be (episode to be released!)
    • Carl Shulman on the challenge of trusting foreign AI models (episode #191, part 2)
    • Plus lots of concrete advice on how to get into this field and find your fit

    Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
    Content editing: Katy Moore and Milo McGuire
    Transcriptions and web: Katy Moore

    Continue reading →

    Preventing an AI-related catastrophe

    We expect there will be substantial progress in AI in the coming years, potentially even to the point where machines come to outperform humans in many, if not all, tasks. This could have enormous benefits, helping to solve currently intractable global problems, but could also pose severe risks. These risks could arise accidentally (for example, if we don’t find technical solutions to concerns about the safety of AI systems), or deliberately (for example, if AI systems worsen geopolitical conflict). We think more work needs to be done to reduce these risks.

    Some of these risks from advanced AI could be existential — meaning they could cause human extinction, or an equally permanent and severe disempowerment of humanity.2 There have not yet been any satisfying answers to concerns — discussed below — about how this rapidly approaching, transformative technology can be safely developed and integrated into our society. Finding answers to these concerns is neglected and may well be tractable. We estimated that there were around 400 people worldwide working directly on this in 2022, though we believe that number has grown.3 As a result, the possibility of AI-related catastrophe may be the world’s most pressing problem — and the best thing to work on for those who are well-placed to contribute.

    Promising options for working on this problem include technical research on how to create safe AI systems, strategy research into the particular risks AI might pose, and policy research into ways in which companies and governments could mitigate these risks. As policy approaches continue to be developed and refined, we need people to put them in place and implement them. There are also many opportunities to have a big impact in a variety of complementary roles, such as operations management, journalism, earning to give, and more — some of which we list below.

    Continue reading →

    The case for AGI by 2030

    In recent months, the CEOs of leading AI companies have grown increasingly confident about rapid progress:

    • OpenAI’s Sam Altman: Shifted from saying in November “the rate of progress continues” to declaring in January “we are now confident we know how to build AGI”
    • Anthropic’s Dario Amodei: Stated in January “I’m more confident than I’ve ever been that we’re close to powerful capabilities… in the next 2-3 years”
    • Google DeepMind’s Demis Hassabis: Changed from “as soon as 10 years” in autumn to “probably three to five years away” by January.

    What explains the shift? Is it just hype? Or could we really have Artificial General Intelligence (AGI) by 2030?

    In this article, I look at what’s driven recent progress, estimate how far those drivers can continue, and explain why they’re likely to continue for at least four more years.

    In particular, while in 2024 progress in LLM chatbots seemed to slow, a new approach started to work: teaching the models to reason using reinforcement learning.

    In just a year, this let them surpass human PhDs at answering difficult scientific reasoning questions, and achieve expert-level performance on one-hour coding tasks.

    We don’t know how capable AI will become, but extrapolating the recent rate of progress suggests that, by 2028, we could reach AI models with beyond-human reasoning abilities, expert-level knowledge in every domain,

    Continue reading →

    Shrinking AGI timelines: a review of expert forecasts

    As a non-expert, it would be great if there were experts who could tell us when we should expect artificial general intelligence (AGI) to arrive.

    Unfortunately, there aren’t.

    There are only different groups of experts with different weaknesses.

    This article is an overview of what five different types of experts say about when we’ll reach AGI, and what we can learn from them (that feeds into my full article on forecasting AI).

    In short:

    • Every group shortened their estimates in recent years.
    • AGI before 2030 seems within the range of expert opinion, even if many disagree.
    • None of the forecasts seem especially reliable, so they neither rule in nor rule out AGI arriving soon.

    Graph of forecasts of years to AGIIn four years, the mean estimate on Metaculus for when AGI will be developed has plummeted from 50 years to five years. There are problems with the definition used, but the graph reflects a broader pattern of declining estimates.

    Here’s an overview of the five groups:

    AI experts
    1. Leaders of AI companies

    The leaders of AI companies are saying that AGI arrives in 2–5 years, and appear to have recently shortened their estimates.

    This is easy to dismiss. This group is obviously selected to be bullish on AI and wants to hype their own work and raise funding.

    However, I don’t think their views should be totally discounted.

    Continue reading →

    Great power war

    Economic growth and technological progress have bolstered the arsenals of the world’s most powerful countries. That means the next war between them could be far worse than World War II, the deadliest conflict humanity has yet experienced.

    Could such a war actually occur? We can’t rule out the possibility. Technical accidents or diplomatic misunderstandings could spark a conflict that quickly escalates. International tension could cause leaders to decide they’re better off fighting than negotiating. And rapidly advancing AI may change the way wars are fought and lead to dangerous geopolitical upheaval.

    It seems hard to make progress on this problem. It’s also less neglected than some of the problems that we think are most pressing. There are certain issues, like making nuclear weapons or military artificial intelligence systems safer, which seem promising — although it may be more impactful to work directly on reducing risks from AI, bioweapons or nuclear weapons directly. You might also be able to reduce the chances of misunderstandings and miscalculations by developing expertise in one of the most important bilateral relationships (such as that between the United States and China).

    Finally, by making conflict less likely, reducing competitive pressure, and improving international cooperation, you might be helping to reduce other risks, like the chance of future pandemics.

    Continue reading →

    Expression of Interest: Podcast Host

    Help shape the most important conversation of our time

    We’re looking for someone exceptional to join 80,000 Hours as a Podcast Host, helping us build a media platform that properly covers recursively self-improving AGI and its implications.

    Why this matters

    Frontier AI systems are developing rapidly, with capabilities progressing much faster than most expected even a few years ago.

    If development continues at its current pace we could see an ‘intelligence explosion’ in which AI takes over AI research and rapidly self-improves, starting in as little as 2-7 years.

    This is the defining issue of our time and decisions made in the next decade could shape humanity’s long-term future in profound ways.

    Yet the public conversation around AGI remains completely inadequate. Mainstream sources neglect the issue entirely for the most part. When AGI is discussed, it’s often covered superficially and fails in one of various ways:

    • naive optimism and hype
    • reflexive dismissiveness
    • focusing only on immediate issues like job displacement or copyright violations
    • uncritically accepting narratives from actors with strong financial interests in the industry.

    The extremely profound implications of developing AI, then AGI, and then superintelligence are basically ignored.

    We believe there’s a critical opportunity to fill this gap by scaling up our podcast from occasional in-depth interviews to a more comprehensive media platform with multiple weekly pieces covering AGI-related developments, risks, governance, and alignment efforts in a substantive way.

    Continue reading →

      Expression of Interest: Podcast Chief of Staff

      Help shape the most important conversation of our time

      We’re looking for someone exceptional to join 80,000 Hours as Chief of Staff for our podcast team, helping us build a media platform that properly covers recursively self-improving AGI and its implications.

      Why this matters

      Frontier AI systems are developing rapidly, with capabilities progressing much faster than most expected even a few years ago.

      If development continues at its current pace we could see an ‘intelligence explosion’ in which AI takes over AI research and rapidly self-improves, starting in as little as 2-7 years.

      This is the defining issue of our time and decisions made in the next decade could shape humanity’s long-term future in profound ways.

      Yet the public conversation around AGI remains completely inadequate. Mainstream sources neglect the issue entirely for the most part. When AGI is discussed, it’s often covered superficially and fails in one of various ways:

      • naive optimism and hype
      • reflexive dismissiveness
      • focusing only on immediate issues like job displacement or copyright violations
      • uncritically accepting narratives from actors with strong financial interests in the industry.

      The extremely profound implications of developing AI, then AGI, and then superintelligence are basically ignored.

      We believe there’s a critical opportunity to fill this gap by scaling up our podcast from occasional in-depth interviews to a more comprehensive media platform with multiple weekly pieces covering AGI-related developments, risks, governance,

      Continue reading →

        Understanding trends in our AI job postings

        This week, let’s review key trends in the jobs we’ve found that may help mitigate AI risk, including:

        • Growth in the number of postings in the field
        • The types of organisations that are hiring
        • The most in-demand skills
        • The experience level required for these roles

        We’ve ranked catastrophic risks from AI as the world’s most pressing problem since 2016, but it’s only in the last few years that the topic has really hit the mainstream.

        As AI has advanced rapidly and the risks have become more salient, we’ve seen many more jobs available to help mitigate the dangers.

        The number of AI-related jobs we posted on our job board rose throughout 2023 and then plateaued in 2024. But in January 2025, we posted the most AI-relevant jobs yet!

        Total AI roles

        In 2023, we posted an average of 63 AI-related roles per month. In 2024, the average rose to 105 — a 67% increase.

        Over this time, nonprofit jobs have been the most common, though they were briefly overtaken by both company and government jobs in early 2024.

        By org type

        This trend could reflect our vantage point. As a nonprofit that works closely with other nonprofits, we may be best positioned to find and assess high-impact roles in this sector while potentially missing other great roles in sectors more opaque to us.

        Continue reading →

        Why AGI could be here soon and what you can do about it: a primer

        I’m writing a new guide to careers to help artificial general intelligence (AGI) go well. Here’s a summary of the bottom lines that’ll be in the guide as it stands. Stay tuned to hear our full reasoning and updates as our views evolve.

        In short:

        • The chance of an AGI-driven technological explosion before 2030 — creating one of the most pivotal periods in history — is high enough to act on.
        • Since this transition poses major risks, and relatively few people are focused on navigating them, if you might be able to do something that helps, that’s likely the highest-impact thing you can do.
        • There are now many organisations with hundreds of jobs that could concretely help (many of which are non technical).
        • If you already have some experience (e.g. age 25+), typically the best path is to spend 20–200 hours reading about AI and meeting people in the field, then applying to jobs at organisations you’re aligned with — this both sets you up to have an impact relatively soon and advance in the field. If you can’t get a job right away, figure out the minimum additional skills, connections, and credentials you’d need, then get those. Alternatively, contribute from your existing position by donating, spreading clear thinking about the issue, or getting ready to switch when future opportunities arise.
        • If you’re at the start of your career (or need to reskill),

        Continue reading →

        #213 – Will MacAskill on AI causing a "century in a decade" — and how we're completely unprepared

        The 20th century saw unprecedented change: nuclear weapons, satellites, the rise and fall of communism, third-wave feminism, the internet, postmodernism, game theory, genetic engineering, the Big Bang theory, quantum mechanics, birth control, and more. Now imagine all of it compressed into just 10 years.

        That’s the future Will MacAskill — philosopher, founding figure of effective altruism, and now researcher at the Forethought Centre for AI Strategy — argues we need to prepare for in his new paper “Preparing for the intelligence explosion.” Not in the distant future, but probably in three to seven years.

        The reason: AI systems are rapidly approaching human-level capability in scientific research and intellectual tasks. Once AI exceeds human abilities in AI research itself, we’ll enter a recursive self-improvement cycle — creating wildly more capable systems. Soon after, by improving algorithms and manufacturing chips, we’ll deploy millions, then billions, then trillions of superhuman AI scientists working 24/7 without human limitations. These systems will collaborate across disciplines, build on each discovery instantly, and conduct experiments at unprecedented scale and speed — compressing a century of scientific progress into mere years.

        Will compares the resulting situation to a mediaeval king suddenly needing to upgrade from bows and arrows to nuclear weapons to deal with an ideological threat from a country he’s never heard of, while simultaneously grappling with learning that he descended from monkeys and his god doesn’t exist.

        What makes this acceleration perilous is that while technology can speed up almost arbitrarily, human institutions and decision-making are much more fixed.

        Consider the case of nuclear weapons: in this compressed timeline, there would have been just a three-month gap between the Manhattan Project’s start and the Hiroshima bombing, and the Cuban Missile Crisis would have lasted just over a day.

        Robert Kennedy, Sr., who helped navigate the actual Cuban Missile Crisis, once remarked that if they’d had to make decisions on a much more accelerated timeline — like 24 hours rather than 13 days — they would likely have taken much more aggressive, much riskier actions.

        So there’s reason to worry about our own capacity to make wise choices. And in “Preparing for the intelligence explosion,” Will lays out 10 “grand challenges” we’ll need to quickly navigate to successfully avoid things going wrong during this period.

        Will’s thinking has evolved a lot since his last appearance on the show. While he was previously sceptical about whether we live at a “hinge of history,” he now believes we’re entering one of the most critical periods for humanity ever — with decisions made in the next few years potentially determining outcomes millions of years into the future.

        But Will also sees reasons for optimism. The very AI systems causing this acceleration could be deployed to help us navigate it — if we use them wisely. And while AI safety researchers rightly focus on preventing AI systems from going rogue, Will argues we should equally attend to ensuring the futures we deliberately build are truly worth living in.

        In this wide-ranging conversation with host Rob Wiblin, Will maps out the challenges we’d face in this potential “intelligence explosion” future, and what we might do to prepare. They discuss:

        • Why leading AI safety researchers now think there’s dramatically less time before AI is transformative than they’d previously thought
        • The three different types of intelligence explosions that occur in order
        • Will’s list of resulting grand challenges — including destructive technologies, space governance, concentration of power, and digital rights
        • How to prevent ourselves from accidentally “locking in” mediocre futures for all eternity
        • Ways AI could radically improve human coordination and decision making
        • Why we should aim for truly flourishing futures, not just avoiding extinction

        This episode was originally recorded on February 7, 2025.

        Video editing: Simon Monsour
        Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
        Camera operator: Jeremy Chevillotte
        Transcriptions and web: Katy Moore

        Continue reading →

        Emergency pod: Judge plants a legal time bomb under OpenAI (with Rose Chan Loui)

        When OpenAI announced plans to convert from nonprofit to for-profit control last October, it likely didn’t anticipate the legal labyrinth it now faces. A recent court order in Elon Musk’s lawsuit against the company suggests OpenAI’s restructuring faces serious legal threats, which will complicate its efforts to raise tens of billions in investment.

        As nonprofit legal expert Rose Chan Loui explains, the court order set up multiple pathways for OpenAI’s conversion to be challenged. Though Judge Yvonne Gonzalez Rogers denied Musk’s request to block the conversion before a trial, she expedited proceedings to the fall so the case could be heard before it’s likely to go ahead. (See Rob’s brief summary of developments in the case.)

        And if Musk’s donations to OpenAI are enough to give him the right to bring a case, Rogers sounded very sympathetic to his objections to the OpenAI foundation selling the company, benefiting the founders who forswore “any intent to use OpenAI as a vehicle to enrich themselves.”

        But that’s just one of multiple threats. The attorneys general (AGs) in California and Delaware both have standing to object to the conversion on the grounds that it is contrary to the foundation’s charitable purpose and therefore wrongs the public — which was promised all the charitable assets would be used to develop AI that benefits all of humanity, not to win a commercial race. Some, including Rose, suspect the court order was written as a signal to those AGs to take action.

        And, as she explains, if the AGs remain silent, the court itself, seeing that the public interest isn’t being represented, could appoint a “special interest party” to take on the case in their place.

        This places the OpenAI foundation board in a bind: proceeding with the restructuring despite this legal cloud could expose them to the risk of being sued for a gross breach of their fiduciary duty to the public. The board is made up of respectable people who didn’t sign up for that.

        And of course it would cause chaos for the company if all of OpenAI’s fundraising and governance plans were brought to a screeching halt by a federal court judgment landing at the eleventh hour.

        Host Rob Wiblin and Rose Chan Loui discuss all of the above as well as what justification the OpenAI foundation could offer for giving up control of the company despite its charitable purpose, and how the board might adjust their plans to make the for-profit switch more legally palatable.

        This episode was originally recorded on March 6, 2025.

        Video editing: Simon Monsour
        Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
        Transcriptions: Katy Moore

        Continue reading →

        Ask a career advisor: how to compete with AI models in the job market and whether to stay in the US government

        This week, we answer more of our newsletter subscriber’s career questions!
        I’ll tackle:

        Laura: This is a difficult question, but I think it’s important to ask. I certainly don’t have all the answers, because the future is uncertain — perhaps especially now.

        That said, I do think there are sensible steps to take to navigate the trajectory of AI and its global impact

        First, let’s address the immediate concern about the job market. While AI is already changing the landscape of work, entry-level jobs haven’t disappeared. The 80,000 Hours job board has hundreds of entry-level openings, and other job sites will have many more. I expect there will still be plenty of jobs available when you graduate, even as AI continues to advance quickly.

        As technology becomes better at the tasks that human workers perform, there are typically significant delays before it is widely adopted. Still, the economy will likely change as AI gets more powerful, and that may start happening in just a few years. So, my advice is to stay ahead of the adoption curve. Master prompting and evaluation techniques to ensure the quality of outputs. Learn to set up AI task automation systems. Experiment with new tools and applications.

        Continue reading →