Preventing an AI-related catastrophe

We expect there will be substantial progress in AI in the coming years, potentially even to the point where machines come to outperform humans in many, if not all, tasks. This could have enormous benefits, helping to solve currently intractable global problems, but could also pose severe risks. These risks could arise accidentally (for example, if we don’t find technical solutions to concerns about the safety of AI systems), or deliberately (for example, if AI systems worsen geopolitical conflict). We think more work needs to be done to reduce these risks.

Some of these risks from advanced AI could be existential — meaning they could cause human extinction, or an equally permanent and severe disempowerment of humanity.1 There have not yet been any satisfying answers to concerns — discussed below — about how this rapidly approaching, transformative technology can be safely developed and integrated into our society. Finding answers to these concerns is neglected and may well be tractable. We estimated that there were around 400 people worldwide working directly on this in 2022, though we believe that number has grown.2 As a result, the possibility of AI-related catastrophe may be the world’s most pressing problem — and the best thing to work on for those who are well-placed to contribute.

Promising options for working on this problem include technical research on how to create safe AI systems, strategy research into the particular risks AI might pose, and policy research into ways in which companies and governments could mitigate these risks. As policy approaches continue to be developed and refined, we need people to put them in place and implement them. There are also many opportunities to have a big impact in a variety of complementary roles, such as operations management, journalism, earning to give, and more — some of which we list below.

Continue reading →

Why AGI could be here by 2028

In recent months, the CEOs of leading AI companies have grown increasingly confident about rapid progress:

  • OpenAI’s CEO shifted from saying in November 2024 he expects “the rate of progress we’ve made over the last three years continues” to declaring just one month later “we are now confident we know how to build AGI as we have traditionally understood it.”
  • Anthropic’s CEO stated in January 2025: “I’m more confident than I’ve ever been that we’re close to powerful capabilities…A country of geniuses in a data centre…that’s what I think we’re quite likely to get in the next 2-3 years.”
  • Even Google DeepMind’s more cautious CEO shifted from saying “as soon as 10 years” in the autumn, to “I think we’re probably three to five years away” by January.

What explains the shift? Could they be right, or is this just hype? Could we really have Artificial General Intelligence (AGI) by 2028?

In this article, I look at what’s driven recent progress, estimate how far those drivers can continue, and explain why they’re likely to continue for at least four more years.

In particular, while in 2024 progress in LLM chatbots seemed to slow, a new approach started to work: teaching the models to reason using reinforcement learning. In just a year, this let them surpass human PhDs at answering difficult scientific reasoning questions and achieve expert-level performance on one-hour coding tasks.

Continue reading →

When do experts expect AGI to arrive?

I’ve argued elsewhere that it’s plausible AGI arrives before 2030. That’s a big claim.

As a non-expert, it would be great if there were experts who could tell us what to think.

Unfortunately, there aren’t.

There are only different groups with different weaknesses. Here’s an overview.

1. Leaders of AI companies

The leaders of AI companies are saying that AGI arrives in 2–5 years.

This is easy to dismiss. This group is obviously selected to be bullish on AI and wants to hype their own work and raise funding.

However, I don’t think their views should be totally discounted. They’re the people with the most visibility into the capabilities of next-generation systems.

And they’ve also been among the most right about recent progress, even if they’ve been too optimistic.

Most likely, progress will be slower than they expect, but maybe only by a few years.

2. AI researchers in general

One way to reduce selection effects is to look at a wider group of AI researchers than those working on AGI directly, including in academia. This is what Katja Grace did with a survey of thousands of recent AI publication authors.

The survey asked for forecasts of “high-level machine intelligence,” defined as when AI can accomplish every task better or more cheaply than humans. The median estimate was a 25% chance in the early 2030s and 50% by 2047 — with some giving answers in the next few years and others hundreds of years in the future.

Continue reading →

    Great power war

    Economic growth and technological progress have bolstered the arsenals of the world’s most powerful countries. That means the next war between them could be far worse than World War II, the deadliest conflict humanity has yet experienced.

    Could such a war actually occur? We can’t rule out the possibility. Technical accidents or diplomatic misunderstandings could spark a conflict that quickly escalates. International tension could cause leaders to decide they’re better off fighting than negotiating. And rapidly advancing AI may change the way wars are fought and lead to dangerous geopolitical upheaval.

    It seems hard to make progress on this problem. It’s also less neglected than some of the problems that we think are most pressing. There are certain issues, like making nuclear weapons or military artificial intelligence systems safer, which seem promising — although it may be more impactful to work directly on reducing risks from AI, bioweapons or nuclear weapons directly. You might also be able to reduce the chances of misunderstandings and miscalculations by developing expertise in one of the most important bilateral relationships (such as that between the United States and China).

    Finally, by making conflict less likely, reducing competitive pressure, and improving international cooperation, you might be helping to reduce other risks, like the chance of future pandemics.

    Continue reading →

    Expression of Interest: Podcast Host

    Help shape the most important conversation of our time

    We’re looking for someone exceptional to join 80,000 Hours as a Podcast Host, helping us build a media platform that properly covers recursively self-improving AGI and its implications.

    Why this matters

    Frontier AI systems are developing rapidly, with capabilities progressing much faster than most expected even a few years ago.

    If development continues at its current pace we could see an ‘intelligence explosion’ in which AI takes over AI research and rapidly self-improves, starting in as little as 2-7 years.

    This is the defining issue of our time and decisions made in the next decade could shape humanity’s long-term future in profound ways.

    Yet the public conversation around AGI remains completely inadequate. Mainstream sources neglect the issue entirely for the most part. When AGI is discussed, it’s often covered superficially and fails in one of various ways:

    • naive optimism and hype
    • reflexive dismissiveness
    • focusing only on immediate issues like job displacement or copyright violations
    • uncritically accepting narratives from actors with strong financial interests in the industry.

    The extremely profound implications of developing AI, then AGI, and then superintelligence are basically ignored.

    We believe there’s a critical opportunity to fill this gap by scaling up our podcast from occasional in-depth interviews to a more comprehensive media platform with multiple weekly pieces covering AGI-related developments, risks, governance, and alignment efforts in a substantive way.

    Continue reading →

      Expression of Interest: Podcast Chief of Staff

      Help shape the most important conversation of our time

      We’re looking for someone exceptional to join 80,000 Hours as Chief of Staff for our podcast team, helping us build a media platform that properly covers recursively self-improving AGI and its implications.

      Why this matters

      Frontier AI systems are developing rapidly, with capabilities progressing much faster than most expected even a few years ago.

      If development continues at its current pace we could see an ‘intelligence explosion’ in which AI takes over AI research and rapidly self-improves, starting in as little as 2-7 years.

      This is the defining issue of our time and decisions made in the next decade could shape humanity’s long-term future in profound ways.

      Yet the public conversation around AGI remains completely inadequate. Mainstream sources neglect the issue entirely for the most part. When AGI is discussed, it’s often covered superficially and fails in one of various ways:

      • naive optimism and hype
      • reflexive dismissiveness
      • focusing only on immediate issues like job displacement or copyright violations
      • uncritically accepting narratives from actors with strong financial interests in the industry.

      The extremely profound implications of developing AI, then AGI, and then superintelligence are basically ignored.

      We believe there’s a critical opportunity to fill this gap by scaling up our podcast from occasional in-depth interviews to a more comprehensive media platform with multiple weekly pieces covering AGI-related developments, risks, governance,

      Continue reading →

        Understanding trends in our AI job postings

        This week, let’s review key trends in the jobs we’ve found that may help mitigate AI risk, including:

        • Growth in the number of postings in the field
        • The types of organisations that are hiring
        • The most in-demand skills
        • The experience level required for these roles

        We’ve ranked catastrophic risks from AI as the world’s most pressing problem since 2016, but it’s only in the last few years that the topic has really hit the mainstream.

        As AI has advanced rapidly and the risks have become more salient, we’ve seen many more jobs available to help mitigate the dangers.

        The number of AI-related jobs we posted on our job board rose throughout 2023 and then plateaued in 2024. But in January 2025, we posted the most AI-relevant jobs yet!

        Total AI roles

        In 2023, we posted an average of 63 AI-related roles per month. In 2024, the average rose to 105 — a 67% increase.

        Over this time, nonprofit jobs have been the most common, though they were briefly overtaken by both company and government jobs in early 2024.

        By org type

        This trend could reflect our vantage point. As a nonprofit that works closely with other nonprofits, we may be best positioned to find and assess high-impact roles in this sector while potentially missing other great roles in sectors more opaque to us.

        Continue reading →

        Why AGI could be here soon and what you can do about it: a primer

        I’m writing a new guide to careers to help artificial general intelligence (AGI) go well. Here’s a summary of the key ideas. If any surprise you, stay tuned to read the full guide with our reasoning and counterarguments, and for updates as our views evolve.

        In short:

        • The chance of an AGI-driven technological explosion before 2030 — creating one of the most pivotal periods in history — is high enough to act on.
        • If you can help the world navigate that transition, it’s likely the highest-impact thing you can do.
        • Hundreds of jobs tackling this issue are currently available.
        • For most people, the best entry path is completing a crash course in AI or self-study, and applying to organisations you’re aligned with. Our job board and one-on-one advice can help you do that.
        • Alternatively, build a relevant skillset or contribute from your current position by donating and spreading these ideas.

        Why AGI could be here by 2030

        Continue reading →

        #213 – Will MacAskill on AI causing a "century in a decade" — and how we're completely unprepared

        The 20th century saw unprecedented change: nuclear weapons, satellites, the rise and fall of communism, third-wave feminism, the internet, postmodernism, game theory, genetic engineering, the Big Bang theory, quantum mechanics, birth control, and more. Now imagine all of it compressed into just 10 years.

        That’s the future Will MacAskill — philosopher, founding figure of effective altruism, and now researcher at the Forethought Centre for AI Strategy — argues we need to prepare for in his new paper “Preparing for the intelligence explosion.” Not in the distant future, but probably in three to seven years.

        The reason: AI systems are rapidly approaching human-level capability in scientific research and intellectual tasks. Once AI exceeds human abilities in AI research itself, we’ll enter a recursive self-improvement cycle — creating wildly more capable systems. Soon after, by improving algorithms and manufacturing chips, we’ll deploy millions, then billions, then trillions of superhuman AI scientists working 24/7 without human limitations. These systems will collaborate across disciplines, build on each discovery instantly, and conduct experiments at unprecedented scale and speed — compressing a century of scientific progress into mere years.

        Will compares the resulting situation to a mediaeval king suddenly needing to upgrade from bows and arrows to nuclear weapons to deal with an ideological threat from a country he’s never heard of, while simultaneously grappling with learning that he descended from monkeys and his god doesn’t exist.

        What makes this acceleration perilous is that while technology can speed up almost arbitrarily, human institutions and decision-making are much more fixed.

        Consider the case of nuclear weapons: in this compressed timeline, there would have been just a three-month gap between the Manhattan Project’s start and the Hiroshima bombing, and the Cuban Missile Crisis would have lasted just over a day.

        Robert Kennedy, Sr., who helped navigate the actual Cuban Missile Crisis, once remarked that if they’d had to make decisions on a much more accelerated timeline — like 24 hours rather than 13 days — they would likely have taken much more aggressive, much riskier actions.

        So there’s reason to worry about our own capacity to make wise choices. And in “Preparing for the intelligence explosion,” Will lays out 10 “grand challenges” we’ll need to quickly navigate to successfully avoid things going wrong during this period.

        Will’s thinking has evolved a lot since his last appearance on the show. While he was previously sceptical about whether we live at a “hinge of history,” he now believes we’re entering one of the most critical periods for humanity ever — with decisions made in the next few years potentially determining outcomes millions of years into the future.

        But Will also sees reasons for optimism. The very AI systems causing this acceleration could be deployed to help us navigate it — if we use them wisely. And while AI safety researchers rightly focus on preventing AI systems from going rogue, Will argues we should equally attend to ensuring the futures we deliberately build are truly worth living in.

        In this wide-ranging conversation with host Rob Wiblin, Will maps out the challenges we’d face in this potential “intelligence explosion” future, and what we might do to prepare. They discuss:

        • Why leading AI safety researchers now think there’s dramatically less time before AI is transformative than they’d previously thought
        • The three different types of intelligence explosions that occur in order
        • Will’s list of resulting grand challenges — including destructive technologies, space governance, concentration of power, and digital rights
        • How to prevent ourselves from accidentally “locking in” mediocre futures for all eternity
        • Ways AI could radically improve human coordination and decision making
        • Why we should aim for truly flourishing futures, not just avoiding extinction

        This episode was originally recorded on February 7, 2025.

        Video editing: Simon Monsour
        Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
        Camera operator: Jeremy Chevillotte
        Transcriptions and web: Katy Moore

        Continue reading →

        Emergency pod: Judge plants a legal time bomb under OpenAI (with Rose Chan Loui)

        When OpenAI announced plans to convert from nonprofit to for-profit control last October, it likely didn’t anticipate the legal labyrinth it now faces. A recent court order in Elon Musk’s lawsuit against the company suggests OpenAI’s restructuring faces serious legal threats, which will complicate its efforts to raise tens of billions in investment.

        As nonprofit legal expert Rose Chan Loui explains, the court order set up multiple pathways for OpenAI’s conversion to be challenged. Though Judge Yvonne Gonzalez Rogers denied Musk’s request to block the conversion before a trial, she expedited proceedings to the fall so the case could be heard before it’s likely to go ahead. (See Rob’s brief summary of developments in the case.)

        And if Musk’s donations to OpenAI are enough to give him the right to bring a case, Rogers sounded very sympathetic to his objections to the OpenAI foundation selling the company, benefiting the founders who forswore “any intent to use OpenAI as a vehicle to enrich themselves.”

        But that’s just one of multiple threats. The attorneys general (AGs) in California and Delaware both have standing to object to the conversion on the grounds that it is contrary to the foundation’s charitable purpose and therefore wrongs the public — which was promised all the charitable assets would be used to develop AI that benefits all of humanity, not to win a commercial race. Some, including Rose, suspect the court order was written as a signal to those AGs to take action.

        And, as she explains, if the AGs remain silent, the court itself, seeing that the public interest isn’t being represented, could appoint a “special interest party” to take on the case in their place.

        This places the OpenAI foundation board in a bind: proceeding with the restructuring despite this legal cloud could expose them to the risk of being sued for a gross breach of their fiduciary duty to the public. The board is made up of respectable people who didn’t sign up for that.

        And of course it would cause chaos for the company if all of OpenAI’s fundraising and governance plans were brought to a screeching halt by a federal court judgment landing at the eleventh hour.

        Host Rob Wiblin and Rose Chan Loui discuss all of the above as well as what justification the OpenAI foundation could offer for giving up control of the company despite its charitable purpose, and how the board might adjust their plans to make the for-profit switch more legally palatable.

        This episode was originally recorded on March 6, 2025.

        Video editing: Simon Monsour
        Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
        Transcriptions: Katy Moore

        Continue reading →

        Ask a career advisor: how to compete with AI models in the job market and whether to stay in the US government

        This week, we answer more of our newsletter subscriber’s career questions!
        I’ll tackle:

        Laura: This is a difficult question, but I think it’s important to ask. I certainly don’t have all the answers, because the future is uncertain — perhaps especially now.

        That said, I do think there are sensible steps to take to navigate the trajectory of AI and its global impact

        First, let’s address the immediate concern about the job market. While AI is already changing the landscape of work, entry-level jobs haven’t disappeared. The 80,000 Hours job board has hundreds of entry-level openings, and other job sites will have many more. I expect there will still be plenty of jobs available when you graduate, even as AI continues to advance quickly.

        As technology becomes better at the tasks that human workers perform, there are typically significant delays before it is widely adopted. Still, the economy will likely change as AI gets more powerful, and that may start happening in just a few years. So, my advice is to stay ahead of the adoption curve. Master prompting and evaluation techniques to ensure the quality of outputs. Learn to set up AI task automation systems. Experiment with new tools and applications.

        Continue reading →

          Expression of interest: writer-researcher

          About 80,000 Hours

          80,000 Hours’ mission is to get talented people working on the world’s most pressing problems.

          In 2025, we are planning to focus especially on helping explain why and how our audience can help society safely navigate a transition to a world with transformative AI.

          Over a million people visit our website each year, and thousands of people have told us that they’ve significantly changed their career plans due to our work. Surveys conducted by our primary funder, Open Philanthropy, show that 80,000 Hours is one of the single biggest drivers of talent moving into work related to reducing global catastrophic risks.

          Our most popular pieces get over 10,000 unique visitors each month, and are among the most important ways we help people shift their careers towards higher-impact options.

          The role

          In this role, you would:

          • Research and write new articles for the 80,000 Hours website, e.g. articles explaining particular issues within AI risk reduction for a popular audience (e.g. an explainer on misuse, or government abuse of AI technology)
          • Survey and interview experts in the issues we prioritise to inform our career advice
          • Update or rewrite core articles with new information and resources or to appeal to new audiences
          • Write substantive newsletters for an audience of over 500,000 people
          • Continually learn from analytics and user feedback what kinds of framings are most clear and engaging
          • Manage and edit external expert contributors to write articles

          Who we’re looking for

          For this role,

          Continue reading →

            Ask a career advisor: how mid-career people can help with AI and how to survive a job hunt

            This week, we’re answering our newsletter subscribers’ career questions.

            Question one: I’ve been working over 20 years as a software engineer, and participated in a few publications as a research engineer. I’m open to getting deeper into research, in fact I’m starting a PhD in computer science. I’m close to 50, but most of the career advice I see from 80,000 Hours seems focused on younger folk.

            I’m wondering about reorienting towards global issues; does it make sense to continue with my current PhD? Drop it for one directly related to (e.g.) AI? Drop the PhD and just apply as a research engineer to appropriate companies with such engineers? Or just leave it to the next generation?

            Sudhanshu: I would pretty much never say “leave it to others” — there might not be many others, and experienced folks like you often have a lot to offer. We are always hearing from impactful organisations all the time who want experienced, mid-career people to join them in key roles.

            With 20 years as a software engineer, I’d be excited if you could apply that experience directly to a pressing world problem, unless you’ve gotten some evidence to suggest that you need some new skills. I think it makes sense to try your hand at just working directly on the problem you think is most important before moving on to other options.

            If you do think working on reducing AI risk is the most important problem,

            Continue reading →

              #212 – Allan Dafoe on why technology is unstoppable & how to shape AI development anyway

              Technology doesn’t force us to do anything — it merely opens doors. But military and economic competition pushes us through.

              That’s how today’s guest Allan Dafoe — director of frontier safety and governance at Google DeepMind — explains one of the deepest patterns in technological history: once a powerful new capability becomes available, societies that adopt it tend to outcompete those that don’t. Those who resist too much can find themselves taken over or rendered irrelevant.

              This dynamic played out dramatically in 1853 when US Commodore Perry sailed into Tokyo Bay with steam-powered warships that seemed magical to the Japanese, who had spent centuries deliberately limiting their technological development. With far greater military power, the US was able to force Japan to open itself to trade. Within 15 years, Japan had undergone the Meiji Restoration and transformed itself in a desperate scramble to catch up.

              Today we see hints of similar pressure around artificial intelligence. Even companies, countries, and researchers deeply concerned about where AI could take us feel compelled to push ahead — worried that if they don’t, less careful actors will develop transformative AI capabilities at around the same time anyway.

              But Allan argues this technological determinism isn’t absolute. While broad patterns may be inevitable, history shows we do have some ability to steer how technologies are developed, by who, and what they’re used for first.

              As part of that approach, Allan has been promoting efforts to make AI more capable of sophisticated cooperation, and improving the tests Google uses to measure how well its models could do things like mislead people, hack and take control of their own servers, or spread autonomously in the wild.

              As of mid-2024 they didn’t seem dangerous at all, but we’ve learned that our ability to measure these capabilities is good, but imperfect. If we don’t find the right way to ‘elicit’ an ability we can miss that it’s there.

              Subsequent research from Anthropic and Redwood Research suggests there’s even a risk that future models may play dumb to avoid their goals being altered.

              That has led DeepMind to a “defence in depth” approach: carefully staged deployment starting with internal testing, then trusted external testers, then limited release, then watching how models are used in the real world. By not releasing model weights, DeepMind is able to back up and add additional safeguards if experience shows they’re necessary.

              But with much more powerful and general models on the way, individual company policies won’t be sufficient by themselves. Drawing on his academic research into how societies handle transformative technologies, Allan argues we need coordinated international governance that balances safety with our desire to get the massive potential benefits of AI in areas like healthcare and education as quickly as possible.

              Host Rob and Allan also cover:

              • The most exciting beneficial applications of AI
              • Whether and how we can influence the development of technology
              • What DeepMind is doing to evaluate and mitigate risks from frontier AI systems
              • Why cooperative AI may be as important as aligned AI
              • The role of democratic input in AI governance
              • What kinds of experts are most needed in AI safety and governance
              • And much more

              Video editing: Simon Monsour
              Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
              Camera operator: Jeremy Chevillotte
              Transcriptions: Katy Moore

              Continue reading →

              Emergency pod: Elon tries to crash OpenAI's party (with Rose Chan Loui)

              On Monday, February 10, Elon Musk made the OpenAI nonprofit foundation an offer they want to refuse, but might have trouble doing so: $97.4 billion for its stake in the for-profit company, plus the freedom to stick with its current charitable mission.

              For a normal company takeover bid, this would already be spicy. But OpenAI’s unique structure — a nonprofit foundation controlling a for-profit corporation — turns the gambit into an audacious attack on the plan OpenAI announced in December to free itself from nonprofit oversight.

              As today’s guest Rose Chan Loui — founding executive director of UCLA Law’s Lowell Milken Center for Philanthropy and Nonprofits — explains, OpenAI’s nonprofit board now faces a challenging choice.

              The nonprofit has a legal duty to pursue its charitable mission of ensuring that AI benefits all of humanity to the best of its ability. And if Musk’s bid would better accomplish that mission than the for-profit’s proposal — that the nonprofit give up control of the company and change its charitable purpose to the vague and barely related “pursue charitable initiatives in sectors such as health care, education, and science” — then it’s not clear the California or Delaware Attorneys General will, or should, approve the deal.

              OpenAI CEO Sam Altman quickly tweeted “no thank you” — but that was probably a legal slipup, as he’s not meant to be involved in such a decision, which has to be made by the nonprofit board ‘at arm’s length’ from the for-profit company Sam himself runs.

              The board could raise any number of objections: maybe Musk doesn’t have the money, or the purchase would be blocked on antitrust grounds, seeing as Musk owns another AI company (xAI), or Musk might insist on incompetent board appointments that would interfere with the nonprofit foundation pursuing any goal.

              But as Rose and Rob lay out, it’s not clear any of those things is actually true.

              In this emergency podcast recorded soon after Elon’s offer, Rose and Rob also cover:

              • Why OpenAI wants to change its charitable purpose and whether that’s legally permissible
              • On what basis the attorneys general will decide OpenAI’s fate
              • The challenges in valuing the nonprofit’s “priceless” position of control
              • Whether Musk’s offer will force OpenAI to up their own bid, and whether they could raise the money
              • If other tech giants might now jump in with competing offers
              • How politics could influence the attorneys general reviewing the deal
              • What Rose thinks should actually happen to protect the public interest

              Video editing: Simon Monsour
              Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
              Transcriptions: Katy Moore

              Continue reading →

              Bonus: AGI disagreements and misconceptions: Rob, Luisa, & past guests hash it out

              Will LLMs soon be made into autonomous agents? Will they lead to job losses? Is AI misinformation overblown? Will it prove easy or hard to create AGI? And how likely is it that it will feel like something to be a superhuman AGI?

              With AGI back in the headlines, we bring you 15 opinionated highlights from the show addressing those and other questions, intermixed with opinions from hosts Luisa Rodriguez and Rob Wiblin recorded back in 2023.

              You can decide whether the views we expressed (and those from guests) then have held up these last two busy years. You’ll hear:

              • Ajeya Cotra on overrated AGI worries
              • Holden Karnofsky on the dangers of aligned AI, why unaligned AI might not kill us, and the power that comes from just making models bigger
              • Ian Morris on why the future must be radically different from the present
              • Nick Joseph on whether his companies internal safety policies are enough
              • Richard Ngo on what everyone gets wrong about how ML models work
              • Tom Davidson on why he believes crazy-sounding explosive growth stories… and Michael Webb on why he doesn’t
              • Carl Shulman on why you’ll prefer robot nannies over human ones
              • Zvi Mowshowitz on why he’s against working at AI companies except in some safety roles
              • Hugo Mercier on why even superhuman AGI won’t be that persuasive
              • Rob Long on the case for and against digital sentience
              • Anil Seth on why he thinks consciousness is probably biological
              • Lewis Bollard on whether AI advances will help or hurt nonhuman animals
              • Rohin Shah on whether humanity’s work ends at the point it creates AGI

              And of course, Rob and Luisa also regularly chime in on what they agree and disagree with.

              Audio engineering: Ben Cordell, Milo McGuire, Simon Monsour, and Dominic Armstrong
              Transcriptions and additional content editing: Katy Moore

              Continue reading →

              Preventing catastrophic pandemics

              Some of the deadliest events in history have been pandemics. COVID-19 demonstrated that we’re still vulnerable to these events, and future outbreaks could be far more lethal.

              In fact, we face the possibility of biological disasters that are worse than ever before due to developments in technology.

              The chances of such catastrophic pandemics — bad enough to potentially derail civilisation and threaten humanity’s future — seem uncomfortably high. We believe this risk is one of the world’s most pressing problems.

              And there are a number of practical options for reducing global catastrophic biological risks (GCBRs). So we think working to reduce GCBRs is one of the most promising ways to safeguard the future of humanity right now.

              Continue reading →

              How quickly could robots scale up?

              This post was written by Benjamin Todd in his personal capacity and originally posted on benjamintodd.substack.com.

              Today robots barely have the dexterity of a toddler, but are rapidly improving.

              If their algorithms and hardware advance enough to handle many physical human jobs, how quickly could they become a major part of the workforce?

              Here’s some order of magnitude estimates showing it could happen pretty fast.

              Robot cost of production

              Today’s humanoid robots cost about $100,000, with perhaps 10,000 units produced annually. But manufacturing costs tend to plummet with scale:

              For solar energy, every doubling of production was associated with a 20% decline in costs. In other industries, we see estimates ranging from 5-40%, so 20% seems a reasonable middle point.

              Solar power cost over time

              That means a 1000x increase in production (10 doublings), should decrease costs 10x to $10,000/unit. That’s around the cost of manufacturing a car.

              However, humanoid robots only use about 10% the materials of a car, so it’s plausible they could eventually become another 10x cheaper, or $1000 each.

              Though, it’s also possible the elements for fine motor control remain far more difficult to manufacture. Let’s add 2x to account for that.

              Robot operating costs

              If a robot costs $10,000 and lasts for three years working 24/7, the hardware costs $0.40 per hour.

              At $2000 each,

              Continue reading →

                It looks like there are some good funding opportunities in AI safety right now

                This post was written by Benjamin Todd in his personal capacity and originally posted on benjamintodd.substack.com.

                The AI safety community has grown rapidly since the ChatGPT wake-up call, but available funding doesn’t seem to have kept pace.

                However, there’s a more recent dynamic that’s created even better funding opportunities, which I witnessed as a recommender in the most recent Survival and Flourishing Fund grant round.

                Most philanthropic (vs. government or industry) AI safety funding (50%) comes from one source: Good Ventures, via Open Philanthropy. But they’ve recently stopped funding several categories of work (my own categories, not theirs):

                In addition, they are currently not funding (or not fully funding):

                • Many non-US think tanks, who don’t want to appear influenced by an American organisation (there’s now probably more than 20 of these)
                • They do fund technical safety non-profits like FAR AI, though they’re probably underfunding this area, in part due to difficulty hiring for this area the last few years (though they’ve hired recently)
                • Political campaigns,

                Continue reading →

                What happened with AI in 2024?

                The idea this week: despite claims of stagnation, AI research still advanced rapidly in 2024.

                Some people say AI research has plateaued. But a lot of evidence from the last year points in the opposite direction:

                • New capabilities were developed and emerged
                • Research indicates existing AI can accelerate science

                And at the same time, important findings about AI safety and risk came out (see below).

                AI advances might still stall. Some leaders in the field have warned that a lack of good data, for example, may impede further capability growth, though others disagree. Regardless, growth clearly hasn’t stopped yet.

                Meanwhile, the aggregate forecast on Metaculus of when we’ll see the first “general” AI system — which would be highly capable across a wide range of tasks — is 2031.

                All of this matters a lot, because AI poses potentially existential risks. We think making sure AI goes well is a top pressing world problem.

                If AI advances fast, this work is not only important but urgent.

                Here are some of the key developments in AI from the last year:

                New AI models and capabilities

                OpenAI announced in late December that its new model o3 achieved a large leap forward in capabilities. It builds on the o1 language model (also released in 2024),

                Continue reading →