Google DeepMind's CEO Demis Hassabis has listed a number of research breakthroughs or new capabilities he believes are needed for artificial general intelligence (AGI). I caught at least two that he mentioned when I listened to a recent episode of Google DeepMind's podcast:

  • Continual learning, a longstanding problem in AI, the ability to learn in every new moment, as in humans do, rather than only learning every few months when a new training run happens (sometimes Hassabis seems to use the term "long-term memory" interchangeably with continual learning, but in the recent podcast he explicitly says continual learning)
  • World models, which, as I understand the term, is somewhere between a philosophical concept and a technical term, referring to some general ability to understand, predict, and model how the real world works, such as the way humans and other mammals have an intuitive understanding of physics

In a January interview, he also mentioned the following missing capabilities:

  • Reasoning, which he also discusses on the podcast, noting a contrast between chatbots' impressive performance on some advanced math problems and a preponderance of elementary mistakes[1]
  • Hierarchical planning, the ability to plan actions that are composed of sub-actions, which are composed of sub-sub-actions, and so on, in a nested hierarchy
  • The ability to creatively generate novel hypotheses or conjectures

Here's the quote from that interview:

The models today are pretty capable, but there are still some missing attributes: things like reasoning, hierarchical planning, long-term memory. There's quite a few capabilities that the current systems don't have. They're also not consistent across the board. They're very strong in some things, but they're still surprisingly weak and flawed in other areas. You'd want an AGI to have pretty consistent, robust behavior across the board for all cognitive tasks.

One thing that's clearly missing, and I always had as a benchmark for AGI, was the ability for these systems to invent their own hypotheses or conjectures about science, not just prove existing ones. They can play a game of Go at a world champion level. But could a system invent Go? Could it come up with relativity back in the days that Einstein did with the information that he had? I think today's systems are still pretty far away from having that kind of creative, inventive capability.

Remarkably, immediately following this, Hassabis says he thinks AGI is "probably three to five years away."

I'm skeptical of such predictions for (at least) two reasons:

1. Many such predictions have come and gone. At very beginning of the field of artificial intelligence in the 1950s, a similarly ambitious goal was set for the first summer of AI research. Anthropic's CEO Dario Amodei predicted in March that AI would take over 90% of coding by September and, well, here we are. (It's not even true at Anthropic!)

2. Hassabis's discussion of the remaining research breakthroughs or missing capabilities seems incongruent with a prediction that those breakthroughs will be made or those capabilities will be developed in such a short time. Hassabis gives the impression he's not even sure he knows the full list yet of what is still missing. And he really thinks the longstanding problems in AI research that he did list will be solved in such a short term? I don't understand what could possibly justify such confidence.

I think it could be an interesting exercise to, rather than (or in addition to) forecasting AGI as a single idea, to decompose AGI into research problems or capabilities like continual learning, world models, reasoning, hierarchical planning, and creative idea generation, and then ask people to forecast those things individually. My guess is that for many or most people, reframing the forecasting question this way would lead to longer timelines for AGI.

I think before venturing a guess, people who want to try forecasting when these research problems will be solved should look into how long AI researchers have been working on them and how much research has already been published. Many of them, in fact I think all of them, are decades old. For instance, I searched Google Scholar and found several papers on hierarchical reinforcement learning from the early 1990s. (Searching for the exact phrase filters out unrelated stuff, but also misses some relevant stuff.) There is more funding now, but it seems like the vast majority of it is being spent on scaling large language models (LLMs) and AI models that generate images and videos, and very little on discovering the new science that Hassabis says is necessary to get to AGI.


  1. A few days ago, I told GPT-5.2 Thinking I had misplaced my AirPods somewhere in my home and asked it I could if I could use Apple's Find Devices feature to make them play a noise while they were closed in the case. It said no, but offered this helpful advice:If they really are in the closed case, force a situation where sound can work. Open the case (lid open) and take at least one AirPod out (even briefly), then retry Play Sound. Other peculiarities I've noticed recently include a pattern of answering "yes" to questions that are not yes or no questions. ↩︎

Comments


No comments on this post yet.
Be the first to respond.
Curated and popular this week
 ·  · 11m read
 · 
Crossposting this essay by my friend Gavriel Kleinwaks, who is approximately the world expert in Far UV deployment. [Edit: Austin is very kind--I am not close to being the world expert in far UV deployment; there are people who run/used to run companies trying to do that, and researchers who work with them very closely, who know more about far-UV deployment, and I'm largely consolidating information from them. -Gavriel] ---------------------------------------- Disclaimer: In the interest of avoiding the problem where a post languishes in drafts forever, I am not attempting to be particularly rigorous, but I’m happy to answer questions to fill in details about how I’ve come to the conclusions I’m writing about. This post is informed by my work experience at 1Day Sooner, but only reflects my personal takes, not the official stance of my organization. I work on indoor air quality to reduce disease burden and pandemic risk, and far-UV is an extremely promising intervention. One of the questions I get most often is, “If this is so promising, why isn’t it everywhere?” The answer requires a lot of breakdown and I get a lot of follow-up questions as soon as I start to answer, so this post is an attempt to run through the entire conversation at once. I’d never actually get around to posting this if I also included a deep dive on far-UV or tried to prove its value, so if you don’t know yet why it’s promising, you can check out this report. Top line takeaway: Far-UV isn’t widely used because no individual organization has the incentive or bandwidth to install it, and the industry is stuck in a vicious cycle, making companies unreliable. Filtration is already widely used, but more user-friendly filters would greatly improve implementation. ---------------------------------------- Q: What are the major roadblocks to widespread use of high-quality filters? A: The market is hard for consumers to navigate; it’s difficult for consumers to be well-informed enough to assess wh
 ·  · 6m read
 · 
I really enjoyed reading the "why I donate" posts in the past week, so much so that I felt compelled to add my reflections, in case someone finds my reasons as interesting as I found theirs. 1. My money must be spent on something, might as well spend it on the most efficient things The core reason I give is something that I think is under-represented in the other posts: the money I have and earn will need to be spent on something, and it feels extremely inefficient and irrational to spend it on my future self when it can provide >100x as much to others. To me, it doesn't seem important whether I'm in the global top 10% or bottom 10%, or whether the money I have is due to my efforts or to the place I was born. If it can provide others 100x as much, it just seems inefficient/irrational to allocate it to myself. The post could end here, but there are other secondary reasons/perspectives on why I personally donate that I haven't seen commonly discussed. 2. Spending money is voting on how the global economy allocates its resources In 2017, I read Wealth: The Toxic Byproduct by Kevin Simler. Surprisingly, I don't think it has ever been posted on this forum. I disagree with some of it, but the core points really changed how I think about wealth, earning, and spending. The post is very well written and enjoyable, but it's 2400 words, so copy-pasting some snippets: > A thought experiment — the Congolese Trading Window: > > Suppose one day you wake up to find a large pile of Congolese francs. [...] A window [...] pushes open to reveal the unfamiliar sights of a Congolese outdoor market [...] a man approaches your window. [...] He's asking if you'd like to buy his grain for 500 francs. > > What should you do? [...] > Your plan is to buy grain whenever you think the price is poised to go up in the near future, and sell whenever you think the price is poised to go down. > [...] > Imagine a particular bag of grain that you bought at T1 for 200 francs, and then sold at T
 ·  · 4m read
 · 
This year, I have given money to a range of EA cause areas. Most of it has either been towards global health and development, or EA infrastructure I believe does or could lead to effective fundraising for global health and development. The following are a list of very selfish personal reasons why I like to do this. I feel the selfless reasons have been adequately covered elsewhere, so I'm intentionally leaving them off. I get to ignore ineffective charity adverts. In order to genuinely convince myself that I am helping, I want to see things like well-regarded cost-effectiveness metrics. I do not like heartstring-tugging advertising or vague statements of "should", particularly to do with orphanages. They make me feel a bit ill. So I am glad that donating effectively gives me a very good justification to ignore them. It is a marker of my politics. I don't believe that poor people I don't know in rich countries are 100× more worthy of my help [i.e. worthy of help that's 100× less cost-efficient] than poor people in poor countries. This is because I don't believe anyone is 100× more worthy than anyone. Choosing to donate based on the cost-effectiveness of helping is making a radical political statement about equality. It's also quite anti-nationalist, and I like that, because I think excessive nationalism is wrecking havoc on my country right now. Giving expresses abundance. I earn approximately the median income in my country (the UK). This is not a lot, relatively, although it's high in absolute terms compared to the rest of the world. I earn less than my siblings, for example, and less than many people I interact with in my hobbies. I do, however, earn something that many people who earn more than me don't: "enough". So much enough, indeed, that 10% less is still "enough". By stepping away from lifestyle creep and the related creeping beliefs of feeling personally put upon by not having enough to spend, I have found a sense of quiet abundance. I live my lif