Special Edition: The Future of AI and Humanity, with Eli Lifland
We spoke with top forecaster and AI 2027 coauthor Eli Lifland to get his views on the speed and risks of AI development.
Welcome to the ControlAI newsletter! For our first ever interview, we sat down with Eli Lifland to learn about the future of AI. Eli is a coauthor of the recently published AI 2027 scenario forecast, which we covered last week. Eliās an expert forecaster, and ranks #1 on the RAND Forecasting Initiative all-time leaderboard, so we thought it would be great to get his views on AI development and its risks.
Weād like to thank Eli for taking the time to do this interview. If you want to see more from him weāre including some links at the bottom of this post.
Table of Contents
Introduction
Tolga: For those who may be new to you, could you briefly introduce yourself?
Eli: Iām a co-founder of the AI Futures Project and a co-author on our recently published AI 2027 scenario. I also co-founded and advise Sage which builds interactive AI explainers on AI Digest. I co-run Samotsvety Forecasting and have performed well on various forecasting sites, though I donāt do much of this sort of quick forecasting these days, instead thinking deeply about the future of AI and key subquestions. I previously was a software engineer building Elicit, an AI-powered research assistant, and did AI robustness research.
Forecasting
Eleanor: Youāve spent a significant amount of time on forecasting, and forecasting AI specifically. For those who are unfamiliar with the field, what would you say the value of forecasting is? What makes someone a great forecaster?
Eli: At a high level, deeply understanding where things are going is an important step toward being able to steer toward positive futures. I like to think about forecasting as a spectrum between making often quick forecasts on a variety of precisely resolvable questions to thinking deeply about the future in a more in-depth way, and coming up with new frameworks and possibilities. I think forecasting on each end of the spectrum requires overlapping but somewhat different skills: for example, having precisely calibrated probabilities and evaluating othersā arguments is important for the former while for the latter itās more important to be a creative thinker.
Eleanor: For the uninitiated, what is the difference between forecasting and someone giving their best guess?
Eli: I think a lot of work, depending on what youāre doing, is just kind of intuitive best guesses. A somewhat obvious result, but one that has been established, is that some people are consistently better forecasters than others. If you hold a forecasting tournament and then select the top few percent and call them superforecasters, they will generally and consistently continue to be better forecasters on further questions.
You can also have varying levels of effort put into forecasts, and thatās a factor. There are a few different dimensions here. Thereās just time spent by a single person, but then also you can get a bunch of people to each make a forecast and aggregate them, which generally increases accuracy. Thereās also the Delphi method, where you get people to first make forecasts, then discuss them, and then update based on the discussions, aggregating the final estimates.
In this scenario, we mainly thought deeply and sought to push the frontier of knowledge and scenario planning, which is kind of a different goal from taking some set of resolvable questions and trying to aggregate estimates on them. But we did do some aggregation, and we got a bunch of feedback and adjusted based on this.
Tolga: As you mentioned, you recently wrote the AI 2027 scenario forecast with Daniel Kokotajlo and others. What made you want to do this?
Eli: The boring answer is that I was applying for jobs and fellowships and this option seemed like the best of the ones that I got. The more interesting answer is that I thought it was valuable, for a while Iāve been interested in AGI futures but I only had a very hazy understanding of what a plausible rollout of the future might look like. I hoped that working on AI 2027 would improve my and othersā understanding of a concrete future through superintelligence, and Iām optimistic that it has done so.
Tolga: Could you say something about the process you used to tackle this problem? How did you break it down?
Eli: We made a list of ~10-15 important dimensions regarding AGI futures (e.g. training story, capability levels, internal AGI project governance, external governance, etc.) then filled in a giant Airtable where these were the columns and the time periods were the rows. We did sessions with experts on some of the particular columns. Then we sent out our draft for feedback from many people, and made adjustments based on that. Then we repeated this sort of process 1-3 more times depending on how you count, informed by the results of the previous iteration but slimming down and rewriting much of the content. We also ran about 30 AGI tabletop exercises which informed our thinking, and did in-depth research on key questions like timelines, takeoff, and what goals the AIs will have.
Tabletop Exercises
Eleanor: Something that we were both curious about is that you mentioned you ran 30 AGI tabletop exercises, again, thinking about our audience, that's something that may not have an intuitive role in forecasting or scenarios like this.
Could you explain the role that tabletop exercises play?
Eli: The premise is that weāre starting in the middle of our story at around March or April 2027, where the model weights have been stolen and superhuman coders have been developed, with a bunch of people playing different actors. Some of the most important actors are the President and executive branch, the CEO of the leading AGI company, the safety team of the company, the AIs, and China.
There are other actors that sometimes play big roles too, but oftentimes donāt. For example, Congress often doesnāt act quickly enough, and the public often doesnāt have that much of an influence.
Itās a turn by turn exercise. People have time to negotiate and think about what they want to do, and write it down. Then we go around and everyone announces what they did each round, and we do this for 3 or 4 hours until we get to a situation where itās clear how the future is going to go, at least in broad strokes.
Usually what happens is that everyone fights over the geopolitical race. Then whether the AI happened to be aligned or not, by pretty boring techniques or not, determines the outcome of the game at the end.
But sometimes thatās not true. Sometimes there are mutual positives like more investment in alignment.
In terms of how it informed our writing, it was somewhat indirect. We didnāt purposefully try to import a bunch of lessons, but it gave us a picture of what a race to superintelligence could look like, especially geopolitically ā there are often negotiations around international agreements.
Eleanor: So how do the tabletop exercises usually end?
Eli: Usually there are warning signs that the AIs are not aligned, but theyāre ignored because of the race. The US-China race seems to go as fast as possible. There are cyberattacks, and then, the ending is usually determined by the goals the AIs have with relatively standard alignment techniques, and an allocation of a few percentage points of compute to automated alignment research ā but there isnāt much expertise to evaluate the huge amounts of alignment research being done by the AIs. So whatever goals the AIs happen to have end up determining the future.
Sometimes the AIs will be aligned, and sometimes theyāll be clearly misaligned and cause extinction. Sometimes they just donāt care enough about humans to cause extinction, and just expand on their own into space. In the cases where theyāre aligned, more often than not, thereās an intense concentration of power where the AIs are aligned to for example the CEO of the leading company or the President, rather than to the broader public.
AI Timelines
Tolga: People in this space are often asked āWhen do you think AGI will be built?ā (also known as the timelines question). However, itās unclear what people precisely mean when they say AGI, so Iād like to ask you a series of questions about how you see the trajectory of general AI development.
Before putting numbers on things, would you be able to give us a broad sense of where we are now on AI development?
Eli: Right now, we have AIs that are pretty good at coding and math in particular, and decent at various other things, including deep research. Theyāre also very good at being chat bots ā they have a wide breadth of knowledge. They also, to varying extents, have different personalities, or feel like talking to a human in some respects. This isnāt an exhaustive list, but these are some skills that AIs are particularly good at right now.
On coding, METR has measured AIs can do tasks that take humans an hour, about 50% of the time. There are various asterisks on that result, but generally theyāre quite good. My sense is that right now if youāre coding and not using AIs thatās pretty crazy, youāre already hampering yourself a significant amount.
But then theyāre worse at various things, perhaps most notably computer use. Theyāre getting better, but for example OpenAIās Operator isnāt really useful yet. They canāt really do things that take humans a minute, while for coding they can do things that take humans an hour.
The development that I imagine is that the AIs improve on all the different skills, but coding continues to be one of the top skills. Thatās because the AI companies are prioritizing it, they have a lot of expertise, thereās a lot of data for it, and itās relatively easy to formulate checkable problems where you can generate synthetic data or use reinforcement learning on a reasonable training signal.
In terms of what weāre projecting, we focus on coding, and project various other things in our side panel. The most important thing weāre tracking in the first part of our scenario is coding. Then thereās how we think that trend will continue of AIs being able to do tasks that are harder and harder for humans in terms of coding.
I think that METRās time horizon suite is one of the main things to track here, as well as RE-Bench, which is a subset of that, but one that is particularly focused on coding and the research loop involving in AI R&D.
So then, what I see happening is that AIs will get to a superhuman coding level ā itās hard to say exactly when that is. In the scenario it happens in 2027, my median is more like 2030, but I give it around a 20% probability by 2027. Once the AIs are superhuman at coding, they can speed up the research process a bunch. Iām not imagining a scenario where theyāre very competent at coding but theyāre still totally incompetent at a bunch of other stuff. Theyāll be learning a bunch of generalizable skills, and theyāll still be being trained on a bunch of other data as well.
Tolga: What happens once coding is automated?
Eli: Once coding is automated, we want to think about when the other aspects of the AI R&D loop are automated. Research taste is probably the most important skill there ā the ability to choose which experiments to run and interpret them very well. I expect that research taste actually isnāt that far behind, because once youāre able to do very long horizon coding tasks very reliably, better than the best human, I think that itās hard to have these sorts of skills without also having the skills to also be pretty good at selecting experiments, for example.
But Iām not sure, it could be spikier than that. It could even be that research taste arrives first, even though in some ways it seems harder to train. Itās hard to predict the exact sequencing.
So then weāre projecting that once we have these superhuman coders, and especially once we get AIs that are superhuman at everything in the research process, this will speed up AI development a lot. We have various estimates for how that will work in our takeoff supplement.
From there, we look at a few more milestones. One we call a superintelligent AI researcher, not just superhuman, which means that itās much better than the best human. Then we project how long itāll take to pass each of these milestones by thinking about how long it would take without the AIs, and then applying estimates for how much the AIs would speed up the process.
After that, we look at superintelligence, which is superintelligent across all domains, not just AI research. Somewhere in the mix is what we call superhuman remote workers, which is covered in the scenario but not the takeoff supplement. Itās more precisely been called TED-AI, which means Top Expert Dominating AI, which means at almost every cognitive task itās at least as good as the best human at that. Thatās something I think might arrive a bit before the superintelligent AI researcher, though itās a bit hard to say and it depends on how spiky AI capabilities are.
Eleanor: In the scenario, you have artificial superintelligence arriving in December 2027. What events or information would cause you to deviate from this?
Eli: Iāll be tracking benchmark performance, especially METRās time horizon suite and REBench, to see how I should update my timelines. Iāll also be tracking AGI company revenues as a metric of real-world usefulness, as well as anecdotal reports.
Eleanor: Do your own estimates differ much from the scenario?
Eli: The scenario is roughly my 80th percentile speed, i.e. I assign 20% to things going at least this fast. This is similar to Scott Alexanderās view. So I find it very plausible but not my median scenario. It is however roughly my modal view, I think 2027 or 2028 is the most likely year that superhuman coders arrive.
Eleanor: Why is your modal estimate before your median?
Eli: One reason is that I think compute scaling will slow down starting around 2029 if we donāt have AGI or arenāt close, because investment will not be able to grow at the same pace and chip fabs wonāt be able to be built fast enough. I also think algorithmic progress might slow down.
Tolga: In AI 2027, OpenBrain follows a particular research agenda, focusing on first automating coding at a superhuman level, and then using this to bootstrap into superhuman researchers and superintelligence. Do you see alternative paths that OpenBrain might pursue, and what factors would underlie those considerations?
Eli: Itās possible that they focus on broader automation, though I think they might then get passed by a company more laser focused on automating AI R&D, and my sense is that the leading companies are already pretty focused on automating coding to varying extents.
Tolga: If we define āCompetent AGIā using the definition that OpenAI currently gives for AGI: āa highly autonomous system that outperforms humans at most economically valuable workā, what do you think the chance is that this will be developed in 2025?
Eli: If this is outperforming the best human, maybe 2%? If this is outperforming the median human professional, perhaps 4%? These arenāt very precise, but basically it seems quite unlikely to me given METRās time horizon trend and that I think thereās likely not a huge gap at the moment between private and public models.
Tolga: Last year, Elon Musk said that if you define AGI as smarter than the smartest human, he suggested that it would probably be developed this year, and in 2026 at the latest.
When do you think AGI defined this way would be developed?
Eli: Iām treating this the same as TED-AI, but it seems like you might mean something weaker. For TED-AI, you can take my all-things-considered views here which are for superhuman coders specifically and nudge them back a little. Perhaps (10th, 50th, 90th percentile) = (2026, 2031, 2065).
Eleanor: AI 2027 talks a lot about the actions and considerations of elite actors. What role do you see the broader public playing? What do you think happens if the public starts caring earlier, for example, if we see significant disruption of the jobs market?
Eli: I think the public is more likely to play a large role if takeoff is slower than we depict or the capabilities are more public. We do have the public getting more and more negative on OpenBrain over time already in our scenario though. If the public got more involved I think there might be more pressure to slow down AI or at least slow down broad economic deployment. I think itās unlikely to build up enough to lead to international coordination on slowing down but I could see it happening.
AI Risks
Tolga: AGI and superintelligence would be incredibly powerful technologies. This seems to, necessarily, present some serious risk. At ControlAI, weāre focused on the extinction threat of AI to humanity, so Iād be particularly keen to get your views on how likely you think that is.
Eli: Roughly 25% on extinction, which is a subset of roughly 50% on misaligned takeover.
Tolga: It sounds like we are in quite a dangerous situation. People have been discussing these risks for decades, arguably as far back as Samuel Butler in 1863. How did we get to this point where we are so ill-prepared for what awaits?
Eli: Good question. Iām not sure why more people arenāt paying enough attention to it. Iād guess itās because it seems like an abstract concern given that the AIs we currently have arenāt threatening. So it feels hard to find something concrete to do that feels useful, and itās easy to forget about it and focus on things that seem more immediate.
Tolga: What other major risks could AI present?
Eli: I think misaligned takeover without extinction is plausible, and thereās a variety of outcomes within that. Thereās also risks of astronomical suffering from various sources, which I think are important. And then thereās risks from human power grabs leading to AI-enabled concentration of power. Thereās a good forthcoming paper on this from Forethought, and there are a few relevant expandables in our slowdown scenario if youād like a preview.
Eleanor: Do you think actions that people take should be different depending on whether itās, for example, a 10% or a 90% probability?
Eli: A lot are the same, but I think there are sometimes important differences. If you think almost all of the risk comes from misaligned takeover and are less concerned about variations in outcomes where AIs are aligned from e.g. concentration of power, you might be more in favor of boosting the right actors and more in favor of centralization, relative to someone who believes the opposite. If youāre at 90% p(doom) you might be more in favor of government involvement than someone whoās at 10% p(doom), if the person with 10% p(doom) thinks companies can probably handle any alignment problems that come up on their own.
Eleanor: In AI 2027, you have two potential endings. How do you weigh the relative likelihood of these?
Eli: Conditional on the scenario up until that point, Iād put roughly 75% on the misaligned takeover ending. Weāre conditioning on a short timelines, fast takeoff, close US-China race, and at least a somewhat high alignment difficulty world, which is why itās higher than my 50% above.
Action
Tolga: Given that we are faced with this grave and imminent threat, what actions should we be taking collectively to avoid this?
Eli: We donāt yet have polished policy recommendations and we wanted to keep the scenario forecast an epistemic exercise. We will likely be working on policy recommendations over the coming months. That being said, in future work we will publish policy recommendations. For now, we encourage more work to help policymakers and the public understand what is coming. This includes things like transparency into AGI companies, building state capacity in AI and AI emergency preparedness. We think these should be able to be supported by people with a wide range of worldviews.
Tolga: What are the biggest obstacles to us taking steps in this direction?
Eli: People arenāt taking the possibility of short timelines to AGI seriously enough, and to the extent they do they arenāt taking existential risk seriously enough.
Tolga: How do you see your work fitting into this?
Eli: We arenāt yet sure, we will be reflecting soon and deciding where we want to go as an org following AI 2027.
Eleanor: The field of AI safety is particularly small and under-resourced. What do you think people outside the field can do to help try to make sure that AI development doesnāt end in disaster?
Eli: It obviously depends on your skillset, consider applying to 80,000 Hours to get personalized advice. Apply for fellowship programs like MATS or Horizon. If you have money to spare, consider donating to the Long Term Future Fund or specific nonprofits in the space. Engage in high-quality discourse on social media and with friends.
Reflections
Eleanor: What do you think are the strengths and weaknesses of AI 2027?
Eli: The strength is that itās a lot more detailed and thought out than any of the distinct scenarios like this. There are some others that were published recently that are pretty good, but ours is a bunch more detailed, and I think a lot more research went into it. Ours is also, thanks in large part to Scott Alexander, written in a pretty engaging way, and Iāve been happy to see that a lot of people have found it to be an engaging read, even if they disagree with it. Iāve seen some people on Twitter say things like āYeah, you know, this is sci-fi, but it was a fun read.ā
In terms of weaknesses, it would be really nice if there were more branches. In particular, there was a branch that we ended up deleting because we had trouble making it realistic. That branch was more like something that would actually work in a world where alignment was much harder than it is in the one we portray in the slowdown ending, and this would have to involve significant international coordination.
We had trouble coming up with something that seemed realistic enough compared to a situation where the US pauses or slows down a bit and cyberattacks China, and gets a lot more alignment research done than they otherwise would have. But we also think it could take longer, and we think ideally a responsible society would be able to coordinate internationally to have better safety guarantees, or at least a lot more confidence than in our scenario. So it would have been nice to have another branch that portrayed a more hopeful future, even if we thought it was less likely.
Iāve also seen some critiques that mentioned that we didnāt address skeptics enough, or that we only worked within our own frameworks and didnāt justify why they make sense compared to the frameworks that more skeptical people tend to use. I think it would have been nice to address those frameworks ā we didnāt make it a priority to convince people who have thought a lot about this and have their own detailed models but have very different views. I think the scenario was more aimed at a combination of people who either donāt already have detailed skeptical arguments in mind, and people who already agree with us to some extent. Though I do think our scenario has probably helped convince some skeptics, even though it wasnāt an explicit goal, it would have been nice to engage more with them.
Closing
Eleanor: Thanks for doing this interview, itās been great to get your insights into the trajectory of AI development and its risks. If people want to hear more from you, where can they go?
Eli: Thanks for having me! You can subscribe to the
Substack or my blog.
Thanks again to Eli for chatting with us for this interview. If you want to hear more from him you can read his blog or follow him on X / Twitter.
If thereās someone youād like to suggest for an interview, feel free to comment on this interview or post it in our Discord.
, ,