9d ago

Anthropic co-founder Jack Clark estimates 60% chance of AI recursive self-improvement by 2028

0

Anthropic co-founder and Head of Policy Jack Clark forecasts 60% probability that AI systems achieve recursive self-improvement by end of 2028. He defines it as AI autonomously enhancing capabilities and building superior versions. Clark gives 30% chance by end of 2027, based on review of hundreds of public AI development sources. Analysis cites rapid progress in AI coding and long-horizon agents enabling automated AI R&D.

Original post

@alexolegimas @jackclarkSF I’d be interested in more detailed arguments for this point. I imagine there’s a continuum between meat-and-potatoes work and the extremely left-field, with much work somewhere in-between.

7:01 AM · May 4, 2026 View on X
Reposted by

Probably nothing

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
4:01 PM · May 4, 2026 · 4.6K Views

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views

Major essay in Import AI 455, just published online.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
2:48 PM · May 4, 2026 · 1.4M Views

A lot of the conclusion comes from assembling a mosaic out of many distinct data sources. Some examples - progress on CORE-Bench, where the task is implementing other research papers (huge amounts of AI research comes from interpreting and replicating results)

Jack Clark@jackclarkSF

Major essay in Import AI 455, just published online.

2:48 PM · May 4, 2026 · 1.4M Views
2:54 PM · May 4, 2026 · 41.6K Views

What does this even mean? Haven’t we already seen this with Andrej’s small project more or less? It’s a low bar in retrospect!

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
7:55 AM · May 5, 2026 · 9.7K Views

Anthropic co-founder Jack Clark says 60% chance of RSI by end of 2028:

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
7:01 PM · May 4, 2026 · 16.2K Views

Note Clark’s definition of RSI here, from his newsletter, is “a frontier model is able to autonomously train a successor version of itself.”

This is a weaker claim than what I assumed he meant, which was that human researchers would no longer be useful vs. AI ones.

Riley Goodside@goodside

Anthropic co-founder Jack Clark says 60% chance of RSI by end of 2028:

7:01 PM · May 4, 2026 · 16.2K Views
7:49 PM · May 4, 2026 · 3.2K Views

Anthropic co-founder Jack Clark says 60% chance of RSI by end of 2028.

I need a drink.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
6:57 PM · May 4, 2026 · 682 Views

Jack Clark assigning a 60% chance to RSI by 2028 is notable because RSI matters, unlike all other human endeavors which do not.

7:24 PM · May 4, 2026 · 20.3K Views

@roydanroy As with any field this broad, you can set the bar at doing 20% automatically, 80% automatically, 99% automatically, etc and get wildly different timelines for when it will be possible. Also different ways of quantifying what 99% is.

Dan Roy@roydanroy

What does this even mean? Haven’t we already seen this with Andrej’s small project more or less? It’s a low bar in retrospect!

7:55 AM · May 5, 2026 · 9.7K Views
8:27 AM · May 5, 2026 · 1.7K Views

Co-founder of Anthropic, interesting that he refers to public sources when he is also obviously privy to lots of internal sources that he cannot discuss. I assume he sees the same thing at Anthropic.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
4:04 PM · May 4, 2026 · 261.2K Views

And OpenAI

Adrien Ecoffet@AdrienLE

Seems right. (as a reminder, if you think OpenAI disagrees, our stated estimate is that automated AI research will be developed around March 2028)

4:02 AM · May 5, 2026 · 36.3K Views
5:02 AM · May 5, 2026 · 11.8K Views

in this case, let's make RSI beautiful and genuinely good for the world

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
8:40 PM · May 4, 2026 · 4K Views

IMO Jack is right that RSI is imminent, but AI labs are too compute constrained right now for RSI to be a foom risk. Some napkin math:

Anthropic before RSI:

~5000 employees x average ~10 agents / employee building better AI 24x7 == 50K AI agents building better AI at Anthropic

Anthropic after RSI:

5 employees x average 10,000 agents / employee building better AI == 50K AI agents building better AI at Anthropic

If human staff was convertible into more compute/data, maybe RSI would be more important... but if AI agents are already doing 90% of the intellectual labor anyway... shifting that to 100% just doesn't seem like the foom risk that everyone is talking about. Seems more like a risk that Anthropic lays off some of its $750,000/year engineers and drops the price of Claude tokens a bit.

I could see how RSI is a governance risk tho ... probably harder to convince 5000 employees to allow bad AI than it is to convince 5 employees.

So if RSI is a risk... I think it's more from the human-in-the-loop socio-technical governance dynamics moreso than a simple "AI gets crazy smart overnight" risk.

Here we go.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
9:11 PM · May 5, 2026 · 22K Views

Ok... gonna break an X norm here... honestly I wrote a whole different, longer reply and then this model caught my eye (see screenshot). This looks really good to me, especially since it's got that sigmoidal shape. I honestly wasn't expecting that, hadn't seen that part of your work yet, and I think that's right.

Per your prompt, I don' think I can point out anything "wrong" in that way... too much uncertainty predicting the future... I assume the best model we have going forward is the combination of many great perspective / mental models / etc. I'm going to guess that everything I'd say is already in your model, but maybe I can make the case for a re-weighting? Let's see any of these interesting to talk about:

1) There's a big bend at Superhuman AI researcher. This might be warranted, but if we're compute constrained, one could argue that 1 Anthropic employee + 10 agents is already superhuman... and as the ratio gets bigger 1 -> 100, 1->1000,... the human is slowly becoming less relevant.... and everything is compute constrained in that regime. My hunch is that this smoothes out that "kink" in the slope... but you've already got it reasonably smooth here. Most of the charts I've seen have more of a hockey-stick moment post mass-AI-resercher-layoff... and my main doubt is the pointyness of that hockey-stick.

2) The other area I'd be skeptical about is the smoothness of the upward slope after the bend, unless there's an aspect of RSI which leads to faster data/compute growth than we're already seeing. My impression of AI progress is that there are some long, relatively smooth trend lines:

- increasing GPU capacity (more cloud) - increasing GPU output (better chips) - increasing algorithm efficiency (better caching/branching)

And these cause straight-line-like growth (like what you've pictured) until they exhaust their data supplies for a given domain or topic, after which point there's a plateauting for that domain area... followed by a "BURST" when a new 10-10,000x bigger dataset opens up suddenly (e.g. Imagenet, Demis discovers videogames, risk tolerance for scraping book torrents / the internet reaches a tipping point, etc.), after which there's a big burst for a while.

So if you focused on any paricular task... it's probably really bumpy more than smooth... but also your graph is super zoomed out and is some amalgomation of "all tasks".

Honestly the part of your graph I'm most impressed by is the top of the sigmoidal curve. Maybe you're not the first... but you're the first high-profile ASI forecaster I've seen that made a general claim that there is an limit. I'd probably extend the "algorithmic limit" to be an algorithm limit + data novelty limit (drawing on machine learning fundamentals... sample complexity X number of samples). On data in particular, there's probably some panopticon (bad word - and i think a true panopticon should be avoided) level of data saturation wherein an AI system just doenst' seen patterns it's never seen a million times before... and so all the domains are covered... and the intelligence is sortof... rate limited by the lack of novelty within its sensor field. We're not close to that yet (there's a million->billion times more digitzed data than our AIs are presently training on, and the global rollout of more sensors is doubling data every 2 years), but I'm being pretty nit-picky.

Overall, really strong model. Thanks for sending it. Exceeded my expectations. Keep up the great work.

Ryan Greenblatt@RyanPGreenblatt

@iamtrask This argument dismisses: - >>50k parallel workers - Serial speed well beyond humans - Increasing capabilities/taste beyond the human range Where do you think the modeling at https://www.aifuturesmodel.com/ or https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion is wrong?

10:50 PM · May 5, 2026 · 1.2K Views
4:03 AM · May 6, 2026 · 447 Views

IMO Jack is right that RSI is imminent, but AI labs are too compute constrained right now for RSI to be a foom risk. Some napkin math:

Anthropic before RSI:

~5000 employees x average ~10 agents / employee building better AI 24x7 == 50K AI agents building better AI at Anthropic

Anthropic after RSI:

5 employees x average 10,000 agents / employee building better AI == 50K AI agents building better AI at Anthropic

If human staff was convertible into more compute/data, maybe RSI would be more important... but if AI agents are already doing 90% of the intellectual labor anyway... shifting that to 100% just doesn't seem like the foom risk that everyone is talking about. Seems more like a risk that Anthropic lays off some of its $750,000/year engineers.

I could see how RSI is a governance risk tho ... probably harder to convince 5000 employees to allow bad AI than it is to convince 5 employees.

So if RSI is a risk... I think it's more from the human-in-the-loop socio-technical governance dynamics moreso than a simple "AI gets crazy smart overnight" risk.

Here we go.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
9:08 PM · May 5, 2026 · 169 Views

Guys… RSI mostly already happened, you just don’t see foom b/c we’re compute constrained. RSI is a ratio not a binary. 5000 anthropic employees wielding 45,000 agents to build better Claude is 1-(5k/(5k+45k))=90% RSI already. But if Dario fired everyone but him tomorrow it wouldn’t make Anthropic go faster. He’d just have 45,000 workers instead of 50,000 because he can’t turn that 5000 headcount into compute overnight. We’re compute constrained. RSI is a distraction RN compared to other stuff like Mythos.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
10:18 PM · May 5, 2026 · 94 Views

Actually… non-binary RSI mostly already happened, you just don’t see foom b/c we’re compute constrained. If RSI is a ratio not a binary, 5000 anthropic employees wielding 45,000 agents to build better Claude is 1-(5k/(5k+45k))=90% RSI already. But if Dario fired everyone but him tomorrow it wouldn’t make Anthropic go faster. He’d just have 45,000 workers instead of 50,000 because he can’t turn that 5000 headcount into compute overnight. We’re compute constrained. RSI is a distraction RN compared to other stuff like Mythos.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
10:21 PM · May 5, 2026 · 6.8K Views

Ah... ok I see where I may have miscommunicated. My understanding of the normative use of "recursive self improvement" is that it's a specific moment wherein humans leave the loop, and that "humans leaving the loop" would happen somewhere between two thresholds:

Min: AI is capable of doing the same task at a lower price. This is where Anthropic just does a layoff because they want to deploy that capital on something else (e.g., drop the cost of tokens to compete)

Max: the human in the loop is actively reducing AI agent output because they sleep and block saying "yes" to requests (e.g. tool use) which the model is getting right anyway... and so 45,000 agents will go faster when the 5,000 people aren't involved at all.

Maybe this is a controversial statement, but my impression is that these thresholds actually have a very loose relationship with the absolute level of increased/decreased productivity during the process.

For the "MIN" level, AI need only be as capable as the 5000 employees and cheaper. So if current Anthropic employees are 1x their normal capability with AI (so it's not helping them at all)... but they can safely take their hands off the wheel... Anthropic could decide that the AI agents can take on AI research from here... and recursive self improvement continues.

For the "MAX" level, AI need only work longer hours and make reliable enough governance decisions. But even if current Anthropic employees are 1x their normal capability when using AI... and are say... getting 30% more running it at night... Anthropic could still be like "yeah we don't have a reason to pay these people anymore" and handoff future AI research to the machines, with RSI in effect.

That said... candidly... your question (and the other thread) has pulled my headspace into a newer-ish area for me and this particular interpretation is a bit fresh. Could be something big I'm missing here...

Ryan Greenblatt@RyanPGreenblatt

@iamtrask You seem very wrong about how accelerated Anthropic employees are right now. Do you think that Anthropic employees are as productive as if they worked ~10x faster right now? I don't think so (I'd guess more like 1.5x), and neither do people at Anthropic AFAICT.

12:38 AM · May 6, 2026 · 1.5K Views
4:25 AM · May 6, 2026 · 180 Views

Inversely... and now i'm in a really unfamiliar headspace... could be the foom effects of "recursive self improvement" happen while 5000 people remain employeed doing all the prompting... because AI is accelerating their work by some huge margin... but the AI isn't good enough at governance or has jagged enough intelligence to be dependent on the human in the loop for some level of coordination and continuous value alignment.

⿻ Andrew Trask@iamtrask

Ah... ok I see where I may have miscommunicated. My understanding of the normative use of "recursive self improvement" is that it's a specific moment wherein humans leave the loop, and that "humans leaving the loop" would happen somewhere between two thresholds: Min: AI is capable of doing the same task at a lower price. This is where Anthropic just does a layoff because they want to deploy that capital on something else (e.g., drop the cost of tokens to compete) Max: the human in the loop is actively reducing AI agent output because they sleep and block saying "yes" to requests (e.g. tool use) which the model is getting right anyway... and so 45,000 agents will go faster when the 5,000 people aren't involved at all. Maybe this is a controversial statement, but my impression is that these thresholds actually have a very loose relationship with the absolute level of increased/decreased productivity during the process. For the "MIN" level, AI need only be as capable as the 5000 employees and cheaper. So if current Anthropic employees are 1x their normal capability with AI (so it's not helping them at all)... but they can safely take their hands off the wheel... Anthropic could decide that the AI agents can take on AI research from here... and recursive self improvement continues. For the "MAX" level, AI need only work longer hours and make reliable enough governance decisions. But even if current Anthropic employees are 1x their normal capability when using AI... and are say... getting 30% more running it at night... Anthropic could still be like "yeah we don't have a reason to pay these people anymore" and handoff future AI research to the machines, with RSI in effect. That said... candidly... your question (and the other thread) has pulled my headspace into a newer-ish area for me and this particular interpretation is a bit fresh. Could be something big I'm missing here...

4:25 AM · May 6, 2026 · 180 Views
4:27 AM · May 6, 2026 · 152 Views

@RyanPGreenblatt This latter seems like the most likely path for a while, but it also seems like it would get really hard to objectively measure the difference between RSI and just... normal research/engineering progress.

⿻ Andrew Trask@iamtrask

Inversely... and now i'm in a really unfamiliar headspace... could be the foom effects of "recursive self improvement" happen while 5000 people remain employeed doing all the prompting... because AI is accelerating their work by some huge margin... but the AI isn't good enough at governance or has jagged enough intelligence to be dependent on the human in the loop for some level of coordination and continuous value alignment.

4:27 AM · May 6, 2026 · 152 Views
4:32 AM · May 6, 2026 · 84 Views

Nobody really cares about AI fundamentals these days, but if you'd like to hear what AI fundamentals say about the risk of recursive self-improvement... here's a summary:

Non-technical folks can hear the phrase "AI writing code for AI" and it sounds like a meaningful feedback loop leading to infinite AI capability... and indeed RSI seems very likely as AI tools get better... but recursive self-improvement would be a lot more like a tech worker layoff (and corresponding drop in token prices) than a meaningful change in the rate of AI capability growth. Why? Scaling laws.

AI's scaling laws have been very good predictors of AI capability growth. TLDR... you need growth in data AND compute AND talent/algorithms for AI capability to make progress. A bunch of fancy new algorithms are great, but if we don't have 10x the compute and 10x the data to run them, it's not gonna mean much.

Why did deep learning replace Bayesian models (which are more sample efficient... strictly speaking better algorithms than neural networks)?... not enough compute. GPUs forced the switch. And even today... GPUs are holding the world in neural network land for the same reason. Quantum, photonics, or analog computing could undermine the whole playbook there by offering way more compute under a sparse computation model.

Why wasn't Transformer popular the day the paper was published? Not enough data/compute used for the initial experiments.

Why was LSTM (a superior, more expressive algorithm with an infinite context window) replaced by Transformer (simpler, fixed-context window)? Not enough data/compute to justify all the gates... the Transformer is a *simpler* (read: dumber) architecture. We still don't have enough compute to move past it.

Wanna know why Schmidhuber's always so grumpy? It's because a high percentage of today's algorithms were invented decades ago, but they didn't have the data/compute to try them out... were forgotten... and then some Stanford student with some H100s "rediscovers" it.

Yeah sure there's the hail-mary possibility that some boogeyman breakthrough happens because an AI searched a wider search space, but the no-free-lunch theorem and the tradjectory of technical progress of AI for the last 50 years is a pretty strong signal that that's just not how it works. Gotta have data and compute growth... and AI writing its own Jax code doesn't necessarily move the needle there.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
8:34 PM · May 5, 2026 · 122 Views

Now AI fully automating and scaling the full supply chain which extracts rare earth minerals and converts them into fully functional data centers. That would be interesting.

Or new worldwide trust infrastructure which allows the world's doormant 180 zettabytes of data to be collectively used to train AI? (as opposed to the... less than a 100 millionth o that we currently use to train aI)

Combine those ingredients with recursive self improvement and then you've got my attention.

⿻ Andrew Trask@iamtrask

Nobody really cares about AI fundamentals these days, but if you'd like to hear what AI fundamentals say about the risk of recursive self-improvement... here's a summary: Non-technical folks can hear the phrase "AI writing code for AI" and it sounds like a meaningful feedback loop leading to infinite AI capability... and indeed RSI seems very likely as AI tools get better... but recursive self-improvement would be a lot more like a tech worker layoff (and corresponding drop in token prices) than a meaningful change in the rate of AI capability growth. Why? Scaling laws. AI's scaling laws have been very good predictors of AI capability growth. TLDR... you need growth in data AND compute AND talent/algorithms for AI capability to make progress. A bunch of fancy new algorithms are great, but if we don't have 10x the compute and 10x the data to run them, it's not gonna mean much. Why did deep learning replace Bayesian models (which are more sample efficient... strictly speaking better algorithms than neural networks)?... not enough compute. GPUs forced the switch. And even today... GPUs are holding the world in neural network land for the same reason. Quantum, photonics, or analog computing could undermine the whole playbook there by offering way more compute under a sparse computation model. Why wasn't Transformer popular the day the paper was published? Not enough data/compute used for the initial experiments. Why was LSTM (a superior, more expressive algorithm with an infinite context window) replaced by Transformer (simpler, fixed-context window)? Not enough data/compute to justify all the gates... the Transformer is a *simpler* (read: dumber) architecture. We still don't have enough compute to move past it. Wanna know why Schmidhuber's always so grumpy? It's because a high percentage of today's algorithms were invented decades ago, but they didn't have the data/compute to try them out... were forgotten... and then some Stanford student with some H100s "rediscovers" it. Yeah sure there's the hail-mary possibility that some boogeyman breakthrough happens because an AI searched a wider search space, but the no-free-lunch theorem and the tradjectory of technical progress of AI for the last 50 years is a pretty strong signal that that's just not how it works. Gotta have data and compute growth... and AI writing its own Jax code doesn't necessarily move the needle there.

8:34 PM · May 5, 2026 · 122 Views
8:39 PM · May 5, 2026 · 13 Views

@davidad 100% this. I think this is absolutely a better reframe.

davidad 🎇@davidad

Instead of focusing on “AGI” or “RSI”, ask: • When will automated AI R&D be moving too fast and/or become too advanced for humans to review (without conceding speed)? • When will robots accelerate energy infrastructure and datacenter construction? • Wen Drexlerian nanosystems?

4:27 PM · May 6, 2026 · 3.2K Views
4:31 PM · May 6, 2026 · 185 Views

Instead of focusing on “AGI” or “RSI”, ask: • When will automated AI R&D be moving too fast and/or become too advanced for humans to review (without conceding speed)? • When will robots accelerate energy infrastructure and datacenter construction? • Wen Drexlerian nanosystems?

⿻ Andrew Trask@iamtrask

IMO Jack is right that RSI is imminent, but AI labs are too compute constrained right now for RSI to be a foom risk. Some napkin math: Anthropic before RSI: ~5000 employees x average ~10 agents / employee building better AI 24x7 == 50K AI agents building better AI at Anthropic Anthropic after RSI: 5 employees x average 10,000 agents / employee building better AI == 50K AI agents building better AI at Anthropic If human staff was convertible into more compute/data, maybe RSI would be more important... but if AI agents are already doing 90% of the intellectual labor anyway... shifting that to 100% just doesn't seem like the foom risk that everyone is talking about. Seems more like a risk that Anthropic lays off some of its $750,000/year engineers and drops the price of Claude tokens a bit. I could see how RSI is a governance risk tho ... probably harder to convince 5000 employees to allow bad AI than it is to convince 5 employees. So if RSI is a risk... I think it's more from the human-in-the-loop socio-technical governance dynamics moreso than a simple "AI gets crazy smart overnight" risk. Here we go.

9:11 PM · May 5, 2026 · 22K Views
4:27 PM · May 6, 2026 · 3.2K Views

Jack Clark now believes RSI has a 60% chance of happening by the end of 2028.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
2:57 PM · May 4, 2026 · 23.3K Views

From Lisan's post.

Andrew Curran@AndrewCurran_

Sorry Lisan, didn't see you posted first. So much news this morning.

2:58 PM · May 4, 2026 · 3.8K Views
2:59 PM · May 4, 2026 · 2.5K Views

how will we know if it is happening?

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
6:21 PM · May 4, 2026 · 305 Views

realistic moreover, "research taste" is likely solvable with brute force on those humongous clusters too. Just RLVR from millions of exploratory experiments. Might need some clever reward densification to definitely outpace humans, but eh. Just raise more money, buy more GPUs.

2:51 PM · May 4, 2026 · 3.6K Views

Rather unsatisfying that this is how it likely ends, but then, I'd probably see it as glorious if it were to benefit me. Titanic birth of a new species. Transcendence.

Teortaxes▶️ (DeepSeek 推特🐋铁粉 2023 – ∞)@teortaxesTex

realistic moreover, "research taste" is likely solvable with brute force on those humongous clusters too. Just RLVR from millions of exploratory experiments. Might need some clever reward densification to definitely outpace humans, but eh. Just raise more money, buy more GPUs.

2:51 PM · May 4, 2026 · 3.6K Views
2:53 PM · May 4, 2026 · 1.1K Views

when this tweet calms you because jack thinks there’s a 40% chance it takes longer

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
3:08 PM · May 4, 2026 · 6.9K Views

@jackclarkSF AI systems have been capable of building themselves since LISP was invented in the 50s. The question is whether you get increasing or diminishing returns, and so far there’s no evidence of the former.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
10:03 PM · May 4, 2026 · 3.7K Views

I think this claim that there's a ~60% chance of a frontier model 'autonomously training a successor version of itself' by 2028 is misleading. Successor models typically involve scaling up compute and data, which I don't think will be automated, nor does Clark claim it will.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
11:36 PM · May 4, 2026 · 8.6K Views

@RyanPGreenblatt I interpret 'autonomously training a successor version of itself' as autonomously producing a model end-to-end, not just doing one of many specific parts of the process.

Ryan Greenblatt@RyanPGreenblatt

@tamaybes I think he just means fully automated AI R&D and I suspect people will read this as this. It's fair enough if you don't think fully automated AI R&D is a big deal (because you think scaling up compute or data that's specifically from humans is the bottleneck), but Jack disagrees.

1:55 AM · May 5, 2026 · 1.1K Views
2:28 AM · May 5, 2026 · 250 Views

Seems right.

(as a reminder, if you think OpenAI disagrees, our stated estimate is that automated AI research will be developed around March 2028)

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
4:02 AM · May 5, 2026 · 36.3K Views

Feels material and realistic

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
8:19 PM · May 4, 2026 · 55.8K Views

@tamaybes I think he just means fully automated AI R&D and I suspect people will read this as this. It's fair enough if you don't think fully automated AI R&D is a big deal (because you think scaling up compute or data that's specifically from humans is the bottleneck), but Jack disagrees.

Tamay Besiroglu@tamaybes

I think this claim that there's a ~60% chance of a frontier model 'autonomously training a successor version of itself' by 2028 is misleading. Successor models typically involve scaling up compute and data, which I don't think will be automated, nor does Clark claim it will.

11:36 PM · May 4, 2026 · 8.6K Views
1:55 AM · May 5, 2026 · 1.1K Views

I think the chance of AIs capable of fully automating AI R&D by the end of 2028 is around 30%. So I expect things to take a bit longer than Jack does, but not by that much and timelines as fast as Jack is imagining seem totally plausible to me.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
6:47 PM · May 4, 2026 · 10.5K Views

By "AIs capable of fully automating AI R&D", I mean "progress would go faster if you fired all humans than if you could only use 2020 AIs"; the only-AI company is better than the only-human company.

Ryan Greenblatt@RyanPGreenblatt

I think the chance of AIs capable of fully automating AI R&D by the end of 2028 is around 30%. So I expect things to take a bit longer than Jack does, but not by that much and timelines as fast as Jack is imagining seem totally plausible to me.

6:47 PM · May 4, 2026 · 10.5K Views
6:47 PM · May 4, 2026 · 999 Views

I discuss my views more here: https://www.lesswrong.com/posts/dKpC6wHFqDrGZwnah/ais-can-now-often-do-massive-easy-to-verify-swe-tasks-and-i

Ryan Greenblatt@RyanPGreenblatt

It's possible Jack is using a weaker operationalization than I am, "a frontier model is able to autonomously train a successor version of itself" technically occurs somewhat earlier, but I assume Jack also means "competitively with humans".

6:47 PM · May 4, 2026 · 1.1K Views
6:55 PM · May 4, 2026 · 750 Views

@iamtrask This argument dismisses: - >>50k parallel workers - Serial speed well beyond humans - Increasing capabilities/taste beyond the human range

Where do you think the modeling at https://www.aifuturesmodel.com/ or https://www.forethought.org/research/will-ai-r-and-d-automation-cause-a-software-intelligence-explosion is wrong?

⿻ Andrew Trask@iamtrask

IMO Jack is right that RSI is imminent, but AI labs are too compute constrained right now for RSI to be a foom risk. Some napkin math: Anthropic before RSI: ~5000 employees x average ~10 agents / employee building better AI 24x7 == 50K AI agents building better AI at Anthropic Anthropic after RSI: 5 employees x average 10,000 agents / employee building better AI == 50K AI agents building better AI at Anthropic If human staff was convertible into more compute/data, maybe RSI would be more important... but if AI agents are already doing 90% of the intellectual labor anyway... shifting that to 100% just doesn't seem like the foom risk that everyone is talking about. Seems more like a risk that Anthropic lays off some of its $750,000/year engineers and drops the price of Claude tokens a bit. I could see how RSI is a governance risk tho ... probably harder to convince 5000 employees to allow bad AI than it is to convince 5 employees. So if RSI is a risk... I think it's more from the human-in-the-loop socio-technical governance dynamics moreso than a simple "AI gets crazy smart overnight" risk. Here we go.

9:11 PM · May 5, 2026 · 22K Views
10:50 PM · May 5, 2026 · 1.2K Views

@iamtrask You seem very wrong about how accelerated Anthropic employees are right now. Do you think that Anthropic employees are as productive as if they worked ~10x faster right now? I don't think so (I'd guess more like 1.5x), and neither do people at Anthropic AFAICT.

⿻ Andrew Trask@iamtrask

Actually… non-binary RSI mostly already happened, you just don’t see foom b/c we’re compute constrained. If RSI is a ratio not a binary, 5000 anthropic employees wielding 45,000 agents to build better Claude is 1-(5k/(5k+45k))=90% RSI already. But if Dario fired everyone but him tomorrow it wouldn’t make Anthropic go faster. He’d just have 45,000 workers instead of 50,000 because he can’t turn that 5000 headcount into compute overnight. We’re compute constrained. RSI is a distraction RN compared to other stuff like Mythos.

10:21 PM · May 5, 2026 · 6.8K Views
12:38 AM · May 6, 2026 · 1.5K Views

There is a mounting economic pressure coming, to make people obsolete in the training of stronger AI systems. At that point, boy had we better hope that the AI 'wants' what we want, and that we've built oversight systems that don't buckle under the pressure

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
5:46 PM · May 4, 2026 · 1.4K Views

This is as close to an acknowledgment you'll get from the labs that they've already got it (which multiple people from Anthropic told me they had achieved RSI around March)

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
11:56 PM · May 4, 2026 · 1.2K Views

I, unfortunately, have similarly short timelines.

Note that this means clearly superhuman AI capabilities since probably no **single** human today could build a frontier model start-to-finish by themselves anymore.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
7:50 PM · May 4, 2026 · 32.2K Views

Idunno MAYBE just maybe we’re jumping to conclusions with a certainty that isn’t warranted

12:53 PM · May 5, 2026 · 8.7K Views

For people who are extremely skeptical of AI being able to automate AI R&D, what specifically is the magic spark that humans have that AI will never have that's relevant to AI research? If you imagine the typical AI researcher today, what knowledge or abilities are in their brain that AI will never approach? If AI achieved those abilities, we could effectively hire gigantic numbers of AI researchers, way more than the human researchers we have access to.

Maybe you disagree with the timeline, or you believe the bottlenecks in AI research are elsewhere, but I find that a lot of people just treat this scenario as sci fi because it "feels" crazy, but don't often have a solid reason it couldn't happen.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
10:29 PM · May 4, 2026 · 16.2K Views

IDK seems right to me and I did think about it for a really long time!

In general I am kind of confused by this move here. Would you make the same critique of a scientist before the bomb warning someone that if they build the bomb they will uncover destructive ability beyond what anyone previously thought possible and develop a weapon so powerful as to potentially end human civilization?

The scientists knew in advance how big of a deal nukes would be. You can know in advance how these things go.

Austen Allred@Austen

Idunno MAYBE just maybe we’re jumping to conclusions with a certainty that isn’t warranted

12:53 PM · May 5, 2026 · 8.7K Views
4:53 PM · May 5, 2026 · 878 Views

Fully automated AI R&D: ~30% chance by the end of 2027, ~60%+ chance by the end of 2028

Overall, Anthropic's Jack Clark has written a very worthwhile essay: His timeline is that fully automated AI R&D probably won’t arrive in 2026, but we may see a proof-of-concept within 1–2 years where an AI system can end-to-end train a non-frontier successor model, with a much more serious possibility of frontier-level automated AI R&D by 2027–2028.

His headline forecast is: ~30% chance by the end of 2027, ~60%+ chance by the end of 2028 that a frontier AI system can autonomously build its own successor, driven by rapid gains in coding, long-horizon agent work, benchmark saturation, AI-managed subagents, and early signs of models handling core AI research tasks like fine-tuning, kernel optimization, reproducibility, and alignment research.

Chubby♨️@kimmonismus

Anthropics Jack Clarke now believes that recurse self-improvement has a 60% change of happening by end of 2028.

3:28 PM · May 4, 2026 · 68.5K Views
5:51 PM · May 4, 2026 · 35.2K Views

Anthropics Jack Clarke now believes that recurse self-improvement has a 60% change of happening by end of 2028.

Jack Clark@jackclarkSF

I've spent the past few weeks reading 100s of public data sources about AI development. I now believe that recursive self-improvement has a 60% chance of happening by the end of 2028. In other words, AI systems might soon be capable of building themselves.

2:47 PM · May 4, 2026 · 1.2M Views
3:28 PM · May 4, 2026 · 68.5K Views