Opinion

If Mythos actually made Anthropic employees 4x more productive, I would radically shorten my timelines

​Anthropic’s system card for Mythos Preview says:

It’s unclear how we should interpret this. What do they mean by productivity uplift? To what extent is Anthropic’s institutional view that the uplift is 4x? (Like, what do they mean by “We take this seriously and it is consistent with our own internal experience of the model.”)
One straightforward interpretation is: AI systems improve the productivity of Anthropic so much that Anthropic would be indifferent between the current situation and a situation where all of their technical employees magically work 4 hours for every 1 hour (at equal productivity without burnout) but they get zero AI assistance.
In other words, AI assistance is as useful as having their employees operate at 4x faster speeds for all activities (meetings, coding, thinking, writing, etc.) I’ll call this “4x serial labor acceleration”
[1]
(see here for more discussion of this idea
[2]
).
I currently think it’s very unlikely that Anthropic’s AIs are yielding 4x serial labor acceleration, but if I did come to believe it was true, I would update towards radically shorter timelines. (I tentatively think my median to Automated Coder would go from 4 years from now to maybe 1.3 years from now; my median to AI R&D parity would go from 5 years from now to maybe 2.5 years from now.) My best guess is that 4x serial labor acceleration would cause AI progress to go 1.75x faster (see “Appendix: Estimating AI progress speed up from serial labor acceleration”) which is very large and close to the 2x “dramatic acceleration” threshold Anthropic is using for “Autonomy threat model 2: risks from automated R&D”.
[3]

My current best (low confidence, low precision) guess for the serial labor acceleration is ~1.55x (with a higher serial labor acceleration of ~1.75x for just research engineering activities). I currently think that reasonably informed Anthropic employees that have thought about this topic in a decent amount of detail think the serial labor acceleration is closer to 1.5x than 4x.
I think uplift metrics like “serial labor acceleration” at AI companies are some of the most relevant metrics to track when trying to figure out how close we are to key risk-relevant milestones in AI development like full automation of AI R&D.
I also think uplift metrics are among the most relevant metrics for Anthropic’s “Autonomy threat model 2: risks from automated R&D”.
I also think accurately capturing the views of employees, managers, and leadership at AI companies (probably with something like a survey) is currently one of our best ways of assessing serial labor acceleration (or other uplift metrics), especially for AI systems that aren’t publicly deployed.
Thus, I’m pretty unhappy about a situation in which:

Anthropic seemingly claims they are getting 4x productivity uplift, but it’s publicly unclear what they mean by this or how much they believe this.
There is virtually no public information about the details of the survey or how seriously this was done.
Their statements are consistent with a very radical situation.
This is approximately the only direct public evidence we have about the level of acceleration (as in, evidence that doesn’t require doing some kind of extrapolation from our views about prior AIs).
This is all happening for an AI that’s not publicly released, is not going to be publicly released, and appears to be much better than the publicly available frontier (making extrapolation harder).

Some things that would improve the situation in future system cards / risk reports:

Generally saying more about the exact details of the survey. For instance: What was the question? How long did people spend answering?
Insofar as Anthropic doesn’t think this survey is very meaningful or thinks this was a very low effort survey, say so. Alternatively, if they don’t think this sort of survey sheds much light, it would be reasonable to not include this in system cards going forward.
Clarifying their institutional view and the view of key individuals with relatively precise operationalizations. Like, what does Jared Kaplan think the level of serial labor acceleration is? (Or whatever operationalization Anthropic would like to use.)

Ideally, they would also explain why they think this even if the evidence is relatively illegible or some of it needs to be redacted.

Insofar as Anthropic thinks the survey is mostly capturing vibes in a way that doesn’t depend much on the operationalization in the survey, this seems important to note.
Ideally some third party would instead do surveys/interviews of employees, managers, and/or leadership and these results would be reported. It seems like Anthropic isn’t that interested in doing carefully done surveys (fair enough!) and it would be useful to standardize this across AI companies.
I do think it would be possible for Anthropic to collect information or do surveys that do shed light on this question. I’d be most interested in a survey of the views of employees for whom we’re very confident that employee has a detailed understanding of exactly what is meant by different uplift notions and is interested in AI forecasting. (I’m much more interested in capturing the views of employees who have thought a decent amount about this than capturing an unbiased sample.) It would also be possible to do more qualitative and quantitative data collection from various employees and then convert this into acceleration estimates taking into account things like people using AIs to do low value tasks they wouldn’t have done otherwise.

If there is a large disagreement about the current level of uplift, this seems like a particularly tractable empirical crux: I would substantially shorten my timelines if I learned the uplift was much higher than I expect, and I’d guess some people at Anthropic would lengthen theirs if they learned it was significantly lower than they expect. I also expect that various people who are much more skeptical than me of reaching very high levels of AI capability within the next 10 years would update some on credible internal uplift measurements. Getting better empirical information about the level of uplift seems hard but doable.
Additionally, Anthropic claims “We estimate that reaching 2× on overall progress via this channel would require uplift roughly an order of magnitude larger than what we observe.” Insofar as “productivity uplift” is supposed to correspond to something like serial labor acceleration, I’m very skeptical. I think ~40x serial labor acceleration would yield much more than 2x faster progress. My guess (see “Appendix: Estimating AI progress speed up from serial labor acceleration”) is that you’d get 2x overall AI progress at around 5x serial labor acceleration. My understanding is that the AI Futures Project timelines model would indicate that around 8x serial labor acceleration is required. It seems that Anthropic might have their own takeoff speeds / timelines model that differs substantially from current public modeling, produces much less conservative conclusions about the level of concern, and that they are using for decision making. If so, I think they should either publicly write up their modeling (informally would be fine) or get third parties to review it privately. Insofar as they mean “we think we’ll maybe reach 2x overall progress when our survey—that’s mostly capturing vibes and doesn’t have a clear correspondence to any particular notion of uplift—reaches 40x”, fair enough, but it seems good to clarify this.
The current state of our evidence about AI R&D acceleration from Mythos seems extremely limited and AI companies should (and can) do much better going forward.
[4]

Appendix: Estimating AI progress speed up from serial labor acceleration

Suppose we had a serial labor acceleration of X (as in, employees go X times faster) and also increased experiment compute by X. Then, AI R&D progress would go X times faster.

I mean instantaneous progress, putting aside diminishing returns to research effort. Equivalently, the “research effort per unit time” would go up by X.
This is also putting aside parallel compute being worse than serially faster compute, though I think this doesn’t make a huge difference in practice.

So, production is some function of serial labor acceleration and experiment compute. We’re uncertain about the exact function between something more like a CES model and a Cobb-Douglas production function. I happen to think it’s more like Cobb-Douglas than CES for reasons I discuss here.
I tend to think the functional form for just AI R&D progress (like algorithmic progress) is like serial_labor_acceleration^0.55 * compute^0.45.

It might be pretty different as you start growing these values by orders of magnitude (especially if it’s very CES-like), but at least if we’re talking about <30x increases to these variables, I think it’s something like this.
I’m uncertain about the exact constants; serial_labor_acceleration^0.7 * compute^0.3 and serial_labor_acceleration^0.3 * compute^0.7 are somewhat plausible and make a pretty big difference to the bottom line.

So, if you get a serial labor acceleration of 4x, I think this increases AI R&D progress by ~2.15x.
AI R&D is only a subset of AI progress; some of the AI progress is driven by scaling up compute for training runs. I tend to think that ~2/3 of AI progress is algorithms while ~1/3 is from scaling up compute for training runs. This means you get only 2/3 * 2.15 + 1/3 = 1.75x AI progress increase from 4x serial labor acceleration.
To get a 2x increase in the rate of AI progress (assuming these constants), we’d need ~5.3x serial labor acceleration.

This model is basically a simplified version of the AI Futures Project model with somewhat different constants.
Appendix: Different notions of uplift
There are several different concepts that could be meant by “productivity uplift”, and which one we’re talking about makes a huge difference:

Serial labor acceleration: Suppose you could speed up everyone at the company by X but had to use no AI assistance (or only 2020 AIs) in your work. For what X would you be indifferent? (Just taking into account productivity, ignoring safety.)
Parallel labor acceleration: Suppose you could magically grow the company by a factor of X, where the new people would have a similar distribution of skills and knowledge to the current people (including knowledge about the company, etc.), but had no AI assistance. For what X would you be indifferent?
Current work acceleration: If the company did all the same work it did last week, but had no AI assistance, how much slower would it have been?
Fraction of work done by AIs: There are different ways to operationalize this and it’s a bit confusing because humans might be spending a bunch of time gaining context so they can (e.g.) tell AIs what to do, and it’s unclear what fraction of the work to count this as. Things like fraction of lines written by AI would be an example of this and it seems hard to convert this number into a guess at serial labor acceleration (or other notions we might care about).

A parallel labor acceleration of X is much less useful than a serial labor acceleration of X. And depending on the operationalization, the AIs doing 90% of the work is way less useful than a 10x serial labor acceleration. So the choice of concept matters a lot for interpreting any claimed level of uplift.
I’m using a specific name to distinguish from other things we might call “4x productivity uplift” like “if the median employee had to do the tasks they are currently doing without the use of AI, they would be 4x slower”. These notions have strongly different implications as I discuss here and in Appendix: Different notions of uplift. ↩︎

For reference, the speed up modeling I do in that post is out of date with my latest thinking. ↩︎

The update toward shorter timelines is almost entirely from thinking we’re further along in the capability progress than I previously realized, rather than from thinking progress will be faster but we’re starting from a similar point. As in, I both update towards getting more acceleration at a lower level of capability and towards models being closer in capability space towards various high milestones, and I’m mostly updating timelines based on the second of these. ↩︎

I think there are also some other issues in the system card’s assessment of AI R&D acceleration. They seemingly argue that even if Mythos was substantially above trend due to AI acceleration, because this acceleration was done by earlier (less capable!) AIs, this would imply this Mythos-caused acceleration wouldn’t be that high: “This means that even if the slope change were AI-attributable, the model it would implicate is not the one we are assessing.” This seems backward: if less capable AIs yield a large acceleration, then we should expect the effect from more capable AIs to be even larger. To be clear, this seems like a minor/moderate issue, I just thought it was worth mentioning. ↩︎

Discuss ​Read More

​Anthropic’s system card for Mythos Preview says:

It’s unclear how we should interpret this. What do they mean by productivity uplift? To what extent is Anthropic’s institutional view that the uplift is 4x? (Like, what do they mean by “We take this seriously and it is consistent with our own internal experience of the model.”)
One straightforward interpretation is: AI systems improve the productivity of Anthropic so much that Anthropic would be indifferent between the current situation and a situation where all of their technical employees magically work 4 hours for every 1 hour (at equal productivity without burnout) but they get zero AI assistance.
In other words, AI assistance is as useful as having their employees operate at 4x faster speeds for all activities (meetings, coding, thinking, writing, etc.) I’ll call this “4x serial labor acceleration”
[1]
(see here for more discussion of this idea
[2]
).
I currently think it’s very unlikely that Anthropic’s AIs are yielding 4x serial labor acceleration, but if I did come to believe it was true, I would update towards radically shorter timelines. (I tentatively think my median to Automated Coder would go from 4 years from now to maybe 1.3 years from now; my median to AI R&D parity would go from 5 years from now to maybe 2.5 years from now.) My best guess is that 4x serial labor acceleration would cause AI progress to go 1.75x faster (see “Appendix: Estimating AI progress speed up from serial labor acceleration”) which is very large and close to the 2x “dramatic acceleration” threshold Anthropic is using for “Autonomy threat model 2: risks from automated R&D”.
[3]

My current best (low confidence, low precision) guess for the serial labor acceleration is ~1.55x (with a higher serial labor acceleration of ~1.75x for just research engineering activities). I currently think that reasonably informed Anthropic employees that have thought about this topic in a decent amount of detail think the serial labor acceleration is closer to 1.5x than 4x.
I think uplift metrics like “serial labor acceleration” at AI companies are some of the most relevant metrics to track when trying to figure out how close we are to key risk-relevant milestones in AI development like full automation of AI R&D.
I also think uplift metrics are among the most relevant metrics for Anthropic’s “Autonomy threat model 2: risks from automated R&D”.
I also think accurately capturing the views of employees, managers, and leadership at AI companies (probably with something like a survey) is currently one of our best ways of assessing serial labor acceleration (or other uplift metrics), especially for AI systems that aren’t publicly deployed.
Thus, I’m pretty unhappy about a situation in which:

Anthropic seemingly claims they are getting 4x productivity uplift, but it’s publicly unclear what they mean by this or how much they believe this.
There is virtually no public information about the details of the survey or how seriously this was done.
Their statements are consistent with a very radical situation.
This is approximately the only direct public evidence we have about the level of acceleration (as in, evidence that doesn’t require doing some kind of extrapolation from our views about prior AIs).
This is all happening for an AI that’s not publicly released, is not going to be publicly released, and appears to be much better than the publicly available frontier (making extrapolation harder).

Some things that would improve the situation in future system cards / risk reports:

Generally saying more about the exact details of the survey. For instance: What was the question? How long did people spend answering?
Insofar as Anthropic doesn’t think this survey is very meaningful or thinks this was a very low effort survey, say so. Alternatively, if they don’t think this sort of survey sheds much light, it would be reasonable to not include this in system cards going forward.
Clarifying their institutional view and the view of key individuals with relatively precise operationalizations. Like, what does Jared Kaplan think the level of serial labor acceleration is? (Or whatever operationalization Anthropic would like to use.)

Ideally, they would also explain why they think this even if the evidence is relatively illegible or some of it needs to be redacted.

Insofar as Anthropic thinks the survey is mostly capturing vibes in a way that doesn’t depend much on the operationalization in the survey, this seems important to note.
Ideally some third party would instead do surveys/interviews of employees, managers, and/or leadership and these results would be reported. It seems like Anthropic isn’t that interested in doing carefully done surveys (fair enough!) and it would be useful to standardize this across AI companies.
I do think it would be possible for Anthropic to collect information or do surveys that do shed light on this question. I’d be most interested in a survey of the views of employees for whom we’re very confident that employee has a detailed understanding of exactly what is meant by different uplift notions and is interested in AI forecasting. (I’m much more interested in capturing the views of employees who have thought a decent amount about this than capturing an unbiased sample.) It would also be possible to do more qualitative and quantitative data collection from various employees and then convert this into acceleration estimates taking into account things like people using AIs to do low value tasks they wouldn’t have done otherwise.

If there is a large disagreement about the current level of uplift, this seems like a particularly tractable empirical crux: I would substantially shorten my timelines if I learned the uplift was much higher than I expect, and I’d guess some people at Anthropic would lengthen theirs if they learned it was significantly lower than they expect. I also expect that various people who are much more skeptical than me of reaching very high levels of AI capability within the next 10 years would update some on credible internal uplift measurements. Getting better empirical information about the level of uplift seems hard but doable.
Additionally, Anthropic claims “We estimate that reaching 2× on overall progress via this channel would require uplift roughly an order of magnitude larger than what we observe.” Insofar as “productivity uplift” is supposed to correspond to something like serial labor acceleration, I’m very skeptical. I think ~40x serial labor acceleration would yield much more than 2x faster progress. My guess (see “Appendix: Estimating AI progress speed up from serial labor acceleration”) is that you’d get 2x overall AI progress at around 5x serial labor acceleration. My understanding is that the AI Futures Project timelines model would indicate that around 8x serial labor acceleration is required. It seems that Anthropic might have their own takeoff speeds / timelines model that differs substantially from current public modeling, produces much less conservative conclusions about the level of concern, and that they are using for decision making. If so, I think they should either publicly write up their modeling (informally would be fine) or get third parties to review it privately. Insofar as they mean “we think we’ll maybe reach 2x overall progress when our survey—that’s mostly capturing vibes and doesn’t have a clear correspondence to any particular notion of uplift—reaches 40x”, fair enough, but it seems good to clarify this.
The current state of our evidence about AI R&D acceleration from Mythos seems extremely limited and AI companies should (and can) do much better going forward.
[4]

Appendix: Estimating AI progress speed up from serial labor acceleration

Suppose we had a serial labor acceleration of X (as in, employees go X times faster) and also increased experiment compute by X. Then, AI R&D progress would go X times faster.

I mean instantaneous progress, putting aside diminishing returns to research effort. Equivalently, the “research effort per unit time” would go up by X.
This is also putting aside parallel compute being worse than serially faster compute, though I think this doesn’t make a huge difference in practice.

So, production is some function of serial labor acceleration and experiment compute. We’re uncertain about the exact function between something more like a CES model and a Cobb-Douglas production function. I happen to think it’s more like Cobb-Douglas than CES for reasons I discuss here.
I tend to think the functional form for just AI R&D progress (like algorithmic progress) is like serial_labor_acceleration^0.55 * compute^0.45.

It might be pretty different as you start growing these values by orders of magnitude (especially if it’s very CES-like), but at least if we’re talking about <30x increases to these variables, I think it’s something like this.
I’m uncertain about the exact constants; serial_labor_acceleration^0.7 * compute^0.3 and serial_labor_acceleration^0.3 * compute^0.7 are somewhat plausible and make a pretty big difference to the bottom line.

So, if you get a serial labor acceleration of 4x, I think this increases AI R&D progress by ~2.15x.
AI R&D is only a subset of AI progress; some of the AI progress is driven by scaling up compute for training runs. I tend to think that ~2/3 of AI progress is algorithms while ~1/3 is from scaling up compute for training runs. This means you get only 2/3 * 2.15 + 1/3 = 1.75x AI progress increase from 4x serial labor acceleration.
To get a 2x increase in the rate of AI progress (assuming these constants), we’d need ~5.3x serial labor acceleration.

This model is basically a simplified version of the AI Futures Project model with somewhat different constants.
Appendix: Different notions of uplift
There are several different concepts that could be meant by “productivity uplift”, and which one we’re talking about makes a huge difference:

Serial labor acceleration: Suppose you could speed up everyone at the company by X but had to use no AI assistance (or only 2020 AIs) in your work. For what X would you be indifferent? (Just taking into account productivity, ignoring safety.)
Parallel labor acceleration: Suppose you could magically grow the company by a factor of X, where the new people would have a similar distribution of skills and knowledge to the current people (including knowledge about the company, etc.), but had no AI assistance. For what X would you be indifferent?
Current work acceleration: If the company did all the same work it did last week, but had no AI assistance, how much slower would it have been?
Fraction of work done by AIs: There are different ways to operationalize this and it’s a bit confusing because humans might be spending a bunch of time gaining context so they can (e.g.) tell AIs what to do, and it’s unclear what fraction of the work to count this as. Things like fraction of lines written by AI would be an example of this and it seems hard to convert this number into a guess at serial labor acceleration (or other notions we might care about).

A parallel labor acceleration of X is much less useful than a serial labor acceleration of X. And depending on the operationalization, the AIs doing 90% of the work is way less useful than a 10x serial labor acceleration. So the choice of concept matters a lot for interpreting any claimed level of uplift.
I’m using a specific name to distinguish from other things we might call “4x productivity uplift” like “if the median employee had to do the tasks they are currently doing without the use of AI, they would be 4x slower”. These notions have strongly different implications as I discuss here and in Appendix: Different notions of uplift. ↩︎

For reference, the speed up modeling I do in that post is out of date with my latest thinking. ↩︎

The update toward shorter timelines is almost entirely from thinking we’re further along in the capability progress than I previously realized, rather than from thinking progress will be faster but we’re starting from a similar point. As in, I both update towards getting more acceleration at a lower level of capability and towards models being closer in capability space towards various high milestones, and I’m mostly updating timelines based on the second of these. ↩︎

I think there are also some other issues in the system card’s assessment of AI R&D acceleration. They seemingly argue that even if Mythos was substantially above trend due to AI acceleration, because this acceleration was done by earlier (less capable!) AIs, this would imply this Mythos-caused acceleration wouldn’t be that high: “This means that even if the slope change were AI-attributable, the model it would implicate is not the one we are assessing.” This seems backward: if less capable AIs yield a large acceleration, then we should expect the effect from more capable AIs to be even larger. To be clear, this seems like a minor/moderate issue, I just thought it was worth mentioning. ↩︎

Discuss ​Read More

Leave a Reply

Your email address will not be published. Required fields are marked *