antistatic.exchange
Register Log in
Back to feed

As with the Anthropic question, the requirement of an official public report that explicitly mentions a capability threshold being met is reasonable for the sake of a clear resolution, but introduces noise with respect to forecasts about the underlying reality. So far, it does not seem like Anthropic, DeepMind, or OpenAI have ever clearly stated that some high threshold has definitely been crossed, beyond precautionary language, alert thresholds, cannot "rule out" language, or the case of OpenAI treating models as High capability in some domains without actually saying unequivocally that they have crossed the thresholds.

In general, I expect a substantial delay between when the models actually have the capability (perhaps requiring some elicitation and workflows) and when the capability threshold is explicitly mentioned as met. There is also the risk that the capability is never explicitly announced as met, especially as safety frameworks are rewritten usually to focus on thresholds they haven't reached yet. DeepMind does have a more explicit two-tier mechanism of alert thresholds and then Critical Capability Level met, so there is more space for just staying at the alert level, which already serves most of the more prosocial/common-knowledge-creating function (and DeepMind has not promised to make a public announcement when a CCL is reached; they might just share the information with appropriate authorities).

Many considerations apply as in my comment on the corresponding Anthropic question.

The thresholds are quite vague (which creates a large part of the uncertainty), but:

DeepMind's CBRN Uplift 1 seems roughly analogous to Anthropic's CBRN-3, and to OpenAI's High biological/chemical capability. Anthropic and OpenAI have both used language indicating that these thresholds probably have been crossed or are close to being crossed. Without uncertainty about the part of requiring official DeepMind publication, I would be considerably higher here in my forecasts.

Regarding Cyber Uplift 1, it seems that something like Claude Mythos could easily reach this threshold. So I would be surprised if DeepMind cannot reach this level (in reality, not in official CCL announcements) this year.

Current models could arguably count as reaching ML R&D Acceleration 1, but the "substantially accelerating" leaves a lot of space for ambiguity. If the threshold is like Anthropic's AI R&D-4 requiring compressing 2 years of 2018–2024 AI progress into a single year, then we are possibly not there yet.

ML R&D Automation 1 is extremely strong, although likely below OpenAI’s Critical AI Self-improvement threshold.


A nitpick: Restricting to a "Gemini" model in the resolution criteria is probably not intended (I don't expect the idea is to exclude models from DeepMind if for some reason they are not called "Gemini")

0 replies

No replies yet.