Speculating on Secret Intelligence Explosions

Cross-posted from Substack. This is a “takes” post and may well not meet your standard for rigour.


In the next few years, I expect AI companies to build AI systems that dramatically speed up their AI research, creating a period of extremely fast growth in capabilities, where AI systems surpass human intelligence in every interesting domain (i.e. an intelligence explosion).[1]

If this happens, I think it will be viable and potentially desirable for an AI company to keep its intelligence explosion secret, even if the project is not securitised. I’m concerned that people outside of AI companies won’t know that an intelligence explosion is happening, they’ll be cut out of high-stakes decisions, and they won’t be able to orient appropriately to a world with superintelligence.

In this post, I’ll discuss:

  • Why AI companies might want to do an intelligence explosion in secret.

  • Why AI workforces might make it much cheaper to keep large engineering projects a secret.

  • How AI companies could finance a secret intelligence explosion.

  • Whether employees will be able to effectively raise the alarm.

Overall, I think that a secret intelligence explosion is somewhat unlikely, but plausible and concerning.

Why might an AI company want to keep its intelligence explosion secret?

AI companies rarely have internal models that are much better than public models, so why might this change as AI systems become increasingly powerful?

  • AI companies will be less focused on consumers: AI companies will primarily use their intelligence-explosion-enabling AI systems (e.g. automated scientists) to accelerate their AI research. They’ll be less focused on selling their most cutting-edge models to millions of people (though they may still serve less capable or narrow models to generate revenue).

  • Superintelligence will create a huge strategic advantage for the country that builds it first. There’ll be a lot of pressure from the host country to use the AI in the interest of national security, and by default large military projects are secretive.

  • Alternatively, AI companies may not want to be securitised, so may want to hide their progress.[2]

  • It would be particularly bad for adversary states to steal an intelligence explosion enabling AI system, as they could then quickly close the gap to the leading AI company.[3] AI companies may adopt significantly more aggressive security, which may push towards more secrecy around AI capabilities and value of their IP.

  • Companies want their competitors to stop racing with them, and showing that they are in an intelligence explosion may push other companies to double down due to winner-takes-all dynamics.

  • AI companies may be worried about their governments shutting them down or forcing them to slow down if there’s public panic around the catastrophic potential or job loss due to AI systems. Or, they might be worried about being nationalised and losing control of their model to their government, particularly as powerful AI systems could be used to seize power from their government.

  • Even if things stay in roughly the status quo, AI development will likely outpace external oversight, especially if oversight bodies lack access to frontier systems. Even if you think that frontier labs will let AISIs get uplifted by AIs they are testing (or models that are only marginally weaker, it seems plausible that AISIs won’t want to use a model that might be subtly sabotaging their work.

If AIs automating AI research produce extremely powerful AI systems after a few months, then it’s at least plausible that AI companies will want to do it in secret.

Common Objections to the Feasibility of Keeping an Intelligence Explosion Secret

  • Customer interactions might be required to train powerful models, which would necessitate serving an up-to-date model to many customers.

  • Foreign intelligence agencies (particularly in China) would likely discover the development anyway, so there might be less pressure from governments to run a covert project.

  • The risk of employees leaking information or whistleblowing increases with the magnitude of the development and the number of people who are aware of the development. Building superintelligence will require a huge workforce and is a big deal.

  • The massive computational and energy requirements will be difficult to hide. See the following image of a satellite image of the Stargate project. The bright green rectangle in the upper right corner is the size of a football field.

  • AI companies will need to raise capital continually, which is much easier if you are building in public.

One of the Stargate data centre megasites currently under construction from @nunosempere. The green thing in the top right is the size of a football field.

These are all reasonable objections and make me think a secret intelligence explosion is less likely, but I don’t think any of them are damning.

Three can keep a secret if two are AIs

I occasionally speak to people who think AI companies couldn’t do an intelligence explosion secretly, even if they wanted to. I think their model is roughly:

  1. An AI company would need to invest substantial resources to create an intelligence explosion.

  2. So internally, a lot of people would need to be aware that the company is massively accelerating.

  3. AI company employees don’t take secrecy or security particularly seriously, and it’s really hard to keep a secret if hundreds of people are aware of it.

This seems wrong if you think (as I do) that an intelligence explosion will be AI-driven, i.e. making substantive use of drop-in replacements for ML engineers and researchers.

An AI workforce may be better at keeping secrets than a human one

Imagine that you are a researcher in North Korea working on an extremely secret military project as part of a large team.[4] The project’s leadership can take any actions they deem necessary to prevent the US from discovering the project.

Working there is frustrating:

  • You’re unable to leave the facility for the duration of the project, which could last several years.

  • You frequently do polygraphs and are interrogated to make sure that you don’t blow the whistle to the several other interested governments.

  • All of your conversations with other team members are monitored.

  • Anything you write down is sent to a team that tries to figure out whether you are trying to betray the interests of the project.

  • You suspect that the news you hear about other parts of the project is fake to stop you from knowing too much about the overall goals of the project.

I think many autocracies would eagerly adopt these extreme measures if they were affordable and didn’t incur large productivity losses. I expect that all of these measures and more will be relatively inexpensive with AIs because we don’t grant AI systems rights, and because their software nature provides them with many secrecy-preserving features.

For example, AI systems are currently unable to leave their workplace, can be interrogated much more invasively than humans, can be comprehensively monitored, erasing their memories is trivial, and they’ll do whatever kinds of malign or uncomfortable tasks you want.[5][6]

Suppose that an AI lab wanted to keep an intelligence explosion secret, what might they do?

  • Replace a large fraction of their capabilities workforce, particularly junior employees, with AI workers (and assign these workers to teams outside the intelligence explosion, to prevent a suspicious exit of many employees).[7] I think it’ll be viable to have a very small team of research scientists directing thousands of AI workers in sophisticated hierarchies with colossal compute budgets (for spinning up AI workers and model training).

  • Use AI systems to monitor all internal computer use to catch any whistleblowing attempts.

  • Use AI systems to profile employees to make sure that particularly plausible whistleblowers aren’t working on sensitive projects.

  • Silo teams to prevent many lab employees having situational awareness, whilst maintaining productivity by giving employees AI assistants which can cheaply and reliably elevate employee information permissions while keeping their baseline permissions and general knowledge about capability progress low.[8]

I’m sure that there are many other creative ways you could use the features of an AI workforce to increase secrecy. If you’re struggling to come up with strategies, try prompting an LM for ideas.

How will AI companies keep the GPUs on?

So why aren’t AI companies building in secret right now? Some hypotheses:

  1. Founder preferences—I think Sam Altman and Mark Zuckerberg are genuinely into building in public, moving fast, shipping cool stuff etc.

  2. Hiring—having a public and successful product makes it much easier to hire people. The talent density at AI companies is really high; most people can get a decent salary elsewhere if they want to, but there’s a kind of status that working at a company building something important and well-known gets you. A more straightforward explanation is that people are just more likely to remember to have a look at the jobs board when they are looking to move on.

  3. Fundraising—investors care a lot about user growth and revenue.[9] Most AI companies are hoping to raise billions of dollars. It’s hard to have 100M users if your product is a secret.

I’m not sure what will happen to founder preferences at the start of an intelligence explosion. I think it’s likely that founders change their minds, given the large pressures to be more secretive and the allure of potentially moving more quickly if they adopt a more secretive posture. Hiring will be less important if increasingly complex tasks are able to be delegated to AI systems, which leaves fundraising. I think the story here is more confusing.

Governments can clearly finance megaprojects when they are sufficiently bought in (e.g. bridges, nuclear power plants). If the US government wants to bankroll an ASI megaproject, I expect they’d be able to find the capital. If there’s sufficient justification to do this, then they’ll likely also be sufficient justification to secure the project and get some more insight into its operations, which will result in the public and some parts of government having some visibility into the AI project. That said, without significant action, I’d be surprised if all important government decision-makers were well-informed, particularly if the project is securitised.

So, how viable will it be for companies to raise billions of dollars in private capital, without releasing or publicly disclosing their best models?

I’m not sure how much an intelligence explosion will cost, but prices in the range of $70M to $2T seem quite plausible to me from holding training costs fixed and naively extrapolating. Let’s take $10B-100B, which is somewhere in the middle (in log space), as our mainline scenario.

Two key questions are:

  • How convinced will investors be by the value proposition?

  • How much money can one raise from private investors for ambitious mega projects, given investor conviction?

I had a quick look at privately funded megaprojects, some of the larger infrastructure projects included Global telecommunications networks ($50+ billion each),[10] the UK-France Channel Tunnel ($21 billion adjusted to current value),[11] and the Gorgon LNG project ($54 billion).[12] Some record-breaking financing includes Saudi Aramco IPO ($30B raised) and Verizon’s 2013 mega-bond ($3B).[13]

Whilst the value proposition is less clear than traditional infrastructure projects, it’s not a mystery either. Millions of people pay for AI products, and code assistants are significantly improving the productivity of software engineers. It’s perhaps not so surprising that Open AI, which released the fastest-growing app of all time, were able to raise $40B in March this year.

They could, of course, continue to charge for less capable models, which might be much cheaper to run and require far lower levels of security than intelligence explosion enabling models. Alternatively, they could partner with industries that have an extremely high ROI on intellectual labour per person (e.g., quantitative finance) and provide them with limited access to frontier models or specialised but less generally capable models, or build products or even compete directly in those industries.

How easily can an employee raise the alarm?

It may be that staff at AI companies are so wrapped up in NDAs and social incentives (particularly if AI is taken seriously as a threat or asset to national security) that they don’t blow the whistle, but let’s suppose that they do want to blow the whistle.

Suppose that an AI researcher tweets, “The intelligence explosion has started, this is crazy, I’ve just seen x”. What are some candidates for x that would convince interested members of the public? I could see claims like “our AI can now design bioweapons” or “our AI is capable of massively speeding up research” to be brushed aside as AI hype, given the steady stream of people overhyping AI capabilities.[14] I’m a little more optimistic about claims like “our AI caused harm in the real world” being taken seriously. Claims like “look at the massive data centers being built” sound compelling, but massive data centres are already being built, I’m not sure what we expect to see change—maybe someone crunches the numbers and notices that there is way too much inference compute to justify the amount of publicly served models?[15]

If the intelligence explosion only needs to last a few months, will society really notice and reorient in time?[16] Most of the data centre build-out would happen before the intelligence explosion starts. The data centres used for serving weak models to consumers and the data centres used for powerful internal AI agents look very similar. Maybe you think that evals will make it very clear the intelligence explosion has started, but AI companies might simply choose not to run them, or sandbag on them, and there likely won’t be a tonne of public outrage if the model hasn’t been publicly released—how would we even know to be outraged?

  1. ^

    In this post I’m not arguing that there will be an intelligence explosion in the next few years, the main take is if there is an intelligence explosion in the next few years it may well be done secretly. I like this detailed illustration of intelligence explosion happening in the next few years.

  2. ^

    E.g. because of preferences of the CEO or employees. Note that CEO preferences also sometimes push towards transparency.

  3. ^

    So long as they have enough inference compute to run the model.

  4. ^

    To be clear this is just a hypothetical, I know very little about how covert and complex military technologies are developed in North Korea.

  5. ^

    In the sense that all inputs, internal states, and outputs can be stored and looked at, not that we really understand what’s going on in their minds.

  6. ^

    Insofar as you can check they are “actually” doing those tasks.

  7. ^

    Even if people are just fired I’m not sure that that intelligence explosion alarm is raised when there are easily preventable counter-narratives like “we wanted to strip our team back and increase talent density like Elon did at Twitter/​X”.

  8. ^

    One reviewer was sceptical because they thought that labs would be respectful of employee preferences, and companies have high inertia preventing them for pivoting to this agent focussed strategy. I think these are reasonably counterarguments but they are fairly small updates for me on the viability of setting up a large AI agent powered team with a substantial fraction of total compute if senior AI company leadership want to do that.

  9. ^
  10. ^

    Major telecommunications companies (AT&T, Verizon, and Vodafone) built extensive global networks primarily financed through corporate debt, equity offerings, and reinvested company profits.

  11. ^

    One of history’s most ambitious privately funded infrastructure endeavors funded by equity investments and commercial bank loans. Developed without direct government funding, though government guarantees were provided to support the project.

  12. ^

    Entirely funded by energy companies (Chevron, ExxonMobil, and Shell) via corporate resources and private debt arrangements. No direct government financial contribution.

  13. ^

    I haven’t looked into any of these examples deeply, maybe a subject for a future post.

  14. ^

    To be clear, I don’t think that AI is overhyped on net, just that it’s easy to dismiss comments that are very bullish on AI.

  15. ^

    Note that I think this will be very hard as the public would need to know how much inference compute models need to run, how much models are being run for public use. If a company is doing a secret intelligence explosion presumably they are also able to significantly push the frontier and serve small, cheap to run models that are competitive with public frontier models and use most of their inference compute for an intelligence explosion.

  16. ^

    Note that there are a lot of simple strategies like distilling a model to save inference compute and serving that model, just saying that your model is being very reward hacky so you’ve decided not to serve it for a while until you get more of a handle on it etc.

OSZAR »