Tag: SpaceX

  • Elon Musk Isn’t Building the Everything App—He’s Building the Everything App’s Power Grid

    The Pivot in One Sentence
    xAI has merged into SpaceX and leased its Colossus 1 supercluster—220,000 NVIDIA GPUs, 300 megawatts of compute—entirely to Anthropic, while simultaneously targeting 2 gigawatts of total capacity at Memphis. Elon Musk is no longer primarily trying to win the AI model race. He’s becoming the AI industry’s infrastructure landlord.

    Earlier in this series, we asked whether Grok and xAI were building the everything app through X—the social-financial superapp thesis. The answer we arrived at was: maybe, but with real limitations on the model quality and consumer trust needed to pull it off.

    Then something happened that reframed the entire question. In early May 2026, xAI merged into SpaceX. Days later, Anthropic—one of xAI’s most direct AI competitors—announced it was renting the entire compute capacity of Colossus 1. All 220,000 GPUs. All 300 megawatts. For Claude. For a reported $3 to $6 billion per year.

    Musk’s comment when asked about leasing infrastructure to a competitor: “No one set off my evil detector.”

    That’s the tell. When you’re building the everything app, you don’t rent your most powerful asset to your rivals. You use it. The fact that Musk is doing exactly that reveals a strategic logic that the Grok-as-everything-app frame completely misses.

    The pivot isn’t from everything app to compute landlord. It’s the recognition that owning the power grid is more valuable than owning any single app that runs on it.

    What Colossus Actually Is

    Colossus is not a single data center. It’s a multi-building supercomputing complex in Memphis, Tennessee—and it is currently the largest single-site AI training installation in the world.

    Colossus 1, the original facility, holds H100, H200, and GB200 accelerators across more than 220,000 GPU units. That is the cluster Anthropic is now renting entirely.

    Colossus 2, the expansion xAI is keeping for its own Grok development, has already expanded to 555,000 NVIDIA GPUs with approximately $18 billion in hardware investment and 2 gigawatts of target power capacity—reached in January 2026 with the purchase of a third Memphis building. Musk’s stated goal: one million GPUs at the Memphis complex, with more AI compute than every other company combined within five years.

    As a point of reference: most frontier AI labs operate training clusters in the tens of thousands of GPUs. Microsoft’s Azure AI infrastructure, the largest hyperscaler allocation for AI, operates in the hundreds of thousands across distributed global regions. Colossus at 555,000+ GPUs in a single complex is a different category of infrastructure entirely.

    And Musk has publicly noted that xAI is only using about 11% of its available compute for Grok. The rest is—in his framing—available. Available to sell. Available to rent. Available to become the compute backbone of the AI industry whether xAI wins the model race or not.

    The xAI-SpaceX Merger: What It Actually Means

    The May 2026 merger of xAI into SpaceX as an independent entity is more than an org chart change. It’s a signals-to-strategy reveal.

    SpaceX has three things xAI needs at scale: capital (SpaceX generates billions in launch revenue annually), real estate and construction expertise (SpaceX builds rockets and factories at speed), and most critically—rockets. Starship can put mass into orbit economically in a way no other launch vehicle can. SpaceX is already moving toward a Starlink constellation of thousands of satellites. The infrastructure to extend that into orbital data centers is not theoretical.

    Anthropic’s announcement noted not just the Colossus 1 ground lease—it also expressed interest in working with SpaceX to develop multiple gigawatts of compute capacity in space. Orbital data centers. Satellite-delivered AI compute. The kind of infrastructure that has zero latency for any application that needs compute without a physical data center address.

    Musk has discussed launching a million data-center satellites as a longer-term infrastructure play. That number sounds unreasonable until you consider that SpaceX already operates over 7,000 Starlink satellites and is building Starship specifically for high-volume orbital delivery. The orbital compute thesis isn’t science fiction for SpaceX. It’s a product roadmap.

    What the xAI-SpaceX merger does is remove the pretense that these are separate businesses. They’re one integrated infrastructure play: ground-based GPU superclusters plus orbital compute capacity, connected by the world’s only commercially viable heavy-lift reusable rocket.

    The Anthropic Deal: A Strategic Reading

    Let’s be specific about what this deal represents for both sides.

    For Anthropic, the deal addresses an acute bottleneck. Anthropic’s annualized revenue grew from roughly $9 billion at end of 2025 to approximately $30 billion by early April 2026—a trajectory that implies an 80-fold increase in usage in Q1 alone. Claude Pro and Claude Max subscriber growth is outpacing Anthropic’s ability to provision compute fast enough. Renting Colossus 1 immediately unlocks 300 megawatts of capacity that would take 18-24 months to build from scratch. For Anthropic, this is a compute emergency solution with strategic upside.

    For xAI, the deal is more nuanced. Colossus 1 was already built and operational. xAI is keeping Colossus 2 for Grok development. Renting Colossus 1 generates—depending on which analyst estimate you use—between $3 billion and $6 billion annually in revenue while the asset runs at capacity rather than sitting idle. That revenue funds Colossus 2 expansion, Colossus 3, and whatever comes next. The compute landlord model is self-funding.

    The strategic implication: xAI doesn’t need Grok to win the model race for this business model to work. If Claude dominates, Anthropic needs more compute and pays xAI for it. If GPT dominates, OpenAI and its partners need more compute. If Gemini dominates, Google builds its own, but every smaller lab comes to whoever has available capacity. xAI wins in every scenario except the one where everyone else simultaneously builds their own supercomputing megacomplexes—which requires the capital and construction expertise that most AI labs don’t have.

    The Grok Situation: Honest Assessment

    The Anthropic deal does raise real questions about Grok’s trajectory. Grok app downloads have reportedly declined significantly in 2026 as ChatGPT and Claude have gained consumer mindshare. In April 2026, Elon Musk testified in the ongoing OpenAI litigation that xAI trained Grok on OpenAI model outputs—a revelation that raised questions about Grok’s training methodology and original capability claims.

    If xAI is using only 11% of its compute for Grok and is renting the rest to a competitor, the implicit message is that xAI is not currently running a max-effort campaign to win the frontier model race. It’s building infrastructure and waiting—or pivoting to a business model where the model race outcome matters less.

    This is not necessarily a failure. It may be a more durable strategy. The history of technology infrastructure is full of examples where the company that built the picks and shovels during a gold rush outlasted the miners. AWS didn’t win by building the best e-commerce site. It built the infrastructure that every e-commerce site ran on. The question is whether xAI’s compute infrastructure can fill that role for AI—and the Anthropic deal is the first real evidence that the answer might be yes.

    The “Everything App Ability” Thesis

    Here’s the reframe that this pivot suggests: maybe the right question isn’t which company will build the everything app. Maybe the right question is which company will own the infrastructure that makes the everything app possible for everyone else.

    Every company in this series—Microsoft, Google, Notion, OpenAI, Perplexity, Mistral, Zapier—needs compute. Massive, reliable, cost-effective GPU compute. The frontier model companies are burning through capital building their own clusters because the alternative is depending on hyperscalers (AWS, Azure, GCP) that charge premium rates and may eventually compete directly.

    xAI with Colossus is offering a third option: AI-native compute infrastructure, built by a company that doesn’t directly compete on most application layers, at a scale that’s difficult to replicate, at a location (Memphis) with power grid access that many coastal data center markets can’t match.

    If you’re building the everything app and you need the compute to run it—Colossus may become the place you go when AWS is too slow, Google is a competitor, and building from scratch takes two years you don’t have.

    That’s not the everything app. That’s the everything app’s power grid. And historically, the entity that owns the power grid captures durable, compounding value regardless of which specific applications win the consumer layer.

    Space: The Long Game

    The orbital compute angle deserves more than a footnote because it’s where this thesis could either collapse into fantasy or become genuinely transformative.

    The practical case for orbital data centers is latency equalization: compute in low Earth orbit can serve any point on the Earth’s surface within milliseconds, without the geographic concentration that makes terrestrial data centers vulnerable to regional power outages, natural disasters, or regulatory shutdown. For AI applications that need global deployment at consistent latency—real-time translation, autonomous vehicle coordination, financial systems—orbital compute offers something no ground-based data center geography can.

    SpaceX’s Starship dramatically changes the economics of getting mass to orbit. Current launch costs for payloads are measured in thousands of dollars per kilogram. Starship’s target is hundreds of dollars per kilogram—an order-of-magnitude reduction that makes orbital infrastructure financially viable in a way it never was before. The satellite internet analogy is instructive: Starlink was also considered impractical until SpaceX dramatically reduced launch costs, then deployed at a scale that changed the calculus entirely.

    Anthropic’s stated interest in orbital compute capacity with SpaceX isn’t a polite corporate gesture. It’s Anthropic hedging its long-term compute dependency on a technology only SpaceX can currently deliver. If even a fraction of that orbital compute vision materializes, xAI/SpaceX’s infrastructure moat becomes essentially unreplicable by any company that doesn’t own a heavy-lift reusable rocket program.

    What This Means for the Everything App Race

    The xAI infrastructure pivot doesn’t remove Grok and X from the everything app conversation entirely. X still has the distribution, the data firehose, the financial services ambitions, and the brand. Those don’t disappear because Colossus 1 is now running Claude.

    But it does add a second thesis that may ultimately matter more: xAI as the infrastructure layer beneath the entire AI economy. Not the everything app—the everything app’s foundation.

    In the history of platform technology, the company that owns the infrastructure layer almost always captures more durable value than the company that owns any individual application. TCP/IP outlasted every early internet application. AWS became more valuable than most of the businesses it hosts. The cloud didn’t belong to any one software company—it belonged to the infrastructure providers who made software deployment cheap and fast.

    If the AI era follows the same pattern, the question isn’t who builds the best everything app. It’s who builds the infrastructure that makes every everything app possible. And as of May 2026, the most credible answer to that question involves 555,000 GPUs in Memphis, a rocket program that can reach orbit, and a business model that profits whether Grok wins or loses.

    Key Takeaway

    Elon Musk pivoted xAI from model competitor to infrastructure landlord. By merging into SpaceX, leasing Colossus 1 to Anthropic, and targeting 2 gigawatts of Memphis compute capacity plus orbital data centers, xAI is positioning to capture value from the AI economy regardless of which application layer wins—the power grid, not the appliance.

    Related Reading

    This article grew out of our everything app series. If you’re tracking where AI consolidation is heading, the full series maps the competitive landscape from nine angles:

    Frequently Asked Questions About xAI, Colossus, and the Compute Landlord Pivot

    Why did xAI merge into SpaceX?

    xAI merged into SpaceX in May 2026 as an independent entity within the broader Musk enterprise. The merger combines xAI’s AI development capabilities with SpaceX’s capital generation, construction expertise, and—critically—rocket launch capabilities. This integration enables the orbital compute strategy: deploying data center satellites via Starship at dramatically lower cost than any competitor could achieve.

    What is the Anthropic-Colossus deal?

    In May 2026, Anthropic agreed to rent the entire compute capacity of Colossus 1—xAI’s first Memphis supercluster, comprising 220,000+ NVIDIA GPUs and 300 megawatts of power. The deal directly addresses Anthropic’s acute compute shortage during a period of explosive Claude usage growth. Anthropic’s annualized revenue grew from roughly $9 billion at end of 2025 to approximately $30 billion by April 2026. Analysts estimate the deal generates between $3 billion and $6 billion annually for xAI/SpaceX.

    How large is the Colossus supercomputer complex?

    As of early 2026, the Colossus complex in Memphis spans three buildings and targets 2 gigawatts of total compute capacity. Colossus 2 (kept by xAI for Grok development) has reached 555,000 NVIDIA GPUs with approximately $18 billion in hardware investment. Long-term targets include one million GPUs at the Memphis site. It is currently the largest single-site AI training installation in the world.

    What are orbital data centers and why does xAI/SpaceX care about them?

    Orbital data centers are computing facilities deployed in low Earth orbit, delivered by rocket. They offer latency equalization (serving any point on Earth within milliseconds), elimination of geographic concentration risk, and compute capacity outside any single regulatory jurisdiction. SpaceX’s Starship reduces launch costs by an order of magnitude compared to existing vehicles, making orbital compute economically viable for the first time. Anthropic’s participation in the deal included expressed interest in developing multiple gigawatts of orbital compute capacity with SpaceX.

    Does the compute landlord strategy mean xAI is giving up on Grok?

    Not necessarily, but the signals are mixed. xAI is reportedly using approximately 11% of its available compute for Grok development—the rest is available to lease. Grok app downloads have declined in 2026, and April 2026 litigation revealed Grok was trained on OpenAI model outputs. The Colossus 1 lease to Anthropic is the clearest evidence that xAI is not running a maximum-effort campaign on frontier model development and is instead diversifying into infrastructure revenue.

    How does the xAI infrastructure play relate to the everything app thesis?

    The xAI pivot suggests a reframe of the everything app question. Rather than competing to be the app users interact with daily, xAI/SpaceX is positioning to own the compute infrastructure that powers any everything app—what we’re calling the “everything app’s power grid.” Historically, infrastructure layer companies (AWS, TCP/IP, electricity grids) capture more durable value than any individual application running on top of them. The Anthropic deal is the first concrete evidence that this model may work at AI scale.