You can't master something that you can't even measure. And everyone has been measuring the cost of compute wrong. This has led to the denouncement of Data Centers in Space. In this whitepaper, I break down how Orbital Compute versus ground data centers through the mechanism of a new standard, the Levelized Cost of Compute.
But first, let's take a step back. Data Centers in Space sounds jarring. Let's break it down into first principles. What are the inputs for a data center on the ground?
To deal with this massive rift in constraints and situations, we need to create a new standard to measure setting up data centers against. The Levelized Cost of Compute (LCOC). This is inspired by a metric used in the energy industry, the Levelized Cost of Energy. It's a standard metric that compares generation/storage assets in power that have different cost structures. It helps you measure a natural gas plant (lower capex, volatile fuel) versus a solar + battery farm (higher upfront investment, lower marginal cost) versus a nuclear plant (huge capex but a potential much larger operating life). LCOE accounts for construction, financing, fuel, O&M, capacity factor, and the rate of return.
With the LCOC we take it a step further, factoring in the time value of money. The current metrics — $/GPU-hour, $/rack/month, $/W capture cost at a certain moment for a certain project structure. The LCOC is a measure of $/delivered GPU-hour with a standard SLA (service level agreement) with 99.9% uptime. This means we factor the capital expenditure (servers, power systems, cooling infrastructure, land/launch costs), operating expenses (power, maintenance, staffing, replacement cycles), financing costs (cost of capital, debt service), and depreciation schedules. Time is money and LCOC factors that in. The denominator is actual useful output after accounting for utilization rates, cooling overhead, and downtime. LCOC is workload agnostic, just like LCOE. You can use it on inference or training, the same way you can use energy for lights or heat pump. With LCOC, we can compare apples to apples in radically different architectures.
Where:
Ground compute costs are flattening, while a lot of the easy sites are already taken. Remaining capacity requires longer timelines for interconnection, more expensive land, and tighter water constraints. At the same time, orbital costs are rapidly declining from launch costs and manufacturing.
I predict crossover sometime in the early 2030s or late 2020s. Reaching this conclusion requires acknowledging that compute demand is insatiable. If you believe in compute growing at anywhere near the scale these hyperscalers are projecting, all you can do is bear witness to the havoc scarcity is going to wreak on ground data centers.
If you don't believe in space data centers, you don't believe in compute growth.
This is the largest engineering challenge to put data centers in space. You can only get rid of heat in space by radiating it away. You don't have the same luxury you have on the ground of moving it through mediums and using convection. Every satellite will need ISS-level, and eventually better, thermal management. However, there are very promising technologies such as deployable structures, droplet radiators, and more that can solve this problem.
This one is still up in the air. The results that will come from Starcloud's deployment will be incredibly important — the entire industry is watching. There are two things you have to deal with in space when it comes to radiation engineering: SEU (single event upsets) and TID (total ionizing dose). Single event upsets are probabilistic and something that you can only protect against with architecture, redundancy, and error correcting codes. Shielding doesn't work here. In TID, you can use shielding but, to heavily simplify the physics, shielding can act as armor. Armor that protects you, but when it breaks it inverts and pierces you right in the heart. Shielding can sometimes hurt a lot more than it helps.
You can have a lot of compute going on in space, but beaming that information down to the ground is really hard.
You're going to have to have autonomous drones that go into space and clear up debris, swap out racks, and do other jobs that are prohibitively expensive with current technologies.
Unless you're in DFW, NOVA or Memphis (where these hyperscalers are building campuses) then you're going to experience lower latency for satellites in LEO. However, as we expand to different shells, we add more and more latency, which restricts certain applications to certain shells.
Space data centers are the obvious next step. If you don't believe me, go to the interactive simulator and play with the sliders yourself. The source code is available on GitHub.