Inside the AI build-out: what CoreWeave just told investors
Ruben Dalfovo
Investment Strategist
Key takeaways
Demand is intact. Backlog sits around USD 55.6 billion, delays shift revenue timing rather than cancel contracts.
Bottlenecks define pace. Power, GPUs and construction limit delivery, pushing capex from 2025 into 2026.
Profit quality is a mix story. More inference and a broader customer base beyond Microsoft should lift utilisation and smooth margins.
Artificial intelligence needs two things in bulk: cutting-edge chips and places to run them. Nvidia supplies the engines. Big platforms like Microsoft, Alphabet and Meta rent and build the garages. Software players such as Palantir turn that raw compute into useful workflows for banks, factories and governments. CoreWeave sits in the middle. It is a specialist cloud that rents graphics processing unit power for AI training and inference to customers that prefer speed and flexibility over owning data centres.
Think of CoreWeave as a focused “neo-cloud.” It packages the latest GPUs, power and networking into ready capacity, then scales sites as demand lands. That makes it a useful barometer for the AI build cycle. When CoreWeave accelerates, it usually means new models, bigger datasets and fresh use cases are moving from pilots to production across the ecosystem.
What just happened
CoreWeave arrived as one of the last AI infrastructure names to report this season, following heavyweight updates from Microsoft, Alphabet, Amazon and Meta that kept pointing to high AI spend and longer build cycles. That backdrop matters because specialist clouds like CoreWeave live where big-tech demand meets real-world constraints such as chips, power and construction. Microsoft even flagged a shift toward more “short-lived assets” in 2026 to better match revenue visibility, which supports selective leasing from third parties.
Against that context, CoreWeave delivered third-quarter revenue of USD 1.36 billion, above the USD 1.29 billion consensus, and a net loss of USD 110 million, or 22 cents per share. The company trimmed full-year 2025 revenue guidance to USD 5.05–5.15 billion from USD 5.15–5.35 billion after a third-party data-centre developer fell behind schedule, pushing some capacity and revenue to later periods. Shares fell about 6% after hours. Management stressed the contract value is preserved even as timing slips.
The backlog tells the demand story. Remaining contracted revenue rose to about USD 55.6 billion, helped by new multi-year deals, including around USD 14.2 billion with Meta and USD 6.5 billion with OpenAI. That aligns with the wider AI build where orders outrun energised capacity and validates a pipeline that can support higher spend once sites and power come online.
Demand is durable, supply is lumpy
AI demand remains broad. Nvidia’s newest chips set the pace, while Microsoft, Alphabet and Meta feed steady orders as they train larger models and roll out AI features. Software players like Palantir push more real-world pilots into production, which keeps utilisation high.
Supply is the brake. Construction timelines, grid connections and next-gen GPU deliveries slow how fast capacity goes live. CoreWeave flagged a partner delay that shifted revenue timing, not contract value. That mirrors the wider build cycle where recognition slips, but demand does not.
Capital expenditure (capex) sequencing tells the same story. Reported Q3 capex was about USD 3.3 billion versus USD 4 billion expected. The lower 2025 capex range of USD 12–14 billion, down from USD 20–23 billion, implies more spend in 2026. Across the industry, the cash curve often lags the contract curve. Backlogs validate demand. Cranes, chips and power unlock revenue.
Concentration and workload mix matter
Customer mix is a key risk across specialist clouds. Hyperscalers can rebalance spend or insource capacity. CoreWeave has added logos like Meta and OpenAI, but Microsoft still matters. The same pattern shows up elsewhere as platforms weigh buy versus build to match product roadmaps and cost targets.
Workload mix shapes earnings quality. Training is spiky and tied to chip cycles and model launches. Inference is steadier and tracks end-user adoption in search, ads, office software and customer support. CoreWeave says inference is growing, which matches the ecosystem shift. The glide path to smoother margins depends on how fast inference scales and how hard hyperscalers compete on price and bundles.
Power, chips, and financing define the moat
In AI infrastructure, the scarce inputs are top-tier GPUs and megawatts. CoreWeave has contracted roughly 1.3 gigawatts of power, enough to host multiple Blackwell-class clusters. The fight for energy-rich sites is intense as hyperscalers, colocation firms and specialist clouds chase the same substations, transformers and fibre routes. Location, interconnect and grid upgrade rights become strategic assets.
Financing matters too. Many providers use hardware-backed debt and vendor partnerships to accelerate builds. CoreWeave’s structure with Nvidia fits that pattern. It speeds deployment, but it also raises sensitivity if GPU rental prices step down as supply loosens. Across the stack, execution on energising sites, sustaining high utilisation and managing debt costs will decide who converts scale into durable operating leverage.
What it means for the broader AI and data-centre trade
CoreWeave’s update reads like the industry’s to-do list. Demand signals remain firm, but delivery timelines stretch as power, construction and next-gen chips catch up. This extends the build cycle into 2026, which supports Nvidia’s high-end pipeline while keeping pressure on grid access, transformers and cooling. Real-estate operators and colocation firms feel the pull too, as hyperscalers mix leased space with owned campuses to match their revenue visibility.
The workload mix matters. Training is bursty and capex-heavy. Inference is steadier and closer to end-user demand. As more inference workloads roll out to search, advertising, office software and customer service, utilisation should smooth and pricing should normalise. That is constructive for specialist clouds and for network and power equipment suppliers. The signal for investors is simple. Track energised megawatts, live GPU deliveries and the balance between lease and own. If power connections speed up and inference ramps, today’s lumpy build turns into broader, more predictable spend across the AI stack.
Investor playbook
Track energised megawatts and site go-lives. Power is the gating factor; new MW are the cleanest leading indicator for revenue catch-up. CoreWeave has about 1.3 gigawatts contracted, but competition for energy-rich sites is intensifying.
Watch big-tech capex signals. Guidance from Microsoft, Amazon and others still points to heavy AI outlays into 2026, with a tilt to assets that map closer to revenue. That supports selective leasing to specialist clouds.
Monitor GPU and networking deliveries. Next-gen Nvidia supply and cluster interconnects set how fast capacity becomes billable. Slips usually delay recognition rather than demand.
Follow the mix shift to inference. Training is bursty; inference is steadier and closer to end-user adoption. More inference-heavy wins should smooth utilisation and margins across the AI stack.
Check customer concentration. Diversification beyond Microsoft lowers single-counterparty risk. New hyperscaler logos and rising contribution from Meta and OpenAI matter.
Map lease vs own dynamics. If hyperscalers lean into leased, short-lived assets to match demand, specialist providers can benefit. If they insource, specialists must win on speed, price and power access.
Keep an eye on financing costs. Hardware-backed debt accelerates builds but raises sensitivity if GPU rental pricing eases as supply loosens. Balance-sheet flexibility is part of the moat.
The road ahead
CoreWeave’s update is the AI build-out in miniature: momentum on orders, friction on delivery. The quarter shows that contracts can surge faster than cranes and cables, and that even leaders must navigate power grids, partner delays, and shifting workload economics. For investors, the signal is not about a broken demand story but about sequencing.
Backlog and big-ticket customers point to growth, while capex timing, mix of training versus inference, and customer concentration decide how smooth that growth feels. The next leg will be won by those who unlock power, compress deployment timelines, and broaden the customer base without over-gearing the balance sheet. If that trifecta lands, today’s lumpy quarter becomes tomorrow’s operating leverage.
The instrument(s) referenced in this content may be issued by a partner, from whom Saxo receives promotional fees, payment or retrocessions. While Saxo may receive compensation from these partnerships, all content is created with the aim of providing clients with valuable information and options.