Outrageous Predictions
Executive Summary: Outrageous Predictions 2026
Saxo Group
Investment Strategist
Nvidia’s latest message suggests AI demand is shifting from training models to running them constantly in real time.
That shift makes memory more strategic, because inference needs speed, bandwidth and power efficiency, not just raw compute.
Micron’s results matter as a read-through for the wider memory cycle, not only for one stock.
At Nvidia’s GTC conference on 16 March 2026, Chief Executive Jensen Huang said “the inference inflection has arrived” and raised the company’s view of the AI chip revenue opportunity to at least USD 1 trillion through 2027. That is a sharp step up from the USD 500 billion opportunity it had discussed through 2026. In simple terms, the industry is moving from building AI models to using them, again and again, at scale.
The brain still matters, but memory is the bloodstreamTraining an AI model is expensive and spectacular, which is partly why it gets so much attention. Inference is less glamorous but arguably more commercial. It is what happens when millions of users ask a model to search, summarise, code, recommend or act. That means low latency, high throughput and efficient power use. Those are not just compute problems. They are also memory problems.
High-bandwidth memory, or HBM, is the premium product here. It sits close to the processor and helps feed it data very quickly. If the processor is the engine, HBM is the fuel line. A giant engine without enough fuel flow is just an expensive sculpture.
That is why Micron’s update this week matters. On 16 March 2026, the company said its HBM4 36GB product had entered high-volume production for Nvidia’s Vera Rubin platform, with more than 2.8 terabytes per second of bandwidth and 20% better power efficiency. Micron also highlighted a new memory module that it says can deliver 2.3 times faster “time to first token” for long-context large language model inference. That is industry language for one very simple thing: answers arrive faster.
Samsung is worth a brief mention here too. At GTC, it showcased Nvidia’s new Groq LP30 inference chip made on Samsung’s 4-nanometre process. That matters because it widens the memory and AI hardware story beyond Micron alone. Samsung is also pushing ahead in advanced memory, having started HBM4 shipments and flagged HBM4E samples for later in 2026. In other words, inference is not just lifting one memory name. It is raising the strategic value of the whole high-end memory stack.
Micron’s earnings on 18 March 2026 are not just a company event. They are a temperature check for the whole memory market. This industry has a long history of being cyclical, messy and prone to mood swings. One year it looks like a commodity swamp, the next it looks like a licence to print money. The difference this time is that AI may be making parts of memory less ordinary.
Micron, Samsung and SK Hynix are all struggling to keep up with AI-driven demand, while Applied Materials has partnered with Micron and SK Hynix on next-generation DRAM and HBM development. Samsung sees an “unprecedented supercycle” in chips, with memory shortages still supporting demand, even as higher prices start to pressure personal computer and smartphone markets.
That is the important nuance for investors. Not all memory is equal. Traditional personal computer and smartphone memory can still behave like an old-fashioned cycle. HBM and AI-oriented data centre memory look more structural. They are tied to platform launches, packaging complexity, power constraints and a small group of credible suppliers. That tends to be a nicer neighbourhood for margins.
Micron’s own recent actions support that idea. Alongside its HBM4 announcement, it completed the acquisition of a Taiwan site and said it plans to add another cleanroom there by the end of fiscal 2026. Earlier this year it also broke ground on a new Singapore wafer fabrication facility, framed around long-term AI-driven demand. That does not remove cyclicality, but it does show management acting like demand is not a one-quarter fad.
There are, as ever, a few flies in the semiconductor soup.
First, valuations already reflect a lot of optimism. Micron has rallied hard into earnings, which means good numbers may not be enough on their own. Investors will want evidence that HBM mix, pricing and margins still have room to improve.
Second, supply eventually responds. Memory shortages are wonderful until they become invitations for more capacity. Micron, Samsung and SK Hynix are all expanding or preparing for next-generation production. Great for the long run, less great if everyone arrives at the same party with too many chips.
Third, inference is not a monopoly story. Nvidia remains dominant, but inference faces greater competition from central processing units and custom chips built by companies such as Google. If more workloads move to alternative architectures, the winners inside memory may be the companies with the broadest product mix and the closest customer ties, not simply the loudest stock chart.
Watch Micron’s comments on HBM mix, pricing and supply discipline, not just headline revenue.
Separate AI memory from consumer memory. They can move in the same sector, but not always for the same reasons.
Track Nvidia platform launches and inference workloads as demand signals for premium memory.
Treat this as infrastructure logic, not fashion. Bottlenecks often earn more than buzzwords.
The neat part of the AI story is no longer just who trains the smartest model. It is who helps that model answer billions of times without slowing down, overheating or becoming uneconomic. Nvidia’s latest message points clearly toward inference, and inference needs memory that is faster, denser and more efficient. That makes Micron’s report on 18 March 2026 more than an earnings event. It is a test of whether memory makers are still just cyclical passengers, or whether they are becoming toll collectors on AI’s busiest roads. In this phase of the AI build-out, the glamour chip still gets the headlines, but the memory chip may quietly help decide who gets paid.