Outrageous Predictions
Die Grüne Revolution der Schweiz: 30 Milliarden Franken-Initiative bis 2050
Katrin Wagner
Head of Investment Content Switzerland
Investment Strategist
Google’s TurboQuant shakes storage-heavy chip names more than the premium memory layer closest to AI compute.
The chip trade is shifting from “more hardware at any cost” to “useful efficiency at scale”.
For long-term investors, the key question is not more chips or fewer chips, but which chips stay indispensable.
When Google shows a way to make artificial intelligence cheaper to run, the semiconductor market tends to hit sell first and ask questions later. That is what happened this week. Google unveiled TurboQuant, a technique it says can sharply reduce key-value cache memory needs while speeding up some workloads on Nvidia H100 chips.
The news triggered a broad sell-off across parts of the semiconductor space, especially among memory names that had ridden the idea that every extra AI query would need ever more storage. But the bigger point for investors is not that the AI chip boom is over. It is that the market is starting to reward efficiency, not just scale, in the next phase of the build-out.
The memory aisle now has labelsThe semiconductor market often talks about “memory” as if it were one big drawer. It is not. High-bandwidth memory, or HBM, sits very close to the graphics processing unit and helps feed AI accelerators quickly. Dynamic random-access memory, or DRAM, handles fast working memory more broadly. NAND flash is the cheaper, denser storage layer used in solid-state drives and other capacity-heavy jobs.
TurboQuant appears to target one specific pain point in inference, which is the stage where a trained model answers real user prompts. In simple terms, it helps compress part of the memory burden involved in handling long conversations and large context windows. That is why the selloff has hit storage and flash-linked names harder than the premium memory suppliers closest to the graphics processing unit.
That distinction matters. Investors had started to price many memory names as if all roads led upward together. Flash memory winners had surged far ahead of traditional DRAM leaders before this week’s reversal, and the new Google news has exposed just how different those businesses really are.
This is where the story gets more interesting than a simple selloff. In semiconductors, an efficiency gain often shifts the bottleneck rather than removing it. If running AI becomes cheaper, more companies can afford to deploy it. If more companies deploy it, total demand for compute, networking, advanced packaging and premium memory can still rise.
That is not just theory. Micron said last week that booming demand from AI systems is driving record results and pushed its fiscal 2026 capital spending plan above USD 25 billion. In its own prepared remarks, Micron also said rapid growth in AI inference is creating new architectures and that demand for data-centre NAND remains well above available supply. Meanwhile, Samsung is trying to move major customers onto three-to-five-year contracts, and SK Hynix has agreed to buy USD 7.97 billion of ASML extreme ultraviolet lithography tools through 2027.
In other words, the big industry players are not behaving like demand is about to disappear. They are behaving like the winners may change, the architecture may evolve, and the cost curve may improve, but the need for serious semiconductor capacity is still very real.
There is also a small irony here. A breakthrough designed to reduce memory intensity may end up helping the AI market expand faster. That is because lower costs usually widen adoption. Cheaper inference can make AI more useful in search, software, customer service, coding tools and on-device features. When that happens, one slice of the bill shrinks, but the number of bills often multiplies.
The latest semiconductor news flow suggests the market is moving from a broad AI story to a more selective one. Names tied to the most irreplaceable parts of the stack still look better placed than names that benefited mainly from scarcity and momentum. That is why HBM suppliers, advanced equipment makers and certain logic or accelerator names may hold up better than businesses exposed to more interchangeable storage demand.
This does not make the sector easy. Memory has a long history of making investors feel clever at the top and philosophical at the bottom. But the current cycle is different in one important respect. The winners are not just the companies making more chips. They are the ones sitting at the hardest bottlenecks, signing longer contracts, and spending with enough confidence to prepare for years of demand rather than one lucky quarter.
The first risk is that Google’s approach spreads faster and wider than the market expects. If TurboQuant or similar methods start cutting into more layers of memory demand, the pressure on flash pricing could last longer than a two-day wobble.
The second risk is the old one dressed in AI clothing: overinvestment. Micron is spending more. SK Hynix is ordering tools aggressively. Samsung is talking like a company that sees a supercycle. That is good while demand stays tight. It is less charming if customers pause, models become more efficient faster than expected, or enterprise adoption takes longer than hoped.
The third risk is geopolitics. Export controls, tariffs, power constraints and supply chain frictions still hang over the whole semiconductor industry like a fluorescent office light. Always on, never flattering.
Separate scarce memory from swappable storage. Not every “AI memory” story has the same economics.
Watch contracts and capital spending. They often tell the truth before share prices do.
Follow inference as closely as training. The next winners may serve daily usage, not headline model launches.
Treat semiconductor volatility as part of the package, not as a surprise gift.
Google’s announcement looks, at first glance, like bad news for chips. It is not that simple. It is better understood as a sorting event. TurboQuant does not kill the semiconductor story. It forces investors to ask which part of that story still earns the fattest margin when AI gets cheaper, faster and more widely used.
Some memory layers may lose a little glamour. Others may become even more essential. That is usually how this industry works. One bottleneck eases, another becomes valuable, and the market has to relearn the map. In semiconductors, using less of one thing often ends up needing more of something else.
This material is marketing content and should not be regarded as investment advice. Trading financial instruments carries risks and historic performance is not a guarantee of future results.
The instrument(s) referenced in this content may be issued by a partner, from whom Saxo receives promotional fees, payment or retrocessions. While Saxo may receive compensation from these partnerships, all content is created with the aim of providing clients with valuable information and options.