Morning Overview

Google’s new AI compression could cut demand for NAND, pressuring Micron

A new compression technique from Google Research threatens to shrink the memory footprint of large AI models so dramatically that it could weaken demand for NAND flash storage, one of Micron Technology’s biggest revenue drivers. The method, called TurboQuant, achieves near-optimal data compression at extremely low bit-widths, meaning AI workloads that once required vast amounts of storage could soon need far less. For Micron, which disclosed significant NAND revenue exposure in its most recent quarterly filing, the timing raises hard questions about the durability of storage demand in an AI-driven market.

What TurboQuant Actually Does

TurboQuant is a vector quantization algorithm designed to compress the internal data structures of AI models with minimal quality loss. Detailed in a paper posted to arXiv, the technique targets two of the most storage-hungry components in modern AI systems: the key-value (KV) cache used during inference and the high-dimensional vectors used in nearest-neighbor search tasks. The paper reports experimental results showing that TurboQuant maintains output quality even at ultra-low bit-widths for KV-cache quantization, a finding that, if replicated at production scale, would sharply reduce how much memory each AI query consumes.

The research builds on a line of prior work in AI efficiency. The TurboQuant authors reference earlier work on extreme compression strategies and related quantization schemes that laid the groundwork for their algorithmic approach. What distinguishes TurboQuant is its claim of near-optimal distortion rates, meaning the gap between the compressed and original data is close to the theoretical minimum. For AI operators running billions of inference queries daily, even modest per-query storage savings compound into enormous reductions in total NAND consumption across data center fleets.

Practically, TurboQuant works by mapping high-dimensional vectors (such as those representing tokens in a language model’s KV cache) into much lower bit-width codebooks. Instead of storing each value as a 16-bit or 8-bit number, the algorithm can represent it with just a few bits while trying to preserve the geometry of the original vector space. The closer the compressed representation stays to the original in terms of distance metrics, the more faithfully the model’s behavior is preserved. The paper’s benchmarks suggest that, for several tasks, TurboQuant can push bit-widths down aggressively before accuracy begins to degrade in a way users would notice.

Because KV caches are ephemeral structures that grow with sequence length and batch size, they are a major driver of memory needs during inference. A compression method that slashes KV-cache size without forcing model retraining or major architecture changes is particularly attractive to operators who want to cut infrastructure costs quickly. TurboQuant’s focus on this cache layer, rather than only on static model weights, is one reason its potential impact on storage demand is drawing attention.

Micron’s NAND Exposure in Focus

Micron Technology’s financial dependence on NAND makes it especially sensitive to any structural shift in how much storage AI workloads require. In its most recent Form 10-Q for the quarter ended February 26, 2026, the company breaks out revenue by technology, highlighting both DRAM and NAND contributions. The filing underscores that NAND flash is a core pillar of Micron’s business and notes that rapid technological change in memory and storage markets is a material risk factor.

NAND is the basis for solid-state drives (SSDs) that underpin cloud storage, enterprise databases, and AI data pipelines. In AI clusters, SSDs store training data, model checkpoints, and sometimes entire model weight files that are paged into higher-bandwidth memory as needed. If compression techniques like TurboQuant reduce the size of KV caches and other intermediate data, operators may be able to serve the same or greater number of queries with fewer total bytes of NAND per server. Over time, that could erode the number of terabytes deployed per rack, even as the number of racks grows.

Micron’s risk disclosures already caution that demand for its products can be “highly volatile” and influenced by changes in customer architectures and technology transitions. TurboQuant-style compression fits squarely within that category: it is not a new competing memory technology, but a software-side innovation that changes how efficiently existing hardware is used. For a commodity-like product such as NAND, where revenue is heavily driven by bit shipments, any improvement in bits-per-query economics on the customer side is a direct headwind.

Why Most Coverage Gets the Demand Story Wrong

Investor commentary around AI and memory has largely emphasized explosive demand. As models scale from billions to trillions of parameters and context windows expand, the intuitive expectation is that storage needs must rise in lockstep. That intuition underpins bullish theses on memory suppliers, including Micron, with the assumption that the AI boom will translate into a supercycle for both DRAM and NAND.

What this narrative often misses is the intensity of the countervailing push toward efficiency. Across the industry, researchers and engineers are racing to squeeze more performance out of fewer bits: quantizing model weights, pruning redundant neurons, distilling large models into smaller ones, and now compressing KV caches and embeddings with advanced vector quantization. TurboQuant is emblematic of this trend, not an outlier.

The economic logic is straightforward. Cloud providers and AI startups operate under tight cost constraints, with memory and storage among the largest line items in serving large models. If a new algorithm cuts memory usage per query by, say, 40% while preserving quality, competitive pressure will push major platforms to adopt it quickly. As these techniques diffuse, the amount of NAND required to support a given volume of AI traffic could fall, even if the traffic itself grows rapidly.

This sets up a race between two curves, growth in aggregate AI workloads and decline in storage needed per workload. Many bullish projections implicitly assume the first curve dominates, but techniques that claim “near-optimal” compression at low bit-widths increase the odds that the second curve catches up. Micron’s own acknowledgment of technological disruption risk suggests the company understands that AI may not be an unalloyed positive for memory demand.

The Hybrid Memory Wrinkle

There is a plausible counterargument that compression will not simply destroy demand for NAND, but reshape it. Modern AI data centers are moving toward tiered memory architectures, high-bandwidth DRAM or HBM close to the compute cores, backed by SSDs that store larger model snapshots, extended context buffers, or retrieval-augmented databases. In such architectures, shrinking the DRAM footprint via TurboQuant could enable operators to fit larger effective models on a given accelerator, potentially increasing reliance on SSDs for spillover storage and retrieval tasks.

In this hybrid scenario, some of the “freed” DRAM demand might reappear as demand for fast, low-latency SSDs tuned for AI inference patterns. Micron could benefit at the margin if it captures design wins in these specialized products. However, the total volume of NAND deployed in these high-performance tiers is typically far smaller, in raw gigabytes, than in bulk storage or archival layers. Even if every AI accelerator node gains a few more terabytes of premium SSD capacity, that may not offset a broad reduction in general-purpose NAND deployed across data center fleets.

Moreover, compression that applies not just to KV caches but also to embeddings and retrieval indices could shrink the very datasets that populate those SSD-backed stores. If vector databases can represent the same corpus with fewer bits per vector while maintaining recall and precision, the net effect is again fewer NAND bits required per unit of application functionality.

What Investors and Industry Watchers Should Track

Whether TurboQuant and similar techniques translate into real pricing pressure for Micron will depend on several observable signals over the next few years.

First, deployment timelines are crucial. The arXiv paper demonstrates algorithmic promise, but integrating TurboQuant into production systems like large-scale language model serving stacks requires extensive engineering, testing, and monitoring. Investors should watch for commentary from large AI operators about KV-cache compression in earnings calls, technical blogs, or conference talks. So far, there have been no public commitments from Google about rolling TurboQuant into flagship products, leaving the adoption timeline uncertain.

Second, competitive dynamics among memory suppliers will shape how any demand shift affects pricing. If Samsung, SK Hynix, and Micron all face similar reductions in NAND growth from AI customers, they may respond with tighter supply discipline, attempting to balance the market and support average selling prices. But if Micron is more exposed to data center NAND than its peers, or lags in pivoting toward higher-value products, it could feel disproportionate pressure on both volumes and margins.

Third, Micron’s own disclosures will offer early clues. The company’s quarterly and annual reports already break out revenue by technology and discuss end-market trends. Close readers should watch for subtle changes in how management describes AI-related demand, including any mention of customers optimizing memory footprints, experimenting with new compression techniques, or slowing capacity additions in cloud storage. Even small shifts in language in future filings could signal that software-side efficiency gains are beginning to bite.

Finally, the broader research pipeline bears monitoring. TurboQuant is one step in a rapidly evolving field where each generation of algorithms tends to be more aggressive than the last. If follow-on work pushes compression further while maintaining quality, or makes it easier to deploy such techniques without retraining models, the structural headwinds for NAND could strengthen. For Micron, the AI era may still be a growth story, but one in which the most important innovations are happening in code, not silicon.

More from Morning Overview

*This article was researched with the help of AI, with human editors creating the final content.