The key strategic question facing investors is whether flash memory will replace DRAM in AI applications, a development that could dramatically shift SanDisk’s valuation. As SanDisk makes a transformative leap from thumb drives to a central role in AI infrastructure, its $85 billion market cap is teetering on a two-year gamble. This article examines the factors likely to shape SanDisk’s trajectory over the next 18 months.
On February 24, 2025, SanDisk began trading on the Nasdaq as an independent company for the first time since its acquisition by Western Digital for $16 billion in 2016. The stock opened at $52.20 and closed at $48.60, resulting in a market capitalization of $5.6 billion. SanDisk, recently separated from a hard drive conglomerate, faced reputational challenges due to a portable SSD data-corruption incident that had damaged its standing among professional photographers and video editors. In early 2025, the company was more frequently associated with legal challenges than with innovation.
By January 30, 2026, SNDK reached an all-time high of $676.69, with a market capitalization of approximately $88 billion -- a 1,500% return in less than a year. SanDisk became the top-performing stock in the S&P 500 for 2025, surpassing more widely anticipated leaders such as Nvidia and Palantir.

The transformation between these two dates reflects a structural repricing of flash memory’s role in the era of artificial intelligence, not a momentum-driven or speculative stock movement. This shift centers on the potential for NAND technology to transition from the storage tier to the memory tier within the AI compute stack, a development embodied by High Bandwidth Flash (HBF). The success or failure of this technology will significantly influence both SanDisk’s future and the architecture of AI inference over the next decade.
The Divorce
Understanding SanDisk’s current trajectory requires examining the circumstances that led to its independence.
Western Digital acquired SanDisk in 2016 for $19 billion (the number varies by source depending on how you account for debt), creating a company that sold both spinning-platter hard drives and NAND flash memory. The logic was diversification. The reality was that HDDs and flash have fundamentally different demand curves, capital cycles, and customer bases. By 2023, the marriage was straining. Activist investor Elliott Management got involved, and in October 2023 Western Digital announced it would spin off its flash business as an independent public company.
The separation took longer than planned. Complex is an understatement. Western Digital and SanDisk had to disentangle shared manufacturing, shared IP, shared support infrastructure, and a 25-year joint venture with Kioxia (the former Toshiba Memory) that operates the world’s largest NAND flash fabs in Yokkaichi and Kitakami, Japan. The split was formally completed on February 21, 2025. WD distributed 80.1% of SanDisk’s shares to its stockholders (one-third of a SanDisk share for each WD share held). WD retained 19.9% for up to 12 months.
Post-split, the division was clean: Western Digital kept the hard drive business and is betting its future on heat-assisted magnetic recording (HAMR) technology for high-capacity data center drives. SanDisk got everything flash -- consumer SSDs, enterprise SSDs, embedded storage, memory cards, USB drives, and crucially, 50% of the output from the Kioxia joint venture fabs. CEO David Goeckeler, who had been running Western Digital, moved to SanDisk.
SanDisk’s emergence as an independent, pure-play flash company coincided with a growing market recognition of a significant memory constraint within AI infrastructure.
The Memory Wall
Over the past three years, discussions about AI infrastructure have primarily focused on computational resources -- the availability and performance of GPUs. The prevailing investment strategy emphasized GPU manufacturers and companies providing cooling and power solutions for data centers housing these processors.
That thesis remains valid, but it is increasingly incomplete. The primary constraint in AI inference is shifting from computational power to memory capacity and bandwidth. Floating-point operations per second (FLOPS) have scaled exponentially in recent years, far outpacing the linear growth in memory capacity. The result is a widening gap where GPUs increasingly spend time waiting for data rather than processing it.
Three specific trends are converging to make this worse.
First, model architectures are becoming more memory-intensive and less compute-intensive. Mixture-of-Experts (MoE) models, which activate only a subset of parameters for each token, reduce compute requirements but increase the volume of parameters that must be accessible in memory. More data needs to be stored close to the GPU, even if less of it is being processed at any given moment.
Second, context windows are exploding. The key-value (KV) cache -- the data structure that stores the context of a conversation or reasoning chain during inference -- grows linearly with context length. As models move from 8K to 128K to million-token context windows, and as agentic AI introduces multi-turn, multi-session reasoning, the KV cache becomes enormous. A context window increase from 8K to 1 million tokens can require roughly 100 times more memory, translating to a KV cache that can easily surpass 800 GB. That exceeds the 192 GB of HBM available on Nvidia’s current Blackwell architecture by a factor of four.
Third, inference is becoming the dominant workload. Training happens once (or periodically). Inference happens continuously, at scale, for every user query. As AI moves from research labs to production deployments, the ratio of inference compute to training compute is shifting dramatically. And inference, particularly the decode phase where tokens are generated one at a time, is overwhelmingly memory-bound. The GPU spends most of its time waiting for data, not computing.
Professor Joung-ho Kim of KAIST -- the engineer widely credited as the father of High Bandwidth Memory -- put it bluntly in a January 2026 presentation: “In AI models, the real bottleneck isn’t computing power -- it’s memory. Most inference and training processes are limited by memory.”
This is the macro thesis behind SanDisk’s repricing. The market is not paying for thumb drives. It is paying for the possibility that NAND flash, which has historically lived at the bottom of the memory hierarchy (cheap, dense, slow), is about to move up several tiers.
The NAND Supercycle
Before we get to the moonshot technology, the near-term financials tell their own story.
SanDisk’s fiscal Q2 2026 (the quarter ended January 2, 2026) was a blowout. Revenue: $3.0 billion, up 61% year-over-year, beating consensus of $2.7 billion. Non-GAAP gross margin: 51.1%, up 18.6 percentage points YoY. Non-GAAP operating income: $1.1 billion, up 386% YoY. Non-GAAP diluted EPS: $6.20, up 404% YoY. Management guided for even higher margins ahead.

By segment: Datacenter revenue was $440 million (+76% YoY), Edge was $1.7 billion (+63% YoY), and Consumer was $907 million (+52% YoY). The mix is shifting toward higher-margin enterprise SSDs, which carry better pricing and stickier demand than consumer products.
For full fiscal year 2025 (ending June), the company reported $7.4 billion in revenue with gross margins expanding from 22% to 30.1%. As of January 2, 2026, SanDisk holds $1.48 billion in cash and reached a net-cash-positive position far earlier than analysts expected.
What is driving this? Three things happening simultaneously.
First is tight NAND supply. Samsung, SK hynix, and Micron collectively scaled back NAND production in the second half of 2025, leading to price increases of 60% or more in some enterprise contracts. Samsung was reportedly seeking 20-30% price hikes in its 2026 supply deals. In contrast, Kioxia/SanDisk and China’s YMTC, neither of which have a DRAM business to cross-subsidize, have been the most aggressive in expanding capacity, with their JV planning a 41% YoY jump in investment to $4.5 billion targeting BiCS8 expansion and BiCS9 R&D.
Second, enterprise SSD demand is structural. Hyperscalers are building out AI inference infrastructure at an unprecedented rate, requiring enormous amounts of fast, dense storage. Every GPU node needs local NVMe SSDs, every rack requires shared storage, and every pod needs a KV cache tier. This is not cyclical smartphone demand that waxes and wanes but a capex-driven infrastructure buildout with multi-year visibility.
Third, SanDisk’s manufacturing economics are improving. The Kioxia JV extension announced January 29, 2026, locks in supply through 2034. SanDisk will pay Kioxia $1.165 billion in installments from 2026 to 2029 for manufacturing services. The JV’s Yokkaichi and Kitakami fabs are producing BiCS8 (218-layer) 3D NAND today, with BiCS9 entering production at the end of fiscal 2025 and BiCS10 (332-layer) pulled forward to 2026 -- originally scheduled for H2 2027. That 332-layer technology delivers 59% higher bit density and a 4.8 Gbps interface speed, which matters enormously for both enterprise SSDs and HBF.
These developments are tangible and have contributed to the company’s stock performance. However, an $85 billion market capitalization for a company generating approximately $12 billion in annualized revenue from commodity NAND products suggests that investors are pricing in something beyond cyclical growth. They are pricing in a fundamentally different business model.
High Bandwidth Flash represents that bet.
High Bandwidth Flash: The Moonshot
High Bandwidth Flash is SanDisk’s attempt to create an entirely new tier in the memory hierarchy -- one that sits between HBM (fast, expensive, small) and NVMe SSDs (slow, cheap, big).

The concept is architecturally similar to HBM. In HBM, multiple thin DRAM dies are stacked vertically and interconnected using through-silicon vias (TSVs), with a logic die at the base handling I/O. This gives HBM its extraordinary bandwidth -- data flows through thousands of parallel connections simultaneously. The trade-off is capacity. Nvidia’s Blackwell GPU has 192 GB of HBM3e across eight stacks. That sounds like a lot until you realize that a single large language model’s KV cache, at million-token context lengths, can easily blow through it.
HBF applies the same stacking-and-TSV approach, but replaces DRAM dies with NAND flash dies. SanDisk’s proprietary CBA (CMOS directly Bonded to Array) wafer bonding technology enables it to bond a 3D NAND memory array directly on top of an I/O logic die, creating a tightly integrated package. Stack multiple CBA-based NAND dies with TSVs, and you get a memory module that offers something DRAM fundamentally cannot: massive capacity at a fraction of the cost.
The numbers SanDisk is claiming: 8 to 16 times the capacity of HBM at similar cost points. HBM today costs around $8-10 per GB and offers bandwidth of up to 410 GB/s per stack. First-generation HBF could enable up to 4 TB of GPU-addressable memory in a single package, with read bandwidth within 2.2% of hypothetical unlimited-capacity HBM based on internal simulations running Llama 3.1 (405B parameters). Future generations target 2+ TB/s read bandwidth and stack capacities of 1 to 1.5 TB per module with improved power efficiency.
The product timeline: first HBF samples in the second half of 2026. First AI inference devices with HBF sampling in early 2027. Mass production by 2027-2028.
The ecosystem is forming. In August 2025, SanDisk and SK hynix signed an MoU to jointly define the HBF specification and push standardization. SK hynix -- the world’s top HBM supplier and Nvidia’s closest memory partner -- unveiled its “AIN B” HBF-based product line at the OCP Global Summit in October 2025. Samsung has reportedly begun early concept design work. Kioxia prototyped a 5 TB HBF module in August 2025. SanDisk formed a Technical Advisory Board featuring David Patterson (co-inventor of RISC architecture, Turing Award winner) and Raja Koduri (former Intel graphics chief, now running his own AI startup). Both Samsung and SanDisk are reportedly planning to integrate HBF into products for Nvidia, AMD, and Google by late 2027 or early 2028.
Professor Kim offered a vivid mental model for the emerging hierarchy. SRAM inside the GPU: the notebook on your desk, fastest and smallest. HBM: the bookshelf nearby, rapid access for active computation. HBF: the underground library, storing deep AI knowledge, continuously feeding data to HBM. Cloud storage: the public library, connected by networks, durable but slow.
Securities firms project the HBF market growing from $1 billion in 2027 to $12 billion by 2030. Kim himself predicts HBF demand will surpass HBM by 2038.
If you squint, this looks like the early days of HBM itself -- which was invented around 2015 and took seven to eight years to achieve mass adoption. The AI demand environment is arguably more favorable for HBF than the gaming/HPC environment was for HBM in its infancy.
But there is a problem. And the problem’s name is Jensen Huang.
Nvidia has consistently captured significant value across every tier of the AI infrastructure stack, maintaining gross margins around 60% through its ability to control the platform and commoditize the components beneath it. If Nvidia decides to own the KV cache expansion through its own architecture, the prospects for HBF -- and therefore for SanDisk’s premium valuation -- become uncertain.
The Nvidia Problem
At CES 2026, Nvidia announced the Inference Context Memory Storage Platform (ICMSP), powered by its BlueField-4 data processing unit. ICMSP is Nvidia’s answer to the same problem HBF is trying to solve: KV cache capacity exceeding GPU HBM. But it solves it in a fundamentally different way.
Instead of creating a new memory type that sits physically adjacent to the GPU (which is what HBF does), ICMSP extends the KV cache to standard NVMe SSDs connected through BlueField-4 DPUs and Spectrum-X Ethernet networking. Nvidia’s architecture creates what it calls a “G3.5 tier” -- a pod-level, Ethernet-attached flash tier that sits between local SSDs (G3) and shared enterprise storage (G4), optimized specifically for ephemeral inference context.
The performance claims are significant: up to 5x higher tokens-per-second and 5x greater power efficiency compared to traditional storage approaches. The architecture uses RDMA (Remote Direct Memory Access) for low-latency, high-bandwidth access to KV cache data across the pod.
A Vera Rubin SuperPod -- Nvidia’s next-generation AI infrastructure building block -- includes dedicated ICMSP racks with 4 BlueField-4 DPUs per enclosure, each connected to 150 TB of NVMe SSD capacity. That is 600 TB per enclosure of dedicated KV cache storage, using off-the-shelf enterprise SSDs.
The partner ecosystem is already substantial. AIC, Cloudian, DDN, Dell, HPE, Hitachi Vantara, IBM, Nutanix, Pure Storage, Supermicro, VAST Data, and WEKA have all signed on. BlueField-4 ships in the second half of 2026 -- the same timeline as HBF samples.
Here is why this matters for SanDisk’s valuation. ICMSP and HBF are competing architectural visions for the same problem. They are not mutually exclusive -- you could theoretically deploy both -- but they represent very different value capture models.
In an HBF world, SanDisk and other NAND manufacturers create a new class of memory products with defensible IP, high margins, and deep integration into GPU architectures. HBF sits on the same interposer as the GPU, shares the same electrical interface as HBM (with minor protocol changes), and becomes a first-class citizen in the memory hierarchy. GPU vendors would need to design their silicon to accommodate HBF. SanDisk would be selling a differentiated, high-margin product into an architecturally locked-in slot.
In an ICMSP world, the intelligence lives in the DPU and the software stack (DOCA, Dynamo, NIXL). The SSDs are commodity components. They are fast, they are dense, but they are interchangeable. Solidigm, Samsung, Micron, Kioxia, and SanDisk would all compete to fill those 150 TB slots behind each BlueField-4. The value accrues to Nvidia (which sells the DPU, the networking, and the software) rather than to the memory vendor.
Both paths are good for SanDisk’s top-line revenue. Enterprise SSD demand grows in either scenario. But the margin profile, the competitive moat, and the justifiable market cap are radically different.
There is a critical detail that many HBF bulls overlook. SK hynix -- SanDisk’s HBF partner -- is simultaneously working with Nvidia on the AIN-P, a 100-million-IOPS AI SSD designed for exactly the ICMSP use case. If that SSD, connected to Nvidia’s BlueField-4 via ICMSP, delivers adequate performance for KV cache offload, the engineering case for HBF weakens considerably. As Blocks & Files noted in January 2026: “If there is a general SSD industry HBF standard and if Nvidia adopts HBF as a technology direction, then HBF technology has a future. Absent these two things it will struggle.”
Nvidia has not publicly endorsed HBF. Neither has AMD. A couple of Google engineers have discussed it, but no major GPU vendor has committed to designing HBF into their next-generation silicon.
And there is the fundamental physics issue. NAND flash has inherently higher latency than DRAM. SanDisk is engineering around this with parallelism (reading many sub-arrays simultaneously through thousands of pins), but HBF will never match HBM on per-bit latency. For read-intensive inference workloads with batch-friendly access patterns, this may be tolerable. For workloads requiring rapid random access to small data structures, it may not. SanDisk has also been notably quiet about write endurance -- NAND has a finite lifespan -- and write speeds, which lag DRAM significantly.
None of this makes HBF a bad idea. It makes it an unproven idea with a plausible but contested path to adoption. The question investors should be asking is what evidence would move HBF from possible to probable -- and that distinction is the difference between an $85 billion market cap and a $30 billion one.
Who Captures the Value?
Here is how to think about SanDisk’s future as a framework rather than a prediction.
Scenario 1: HBF wins. Nvidia, AMD, or a major hyperscaler adopts HBF as a standard memory tier. GPU architectures are redesigned to accommodate HBF alongside HBM. The HBF specification is standardized (likely through JEDEC or a similar body, with SK hynix and Samsung as co-developers). SanDisk, as the technology originator with the most advanced CBA wafer bonding IP, captures outsized margins on a product category that grows from zero to $12+ billion by 2030. The stock is undervalued even at $85 billion.
Scenario 2: ICMSP wins. Nvidia’s BlueField-4 + NVMe SSD architecture becomes the dominant approach to inference memory scaling. HBF remains a niche or research technology. SanDisk still benefits enormously from enterprise SSD demand -- every ICMSP enclosure needs 600 TB of drives -- but competes as a commodity NAND supplier alongside Samsung, Micron, SK hynix, and Kioxia. Gross margins revert toward historical NAND averages (25-35%) rather than the 51% currently enjoyed. The stock is overvalued at $85 billion, though possibly fairly valued in the $30-50 billion range.
Scenario 3: Both coexist. The memory hierarchy becomes more granular, with HBF serving a specific role (GPU-adjacent, latency-tolerant, high-capacity inference workloads) while ICMSP handles a different one (pod-level, network-attached, shared KV cache). This is the Professor Kim model -- the underground library alongside the networked bookshelf. SanDisk sells both HBF modules and enterprise SSDs. Valuation depends on the relative size of each market and SanDisk’s share.
Scenario 4: Something else entirely. DeepSeek’s recent research on decoupling compute from memory pools, Kioxia’s 100-million-IOPS super-SSD in development with Nvidia, or an as-yet-unannounced technology leapfrogs both HBF and ICMSP. The semiconductor industry is not short on surprises.
The way to think about the current $85 billion valuation is as a real option. The enterprise SSD business provides a floor -- even in the bear case, SanDisk is a $30-50 billion company with structural demand tailwinds. HBF is the call option layered on top: asymmetric upside if adopted, limited incremental downside if it isn’t. The market is pricing in significant probability of Scenario 1 or 3. Whether that probability is warranted depends on signposts that won’t arrive until late 2026 at the earliest.
The honest answer is that no one -- not SanDisk’s management, not Nvidia’s architects, not the analysts with $1,000 price targets -- knows which scenario plays out.
The near-term case (12-18 months) is strong regardless of HBF’s fate. Enterprise SSD demand is structural, NAND pricing is favorable, SanDisk’s Kioxia JV gives it cost-competitive manufacturing locked in through 2034, and the company’s BiCS technology roadmap (BiCS9 now, BiCS10 in 2026 with 332 layers and 59% higher bit density) is aggressive and credible. The 51% gross margins won’t last forever -- NAND is still ultimately a commodity -- but the current cycle has legs. Key events to watch: the Kioxia JV’s BiCS10 production milestone expected mid-2026, the BlueField-4 ship date in late 2026, and any indication from Nvidia or AMD regarding HBF integration in next-generation GPU silicon.
The long-term case (3-5 years) depends almost entirely on whether HBF creates a defensible new product category or remains an impressive technology demo. The key signpost is not SanDisk’s next earnings call. It is whether Nvidia’s Vera Rubin architecture, or its successor, includes an HBF slot. If it does, SanDisk’s moat deepens dramatically. If it doesn’t, the $85 billion valuation is pricing in a future that may not arrive.
The Architecture Story
There is a pattern in technology investing that repeats with remarkable consistency. The market initially pays for the obvious bottleneck (GPUs), then reprices the adjacent bottleneck (memory), then reprices the infrastructure around it (networking, power, cooling). Each repricing feels surprising in the moment but obvious in retrospect. The transition from CPU to GPU dominance followed the same arc -- companies that positioned themselves at the new bottleneck saw valuations transform as the market caught up to the architectural reality.
SanDisk’s 1,500% run is the market repricing memory. The question now is whether the repricing is complete or whether it has barely started.
The bull case is seductive: AI inference is memory-bound, context windows will keep growing, KV caches will keep expanding, and the memory hierarchy will keep getting more complex. SanDisk, with its HBF technology, its 25-year manufacturing JV, its BiCS roadmap, and its partnerships with SK hynix and Samsung, is positioned at the center of this shift.
The bear case is uncomfortable but intellectually honest: NAND is still NAND. It is dense and cheap, but it is slow and it wears out. Nvidia has historically captured the value in every tier of AI infrastructure it touches, and ICMSP is Nvidia doing what Nvidia does best -- controlling the stack and commoditizing the components. HBF is a multi-year R&D bet that requires GPU vendors to redesign their silicon, something no GPU vendor has committed to.
This is not a storage story. It is an architecture story. The AI inference memory hierarchy is being redesigned in real time, and decisions made in 2026 and 2027 about where flash sits in that hierarchy will shape AI infrastructure for the next decade.
Architecture stories are where fortunes are made or lost. HBM was an architecture bet. So was NVMe. So was PCIe. Some architecture bets create trillion-dollar ecosystems. Others end up as footnotes in IEEE proceedings.
SanDisk is making the biggest architecture bet in the flash memory industry’s 37-year history. Whether it pays off is not a question of technology -- the technology works. It is a question of ecosystem adoption, standards politics, and whether Nvidia decides to let NAND sit at the table. That is the only question that matters for the $85 billion valuation, and no one can answer it yet.
Disclosure: This analysis is for informational purposes only and does not constitute investment advice. The author is not a financial advisor. https://zke-solutions.com