Macroplane › Investment Theses › Mass-capacity HDDs are the AI cold-storage / training-data layer

Mass-capacity HDDs are the AI cold-storage / training-data layer

AI supply-chain thesis — mapping bottlenecks, focus companies, and supply-chain exposure for investors.

**Bottleneck theme:** Storage **Focus:** $STX — Seagate Technology Holdings plc Mass-capacity HDDs are the AI cold-storage and training-data layer — the unsexy commodity tier holding the petabyte-scale data lakes that frontier models train on, plus model checkpoints, inference logs, and the cold-tier of vector databases. Seagate is the duopoly leader (with $WDC) in nearline HDD and the most aggressive HAMR ramp story in the industry: 30TB shipping today, 36-50TB on the technology roadmap, with hyperscaler qualifications well advanced. Lead times have stretched into 2027, prices are firming, and supply discipline is tighter than at any point since 2018. The thesis is dual-engine: cyclical recovery from the 2022-2023 inventory correction, plus a secular re-rating as the market finally values nearline HDD as core AI infrastructure rather than a melting ice cube. Higher-capacity HAMR drives also carry better margins than legacy CMR SKUs, so even flat unit shipments deliver meaningful EPS leverage. Pair with $WDC for duopoly exposure or own as the higher-conviction technology leader (HAMR is two years ahead of $WDC's roadmap).

Focus companies in this thesis (1)

  • Seagate Technology Holdings plc (STX)

Supply-chain categories covered

  • HDD — Hard disk drives providing high-capacity, cost-effective storage for AI training data, cold storage, and checkpoints.
  • Storage — Investment-thesis bucket from bottlenecks.app: Storage
  • Hyperscalers — Major cloud operators (AWS, Azure, GCP, Meta, Oracle, Alibaba, Tencent, Baidu, Naver) and tier-2 / neocloud providers (DigitalOcean, OVHcloud, Rackspace, Kingsoft) tracked as a demand signal across multiple theses (photonics, HBM, AI accelerators, power, cooling). Excludes SaaS apps, telcos, REITs, and IT services firms.
  • Memory & Storage Chips — DRAM, NAND, and HBM memory essential for AI training and inference workloads in data centers.
  • Hyperscale Cloud Providers — End customers (CSPs) deploying AI data centers with massive capex.

Thesis milestones & bottleneck markers

  • $STX 30TB HAMR shipments — STX
  • $STX gross margin expansion — STX
  • HDD content of AI data centers >20%
  • $STX exabyte shipped to hyperscalers — STX

Browse all AI supply-chain theses · Macro trends · Industries · Product categories