AI storage bottleneck

Webcast

Your GPUs Shouldn’t Wait on Storage

Your GPUs Shouldn’t Wait on Storage: Why AI Workloads Need Always-Hot Infrastructure

Picture of DataStorage Editorial Team

DataStorage Editorial Team

Watch the Full Webcast

🎙 DataStorage.com x Backblaze: Scaling Always‑Hot Storage for AI, Media, and SaaS Workloads
Hear how cloud storage is evolving to support high‑throughput AI, media, and SaaS infrastructure.

Table of Contents

When Storage Becomes the Bottleneck

As AI workloads scale, many teams discover that compute is no longer the limiting factor. Storage becomes the constraint.

During the conversation, Gleb Budman, CEO of Backblaze, explained that teams often over‑invest in GPUs without addressing the data layer.

“You can throw GPUs at the problem, but if you can’t feed them fast enough, they sit idle. That’s wasted money. Storage throughput becomes the bottleneck.”

Legacy object storage systems were designed for backup and archival use cases, not for sustained, high‑throughput access. When training models or running large‑scale inference, latency at the storage layer translates directly into higher infrastructure costs.

Always‑Hot Storage: Real‑Time Architecture for Modern Workloads

Backblaze B2 Cloud Storage was built with a different assumption. Data should always be available, without tiering or retrieval delays.

“Our customers are using B2 to feed training pipelines, serve media, or move large volumes of data continuously,” Budman said.
“The expectation is that it’s always available.”

Instead of moving data between hot, warm, and cold tiers, the platform keeps everything accessible at all times. That simplicity removes friction for AI teams, media workflows, and SaaS platforms that depend on fast access to large datasets.

Performance Gains When Storage Stops Slowing Compute

Budman described how storage throughput directly affects GPU efficiency.

“If you’re training a vision model and every image fetch adds latency, that time stacks up very quickly.”

By removing storage delays, teams reduce idle GPU time and complete training runs faster. While the exact gains vary by workload, the principle is consistent: faster data access leads to better utilization of expensive compute resources.

Why “Cold Data” No Longer Exists in AI Pipelines

One of the strongest themes from the discussion was how AI is changing the value of older data.

“There’s all this data people thought wasn’t useful anymore,” Budman said.
“But now with AI, they’re realizing it’s incredibly valuable, and they need access to it.”

Media archives, historical logs, and long‑tail datasets are now actively reused for training, analytics, and multimodal models. Storage systems that assume infrequent access create delays and unnecessary cost when that data suddenly becomes critical.

What to Look For in AI‑Optimized Object Storage

As storage becomes performance‑critical, teams should evaluate platforms based on real workload needs:

  • High sustained throughput
  • Low‑latency access without retrieval delays
  • Always‑hot availability without tier transitions
  • Amazon S3 compatibility for existing tools
  • Predictable pricing, especially around egress
  • Proven scalability to large active datasets

“It’s not just about how much data you can store,” Budman said.
“It’s about how quickly you can use it.”

Dig Deeper

🎧 In the full webcast, John Kosturos and Gleb Budman discuss:

  • How egress fees shape infrastructure decisions
  • Why hybrid cloud architectures are becoming the norm
  • How AI, media, and SaaS teams are moving away from legacy storage tiers

Share this article

🔍 Browse by categories

🔥 Trending Articles

Newsletter

Stay Ahead in Cloud
& Data Infrastructure

Get early access to new tools, insights, and research shaping the next wave of cloud and storage innovation.