Want to run AI on your PC? You’re gonna need a bigger hard drive

You May Be Interested In:LG’s smart washer & dryer solved my problems. But now I’m trapped


AI models will need more storage soon

When people talk about the “size” of an AI model, they’re referring to the number of “parameters” it contains. A parameter is one variable in the AI model that determines how it generates output, and any given AI model can have billions of these parameters.

Also referred to as model weights, these parameters occupy storage space to operate properly — and when an AI model has billions of parameters, storage requirements can quickly balloon.

Here are a few current examples:

  • Google Gemma 2 — 1.71GB (2 billion parameters)
  • Meta Llama 3.1 — 4.92GB (8 billion parameters)
  • Qwen 2.5 Coder — 8.52GB (14 billion parameters)
  • Meta Llama 3.3 Instruct — 37.14GB (70 billion parameters)
  • DeepSeek V3 — 221.25GB (671 billion parameters)

As you can see, the storage space consumed by an LLM increases with the size of its parameters. The same is true for other types of generative AI models, too. For example, Stable Diffusion v1.4 takes up about 4GB while Stable Diffusion XL consumes over 13GB. OpenAI’s Whisper, a speech-to-text model, consumes between 1GB and 6GB depending on whether you choose a lightweight variant or the full-fat model.

Matt Smith / Foundry

But there’s something else that also increases with the size of an AI model’s parameters: its intelligence. Llama 3.1 405B, the biggest version of Meta’s AI model, scored 88.6 on the MMLU benchmark, meaning it correctly answered 88.6 percent of questions in the benchmark. Meanwhile, the lesser Llama 3.1 70B model scored 86 and the even lesser Llama 3.1 8B scored 73.

To be clear, there are ways to improve the intelligence of AI models without increasing parameters. That was a key innovation of DeepSeek-V3, the Chinese AI model that emerged from nowhere in December 2024 with performance that rivaled OpenAI’s best.

Still, the very best AI models are massive. DeepSeek-V3 is presumed to be smaller than OpenAI’s models — I say “presumed” because OpenAI doesn’t disclose parameter counts for its cutting-edge models — and yet it’s still over 220GB. Most apps won’t require a local installation of DeepSeek-V3 to power AI features, but they might want to use mid-sized models like Qwen 2.5 14B (8.52GB) or Llama 3.3 70B (37.14GB).

And it gets worse for your hard drive, because different AI models have different strengths. Some are better for writing in natural language while others are optimized for logical reasoning or step-by-step planning. That means you can’t just get by with one locally installed AI model. You’ll likely want a few. Perhaps dozens. 

share Paylaş facebook pinterest whatsapp x print

Similar Content

Justin Trudeau
MEJIA, ALIAKBARI: Canada should know costs of Trudeau’s emission plan
Elon Musk’s extensive ties to China, explained
Elon Musk’s extensive ties to China, explained
Natural Gas facility
FINLAYSON: Canada could capitalize as global electricity demand rises
Deepseek logo large
Microsoft ports DeepSeek’s AI to Copilot+ PCs, and their NPUs
Décomplexé, tumultueux, vengeur: qu’attendre du second mandat de Donald Trump?
Décomplexé, tumultueux, vengeur: qu’attendre du second mandat de Donald Trump?
Dante Fabbro réclamé... mais pas par le CH
Dante Fabbro réclamé… mais pas par le CH
The News Frontier | © 2025 | News