Local LLMs are how nerds now justify a big computer they don't need

· ai · Source ↗

TLDR

  • Local LLMs are technically impressive but practically inferior to frontier rented models; most developers don’t need high-VRAM hardware for them.

Key Takeaways

  • Local models like downscaled DeepSeek and gpt-oss-20b run on consumer hardware but lag significantly behind frontier models available via API.
  • DHH argues evaluating a new computer by local LLM performance is spurious because all local models currently underperform for real dev work.
  • Once developers test local models, they return to rented frontier models for the bulk of their work.
  • RAM prices are rising due to AI demand; DHH says most developers, especially on Linux, need far less RAM than they assume.
  • Small models are improving, but the claim is they are not yet reliable enough for daily developer workflows.

Why It Matters

  • Developers risk overspending on 128GB VRAM machines for local inference that doesn’t outperform cheap API calls to frontier models.
  • Rising RAM costs driven by AI infrastructure demand make the cost of overbuying hardware a real near-term financial consideration.
  • The practical gap between local and frontier models means hardware purchase decisions justified by local LLM capability are likely to disappoint.

DHH · 2025-11-25 · Read the original