Let's Talk about LLMs

· ai coding · Source ↗

TLDR

  • Blog post applies Brooks’ No Silver Bullet argument to LLMs: code generation speed addresses only accidental difficulty, leaving essential complexity untouched.

Key Takeaways

  • Brooks’ math holds: if accidental difficulty is under 90% of total effort, eliminating it cannot yield a 10x productivity gain.
  • Brooks estimated 5/6 of software time is non-coding; LLMs speeding up coding leaves that majority unaffected.
  • Rails scaffolding did auto-generated CRUD skeletons 20 years ago; LLMs extend this but don’t change the underlying constraint.
  • The Tailscale CEO framing applies: faster code generation doesn’t drain the review queue faster or compress spec/design/testing time.
  • “Automatic programming” critiqued by Brooks as just a higher-level language; LLMs fit that pattern, with diminishing returns at each abstraction layer.

Hacker News Comment Review

  • Commenters expressed outright fatigue with LLM discourse, treating another “we’re not sure what impact LLMs will have” post as noise rather than signal in mid-2026.

Original | Discuss on HN