Agents need control flow, not more prompts

· ai-agents coding · Source ↗

TLDR

  • Reliable agents require deterministic control flow encoded in software, treating the LLM as a component rather than the system.

Key Takeaways

  • Prompt chains are non-deterministic and weakly specified; once you write MANDATORY or DO NOT SKIP, you have hit the prompting ceiling.
  • Software scales through recursive composability; prompt chains lack this property and collapse under complexity.
  • Deterministic scaffolds with explicit state transitions and validation checkpoints are the proposed alternative to elaborate prompt chains.
  • Without programmatic verification, teams are left with three failure modes: Babysitter (human-in-loop), Auditor (exhaustive post-run checks), or Prayer (vibe-accepting outputs).

Hacker News Comment Review

  • Commenters broadly agree on the diagnosis but split on the fix: one prominent view is that LLMs should move entirely to code-generation at design time, shrinking their runtime role to input validation.
  • The tradeoff is real: adding control flow introduces its own edge cases, and no universal framework has convincingly solved dynamic adaptability alongside reliability.
  • The AI coding breakthrough is cited as practical evidence: gains came from moving process execution into the harness, not from smarter prompts.

Notable Comments

  • @bwestergard: argues LLMs should shift to writing software at design time, with runtime role shrinking to helping users choose compliant inputs.
  • @apalmer: “the breakthrough in ai coding was not that AI intelligence increased” but that execution moved into the harness.

Original | Discuss on HN