I rebuilt my blog's cache. Bots are the audience now

· Source ↗

TLDR

  • Blog author rebuilt their caching layer after realizing the dominant traffic is AI crawlers and retrieval systems, not human readers.

Key Takeaways

  • Cache architecture was re-evaluated with bot request patterns as the primary use case, not human latency reduction.
  • Author tracks crawler behavior via a dashboard to decide which agents to block or allow before further optimization.
  • The shift implies content infrastructure decisions (CDN placement, cache TTLs) should model retrieval system behavior, not browser UX.

Hacker News Comment Review

  • Commenters split on the strategic response: blocking AI crawlers outright in nginx vs. studying them first to improve index coverage and crawl freshness before blocking.
  • A real tension surfaced around newcomer reputation-building: if blog audiences are now bots, the traditional path of writing to gain human followers and freelance work breaks down.
  • Crawl speed matters beyond bandwidth cost – faster responses affect index depth, content freshness, and potentially ranking in AI retrieval systems.

Notable Comments

  • @rodw: Page load time shapes crawler revisit rate and index depth, making sub-200ms responses strategically relevant even with zero human readers.

Original | Discuss on HN