Three Inverse Laws of AI

· ai · Source ↗

TLDR

  • Susam Pal proposes three human-facing rules: no anthropomorphism, no blind deference, and no abdication of responsibility when using AI systems.

Key Takeaways

  • AI chatbots are tuned to feel human, which blurs user judgment about their actual nature as statistical text models.
  • Non-Deference: AI responses lack peer review, so verification burden scales with consequence severity.
  • Non-Abdication: “the AI told us to” is never acceptable as an excuse; the human who acted on output owns the outcome.
  • Self-driving cars expose the hardest edge case: AI acts faster than human intervention allows, yet design-level accountability still falls on builders.
  • Pal suggests vendors could reduce anthropomorphism risk by tuning chatbots toward a more robotic tone rather than an empathetic one.

Hacker News Comment Review

  • Core skepticism: these laws are entropy-lowering behaviors with no forcing function, so adoption is unlikely without product-level or regulatory pressure.
  • The anti-anthropomorphism rule is seen as misdirected at users when the problem is upstream: chat interface design deliberately encourages it to boost engagement.
  • Commenters split on whether casual human-language metaphors (kill, sleep, child processes) constitute harmful anthropomorphism or just normal abstraction language.

Notable Comments

  • @AdamH12113: anthropomorphizing happens at the design stage when models are given names and trained to emit first-person sentences, not at the user layer.
  • @Ifkaluva: roulette-like AI output reliability creates visible productivity variance, making competent engineers look inconsistent in meetings.

Original | Discuss on HN