HATS (rockcat/HATS) assigns multiple AI personas to argue opposing positions, using structured debate between agents to improve decision output.
Key Takeaways
The project creates distinct AI personas that disagree with each other rather than running a single model pass on a problem.
Targets decision quality improvement through adversarial multi-agent argumentation.
The GitHub repo includes lip-syncing avatar visuals for each persona alongside the core debate logic.
No architecture or benchmark details available from the repo preview; scope and depth are unclear from surface information alone.
Hacker News Comment Review
Skeptics question whether significant engineering effort went into avatar lip-syncing rather than the argumentation mechanism itself, raising a polish-vs-substance concern.
One commenter frames HATS as a slower, less efficient variant of mixture-of-experts, arguing the novelty claim is weak against existing ensemble approaches.
Notable Comments
@oldsecondhand: “less efficient version of the mixture of experts approach” – challenges the core differentiation claim directly.
@zby: Finds the idea interesting but flags lip-syncing avatars in the repo and asks how much of the effort is marketing.