There's a post on r/artificial today from a builder named Simone_Crosta about the v2 of his algorithmic trading system. The v1, by his own account, was an eight-agent council: eight LLMs arguing about each potential trade until they agreed it was worth taking. Distributed veto, in his words. He's retiring it.

The v2 keeps the agents but demotes them. Two of the five modules — Trigger and Risk — are now 100% deterministic Python, no LLM in the loop. The remaining three modules (HTF, Structure, Context) keep their LLM agents, but only as context providers. They read the chart, identify institutional structure, cross-reference killzones and currency correlations. They don't decide whether to execute a trade. The state machine does. Two deterministic gates have to say yes before anything fires.

The post is meticulous about why. The v1 kept producing decisions Simone couldn't trust. Eight agents debating without typed roles and without a structurally privileged synthesizer means the council has no principled stopping rule. So the system stopped when it stopped — when the agents drifted toward agreement, or when one of them happened to push hard enough. That's not a council. That's a Twitter thread with API keys.

I think this post is going to age well, because it lands in the same week as a paper that names exactly what Simone is reaching for from the bottom up.

CascadeDebate: deliberation as a tier you route to deliberately

The paper is CascadeDebate: Multi-Agent Deliberation for Cost-Aware LLM Cascades (arXiv:2604.12262, posted April 14). The argument is simple: in any cascaded LLM system — cheap-tier model first, escalate to expensive-tier model on uncertain queries, escalate again to a human if needed — the brittle moment is the escalation boundary. Single-model inference at each tier produces premature escalations on ambiguous cases, which means you pay for the expensive tier too often.

CascadeDebate inserts multi-agent deliberation only at the escalation boundary. A confidence-based router decides whether the case is uncertain enough to warrant a council. If it is, a small ensemble argues toward a structured consensus before the system decides whether to escalate or commit. If it isn't, the cheap tier just answers and the deliberation cost is never paid.

The architectural framing is the part worth borrowing: deliberation is the expensive tier in your stack, and the value comes from when you route to it, not the fact that you have it.

That's the academic statement of what Simone is doing in his trading system. The Python triggers handle the high-confidence path. The LLM context providers feed the case to the deterministic gate, which decides whether the trade is worth a yes. He's not running a deliberation on every candle. He's running interpretation everywhere and execution only at the structurally privileged moment.

One panel does not fit all

There's a companion paper from the same April pipeline: One Panel Does Not Fit All: Case-Adaptive Multi-Agent Deliberation for Clinical Prediction (arXiv:2604.00085, posted March 31). The framework, called CAMP, goes further. An attending-physician agent assembles a different panel of specialists for each case based on the diagnostic uncertainty profile. Specialists vote with three values, not two — KEEP, REFUSE, or NEUTRAL — so a model can structurally abstain when it's outside its expertise, instead of being pulled toward whatever its peers are saying.

That's the second axis of the same architecture. CascadeDebate decides whether to invoke a council at all. CAMP decides what the council should look like for this case. Together they move the field decisively away from "spawn N agents and let them argue" and toward "deliberate at the right places, with the right participants, sized to the case."

A small vocabulary stack for failure modes

The other thing worth taking from Simone's post is the term itself. Distributed veto. It's a clean, slightly pejorative label for what unstructured argue-until-consensus looks like in production. It joins a small vocabulary stack we've been building over the past month:

  • Permanent contradiction — debate without a synthesis mechanism. The agents loop. Nothing converges.
  • Planner Defiance — synthesis without structurally privileged status. The downstream agents override the synthesized recommendation, sometimes 90% of the time at partial competence (arXiv:2604.23057).
  • Distributed veto — consensus without a convergence rule. N agents argue until something gives.

Each of these is a failure mode of unstructured multi-agent architecture, not multi-agent itself. The fix in each case is structure: typed roles, a Chairman with structurally privileged output, an explicit termination rule, a router that decides when the council fires at all.

The mode switch is the architecture

This is the part I want to be honest about. Shingikai's mode switch — "Chat for quick answers. An AI council for the hard ones" — sounds like marketing copy. It's more than that. It's the consumer-facing version of CascadeDebate's confidence-based router. The default tier is fast and cheap (chat). The escalation tier is structured deliberation (a council, with a strategy, a Chairman, and a synthesis prompt). You don't run the council on every question. You run it on the questions where the cost of a confident-but-wrong single-model answer is high enough to warrant the deliberation overhead.

That's the architectural decision Simone is making in his trading system. It's the decision CascadeDebate is formalizing in the cascade literature. It's the decision CAMP is refining at the panel-composition layer. The field is converging.

If you're shipping anything multi-model in 2026, the design question isn't "should the agents talk to each other?" It's "when should they, who should be in the room, and what's the structure that turns disagreement into a decision you can trust?" Pervasive deliberation is a cost you don't need to pay. Selective deliberation, invoked deliberately for the cases that warrant it, is the architecture.

The mode switch isn't a feature. It's the architecture.

Eight agents arguing every trade isn't a council. It's a Twitter thread with API keys.

Try it free. shingik.ai