Something interesting happened over the last few months. Andrej Karpathy — the most-followed AI researcher on the internet — shipped a weekend project called llm-council. It dispatches your question to a panel of LLMs, has them anonymously rank each other's answers, and uses a "Chairman LLM" to synthesize the final response. He called it a "fun Saturday vibe code project" and explicitly said he wouldn't maintain it. Two months later, Perplexity launched the same idea as a feature called Model Council — available only to $200/month Max subscribers, running your query across Claude Opus 4.6, GPT-5.2, and Gemini 3.0 simultaneously.
The gap between those two endpoints is more revealing than either product. Karpathy proved the concept works and open-sourced it. Perplexity productized it behind an expensive paywall. And in between, a community of builders started forking the repo, wrapping it in MCP servers, building domain-specific versions for biomedical literature, trading setups, and governance simulations. VentureBeat called Karpathy's hack "the missing layer of enterprise AI orchestration." They're right — it's just that nobody has built that layer yet. That's what we're building.
Why This Moment Matters
The "council" framing has a name now. It's not just a research pattern from a 2023 paper on multi-agent debate — it's a user-facing product category. When Perplexity names a feature "Model Council," they're not being cute. They're betting that users will understand the value proposition immediately: multiple perspectives, checked against each other, synthesized into a more trustworthy answer.
What we're watching is the same pattern that played out with search, then with chat: a proof-of-concept from a researcher, adoption by a consumer product, and then a long period where nobody has built the serious version for actual decision-making. The council framing has crossed from research into product. What hasn't happened yet is anyone building it for the problems that actually need it — complex decisions, high-stakes analysis, situations where getting the wrong answer has real consequences.
What We're Seeing This Week
The governance question is getting serious. Researcher Andy Hall ran a controlled experiment testing four different council governance structures — simple majority vote, deliberation then vote, deliberation with a chairman, and evaluation with a chairman. All four outperformed individual models. But the important finding wasn't which one won — it was that the structure of deliberation matters. How models are asked to engage with each other, whether they can see each other's reasoning before or after forming their own view, and who has final authority all affect output quality in measurable ways. This is what Shingikai has been building: not just a way to query multiple models, but a deliberation architecture.
The "disagreement machine" framing is emerging. A builder on Reddit recently described their multi-agent trading council as a "disagreement machine" — a system designed to force justification before action, not to automate decisions. We think this framing is underrated. The value of a council isn't that it gives you a more confident answer. It's that it surfaces the places where you shouldn't be confident, and makes you defend your reasoning before you act. That's a fundamentally different product than a chatbot.
Community builders keep arriving at the same architecture independently. A builder on r/ClaudeAI built a governance simulation with 12 specialized minister roles (Initiator, Guardian, Analyst) debating policy questions and annotating each other's reasoning. A biomedical RAG builder created a "Research Council" because single-model RAG was giving confident answers on topics where the literature genuinely contradicts itself. Neither knew about Shingikai. Both independently discovered that role specialization, structured deliberation, and explicit synthesis produce better outcomes than asking one smart model.
The Shingikai Angle
This is exactly the problem Shingikai was designed for — not the toy version, but the serious one. Karpathy's hack shows you what council looks like. Perplexity's Model Council shows you that users will pay for it. What neither does is let you configure the council, tune its deliberation mechanics, or apply it to decisions that have stakes. The "council" pattern Karpathy sketched is a reference architecture. Shingikai is the implementation.
We've been watching the governance mechanics question emerge in real-time in the research community, and it confirms something we learned early: the chairman model matters less than the deliberation structure. How you ask models to critique each other — whether they see each other's answers, whether they're given distinct roles or identical prompts, whether the synthesis step is constrained or open-ended — determines whether you get genuine cross-model insight or just averaging. We've built deliberation mechanics that reflect this.
Where This Is Going
The next frontier isn't more models on the council — it's smarter deliberation. The multi-agent debate literature is starting to identify a real risk: a single confident wrong agent can drag the whole council toward an incorrect consensus. The adversarial robustness of council architectures is becoming a research topic, and it will become a product requirement. Watch for council platforms to start competing on deliberation quality, not just model diversity. That's the race worth running.
If you're thinking about decisions where getting the answer from one model isn't enough — we'd love to show you what we've built. Try Shingikai at shingikai.ai.