Why I Stopped Asking ChatGPT to Think for Me
A single LLM gravitates toward consensus. Real analysis requires friction.
One evening, I was preparing a strategic analysis for a client. Three hours circling the problem, I decided to ask ChatGPT. The answer came back in seconds — well-written, well-structured. Completely useless.
It was a synthesis of what I already knew. No new angle, no contradiction, no friction. Just the weighted average of everything written on the subject, delivered with confidence.
The problem isn't the model's quality. It's its architecture.
Classic LLMs are designed to converse, not to analyze. By predicting the most probable next word, a single model structurally gravitates toward average and consensus. This is called mode collapse, a phenomenon amplified by alignment methods like RLHF, which pushes the model to favor smooth responses at the expense of divergent perspectives.
For a consultant, this is a dead end. I'm not paid to give the most probable answer. I'm paid to identify blind spots, challenge assumptions, and produce thinking the client wouldn't have reached without me.
Multiple Intelligences Beat a Single One
During my years in consulting, what consistently produced the best analyses was confrontation. A market expert contradicting a technical expert. A lawyer pointing out what the strategist had missed. The value wasn't in stacked knowledge — it was in the friction between perspectives.
That's where the intuition behind Colecia came from. Rather than a single model simulating multiple voices, multiple genuinely independent intelligences that dialogue and challenge each other.
In practice, depending on the question, between 2 and 8 specialized experts emerge automatically. For a question on solid-state batteries: a materials researcher, a patent analyst, a technology strategist. For a competitive analysis in agentic AI, different experts convene. No fixed roles — an analyst team assembled on demand for each brief.
The core of the system is the confrontation mechanism. Agents don't work in silos — they debate. Recent studies show that multi-agent debate frameworks achieve superior accuracy compared to a single LLM (Can LLM Agents Really Debate?). By institutionalizing intellectual conflict, we counter the structural groupthink of monolithic models.
What Changes for a Decision-Maker
Two things, concretely.
Robustness first. Agents consult verifiable external sources (patents, sector reports, scientific publications) and challenge each other on the validity of their assertions. This transforms a probabilistic output into an auditable synthesis.
Traceability second. Every assertion can be traced back to its source. You no longer ask the AI "what do you think about this?" — you ask it to "analyze this from every angle, cite your sources, and surface the risks."
Colecia doesn't sell writing time savings. It doesn't replace human decision-making. It provides the deliberative raw material that executive committees need.
Open Beta
Colecia is in open beta as of today. The approach is simple: move from content generation to analysis generation.
If you work in R&D, innovation, or strategy and recognize the frustration of flat ChatGPT outputs, I'm looking for demanding early users — those who will push the limits, flag what works and what breaks. Try it at colecia.com
Coming soon
Try Colecia yourself
We're looking for R&D, strategy and innovation teams ready to explore multi-agent AI.