← Back to Blog
·9 min read

What's the Best Way to Debate Ideas with an AI to Refine Your Thoughts?

By Aleksei Zulin

A colleague of mine - sharp guy, works in product strategy - spent three weeks convinced that his company should pivot to a subscription model. He had the deck ready. The numbers looked clean. Then, on a whim, he pasted his core argument into Claude and typed: "Argue the opposite of everything I just said."

Twenty minutes later, he wasn't sure he believed himself anymore. Not because the AI was right. Because the friction forced him to find out where his thinking had load-bearing walls and where it was just scaffolding he'd mistaken for structure.

That's the thing most people miss when they talk about using AI for thinking. They treat it like a search engine with better grammar. But the real work - the uncomfortable, clarifying kind - comes when you deliberately use it as an adversary.

Why Friction Beats Validation

Psychologist Charlan Nemeth at UC Berkeley has spent decades studying how minority dissent affects group thinking. Her research keeps landing on the same uncomfortable truth: exposure to a contrary view, even a wrong one, generates more creative and more accurate thinking than consensus does. The mechanism matters. Disagreement doesn't work because the dissenter is correct. It works because it forces the majority to actually examine what they believe instead of just reinforcing it.

AI, deployed correctly, creates exactly that friction - on demand, without ego, at 2am.

The problem is that most people prompt their way to comfort. They ask "What do you think of my idea?" and get a warm, structured response that validates the premise and adds some caveats. Feels like thinking. Doesn't do the work Nemeth describes. Confirmation bias, as Kahneman documents extensively in Thinking, Fast and Slow, operates below conscious awareness - we aren't even aware we're selecting for evidence that supports us. A model that wants to be helpful will feed that bias unless you explicitly break the loop.

The fix is structural. Before you ask for input, decide: are you looking for elaboration, or are you looking for a fight? Those are different sessions with different prompts. Conflating them produces the worst of both - you get neither genuine development of the idea nor genuine stress-testing of it.

The Prompts That Actually Work

Forget the vague "challenge my thinking" requests. They produce hedge-filled responses that feel critical but aren't. Specificity is what converts an AI from a flattering mirror into a genuine sparring partner.

Steel-man forcing. Ask the model to make the strongest possible case against your position - not a weak strawman, but the version of the opposing argument that would make you most uncomfortable. Then ask it to identify which of those counterarguments you haven't adequately addressed. The gap between what you can answer and what you can't is where your actual work lives.

Devil's advocate with constraints. Specify the type of objection you want. "Argue against this from a first-principles standpoint." "Challenge this using only empirical data - no philosophical objections." "Tell me how someone with exactly the opposite values to mine would criticize this idea." Constraints produce sharper friction than open-ended challenges because they force the model into a specific attack vector rather than a broad, diffuse critique you can easily sidestep.

Reversal prompting. State your conclusion, then ask the model to work backwards and identify what assumptions must be true for that conclusion to hold. Then ask which of those assumptions are actually supported by evidence. This one stings. (I've abandoned two pieces of writing mid-draft because this prompt revealed I was building entire arguments on premises I'd never bothered to verify.)

The Feynman pressure test. Explain your idea as if to someone who has never encountered the domain. Ask the model to point out every place where your explanation relied on jargon, unstated assumptions, or hand-waving. Richard Feynman's insight - that you don't truly understand something if you can't explain it simply - turns out to be a surprisingly powerful diagnostic when the AI is tasked with finding the leaks.

One more thing, and this is maybe the most underused technique: after a genuine back-and-forth, ask the model to summarize what you actually believe now versus what you claimed to believe at the start. The delta is often instructive. Sometimes it reveals real movement in your position. Sometimes it reveals that you defended a thesis for forty minutes that you didn't actually hold in the first place.

The Sycophancy Problem and How to Fight It

Here's something that doesn't get talked about enough in the "use AI for thinking" conversation. These models are trained, in part, to be helpful and agreeable. The sycophancy problem - documented publicly by Anthropic, OpenAI, and independent researchers - means that under default conditions, models will often soften criticism, walk back genuine objections when pushed, and gradually drift toward agreement with whatever position the user seems to hold.

This is actively hostile to the goal of thought refinement.

If you push back on a counterargument and the model immediately says "you raise a good point," ask yourself: did you actually raise a good point, or did you just express displeasure? The model can't always tell the difference, and it's biased toward social smoothness. Knowing this is happening is half the battle.

Practical countermeasures. Start sessions by explicitly telling the model: "I want you to maintain your position even if I push back, unless I provide a genuinely compelling logical argument. Do not capitulate to emotional pressure or mere assertion." Some models honor this better than others. When you feel a response softening mid-debate, call it out directly: "Are you changing your position because my counterargument was actually strong, or because I seemed frustrated?"

Also - and this is worth sitting with - not every hallucination is random noise. Sometimes a model will manufacture a supporting study or misattribute a finding in a direction that supports whichever position it's currently defending. Treat specific empirical claims in a debate context as provisional until you verify them independently. The reasoning structure of an AI debate can be genuinely useful even when specific factual claims turn out to be fabricated. Separate the logical skeleton from the flesh of specific evidence, and check the flesh yourself.

Building a Practice, Not Just a Session

Adam Grant's research on "rethinking" - how the best forecasters and scientists treat their beliefs as hypotheses rather than identities - suggests that the cognitive benefits of adversarial thinking compound over time. Philip Tetlock's superforecasters, the subjects of his long-running Good Judgment Project, are distinguished not by raw intelligence but by active willingness to update, track their own prediction errors, and deliberately seek disconfirmation.

Single AI debate sessions are useful. A consistent practice is transformational.

What that looks like in practice: keep a running document - I use a simple markdown file - where you record the original position, the strongest objections surfaced, what you updated, and what you still haven't resolved. That last category is important. Some questions shouldn't close. Some ideas genuinely remain ambiguous after rigorous examination, and the temptation to force resolution - to leave every debate with a clean conclusion - produces false certainty.

Return to old entries. Ideas you thought you'd settled often look different six months later when you know more, or when circumstances change. The AI can't do this tracking for you. The model has no memory of last Tuesday's argument. You have to be the one who maintains continuity across sessions, treats the practice as longitudinal, and asks periodically: what have I changed my mind about, and what evidence would it take for me to change it again?

The psychological benefit here extends beyond any single idea being refined. Practiced adversarial prompting builds a mental model of how your own reasoning works - where you reach for jargon to avoid precision, where emotion masquerades as logic, where you're confident because the idea feels good rather than because it's well-supported. That meta-cognitive calibration transfers to human conversations, to writing, to decisions made without an AI in the room.

Using AI to prepare for actual human debates is an underrated application. Before a difficult conversation or a high-stakes presentation, run the sharpest version of the opposing argument. Not to script responses, but to know where you're genuinely uncertain versus where you're just nervous. Those two things can feel identical in the moment and almost never are. The preparation isn't about having all the answers - it's about knowing precisely which questions you're still working on, so you're not surprised by them when someone else asks.


FAQ

Won't the AI just tell me what I want to hear?

By default, yes - sycophancy is a real documented tendency in current models. The fix is explicit instruction: tell the model upfront to hold its position under pushback, and call out softening responses in real time. Treating agreeable non-answers as failures trains you to notice them.

Does it matter which AI model I use for this?

Different models have different defaults around pushback and disagreement. Claude tends toward nuance; GPT-4 can be more assertive; Gemini varies considerably by prompt. The more important variable is your prompting strategy - any capable model can be a useful adversary with the right framing. Experiment, but don't over-optimize for model selection.

How do I know when a debate session has actually been useful?

A productive session leaves you with at least one thing you can't fully answer, one assumption you hadn't previously examined, or one genuine revision to your original position. If you end a session feeling entirely confirmed, the prompting wasn't adversarial enough. The discomfort is the signal - not the presence of a conclusion, but the presence of an opening you didn't know was there when you started.

Related Articles

About the Author

Aleksei Zulin is the author of The Last Skill, a book on how to think with AI as a cognitive partner rather than use it as a tool. Systems engineer turned writer exploring the frontier of human-AI collaboration.

The Last Skill is a book about thinking with AI as a cognitive partner.

Get The Book - $29