← Back to Blog
·9 min read

How to Brainstorm Ideas Using AI Prompts (Without Outsourcing Your Thinking)

By Aleksei Zulin

…and that's the moment most people get it wrong. They open a chat window, type "give me ideas for X," get a list of twelve mediocre suggestions, and conclude that AI brainstorming doesn't work. What they missed isn't the technology. It's the conversation structure.

Brainstorming with AI is a dialogue architecture problem. The quality of what comes back is almost entirely determined by how much cognitive work you do before you hit send. Which means the first question worth asking isn't "what prompt should I use?" but "what do I actually need to figure out here?"

That reframe changes everything.


Why Most AI Brainstorming Prompts Fail Before They Start

Researchers studying divergent thinking - the cognitive process that generates novel ideas - have found that the richest ideation happens at the intersection of constraints and freedom. Sarnoff Mednick's associative theory of creativity suggests that the most creative people aren't those with the fewest mental blocks, but those whose associative hierarchies are flat: they connect remote concepts easily. An AI model is, structurally speaking, a very flat associative hierarchy. It has read everything. It can connect anything.

The problem is that flatness without direction produces mush.

When you ask "give me startup ideas," you're giving the model maximal freedom and minimal constraint. You get the median of the internet's startup ideas. Predictable. Boring. When you say "I'm building something for forensic accountants who hate Excel but are required to use it by their firms - what friction points could become product opportunities?" you've introduced enough constraint that the model has to actually think sideways. The gap between those two prompts is the entire skill of AI-assisted brainstorming.

There's also a bias problem nobody talks about enough. Large language models are trained on text that over-represents certain demographics, industries, and problem types. If you ask for "marketing campaign ideas," you'll skew toward B2C consumer goods - because that's where most marketing case studies live. Scientific research brainstorming, infrastructure engineering, public policy design - these are underserved by default model behavior. You have to explicitly pull the model toward those domains, or you'll get the internet's loudest answers, not the most relevant ones.


The Prompting Frameworks That Actually Produce Novel Ideas

Start with constraint injection. Before asking for ideas, give the model three things it cannot do, three things it must do, and one wild card. A product manager I know uses this for feature ideation: "We cannot add another onboarding step, we cannot require a new user permission, we must improve first-session retention - and one idea must be completely counterintuitive." The last clause does disproportionate work. It gives the model explicit permission to be weird, which it otherwise suppresses in favor of plausibility.

Perspective forcing is the second move. Ask the model to brainstorm as a specific person who would hate your idea. Not a critic - a specific archetype with specific objections. "Generate ten objections a 60-year-old manufacturing plant manager would have to this IoT sensor implementation, then suggest how each objection could become a product feature." This is essentially cognitive reframing at scale, and it surfaces constraints you didn't know existed.

Then there's what I call iterative pressure - and this one took me a while to understand (I kept stopping too early). The first response from any LLM is always the most conventional. It's the model's prior. Your job is to apply pressure: "These are obvious. Go deeper. Assume the first three answers are already taken. What would a contrarian engineer propose?" Each round of pressure moves you further from the mean of existing ideas toward something genuinely unusual. Stanford's d.school uses a similar forcing function in human brainstorming sessions - after the obvious ideas are exhausted, the room gets more interesting. Same principle applies here.


Measuring Whether the Ideas Are Actually Good

Nobody asks this question, which is strange, because it's the whole point.

Quality in brainstorming has two axes that are often conflated: novelty and viability. A completely original idea with zero implementation path scores high on novelty and zero on viability. A safe incremental improvement is the inverse. What you want for most practical purposes - product development, research hypotheses, content strategy - is ideas in the viable novelty quadrant: different enough to matter, grounded enough to pursue.

One useful heuristic comes from Paul Romer's thinking on combinatorial innovation (though he applied it to economic growth, not brainstorming). Novel value comes from new combinations of existing elements. When you evaluate AI-generated ideas, ask: what two things is this combining that aren't usually combined? If the answer is obvious, the idea is conventional. If the combination is surprising but coherent, you've got something.

A practical scoring approach: after a brainstorm session, take the output and rate each idea on a 3x3 matrix - feasibility (can we actually do this?), differentiation (is anyone already doing this?), and resonance (does this connect to real user pain?). Only ideas that score 2 or above on all three get moved to the next stage. Run this evaluation with the AI - ask it to stress-test each idea against the criteria. The model is often better at poking holes than generating the initial concept.


Human-AI Hybrid Brainstorming for Teams

Most brainstorming advice is written for individuals. Enterprise reality looks different.

When you're running a brainstorming session for a team of twelve people, AI doesn't replace the group dynamic - it changes the preparation and synthesis layers. Before the session, use AI to generate a stimulus document: fifteen unusual analogies for the problem, five examples from completely unrelated industries, three contrarian positions. Don't share where the document came from. Just drop it in Slack the morning of the session. Research on incubation effects (Dijksterhuis and Meurs, 2006) suggests that unconscious processing of unusual stimuli before a creative task improves outcome quality. You're seeding the room.

During the session, let humans generate. After the session, use AI to synthesize, cluster, and stress-test. This hybrid structure respects something important: group brainstorming is partly a social and trust-building exercise, not purely an idea-generation exercise. Replacing it with AI outputs misses that function entirely.

The scaling question gets harder at enterprise level - 50 people, multiple geographies, asynchronous participation. Collaborative AI tools like Miro's AI features or Notion AI can help aggregate ideas across async sessions, but the prompt architecture still needs a human curator. Someone has to decide what constraints get injected, whose perspective gets forced, what counts as viable novelty. That's a skill, not a feature.


Prompts for Fields Where "Be Creative" Is the Wrong Ask

Technical problem-solving and scientific research brainstorming follow different rules. Creativity in engineering isn't divergence for its own sake - it's constraint satisfaction under novel conditions. A research hypothesis isn't interesting because it's unusual; it's interesting because it's falsifiable, plausible, and not yet tested.

For technical problem-solving, the most effective prompts I've found follow this structure: describe the system you're working in, describe what the system currently does that you wish it didn't, then ask for failure mode analogies from adjacent engineering domains. "This API gateway introduces 200ms latency under load - what failure patterns in distributed manufacturing systems have this signature, and how were they resolved?" You're using AI as a cross-domain pattern matcher, which is what it does well.

For scientific research brainstorming, the productive move is hypothesis generation with automatic red-teaming. "Generate five falsifiable hypotheses about the relationship between X and Y. For each one, immediately describe what result would disprove it." This catches the common failure mode where AI (and humans) generate hypotheses that sound scientific but aren't actually falsifiable. The red-team clause forces rigor into the structure of the prompt itself.

One thing worth sitting with: the best AI-generated research directions aren't usually in the center of the domain. They're at the edges - where the vocabulary of one field maps unexpectedly onto the problems of another. Which means the most useful prompt for a biologist might start with asking an AI what physicists say about emergence, not what biologists say about cellular systems. I haven't fully worked out the implications of this for how research teams should be structured - but it feels significant.


FAQ

How do I know when to keep pushing the AI for better ideas versus accepting what it gave me?

A reliable signal is whether the ideas surprise you. If everything the model generates feels predictable or slightly reworded from your original prompt, apply one more round of pressure with an explicit constraint: "Assume every idea you've given me is already in production somewhere. What would a late-entrant do differently to compete?" If the second pass also feels flat, the constraint structure of your prompt needs rethinking, not more iteration.

What's the difference between brainstorming with GPT-4 versus Claude versus Gemini?

The differences are real but often overstated for brainstorming purposes. Claude tends toward longer, more elaborations on ideas, which helps in iterative refinement. GPT-4 is more list-oriented, which speeds initial divergence but flattens depth. For scientific or technical domains, the model's training emphasis matters more than general capability rankings. Test the same prompt structure across two models and pick the one whose first response surprises you more - that's your proxy metric.

How do I prevent AI brainstorming from just reinforcing my existing biases?

Explicitly name your assumptions at the top of the prompt and ask the model to generate ideas that contradict each one. If you assume your customers want simplicity, instruct the model to generate ideas assuming they actually want complexity and control. Also, actively vary the personas you ask the model to adopt - a skeptic from a different cultural context will surface biases a like-minded collaborator misses entirely.

Can AI brainstorming work for highly specialized technical fields where the model might lack deep expertise?

Yes, but the value shifts. In deep specialization, you're not using AI for domain expertise - you're using it for structural reasoning and cross-domain analogy. Prompt the model to reason about the shape of your problem rather than the content. "What does a system that has this constraint pattern typically look like in other engineering contexts?" The model's breadth compensates for its lack of specialized depth if you frame the task correctly.

Related Articles

About the Author

Aleksei Zulin is the author of The Last Skill, a book on how to think with AI as a cognitive partner rather than use it as a tool. Systems engineer turned writer exploring the frontier of human-AI collaboration.

The Last Skill is a book about thinking with AI as a cognitive partner.

Get The Book - $29