Why Think About AI's Risks Alongside Its Practical Applications?
By Aleksei Zulin
Deploying AI without mapping its risks in parallel isn't bold strategy. It's structural negligence - and the organizations doing it will eventually pay for it in ways that make the original efficiency gains look embarrassing.
I say this not as a critic of AI but as someone who spends considerable time thinking about how to use it well. The case for AI's practical value is overwhelming. Diagnosis support in radiology. Contract review that takes minutes instead of days. Code generation that strips away the worst of repetitive work. The applications are real, the productivity gains are measurable, and the competitive pressure to adopt is legitimate.
But the question isn't whether to use AI. The question is whether you can afford to think about the application without simultaneously thinking about what it might break.
The Separated Thinking Problem
Most organizations approach AI in one of two modes. Either they build an applications team focused purely on deployment and ROI, or they build a risk and ethics team focused on compliance and liability. Both teams exist. They rarely talk.
This structural separation feels like due diligence. It isn't. When application thinking and risk thinking happen in separate rooms, the application team moves fast and the risk team reviews after the fact - which turns risk assessment into a post-hoc rationalization rather than a design input. Stuart Russell, in Human Compatible, describes this as the problem of building systems that optimize for a proxy goal without fully specifying what we actually value. The same logic applies at the organizational level. When you separate the "what can this do" conversation from the "what might this damage" conversation, you end up optimizing for a proxy metric - deployment speed, feature count, demo readiness - while leaving value destruction unexamined.
The integrated approach isn't slower. Done well, it's faster, because it catches category errors early.
History Has Seen This Before
Nuclear energy is the obvious parallel, and it's worth taking seriously rather than dismissing as overdramatic.
In the 1950s, the dominant framing was "atoms for peace" - a genuine and justified enthusiasm for what fission could do for energy, medicine, and materials science. The applications were real. Risk conversations were happening in parallel, but not always in the same room as the applications conversation. The result was a technology that generated enormous value and also produced a trail of contamination, near-misses, and institutional failures that took decades to surface. Three Mile Island. Chernobyl. The slow leaching from weapons testing sites that we're still measuring.
Pharmaceuticals learned a similar lesson, more deliberately. Post-thalidomide, regulators mandated that efficacy data and safety data had to be developed together, not sequentially. You couldn't prove a drug worked and then go figure out if it was safe. The two questions had to be answered simultaneously, with the same rigor, drawing on the same investment of resources. That structural change - painful and expensive at the time - is now understood as a prerequisite for trustworthy medicine.
AI will get there. The question is whether it gets there through imposed regulation or through organizations choosing to build the parallel capacity themselves.
The Economic Case Nobody Is Making Loudly Enough
The case for thinking about AI risks alongside applications isn't primarily moral. It's financial.
Joy Buolamwini's work at MIT on algorithmic bias in facial recognition systems - documented in her research and the subsequent Coded Bias documentary - didn't just reveal a fairness problem. It revealed a product liability problem. When a system misidentifies someone at a rate that correlates with their race, that isn't an abstraction. That's a lawsuit. That's regulatory action. That's a customer trust collapse that can take years to rebuild.
Amazon scrapped an internal AI hiring tool in 2018 because it had learned to systematically downgrade resumes from women. The cost wasn't just the development investment. It was the reputational exposure, the regulatory scrutiny, and the lost time on a tool that had to be abandoned rather than deployed.
Arvind Narayanan at Princeton has been methodical in cataloguing the gap between AI capability claims and actual performance in high-stakes settings - and the pattern he documents is consistent. Systems deployed without rigorous failure mode analysis fail in predictable, expensive ways. The failures aren't random. They cluster around distribution shift, around edge cases that were underrepresented, around feedback loops that the application team didn't model because nobody was looking for them.
Early risk integration functions like an insurance policy that pays back many times over. Frame it that way in budget conversations and the resistance tends to soften.
What Integrated Frameworks Actually Look Like
Some organizations have started building this capacity in ways worth studying closely.
Google DeepMind's approach to model evaluation has - at least in published work - moved toward what they call "model cards," structured documentation of where a model performs well and where it doesn't, deployed alongside the model itself. The framing matters. A model card isn't a disclaimer buried in legal text. It's a technical artifact that the application team has to engage with before deployment. Risk information gets embedded in the development artifact rather than siloed in a compliance review.
The EU AI Act, which came into force in 2024, takes a different approach - regulatory rather than voluntary - by requiring risk classification before deployment, with higher-risk applications subject to more intensive review. Whatever you think of the specific regulatory framework, the structural insight holds. Risk classification has to happen before deployment decisions are locked in.
The most practical version of integrated thinking works at the level of the individual product decision. Before deploying an AI feature, a team asks four questions together as a single conversation rather than as sequential sign-offs. What is this system supposed to do? What happens when it's wrong? Who bears the cost of those errors? And do the people affected have any recourse?
The "who bears the cost" question is particularly underused. When an AI system makes an error in a B2B SaaS product, the cost typically falls on the customer's customer - someone who has no relationship with the company that built the model. That distance is exactly what makes it easy to ignore the risk. Forcing the question explicitly closes that distance, and that simple act of closing tends to change the conversation.
The Philosophical Problem Worth Sitting With
Yoshua Bengio, one of the architects of modern deep learning, has spent recent years publicly arguing that the field has moved faster on capability than on alignment - on making systems more powerful than on ensuring that power serves human values reliably. His position has become more urgent as capabilities have accelerated.
Slowing down AI development isn't the answer - I'm not sure that's even possible given the competitive dynamics, and I'm not convinced it would be the right move even if it were. But Bengio's concern points at something that doesn't fully resolve with better risk frameworks or smarter regulation.
When we deploy AI systems that make decisions affecting people's lives - hiring, credit, medical diagnosis, legal risk assessment - we're not just deploying software. We're delegating judgment. And delegated judgment carries moral weight, regardless of how efficient the delegation is.
Kate Crawford's Atlas of AI traces the material and political dimensions of this delegation - the labor, the data, the infrastructure, the power asymmetries embedded in systems that present themselves as neutral. Systems aren't neutral. Thinking about applications without thinking about what values those applications encode is a kind of willful blindness. Whether the willful part is intentional or merely convenient varies by organization.
Maybe the practical answer is simpler than the philosophical problem. Build the capacity to ask both questions together. Don't wait for regulation to impose it. And when you find yourself in a room where only one question is being asked - either "what can this do?" or "what might this break?" - treat that as a warning sign worth naming out loud.
The organizations that get this right won't necessarily move faster than the ones that don't. But they'll move more sustainably. In a technology rewriting itself every eighteen months, that matters more than people currently admit.
FAQ
Why should businesses think about AI risks at the same time as its applications?
Because risk assessment conducted after deployment has already locked in architectural decisions that are expensive or impossible to reverse. Organizations that integrate risk thinking early avoid the category of failure where a system works as designed but produces outcomes - discriminatory, legally exposed, trust-destroying - that no one explicitly chose and no one has a clear plan to fix.
What frameworks exist for evaluating AI risks alongside applications?
Model cards, developed at Google DeepMind, document where a model works well and where it doesn't, making risk information a deployment artifact rather than a compliance afterthought. The EU AI Act's risk-tiering system provides a regulatory structure. Pre-deployment checklists that identify who bears the cost of errors before launch offer a leaner internal approach.
Does thinking about AI risks slow down innovation?
Evidence across deployment histories suggests the opposite over any meaningful time horizon. Systems launched without failure mode analysis fail in predictable, expensive ways - Amazon's scrapped hiring tool, repeated facial recognition misidentification cases, recommendation systems amplifying harmful content. Integrated risk thinking catches category errors while they're still cheap to fix.
Related Articles
About the Author
Aleksei Zulin is the author of The Last Skill, a book on how to think with AI as a cognitive partner rather than use it as a tool. Systems engineer turned writer exploring the frontier of human-AI collaboration.
The Last Skill is a book about thinking with AI as a cognitive partner.
Get The Book - $29