Human-in-the-Loop Is the AI Security Community's Favorite Lie
Docker's response to the DockerDash vulnerability tells you everything you need to know about the current state of AI security: we're making the same catastrophic mistakes we made twenty years ago, just with shinier technology.
Their fix? "Explicit confirmation before executing MCP tools."
Sounds secure, right?
Wrong. Dead wrong.
It's security theater—and it's the exact same mistake every AI vendor is currently making.
The Human Approval Trap
Let me be clear: by the time you're asking humans to approve AI actions, you've already lost the security game.
Here's what actually happens in the real world. Humans will click "approve" for one of three reasons:
-
They don't understand what they're approving (the technical details are opaque)
-
They're busy and the prompt is interrupting their workflow
-
The AI made the request look completely legitimate (and it probably is... 99% of the time)
This isn't speculation. We have decades of evidence proving this pattern.
We've Seen This Movie Before
Remember Windows Vista's User Account Control (UAC) prompts? Microsoft's brilliant plan was to interrupt users with security confirmations for every system-level action. The theory was sound: make users conscious participants in security decisions.
The reality? Users trained themselves to click "yes" within weeks.
The security feature became muscle memory, not a decision point. Microsoft had to completely redesign the system in Windows 7 because UAC had become worse than useless—it had become a ritual that gave users a false sense of security while providing none.
The DockerDash "fix" has the exact same UX problem, and Docker isn't alone in making this mistake.
The Fundamental Contradiction
Consider Anthropic's Claude in legal research workflows. The entire value proposition is that it autonomously conducts research, synthesizes information, and prepares materials without constant human supervision. That's not a feature—that's the product. The moment lawyers need to review and approve every action, Claude becomes just an expensive autocomplete.
This is the central contradiction plaguing AI security right now: the value of AI tools comes from removing human bottlenecks, but our security response is to add human bottlenecks back in.
So what actually happens when you deploy these "secure" systems?
Users disable the friction mechanisms. Or they approve everything without reading it. Same security outcome either way—except now you've also destroyed the user experience.
Real Security Is Architectural, Not Procedural
Here's the uncomfortable truth that every AI vendor needs to hear: real security is architectural, not procedural.
You cannot patch a fundamentally insecure architecture with approval workflows. You cannot procedure your way out of a design problem.
The actual fix for vulnerabilities like DockerDash requires rethinking the entire MCP (Model Context Protocol) architecture from the ground up. Every piece of external context needs to be treated as untrusted input—exactly the same way we've treated user input for the past two decades of web application security.
Not as safe configuration. Not as trusted data. As potentially hostile input that needs validation, sanitization, and sandboxing before it gets anywhere near execution.
Yet this pattern of trusting inputs repeats everywhere in the AI ecosystem:
-
AI coding assistants trust the repository context they're given
-
AI meeting transcribers trust the audio sources they process
-
AI email responders trust the message threads they analyze
-
AI search tools trust the websites they scrape
None of them validate before acting. None of them assume adversarial inputs. They're all optimizing for the 99% case while ignoring the 1% case that will destroy your infrastructure.
The Real AI Security Checklist
Your AI security checklist shouldn't start with "add confirmation dialogs."
It should start with "assume every input is adversarial."
This is the boring version of AI security that actually matters. Not the flashy dashboards, not the human-in-the-loop marketing, not the compliance checkboxes.
The unsexy truth is that securing AI requires the same foundational security principles we've known for decades:
-
Input validation at every boundary
-
Principle of least privilege for AI agents
-
Defense in depth, not defense in theater
-
Fail-safe defaults, not fail-open convenience
If you're not doing input validation at the AI layer—treating every external context source as potentially compromised—you're doing security theater. You're putting on a show that makes stakeholders feel better while doing nothing to stop actual attacks.
The Question Every CISO Needs to Answer
Right now, every CISO is being pitched AI tools that promise to "speed up" security operations. AI-powered threat detection. AI-enhanced SOC workflows. AI-driven incident response.
The pitches are compelling. The demos are impressive. The efficiency gains are real.
But here's the question I'm not hearing enough people ask: "Who's validating the AI's assumptions?"
When your AI security tool makes a decision—to quarantine a file, to block a user, to modify a firewall rule—what validates that the context it's operating on hasn't been poisoned? What ensures the training data hasn't been compromised? What checks that the prompt injection attack you can't even see yet isn't happening right now?
Most vendors don't have good answers to these questions. They'll pivot to compliance frameworks, to audit logs, to human oversight processes.
But compliance doesn't stop attacks. Audit logs show you what happened after you've been breached. And human oversight—well, we've already covered that.
The Boring Revolution
Major enterprises are beginning to recognize this gap. When Google announced their Secure AI Framework (SAIF) in late 2023, the most important element wasn't the flashy ML security features. It was the boring commitment to treating AI systems with the same zero-trust principles as any other infrastructure component.
No special exemptions for AI. No "it's too complex for traditional security" exceptions. Just the unglamorous work of applying decades of security knowledge to a new technology domain.
That's the real revolution in AI security: the boring version where we stop treating AI as magical and start treating it as infrastructure that needs to be secured.
It means slower deployments. More architecture reviews. Less "move fast and break things." More "validate inputs and constrain outputs."
It's not sexy. It won't make for good conference demos. But it's the only approach that will actually work.
Your Turn
I already know my answer to the validation question. Every AI system I deploy operates under the assumption that its inputs are adversarial until proven otherwise.
Do you know yours?
More importantly: are you applying the same scrutiny to AI that you'd apply to any other system touching your critical infrastructure? Or are you being dazzled by the technology into accepting security approaches you'd reject in any other context?
The AI security community's favorite lie is that human oversight makes unsafe systems safe. The truth is that architectural security makes AI systems trustworthy enough to deploy at all.
Which version are you building?
More Ai Posts
Why Solo AI Builders Are Your Market Canaries
Solo developers using AI are discovering pricing models and tools enterprises will demand in 2-3 years. Watch them to pr...
Season 1: Masterclass
Dive into the Season 1 Masterclass podcast episode, featuring highlights and diverse perspectives from the past 12 weeks...
Stop Waiting for AI: Your Competition Already Started
AI disruption isn't coming tomorrow—it's happening now. While most companies debate, competitors are shipping. Here's wh...
