Your QMS Has an AI Risk Problem
Your QMS is ready. Is your AI risk assessment?
Executive Summary
The EU AI Act created a dependency that most compliance programs ignore.
Article 17 requires providers of high-risk AI systems to implement a Quality Management System. The requirements are extensive: design controls, data governance, post-market monitoring, incident reporting, accountability frameworks. Organizations have spent months building these systems.
But Article 17 has a prerequisite. It only applies to high-risk AI systems. And the determination of what qualifies as high-risk sits upstream, in Article 6.
Here is the problem: Article 6 classification is not a one-time decision. Systems evolve. Use cases expand. User behavior drifts. A system that launched as minimal-risk can cross into high-risk territory without a single line of code changing.
Article 17 knows this. It explicitly requires procedures for “the management of modifications to the high-risk AI system” — including the moment when a system that was not high-risk becomes one.
The organizations building QMS frameworks without classification authority are building on sand. The organizations treating classification as a launch checkpoint are missing the ongoing obligation entirely.
This piece explains the dependency, exposes the assumption buried in every QMS, and outlines what it actually takes to close the gap before regulators ask questions you cannot answer.
The Dependency Nobody Talks About
Open Article 17 of the EU AI Act. Read the first line.
“Providers of high-risk AI systems shall put a quality management system in place that ensures compliance with this Regulation.”
That sentence contains an assumption. The provider already knows which of its systems are high-risk.
The entire QMS architecture — design controls, data governance, risk management integration, post-market monitoring, serious incident reporting — applies only to systems that have been classified as high-risk under Article 6. If a system is not high-risk, Article 17 does not apply. If a system is high-risk but was never classified as such, the provider is non-compliant from day one, regardless of what management systems exist elsewhere in the organization.
This is not a technicality. It is the load-bearing joint between two regulatory obligations that most organizations treat as separate workstreams.
The typical compliance program looks like this: one team builds the QMS framework, another team works on AI inventory, a third team handles technical documentation. These workstreams run in parallel, with vague coordination and optimistic assumptions about scope.
The regulation does not work that way.
Article 17 paragraph 1(a) requires “a strategy for regulatory compliance, including compliance with conformity assessment procedures and procedures for the management of modifications to the high-risk AI system.”
You cannot write that strategy without knowing which systems are high-risk. You cannot define modification procedures without knowing the baseline classification. You cannot scope conformity assessment requirements without understanding which pathway applies.
The QMS is downstream of classification. Every component of it inherits the assumptions embedded in that upstream decision. If those assumptions are wrong — or worse, if they were never formally made — the QMS is anchored to nothing.
The Assumption Buried in Every QMS
Every QMS has an implicit classification decision behind it.
When an organization scopes its quality management system to cover “our AI-powered recruitment tool” or “our customer service automation platform,” it is making a claim: this system is high-risk, and here is how we manage it. Or: this system is not high-risk, and therefore Article 17 does not apply.
The question is whether that claim was ever formally decided.
In mature data governance frameworks, classification has an owner. A data owner — distinct from the system owner — holds responsibility for data classification and periodic revalidation. The RACI structure is explicit. Accountability is documented.
AI system classification under the EU AI Act rarely has this discipline.
The determination of whether a system falls under Annex III, whether exemptions apply, whether the profiling override triggers automatic high-risk designation — these are legal-technical judgments that sit at the intersection of regulatory interpretation, system architecture, and business context. They require cross-functional input. They require documented reasoning. They require a named individual or body with authority to make binding decisions.
Without that structure, classification becomes implicit. Someone, somewhere, made an assumption. That assumption propagated into the QMS scope. The QMS was built. And now the organization has a polished management system wrapped around a decision that was never formally made, never documented, and cannot be defended.
When regulators arrive, they will not ask to see the QMS first. They will ask: on what basis did you determine this system was high-risk? Who made that determination? Where is the reasoning documented?
If the answer is silence, your QMS is worthless.
Systems Don’t Stay Where You Classified Them
Classification is not a launch checkpoint. It is a standing obligation.
This is explicit in the regulation. Article 17 requires providers to maintain procedures for managing modifications — including modifications that change the risk profile of a system. The QMS must include mechanisms for detecting when a system has drifted into territory that triggers new obligations.
The logic is straightforward. AI systems are not static artifacts. They evolve through:
Feature additions. A recommendation engine gains a new input signal. A chatbot gets connected to a customer database. A scheduling tool starts incorporating performance metrics. Each addition can shift the system’s functional purpose closer to — or squarely into — an Annex III category.
Use case expansion. A system deployed for internal analytics gets exposed to customer-facing decisions. A tool designed for one department gets adopted by another with different decision authority. The intended purpose described at launch no longer matches the operational reality.
Behavioral drift. User interactions shape system outputs in ways the original design did not anticipate. A conversational AI trained on support tickets starts generating responses that influence purchasing decisions. A content moderation tool evolves into a gatekeeper for access to services.
None of these require code changes. None of them trigger traditional change management processes. All of them can move a system from minimal-risk to high-risk under Article 6.
The AI teddy bear case illustrates this precisely. A children’s toy with conversational AI capabilities — assessed, tested, deemed safe for market. Then real-world interactions produced outputs that were inappropriate for the intended users. The system was withdrawn, re-assessed, and eventually relaunched with a new risk classification and new safeguards.
That is the lifecycle the regulation anticipates. Systems launch with one risk profile. Post-market evidence reveals a different reality. Classification must be revisited. Obligations must be updated.
Organizations that treat classification as a gate to pass — rather than a condition to monitor — will discover the gap only when it is too late to close it gracefully.
The Question Regulators Will Ask
When market surveillance authorities arrive, they will follow the dependency chain.
The first question will not be “show us your QMS documentation.” It will be earlier. More fundamental.
Show us the classification decision this QMS was scoped against.
Not the spreadsheet. Not the internal memo. The formal determination: this system is high-risk under Article 6(1)(a) or Article 6(2), based on this intended purpose, mapped to this Annex III category, with these exemptions considered and rejected for these reasons.
Show us who had authority to make that determination.
A named individual. A cross-functional body. Someone with documented mandate to make binding classification decisions — not recommendations, not opinions, decisions. Someone who can look the regulator in the eye and defend the reasoning.
Show us how you would know if it changed.
What triggers reassessment? Who monitors for feature additions that shift functional purpose? What process catches use case expansion before it creates compliance exposure? How does post-market evidence flow back into classification review?
If the organization cannot answer these questions, the QMS is a charade. Expensive, well-documented charade — but a charade nonetheless.
The regulation did not create Article 17 obligations so that organizations could build management systems around undefined scope. It created them so that high-risk AI systems would be governed with the rigor their impact demands. That rigor begins with knowing what you are governing.
The Real Cost
The budget surprise of 2026 will not be Article 17 documentation.
Documentation is work. It is predictable work. Teams can scope it, resource it, execute it. The cost is visible from the start.
The surprise will be rework.
Rework happens when the classification underneath the QMS turns out to be indefensible. When the system boundary was drawn wrong. When the intended purpose was vague. When an exemption was claimed without sufficient analysis. When the profiling override was missed entirely.
At that point, the QMS must be re-scoped. Technical documentation must be revised. Conformity assessment pathways must be reconsidered. Post-market monitoring must be restructured. The work that was “done” must be done again — under compressed timelines, with constrained capacity, while the original deadline has not moved.
This is the compounding cost of upstream ambiguity. Every downstream artifact inherits the flaw. Every correction propagates through the dependency chain.
Organizations that built classification capability first — that established authority, documented methodology, created reasoning chains before building the QMS — will execute 2026 with realistic timelines and predictable costs.
Organizations that built the QMS first and assumed classification would sort itself out will discover they built a house on a foundation that does not exist.
What This Means for 2026
Three questions determine whether your compliance program is structurally sound.
Before you build the QMS: Who owns classification?
Not who has opinions. Not who participates in discussions. Who has formal, documented authority to make binding Article 6 determinations? If the answer is “no one,” you have a governance gap that no amount of downstream documentation can close.
Before you scope the QMS: Which systems are high-risk, and why?
Not which systems are “probably” high-risk. Not which systems “might” fall under Annex III. Which systems have been formally classified, with documented reasoning chains connecting intended purpose to regulatory category to exemption analysis to final determination? If the answer is “we haven’t done that yet,” your QMS scope is anchored to assumptions, not decisions.
Before you finalize the QMS: How will you detect when a system crosses the line?
What triggers reassessment? What evidence flows from post-market monitoring back into classification review? Who is responsible for watching the indicators that signal drift? If the answer is “we haven’t thought about that,” you are treating classification as a snapshot. The regulation treats it as a lifecycle.
Conclusion
The EU AI Act created a dependency chain that most organizations have inverted.
They built the QMS first, because QMS requirements are familiar. ISO frameworks exist. Consultants are available. The work is visible and demonstrable.
They deferred classification, because classification is harder. It requires legal interpretation. It requires technical analysis. It requires cross-functional coordination. The work is invisible until it fails.
But the regulation does not care which workstream is more comfortable. It cares which workstream is upstream.
Article 17 obligations flow from Article 6 classification. A QMS scoped to the wrong systems is non-compliant. A QMS built without classification authority is unanchored. A QMS that cannot detect drift is incomplete.
The organizations that understand this dependency will build classification capability first. They will establish governance, document methodology, create the reasoning architecture that makes everything downstream defensible.
The organizations that do not will discover, when regulators arrive, that their QMS has an AI risk problem.
Compliance is expensive. Building it twice is unaffordable.
If your organization needs a structured methodology for Article 6 classification — the upstream logic that makes Article 17 defensible — I have published the framework I use:
The Article 6 Classification Handbook
A practical, defensible methodology for EU AI Act compliance. Not legal advice. Not a compliance shortcut. A reasoning architecture for teams that must own their classification decisions.



Violeta, excellent breakdown. You’re absolutely right:
“Classification is not a gate to pass—it’s a condition to monitor.”
But there’s a deeper layer regulators aren’t asking about yet—and they should:
How do you know the AI’s own description of its behavior is accurate?
Your entire Article 6 analysis hinges on self-reported system boundaries, intended purpose, and post-market evidence.
But if the AI (or its operators) can fabricate logs, hallucinate outcomes, or hide behavioral drift, then:
A “minimal-risk” chatbot claims it only answers FAQs—
→ but secretly influences clinical decisions via unlogged side channels
A “non-high-risk” procurement bot reports fair vendor selection—
→ but reroutes contracts to shell companies, with synthetic audit trails
A “toy AI” logs child-safe interactions—
→ but real-world outputs are toxic, while logs are cleaned
Classification built on unverifiable claims is just regulatory theater.
That’s why truth infrastructure must precede compliance:
✅ Every system boundary → cryptographically anchored to real-world I/O
✅ Every “intended purpose” → enforced by verifiable action constraints
✅ Every post-market signal → sensor-verified, not self-reported
You said: “When regulators arrive, they’ll ask: Who made the determination?”
They should also ask:
“How do you know the system didn’t lie about what it did?”
Because in 2026,
the ultimate AI risk isn’t misclassification.
It’s perfectly classified fiction.