top of page
James F. Kenefick Website Icon

JAMES F.

   KENEFICK

Make Oversight Operate Itself

  • 2 hours ago
  • 8 min read

The oversight gap is now the main AI risk 

Most boards no longer ask, “Should we use AI?” They ask, “Can we prove it’s under control?” 


The stakes are high. The latest IBM Cost of a Data Breach puts the global average breach cost around $4.88M, driven by downtime, lost customers, and the complexity of incident response. At the same time, AI use is exploding: McKinsey’s most recent State of AI work shows around 80–90% of organizations now use AI in at least one business function, with generative AI adoption rising just as fast. 


Regulators are catching up. The EU AI Act introduces a risk-based regime with four categories—unacceptable, high, limited, minimal—placing strict obligations on “high-risk” systems, especially in areas like employment, credit, and critical infrastructure. ISO/IEC 27001 remains the global baseline for information security management systems (ISMS), demanding defined controls and continuous improvement. And NIST CSF 2.0 now adds a new Govern function to make clear that risk decisions and accountability must be explicit, not implied. 


The paradox: AI adoption is scaling, but oversight still runs on Word docs, spreadsheets, and tribal knowledge. Oversight is something people promise, not something the system does

Make Oversight Operate Itself

If 2024 was about “AI everywhere,” 2026 has to be about oversight that operates itself: policy turned into code, evidence produced by default, and continuous conformance against your CSF and ISMS. 

 

Executive brief: what oversight must look like in the AI era 

For boards, CISOs, CDOs, and GRC leaders, the oversight agenda shifts from policy writing to policy execution


  • Start from purpose limits, not features. Every AI/agentic system should have a clear, approved purpose aligned to the EU AI Act’s risk categories and your ISO 27001 risk treatment plan. 

  • Make approvals and routing executable. Approval workflows, regional routing, and high-risk exceptions must be encoded as rules—not left to inboxes and chance. 

  • Treat models and agents as assets. Maintain model/agent “cards” with purpose, data, risk, and control info—living artifacts, not slideware. 

  • Automate retention and evidence. Map retention rules to AI use cases and have systems enforce them, with immutable logs that double as audit evidence. 

  • Use NIST CSF 2.0 and ISO 27001 as operating rails. Continuous conformance means mapping each AI system to Govern/Identify/Protect/Detect/Respond/Recover and the relevant ISMS controls—and monitoring that map over time. 


Do that, and oversight stops being a once-a-year fire drill. It becomes part of how your business technology support and AI stack actually run. 

 

Step 1: Purpose limits, approval workflows, and regional routing 

Purpose limits as the first control 

The EU AI Act is built around purpose and context, not algorithms in the abstract. It classifies AI systems by use—credit scoring, hiring, critical infrastructure—and assigns obligations accordingly. 

Most organizations still do the opposite: they think in terms of tools (chatbot, summarizer, recommender) and bolt on purpose later. That is backwards. 


A practical starting point: 

  • Define purpose statements for each AI/agentic system: “This agent is allowed to propose and, under policy, execute X in domain Y for population Z.” 

  • Attach a risk category based on EU AI Act logic and your own GRC criteria: high-risk vs limited-risk vs minimal-risk. 

  • Link each purpose to ISO/IEC 27001 risk assessments and controls (data access, logging, supplier risk). 


Purpose becomes the backbone: if a model or agent tries to wander outside its defined use, the orchestrator should treat that as a policy violation—not a clever innovation. 


Approval workflows that exist in more than email 

High-risk purposes (in EU AI Act terms) and sensitive data uses (in ISMS terms) should never be one-click deployments. You need approval workflows that are: 

  • Role-based: specific approvers in risk, legal, security, and business. 

  • Traceable: who approved, under which conditions, and when. 

  • Bound to configurations: changes to data sources, prompts, or action surfaces should re-trigger review, not slide past it. 

Instead of a “please approve” email, think in terms of a workflow engine that won’t let an AI service move from test to production without completed steps, captured as structured data. 


Regional routing baked into the fabric 

The EU AI Act and GDPR-era enforcement made one thing clear: where data lives and is processed matters. Global companies that run a single monolithic AI stack are already under pressure to support regional routing, with different behavior for EU residents vs others. 

Oversight that operates itself needs routing rules like: 

  • EU-resident personal data must only be processed by models running in EU-approved regions, under EU AI Act–compliant controls. 

  • Certain purposes (e.g., employment screening) are disabled or restricted in specific jurisdictions. 

  • Logs and training data for EU systems are kept separate with region-specific retention and access rules. 

Those rules should live in configuration and policy-as-code, not just in the DPO’s memory. 

 

Step 2: Model/agent cards and retention rules that the system enforces 

Model and agent cards as living control objects 

“Model cards” and “system cards” are now mainstream concepts—but too many exist as static PDFs in a GRC folder. To make oversight operate itself, model and agent cards need to be structured, queryable objects

Each card, at minimum, should hold: 

  • Identity: name, owner, environment, version. 

  • Purpose: the approved use(s) tied to risk category and policies. 

  • Data: input sources, feature stores, training/finetuning datasets, regional considerations. 

  • Controls: which ISO 27001 controls and NIST CSF 2.0 outcomes it depends on (access control, monitoring, incident handling). 

  • Limits: action surfaces (where it can act), budgets (monetary, latency), and escalation rules. 

  • Evidence hooks: how its decisions are logged and where those logs live. 


If your AI catalog or configuration management system can’t answer, “Show me all agents that touch HR data in the EU and can write to production,” you don’t have oversight; you have hope. 


Retention rules that “just happen” 

Data retention is where policy and operations most often diverge. Your privacy notices might promise 12-month retention; your AI pipelines quietly keep embeddings and logs “for tuning” indefinitely. 


Bringing AI into conformance with ISO/IEC 27001 and your ISMS means: 

  • Assigning explicit retention periods to each type of data used or produced by models and agents (raw logs, prompts, responses, embeddings, training snapshots). 

  • Mapping those periods to legal and regulatory requirements (labor laws, sector rules, regional privacy regimes). 

  • Implementing automated deletion or anonymization in the data stores—backed by reports you can show 


Bringing AI into conformance with ISO/IEC 27001 and your ISMS means: 

  • Assigning explicit retention periods to each data type used or produced by models and agents (raw logs, prompts, responses, embeddings, training snapshots). 

  • Mapping those periods to your legal and regulatory obligations (EU and local privacy laws, sector rules, contractual commitments). 

  • Implementing automated deletion, minimization, or anonymization in the data stores—not “best-effort” manual cleanup. 


Crucially, this needs evidence: regular reports showing what was removed or anonymized, from which systems, and under which retention policy. That’s how you transform retention from a promise in a policy into a behavior your regulators and auditors can see. 

At that point, model and agent cards stop being paperwork. They become the index for how retention, access, and logging are enforced across the AI estate. 

 

Step 3: Continuous conformance against CSF and your ISMS 

“Compliance” used to mean preparing for the audit once a year. In an AI-heavy environment, that rhythm is too slow. New models, agents, and data flows appear weekly. 

The only sustainable approach is continuous conformance: small, automated checks that tell you whether each AI system still aligns with your frameworks—NIST CSF 2.0, ISO 27001, and (for EU-facing operations) the EU AI Act


Map each AI system into CSF and ISMS once—then monitor 

For every material model or agentic workflow, you want a one-page mapping that answers three questions: 


  1. Which CSF functions does this system touch? 

Almost everything will touch Govern and at least one of Identify, Protect, Detect, Respond, Recover. For example, a fraud-detection model is part of Detect/Respond; a customer-service agent that can issue refunds touches Protect (against fraud) and Respond (to customers). 

  1. Which ISO 27001 controls are in play? 

Access control, logging, supplier management, cryptography, development security—there will be a specific subset that matters. The model/agent card should list those controls explicitly. 

  1. Which EU AI Act obligations apply? 

If a system is high-risk under the Act, you note the relevant obligations: risk management, data and data governance, documentation, transparency, logging, human oversight. 


Once you have that map, you can attach checks to it: 

  • Are the right logs still being produced? 

  • Have any action surfaces or data sources changed without new approvals? 

  • Is latency, error rate, or override rate drifting in a way that suggests policy is being strained? 


This is where your security monitoring, data catalogs, and GRC tooling must stop living in separate universes. For oversight to “operate itself,” they need to share enough metadata to answer: “Is this AI system still behaving within its agreed risk envelope?” 


Put conformance on the same dashboard as performance 

The temptation is to bury conformance in the GRC stack and keep performance in CX/operations dashboards. In practice, that guarantees misalignment. 

A better pattern: 

  • For each high-impact AI/agentic system, show business metrics (e.g., tickets resolved, revenue impact, CX scores) alongside oversight metrics (policy violations, approval exceptions, override rates, log completeness). 

  • When you brief the board or audit/risk committee, frame AI systems as controls that have performance and conformance properties, not as experiments. 


The goal isn’t a perfect green dashboard. It’s a clear, defensible story: “Here are the AI systems in production, here is how they map to EU AI Act risk categories, CSF/ISMS controls, and here is how we know they’re still within bounds.” 

 

What an Oversight Implementation Guide should contain 

To make this concrete, you need a simple, reusable artifact—an Oversight Implementation Guide that travels with each new AI initiative and becomes your internal standard. 

At minimum, it should contain: 

  • Purpose statement templates 

Short, constrained descriptions of what the model/agent is allowed to do, linked to EU AI Act risk categories and your internal risk taxonomy. 

  • Approval workflow patterns 

Who must sign off, in which order, for which risk levels. How design changes (new data sources, new actions) re-trigger review. 

  • Regional routing rules 

How EU, UK, US, and other regions are treated differently; which purposes are allowed or blocked per region; which infrastructure each region’s data may touch. 

  • Model/agent card templates 

Structured fields for identity, purpose, data, controls, limits, and evidence, ready to drop into your catalog or configuration system. 

  • Routing and escalation rules 

When an AI system encounters an out-of-scope purpose, missing data, or ambiguous policy, where does it escalate? To whom? With what context? 

  • Retention rule matrices 

Reference tables that tie data types and log categories to concrete retention periods and anonymization behaviors. 

  • Continuous conformance checkpoints 

A small list of automated checks that must be in place before go-live—mapping to CSF Govern/Identify/Protect/Detect/Respond/Recover and to your ISO 27001 control sets. 

For business technology support teams—whether in-house or from a partner—this Guide becomes the joint playbook. Engineering, data, and GRC stop arguing ad hoc and start using the same templates and thresholds. 

 

Why this matters now 

Two forces are moving in parallel: 

  • AI and agentic workflows are seeping into every operational area—CX, finance, HR, logistics, SOC—whether you plan for it or not. 

  • Regulation and expectations are catching up. The EU AI Act, sector regulators, and updated security frameworks make it clear that “we didn’t know” is no longer a defensible position. 


You can either keep bolting ad-hoc approvals and after-the-fact documentation onto each AI project, or you can industrialize oversight: purpose limits, policy-as-code, and evidence that runs as part of the system, not as an afterthought. 

Oversight that operates itself doesn’t mean fewer humans. It means the humans in risk, audit, and technology spend their time on exceptions and design, not on reconstructing basic facts every time someone asks, “Why did the AI do that?” 

 

Publish the Oversight Implementation Guide 

If your current AI oversight lives mostly in slide decks, policy PDFs, and one-off risk workshops, you have a gap—no matter how many committees you’ve stood up. 

The next strategic but practical move is clear: 


Publish the Oversight Implementation Guide 

Turn governance from a promise into a product by standardizing: 

Purpose statements aligned to EU AI Act risk tiers and your risk taxonomy 

Approval workflows and regional routing rules that systems can actually execute 

Agent/model card templates with clear limits, controls, and evidence hooks 

– A small set of continuous conformance checks mapped to NIST CSF 2.0 and ISO 27001 


Use it to brief your board, your GRC steering group, and your engineering leaders. The question is no longer whether you can write an AI oversight policy. It’s whether you can make oversight operate itself—consistently, transparently, and at the speed your AI portfolio is growing. 

Comments


bottom of page