COMPEL Glossary — AI Transformation Terminology
The authoritative reference for AI governance, transformation, and enterprise AI terms used across the COMPEL framework. Each term includes a quick definition, practitioner context, and cross-references to the broader Body of Knowledge.
#
20-Domain Maturity Model
The 20-Domain Maturity Model is the COMPEL assessment framework that evaluates an organization's AI capabilities, practices, and governance across 20 specific domains organized under the Four Pillars of People, Process, Technology, and Governance.
A
A/B test (RCT)
A randomised controlled trial in which units are randomly assigned to treatment (AI feature) and control (no feature or baseline feature).
A/B Testing
A/B testing is a controlled experiment that compares two versions of an AI model, interface, or process by exposing each to a different group of users and measuring which performs better against predefined metrics.
Absorption Capacity
Absorption capacity is an organization's ability to recognize valuable new knowledge and technology, assimilate it into existing operations and mental models, and apply it productively to create business value.
Accountability
Accountability in AI governance means that when an AI system causes harm, there are clear lines of human responsibility.
Accountability Framework
An accountability framework is a structured system that defines who is responsible for AI decisions, how those decisions are documented, what oversight mechanisms exist, and what consequences apply when things go wrong.
Accountability matrix
A mapping of accountable owners to AI risks, controls, systems, and outcomes — a single place the organization can inspect to see "who is accountable for what" in AI governance.
Accuracy
Accuracy is a model performance metric measuring the proportion of all predictions (both positive and negative) that are correct.
Action Research
Action research is a cyclical research methodology where practitioners simultaneously study and improve their own professional practice by iterating through cycles of planning, acting, observing, and reflecting.
Action Space
The action space is the complete set of all actions an AI agent can potentially take, including tool invocations (API calls, database queries, file operations), communication actions (messages to humans or other agents), reasoning actions (internal processing steps), and environmental interactions (network requests, system operations).
Active-User Rate
The percentage of provisioned users who meaningfully engage with an AI system within a defined measurement window (typically weekly or monthly), where "meaningful engagement" is defined per use case with an explicit action threshold.
Actor topology
A multi-agent topology in which agents communicate via message passing — with each agent as an isolated actor owning its own state and mailbox.
Adaptive Learning System
An adaptive learning system modifies its own behavior based on new data without explicit reprogramming, raising governance challenges as decision processes change over time.
Adaptive Management
Adaptive management is a structured, iterative approach to decision-making that explicitly acknowledges uncertainty and adjusts plans based on new information, monitoring results, and changing conditions rather than rigidly following an original blueprint.
ADKAR
Prosci's five-stage individual change model — Awareness, Desire, Knowledge, Ability, Reinforcement — that describes the sequence through which a single person adopts a change.
Adoption metric
A leading or lagging indicator of transformation uptake — including training-completion rate, active-usage frequency, productivity delta, and worker-sentiment score.
Adoption Rate
Adoption rate measures the percentage of intended users who are actively and effectively using an AI-enabled tool or process.
Adoption Review Report
A structured assessment of actual user adoption rates, workflow integration effectiveness, and change management outcomes — measuring whether end users are actually using the AI system as designed and whether the workflow redesign is delivering its intended effects.
Adoption Trap
The adoption trap is a COMPEL-identified anti-pattern describing the illusion of progress created by accumulating AI tools and deployments without building underlying organizational capability.
ADR (architecture decision record)
A short, dated document capturing the context, decision, consequences, and status of an architectural decision — e.g., "we chose hybrid retrieval over naive RAG because X".
Adversarial Attack
An adversarial attack is a deliberate attempt to fool or manipulate an AI system by providing specially crafted inputs designed to cause incorrect outputs.
Advisory Board
An advisory board is a group of external experts and thought leaders who provide non-binding strategic guidance to an organization's AI transformation program, offering perspectives that internal teams may lack.
Advisory Engagement
An advisory engagement is a COMPEL consulting arrangement where the practitioner provides ongoing strategic counsel to client leadership, guiding their internally-led transformation efforts without taking direct delivery responsibility.
Agent Autonomy Classification
A formal classification of every AI agent in scope according to its autonomy level — from level 0 (no autonomy, human executes) through level 4 (full autonomy, agent executes without human involvement) — with corresponding governance requirements, approval boundaries, and monitoring obligations assigned at each level.
Agent autonomy spectrum
The continuum along which agentic AI systems operate — from single-turn assistance through tool-augmented reasoning to fully autonomous multi-step execution — typically described with named gradations (Level 0 through Level 5).
Agent Governance
The specialized governance framework for autonomous and semi-autonomous AI agents that extends traditional AI governance with agent-specific controls: autonomy level classification, tool access controls, data access boundaries, approval boundaries, human-in-the-loop thresholds, auditability requirements, fallback and kill switch mechanisms, escalation rules, simulation testing requirements, and agent risk tiering..
Agent Governance Pack
An executive-grade, living artefact aggregating all governance records for a deployed agent — autonomy classification, delegation and authority chain, oversight design, tool bindings, memory scope, kill-switch wiring, risk tier, incident history, and named owners.
Agent Lifecycle Management
Agent lifecycle management is the end-to-end governance process covering the creation, registration, testing, deployment, monitoring, updating, and eventual retirement of AI agents within an enterprise.
Agent loop
The iterative pattern in which a model plans, calls tools, observes the results, and replans — continuing until a goal is met, an iteration budget is exhausted, or a human intervenes.
Agent memory
The stored context, preferences, observations, and state an agent accumulates across turns or sessions.
Agent observability
The logging, tracing, and evaluation infrastructure that makes an agent's plans, tool calls, memory reads/writes, and decisions auditable after the fact.
Agent Orchestration
Agent orchestration is the coordination of multiple AI agents working together on complex tasks, managing how work is routed between agents, how handoffs occur, how conflicts are resolved, and how the collective output remains coherent.
Agent Registry
An agent registry is a centralized catalog that tracks all deployed AI agents across an organization, documenting their capabilities, permissions, owners, operational status, tool access, and governance compliance.
Agent runtime
The execution substrate that hosts the agent loop, its tool calls, state, and recovery logic — e.g., LangGraph, CrewAI, AutoGen, or OpenAI Agents SDK.
Agent taxonomy
A seven-category classification — conversational, task, workflow, RPA-adjacent, research, code, embodied — used to scope an agent's design, risk profile, and governance controls.
Agent trace
A span hierarchy that captures a multi-step agent execution — loop iterations, tool calls, memory reads and writes — with enough fidelity to reconstruct the agent's full decision path.
Agent-to-agent (A2A) protocol
The communication mechanism between AI agents — specifying message format, authentication, authorization scope, rate limiting, and audit logging.
Agentic AI
Agentic AI refers to artificial intelligence systems capable of taking autonomous actions in the world, making decisions, using external tools, and pursuing multi-step goals with minimal or no human intervention at each step.
Agentic AI system
An AI system that executes tools, loops over multi-step plans, maintains state across steps, and pursues goals semi-autonomously.
Agentic AI Transformation Strategy
Agentic AI transformation strategy is a comprehensive approach to deploying autonomous AI agents within enterprise transformation contexts, encompassing agent architecture design, autonomy level governance (from advisory through fully autonomous), human-in-the-loop oversight patterns, multi-agent orchestration frameworks, kill switch and containment protocols, observability and monitoring infrastructure, and compliance evidence collection for regulatory requirements.
Agentic Failure Taxonomy
An agentic failure taxonomy is a structured classification system that categorizes the types of failures that can occur in agentic AI systems, providing a shared vocabulary for identifying, discussing, and governing AI agent risks.
Agentic platform
A shared enterprise platform providing agent runtime, tool registry, memory stores, safety layer, observability, and evaluation — so individual agentic use cases inherit these components rather than build them.
Agentic RAG
A retrieval pattern in which the agent decides — within its loop — when and what to retrieve, rather than a retrieval step running before generation as in naive RAG.
Agentic risk tiering
The mapping of an agent's autonomy level, operating domain, and potential-harm profile to a risk tier that determines control requirements — approval depth, oversight mode, audit cadence, insurance, and kill-switch provisioning.
Agile
Agile is a set of principles for software development emphasizing iterative delivery, team collaboration, responsiveness to change, and working outputs over comprehensive documentation.
AI Adoption
AI adoption is the act of introducing AI technologies into an organization's operations -- purchasing AI software, deploying pre-built models, or building custom solutions for specific use cases.
AI Ambition Statement
A concise, board-level declaration that defines the organization's strategic intent for AI transformation — what it aims to achieve, at what pace, and with what risk appetite.
AI Bill of Materials (AI-BOM)
A structured inventory of every component that comprises an AI system — including foundation models, fine-tuned variants, training datasets, embeddings, vector stores, prompts, agent tools, third-party APIs, libraries, and runtime dependencies — together with their provenance, licenses, versions, and known risks.
AI business case
A six-part document — hypothesis, investment, benefit, risk profile, financial summary, recommendation — that justifies an AI investment with explicit counterfactual and confidence bands.
AI Capability Center
An AI Capability Center is an organizational unit that concentrates AI expertise, tools, and shared resources to serve the broader enterprise, representing an evolution beyond the traditional Center of Excellence (CoE) model.
AI change management
The discipline of preparing, supporting, and reinforcing people through AI transformation — distinct from project management (which manages tasks) and from the broader organizational change-management tradition (which is not AI-specific).
AI Change Plan
An executive-grade, living artifact aggregating stakeholder landscape, AI-specific resistance map, AI literacy segmentation and enablement plan, communication strategy, role-redesign actions, adoption metrics, reinforcement mechanisms, and portfolio-capacity analysis.
AI Controls
AI controls are the specific technical, procedural, and organizational mechanisms that enforce AI governance policies in practice.
AI data lifecycle
The nine-stage progression from sourcing through retirement used as the reference frame for data readiness: acquisition, preparation, labeling, governance, training use, validation, deployment use, monitoring, retention and retirement.
AI Demand Review Board
An AI Demand Review Board is a governance body responsible for evaluating, prioritizing, and approving incoming AI project requests from across the organization, ensuring each initiative aligns with the enterprise AI strategy and that limited technical resources are allocated to the highest-value opportunities.
AI Due Diligence
AI due diligence is the comprehensive investigation and assessment of AI capabilities, risks, liabilities, and technical assets conducted during mergers, acquisitions, partnerships, or major vendor selections.
AI Environmental Sustainability (D19)
The 19th maturity domain in the COMPEL framework, covering the governance of AI's environmental impact including energy consumption tracking, carbon footprint management, water usage monitoring, model efficiency optimization, and ESG reporting for AI operations.
AI Ethics Board
An AI Ethics Board is a cross-functional body with genuine authority to review, approve, pause, or halt AI initiatives based on ethical criteria.
AI experiment
A structured comparison producing evidence for a decision — about a model version, a prompt, a feature set, a retrieval strategy, or a deployment change.
AI FinOps
AI FinOps (Financial Operations for AI) is the practice of managing and optimizing the financial costs of AI infrastructure, including cloud compute spending, model training expenses, inference costs, data storage, and third-party API usage.
AI Governance
AI governance is the system of policies, roles, processes, oversight bodies, and controls that an organization uses to manage AI systems responsibly across their full lifecycle.
AI incident (for LLMs)
A subtype of the NIST AI RMF MANAGE 1.4 incident concept specific to LLM systems: confident-but-wrong answer, safety bypass, prompt injection success, sensitive data leakage, or policy-violating tool-call execution.
AI Incident Classification
AI Incident Classification is a systematic framework for categorizing AI failures, malfunctions, and harmful outputs by their severity, impact scope, root cause type, and urgency of required response.
AI Literacy
AI literacy is the degree to which individuals across an organization understand AI concepts, capabilities, and limitations well enough to make informed decisions within their domain and participate meaningfully in AI-enabled work.
AI literacy (WCT)
The minimum knowledge, skills, and judgment required to operate AI systems responsibly — legally required under EU AI Act Article 4 from 2 February 2025 for providers and deployers of AI systems.
AI literacy segmentation
Stratified assessment of AI literacy across executive, manager, specialist, and general-employee cohorts, with cohort-appropriate content depth.
AI Maturity
AI maturity is the measured level of organizational capability in adopting, governing, and scaling artificial intelligence across all relevant dimensions — people, process, technology, and governance.
AI Office
The unit within the European Commission established under Regulation (EU) 2024/1689 to coordinate enforcement of the general-purpose AI model provisions, support the AI Board and the national competent authorities, and issue Union-level guidance..
AI Operating Model
An AI operating model defines how an organization structures its people, processes, data, and technology to deploy and govern AI at scale.
AI Operating Model Blueprint
A comprehensive design document that defines how the organization will govern, fund, staff, and operate AI capabilities at scale — covering the Center of Excellence structure, decision rights, team topologies, tooling standards, and operating procedures.
AI Operating Model Readiness
AI operating model readiness measures an organization's preparedness to establish and sustain the governance structures, decision rights, roles, committees, and processes required to operate AI systems at scale.
AI Operating System
A structured, repeatable management system that enables an organization to plan, govern, deliver, measure, and continuously improve AI capabilities across people, process, technology, and governance dimensions.
AI Platform Strategy
AI Platform Strategy is the enterprise-level approach to selecting, building, and integrating the technology foundation that supports all AI development, deployment, and operations across the organization.
AI portfolio scorecard
An executive dashboard showing all AI investments with status, realized value, risk flag, and next decision.
AI Product Manager
An AI product manager is a professional responsible for defining AI use cases, managing stakeholder engagement, translating business requirements into technical specifications, and ensuring that AI solutions deliver measurable business value.
AI Readiness Assessment
An AI readiness assessment is a structured diagnostic that evaluates an organization's preparedness to adopt, govern, and scale AI across key dimensions: leadership alignment, data quality, technical infrastructure, workforce skills, governance frameworks, and regulatory posture.
AI reference architecture
A canonical layered model — client, orchestration, model, knowledge, observability planes — that every AI system maps onto.
AI Risk Champions
AI Risk Champions are designated individuals embedded within business units who serve as local advocates for AI risk awareness and act as liaisons between frontline operations and the central AI risk management function.
AI Risk Governance Board
An AI Risk Governance Board is a senior leadership body responsible for overseeing AI-related risks across the entire enterprise, establishing the organization's AI risk appetite, making decisions about acceptable risk levels for AI deployments, and ensuring that risk management practices are adequate and consistently applied.
AI Risk Register
An AI Risk Register is a documented, maintained inventory of all identified AI-related risks within an organization, capturing each risk's description, likelihood, potential impact, current mitigation measures, assigned owner, and review status.
AI Safety
AI Safety is the field of research and practice dedicated to ensuring that AI systems operate without causing unintended harm to individuals, organizations, or society.
AI Security Architecture
AI Security Architecture is the comprehensive design of security controls and defense mechanisms specifically tailored to the unique threat landscape of AI systems, covering model protection against extraction and poisoning, training data security, adversarial input defense, prompt injection prevention, API access control, supply chain security for AI components, and audit trail integrity.
AI Service Level Management
AI Service Level Management is the practice of defining, measuring, monitoring, and maintaining agreed-upon performance standards for AI services, extending traditional ITIL service management concepts to cover AI-specific metrics such as model accuracy, prediction latency, fairness consistency, drift thresholds, and retraining frequency.
AI Steering Committee
The AI Steering Committee is the senior governance body that provides strategic direction, resolves cross-functional conflicts, approves budgets, and maintains executive accountability for AI transformation outcomes.
AI Supply Chain Governance (D20)
The 20th maturity domain in the COMPEL framework, covering the governance of AI systems procured from or dependent on external parties.
AI system (EU AI Act)
Under Regulation (EU) 2024/1689, a machine-based system designed to operate with varying levels of autonomy and that may exhibit adaptiveness after deployment, which infers from the input it receives how to generate outputs — such as predictions, content, recommendations, or decisions — that can influence physical or virtual environments..
AI System Classification Register
A formal register that classifies every AI system in scope according to risk tier, autonomy level, data sensitivity, regulatory applicability, and criticality — producing a system-level risk profile that determines which governance controls, review processes, and compliance requirements apply.
AI System Impact Assessment
An AI System Impact Assessment is a structured, documented evaluation of how a proposed or existing AI system affects individuals, groups, organizations, and society across dimensions including fundamental rights, safety, privacy, fairness, environmental impact, and labor market effects.
AI System Registry
An AI System Registry is an organizational catalog that documents all AI systems in use, under development, or being evaluated, recording each system's purpose, data inputs, risk classification, ownership, compliance status, deployment environment, and review history.
AI talent pipeline
A six-stage system — source, hire, onboard, develop, retain, transition — for AI-fluent talent across the enterprise.
AI trace
A span hierarchy — client → orchestration → retrieval → model → tool — capturing a single AI request end-to-end, including prompts, responses, tool calls, and token usage.
AI Transformation
AI transformation is the enterprise-wide process of systematically adopting, governing, and scaling artificial intelligence to change how an organization operates, competes, and creates value.
AI transformation readiness
An organisation's demonstrated capacity to sustain, govern, and scale AI — distinct from whether it currently uses AI.
AI Use Case
A specific, bounded application of AI technology to a defined business process or decision point, with a documented value thesis, identified stakeholders, risk classification, and success criteria.
AI value chain
The six-stage path — data, model, inference, decision, action, outcome — that an AI feature traverses to produce business value.
AI workforce transformation
The composite multi-year program that rewires how work gets done, how careers are built, and how decisions are made in an organisation adopting AI at scale.
AI-fluent manager
A manager who can set expectations, coach employees, and evaluate performance in AI-integrated work — including attributing output credit between human and AI, diagnosing AI-related underperformance, and maintaining psychological safety through change.
AI-specific resistance
Resistance patterns distinct to AI transformation — fear of role replacement, distrust of algorithmic opacity, prior adverse experience with automation, and legitimate ethical concern about bias or surveillance.
AIOps
AIOps (Artificial Intelligence for IT Operations) is the application of AI and machine learning to IT operations tasks such as monitoring, anomaly detection, alerting, root cause analysis, and automated incident resolution.
AITB-LAG
LLM Application Governance.
AITB-RCM
AI Regulatory and Compliance Mapping.
AITB-TRA
AI Transformation Readiness Assessor.
AITE-ATS
Agentic Transformation Specialist.
AITE-SAT
AI Solution Architecture for Transformation Leaders.
AITE-VDT
Value-Driven Transformation.
AITE-WCT
Workforce Change for Transformation.
AITF (COMPEL Certified Practitioner)
AITF is the Level 1 COMPEL certification demonstrating foundational mastery of the COMPEL methodology, including the six lifecycle stages (Calibrate, Organize, Model, Produce, Evaluate, Learn), the Four Pillars (People, Process, Technology, Governance), the 20-domain maturity model, and the core principles of AI transformation.
AITGP (COMPEL Certified Consultant)
AITGP is the Level 3 COMPEL certification for professionals who have mastered the skills needed to architect enterprise-level AI transformation strategies, design operating models, build governance frameworks, and mentor specialist-level practitioners.
AITL Lead
AITL Lead is the Level 4 apex certification in the COMPEL framework for professionals who govern portfolios of AI transformation programs across multiple organizations, harmonize governance across organizational boundaries, design enterprise AI operating models, and actively contribute to industry standards development.
AITM-CMD
Change Management for AI Deployment.
AITM-DR
Data Readiness for AI Transformation.
AITM-ECI
AI Experimentation and Continuous Improvement.
AITM-OMR
AI Operating Model Readiness.
AITM-PEW
Prompt Engineering for Transformation Workflows.
AITP (COMPEL Certified Specialist)
AITP is the Level 2 COMPEL certification for practitioners who can independently design, lead, and deliver COMPEL transformation engagements with real client organizations, including advanced maturity assessment, transformation roadmap architecture, multi-workstream execution management, and measurement and evaluation.
Algorithm
An algorithm is a set of step-by-step instructions or mathematical rules that a computer follows to solve a problem or complete a task.
Algorithmic Accountability
Algorithmic accountability is the principle that organizations deploying algorithms must be answerable for the outcomes those algorithms produce, including unintended consequences, discriminatory effects, and errors that affect individuals.
Algorithmic Audit
An algorithmic audit is an independent, systematic examination of an AI system's decision-making processes, data inputs, outputs, and real-world impacts to assess whether the system operates in compliance with legal requirements, ethical standards, and organizational policies.
Algorithmic Bias
Algorithmic bias is systematic and unfair discrimination in AI system outputs, often arising from biased training data, flawed model design, unrepresentative data samples, or proxy variables that encode protected characteristics.
Algorithmic Impact Assessment
An Algorithmic Impact Assessment (AIA) is a formal, structured evaluation conducted before deploying an AI system to identify and quantify potential negative impacts on individuals and communities, particularly regarding fairness, privacy, civil rights, employment, and access to services.
Andragogy
Andragogy is the theory and practice of adult education, distinct from pedagogy (child education), recognizing that adults learn differently and have specific needs including understanding why they are learning something, drawing on their existing experience, exercising self-direction, focusing on immediately applicable knowledge, and being motivated by internal rather than external factors.
Annex I
The Regulation (EU) 2024/1689 annex listing the existing Union harmonisation legislation — machinery, medical devices, in-vitro diagnostics, toys, radio equipment, lifts, pressure equipment, recreational craft, civil aviation, motor vehicles, rail systems, agricultural vehicles — under which AI-containing safety components fall into the Article 6(1) high-risk pathway..
Annex III
The Regulation (EU) 2024/1689 annex listing eight domains of standalone high-risk AI use cases: biometrics; critical infrastructure; education and vocational training; employment and worker management; essential services; law enforcement; migration, asylum and border control; and administration of justice and democratic processes..
Anomaly Detection
Anomaly detection is a technique that identifies data points, events, or patterns that deviate significantly from expected behavior.
Anonymization
Anonymization is the process of irreversibly removing or altering personally identifiable information from datasets so that individuals cannot be re-identified, even by combining the anonymized data with other available information.
Anti-Pattern
In AI transformation, an anti-pattern is a commonly occurring organizational behavior that appears rational in the moment but systematically undermines transformation outcomes.
Anti-pattern (readiness)
A recurring organisational failure signature that predicts readiness weakness — examples include pilot paralysis, shadow AI, governance theatre, fragmented centre-of-excellence, and tooling-first procurement.
API (Application Programming Interface)
An API is a set of rules and protocols that allows different software systems to communicate with each other in a standardized way.
Architecture Review Board
An Architecture Review Board (ARB) is a governance body that evaluates proposed technology designs, platform changes, and system integrations against enterprise architecture standards, ensuring consistency, scalability, security, and strategic alignment before implementation begins.
Architecture runway
The reusable platform components — inference infrastructure, retrieval stack, observability, policy engine, evaluation harness — that future AI use cases inherit rather than re-build.
Article 6(3) derogation
Under Regulation (EU) 2024/1689, a conditional classification that allows an AI system falling within Annex III to be treated as non-high-risk when it does not pose a significant risk of harm to health, safety, or fundamental rights.
Artifact
In the COMPEL framework, an artifact is a formal document, record, or deliverable produced during the lifecycle that provides evidence of governance activities, decisions, and outcomes.
Artificial General Intelligence (AGI)
Artificial General Intelligence is a theoretical form of AI with human-level cognitive ability across all intellectual domains -- the ability to reason, learn, and adapt to any task a human can perform.
Artificial Intelligence (AI)
Artificial Intelligence is a broad field of computer science focused on building systems that can perform tasks typically requiring human intelligence, such as understanding language, recognizing patterns, making decisions, and generating content.
Artificial Narrow Intelligence (ANI)
Artificial Narrow Intelligence describes AI that performs a specific task within a defined domain, such as detecting fraud, translating languages, or playing chess.
Assessment-Only Engagement
An assessment-only engagement is a COMPEL consulting arrangement focused exclusively on diagnosing an organization's AI maturity using the 20-domain maturity model, typically lasting four to eight weeks with a smaller team and lower commercial risk than full transformation engagements.
Assurance
Assurance is the process of providing justified confidence to stakeholders that AI systems, processes, and governance mechanisms are operating effectively, safely, and in compliance with stated requirements and standards.
Asynchronous kill-switch
A kill-switch variant that terminates the agent immediately — regardless of in-flight operations — via an out-of-band interrupt signal.
Attack Surface
The attack surface of an AI system encompasses all the points where an unauthorized actor could attempt to access, manipulate, or extract data from the system, including model API endpoints, training data pipelines, model weight storage, user interfaces, supply chain components, and the inference process itself.
Attestation
Attestation is a formal declaration by an authorized person or body that an AI system, process, or governance practice meets specified requirements or standards at a particular point in time.
Attribution in AI-integrated performance
Rules for distinguishing human contribution from AI-assisted output in performance measurement — "did the analyst find that insight, or did the tool?" Has direct consequences for promotion, compensation, and public narrative.
Attribution model
A rule set for assigning credit to an AI feature when outcomes involve many touchpoints — e.g., first-touch, last-touch, linear, time-decay, or data-driven multi-touch.
Audit Preparedness
Audit preparedness is the continuous operational discipline of ensuring that governance activities produce the documentation, evidence trails, and records that auditors and regulators can verify on demand.
Audit Trail
An audit trail is a chronological record of all activities, decisions, and changes related to an AI system, maintained to support accountability, compliance verification, and regulatory examination.
Augment / assist / automate / arbitrate
Named collaboration patterns where AI respectively enhances human work (augment), supports human decisions (assist), executes bounded work under oversight (automate), or adjudicates between conflicting human inputs (arbitrate).
Authority chain
The traceable lineage from an organization's decision-rights authority — through any delegating humans — to the agent executing the action.
Auto-scaling
Auto-scaling is the automatic adjustment of computing resources, such as servers, containers, or GPU instances, based on real-time demand patterns.
Auto-Unlock
The automatic granting of a COMPEL professional certification when a recognized external credential is submitted and verified, without requiring the holder to complete the professional certification assessment.
Automation vs. augmentation
A strategic choice for each workflow between removing the human ("automation") and changing what the human does ("augmentation").
Autonomy Calibration
Autonomy calibration determines appropriate AI autonomy levels for specific tasks, balancing efficiency against risks and regulatory requirements.
Autonomy Spectrum
The autonomy spectrum is a classification framework that describes how independently an AI agent can operate, ranging from Level 0 (no autonomy -- fully deterministic instructions) through Level 5 (full autonomy -- self-directed goal setting).
B
Badge Tier
The visual and structural tier assigned to a credential in the lattice, indicating its relative level.
Balanced Scorecard
A Balanced Scorecard is a strategic performance measurement framework that tracks metrics across four complementary perspectives: financial results, customer satisfaction, internal process efficiency, and organizational learning and growth.
Baseline
A baseline is a documented measurement of current performance, capability, or conditions taken before an AI initiative begins, providing the reference point against which progress, improvement, and ROI are measured.
Batch Inference
Batch inference is the practice of running an AI model's predictions on a large collection of data items simultaneously, rather than processing them one at a time in real time.
Batch Processing
Batch processing involves running AI model predictions on large volumes of data at scheduled intervals rather than in real time.
Belonging and equity
A measure of who has voice in AI decisions, who gets which AI-augmented roles, and who is protected from displacement in AI transformations.
Benchmark
A benchmark is a standardized test, dataset, or reference point used to evaluate and compare AI model performance against a common standard.
Benchmark contamination
The presence of benchmark test data in foundation-model training corpora — whether through web crawling or deliberate inclusion — inflating reported benchmark scores and breaking the comparability of benchmark results across models.
Benchmark Update Report
A report that updates the organization's AI performance and maturity benchmarks based on actual results from the current evaluation cycle — revising baselines, adjusting target ranges, and incorporating new external benchmarks to ensure that future KPI targets remain meaningful and appropriately ambitious.
Benefits Register
A benefits register catalogs expected and realized benefits specifying description, magnitude, measurement method, owner, and evidence.
Benefits Tracking
Benefits tracking is the systematic, ongoing process of measuring and documenting the actual value delivered by an AI transformation program against the projected benefits that justified the original investment.
Bias Auditing
Bias auditing is the systematic review of AI training data and model outputs to identify and measure unfair biases.
Bias Delta
The measured difference in a model's performance or outcome distribution across protected groups, expressed against a defined fairness threshold.
Bias Detection
Bias detection is the process of systematically identifying unfair patterns in AI systems, examining training data for historical prejudices, model outputs for discriminatory patterns, and real-world impacts for disproportionate effects on particular demographic groups.
Bias-relevant variable
A feature whose inclusion, exclusion, or proxy-behavior affects fairness across protected groups — a direct sensitive attribute (race, gender) or an indirect proxy (postal code, device type).
Binding Corporate Rules
Binding Corporate Rules (BCRs) are internal data protection policies adopted by multinational organizations and approved by data protection authorities that allow the lawful transfer of personal data between entities within the corporate group across different countries.
Blameless Post-Mortem
A blameless post-mortem is an incident review methodology that deliberately focuses on understanding systemic causes, process failures, and improvement opportunities rather than assigning personal blame to individuals involved in an AI system failure or incident.
Bloom's Taxonomy
Bloom's Taxonomy is a hierarchical framework for classifying educational learning objectives into six levels of increasing cognitive complexity: Remember, Understand, Apply, Analyze, Evaluate, and Create.
Board-grade reporting
Format and cadence rules for reporting AI value and risk to audit committees, investors, and regulators — disciplined to survive external-audit scrutiny.
Board-Level Governance
Board-level governance refers to the oversight, strategic direction, and fiduciary responsibility that an organization's board of directors exercises over AI transformation, including setting risk appetite for AI initiatives, approving AI strategy, allocating transformation investment, and holding the executive team accountable for responsible AI practices.
Body of Knowledge
A Body of Knowledge (BoK) is the complete, structured set of concepts, terms, theories, practices, tools, and standards that define a professional discipline and are required for competent practice within it.
Bridge Credential
A bridge credential is a credential type specifically designed to connect external technical training (such as partner bootcamps in data science, LLM engineering, or agentic AI) with COMPEL transformation methodology.
Bridges Transition Model
William Bridges' three-phase psychological transition model — Ending (letting go of the old), Neutral Zone (disorientation and exploration), New Beginning (adoption of the new).
Brussels Effect
The Brussels Effect describes the tendency of European Union regulation to become the de facto global standard because multinational organizations find it more efficient to adopt a single, stringent standard globally than to maintain different compliance practices for different jurisdictions.
Buffer Management
Buffer management is the deliberate practice of building time and resource margins into project schedules and dependency chains to absorb inevitable delays, unexpected complications, and minor failures without triggering cascading schedule disruptions across connected workstreams.
Build-buy-partner-borrow framework
Four sourcing modes for AI capability: build internally, buy from the market, partner with an ecosystem provider, borrow via contingent labor.
Business Case
A business case is a structured document or analysis that provides the financial and strategic justification for an AI investment by quantifying expected costs, benefits, risks, timelines, and alternative options.
Business Continuity
Business continuity planning ensures that an organization's critical functions can continue during and after disruptions, including AI system failures, data breaches, infrastructure outages, and vendor collapses.
C
C-Suite Advisory
C-suite advisory is the practice of providing strategic counsel to an organization's most senior executives, including the CEO, CTO, CFO, CISO, CDO, and other C-level leaders, on AI transformation strategy, organizational implications, risk landscape, and investment priorities.
Calibrate (COMPEL Stage)
Calibrate is the first of the six COMPEL stages, focused on producing an honest, evidence-based assessment of where the organization stands today in its AI transformation journey.
Calibrate Stage
The Calibrate stage is the first stage of the COMPEL lifecycle where an organization's current AI maturity is systematically assessed across the 20-domain maturity model, establishing a data-driven baseline that informs all subsequent transformation planning.
Canary Deployment
Canary deployment is a risk-mitigation release strategy where a new version of an AI model or system is first deployed to a small, carefully selected subset of production traffic, and its performance is monitored closely before gradually expanding the rollout to the full user base.
Capability Baseline
A grounded inventory of the AI, data, platform, and talent capabilities currently in place across the organization, captured at a point in time.
Capability Compounding
Capability compounding is the principle that AI capabilities build upon each other, with each making subsequent ones easier and more valuable.
Capability map
A hierarchical decomposition of the organization's capabilities — business, shared, and enabling — annotated with AI-impact flags indicating where AI reshapes, augments, or disrupts the capability.
Capability Maturity Model Integration (CMMI)
CMMI is a process improvement framework originally developed at Carnegie Mellon University that defines maturity levels for organizational processes.
Capital Allocation
Capital allocation is the strategic process of distributing financial resources across a portfolio of AI transformation initiatives based on strategic priorities, expected returns, risk profiles, and organizational readiness.
Capstone Portfolio
The capstone portfolio is the comprehensive collection of artifacts, analyses, strategy documents, governance frameworks, and reflective narratives that Level 4 AITL Lead candidates assemble to demonstrate mastery across portfolio leadership, cross-organizational governance, operating model design, framework interoperability, and industry standards contribution.
Capstone Project
A capstone project is a comprehensive, integrative assessment in the COMPEL certification program where candidates must demonstrate mastery by applying the full methodology to a real or simulated enterprise scenario.
Carbon Attribution
Grams of CO2-equivalent emissions attributed to AI model inference, typically reported per 1,000 calls and rolled up per model, workload, and tenant.
Career lattice
A career architecture that allows lateral movement alongside vertical progression — so employees can grow by expanding scope, changing domain, or deepening expertise rather than only by promotion.
Cascading Failure
A cascading failure is a chain reaction where one component's malfunction triggers failures in dependent components, which in turn cause further failures, potentially resulting in widespread or total system collapse.
Catastrophic Forgetting
Catastrophic forgetting occurs when AI models trained on new data lose previously acquired knowledge.
CCPA
The California Consumer Privacy Act (CCPA) is a US state data privacy law that grants California residents specific rights over their personal data, including the right to know what personal information is being collected, the right to request deletion, the right to opt out of data sales, and the right to non-discrimination for exercising these rights.
CE Credit
Continuing Education credits earned by completing credentials, attending events, or maintaining professional activity within the COMPEL ecosystem.
CE Credit (Continuing Education)
Continuing Education credits are the unit of measurement for ongoing professional development required to maintain COMPEL certifications.
CE Credit Grant
Continuing Education credits awarded to holders of recognized external credentials that count toward COMPEL certification renewal or progression requirements.
Center of Excellence (CoE)
The AI Center of Excellence is the operational nucleus of AI transformation -- the organizational structure that transforms individual AI capability into enterprise AI capacity.
Centralized archetype
Operating-model archetype in which all AI capability sits in a single central team serving the organization — high consistency and quality, low business-proximity.
Certification Body
A certification body is an organization authorized to assess and formally certify that individuals, systems, products, or organizations meet the requirements defined by specific standards or qualification frameworks.
Chain-of-thought (CoT)
A prompt pattern that elicits intermediate reasoning steps before the final answer — either zero-shot ("Let's think step by step") or few-shot with example reasoning chains.
Change Architecture
Change architecture is the deliberate, comprehensive design of how organizational change will be structured, sequenced, resourced, and governed across an enterprise-scale AI transformation.
Change capacity
The organisation's currently available bandwidth to absorb transformational change, measured through active-initiative count, leader-attention budget, and employee-fatigue signals.
Change Capacity Management
Change capacity management is the assessment, monitoring, and deliberate management of how much organizational change the workforce and leadership can absorb at any given time without experiencing change fatigue, disengagement, or active resistance.
Change Management
Change management is the structured approach to transitioning individuals, teams, and organizations from a current state to a desired future state.
Change Network
A change network is a distributed group of change advocates and champions embedded across the organization's departments and levels who support AI transformation by communicating key messages, coaching colleagues through transitions, surfacing feedback and concerns from the front lines, and modeling new behaviors in their daily work.
Change portfolio
The set of concurrent change initiatives — AI and non-AI — competing for finite organizational change capacity, leader attention, and employee bandwidth.
Change Saturation
Change saturation is the practical limit on how much simultaneous organizational change a team, department, or enterprise can absorb effectively.
Change saturation index
A measurable index of current change absorption in the organisation — used to trigger pacing adjustments when the population is approaching its capacity ceiling.
Chaos Engineering
Chaos engineering is the discipline of deliberately introducing controlled failures, disruptions, and adverse conditions into a system's production or staging environment to test its resilience and discover weaknesses before they cause real incidents.
Chargeback Model
A chargeback model is a financial governance mechanism where business units are billed for their actual consumption of shared AI services, infrastructure, compute resources, and platform capabilities, creating cost transparency and incentivizing efficient resource use.
Chief AI Officer
The Chief AI Officer (CAIO) is a C-suite executive responsible for an organization's AI strategy, governance, and transformation.
Chief Data Officer (CDO)
The Chief Data Officer is a C-suite executive responsible for enterprise data strategy, data governance, data quality, and data infrastructure.
Chunking
The process of dividing documents into units — typically fixed-token windows or paragraph-level segments — suitable for embedding and retrieval.
Churn Prediction
Churn prediction is an AI application that predicts which customers are likely to stop using a product or service within a defined timeframe, enabling proactive retention efforts before the customer leaves.
CI/CD Pipeline
A CI/CD (Continuous Integration/Continuous Deployment) pipeline is an automated workflow that builds, tests, and deploys software changes through a series of stages, catching errors early and enabling rapid, reliable releases.
Circuit Breaker
A circuit breaker is a resilience design pattern that automatically stops an AI system from sending requests to a failing downstream service or component when it detects a pattern of errors, preventing cascading failures and giving the failing component time to recover.
Circuit breaker (agent tool)
A circuit-breaker pattern applied specifically to agent tool calls: when a tool's failure rate exceeds a threshold, the breaker opens and blocks further calls to that tool.
Classification
Classification is a supervised learning task that assigns inputs to discrete categories.
Client Discovery
Client discovery is the initial investigative phase of a COMPEL engagement where the practitioner gathers comprehensive information about the prospective client's business context, AI aspirations, organizational constraints, stakeholder landscape, budget parameters, previous transformation attempts, and current technology environment.
Cloud Computing
Cloud computing is the delivery of computing services -- servers, storage, processing power, databases, networking, and software -- over the internet on a pay-as-you-go basis rather than owning and maintaining physical infrastructure.
Cloud-Native Architecture
Cloud-native architecture refers to systems designed specifically to leverage cloud computing capabilities such as elastic scaling, distributed processing, managed services, containerized deployment, and microservice decomposition.
Clustering
Clustering is an unsupervised learning technique that groups similar data points together based on shared characteristics, without requiring pre-labeled categories.
Co-Development Agreement
A co-development agreement is a contractual arrangement where two or more parties jointly develop AI capabilities, clearly specifying how intellectual property ownership, development costs, risks, access rights, and commercial benefits are shared between the parties.
Coaching cadence
The rhythm of manager-employee conversations that sustains behaviour change after training events — typically weekly-to-monthly for the first two quarters of an AI rollout, then quarterly.
Coalition Analysis
Coalition analysis is a stakeholder management technique that maps the formal and informal alliances, power relationships, shared interests, and competing agendas among individuals and groups within an organization to understand who can be brought together to support transformation and who may collectively resist it.
COBIT
COBIT (Control Objectives for Information and Related Technologies) is an IT governance and management framework developed by ISACA that provides a comprehensive set of controls, processes, and metrics for governing enterprise information and technology.
CoE Charter
The formal governance document that establishes the AI Center of Excellence — defining its mandate, scope, membership, decision rights, funding model, service catalogue, and success metrics.
Cognitive Load Management
Cognitive load management is the deliberate practice of controlling the mental effort required for learning, comprehension, and task performance, ensuring that training materials, communications, and governance processes do not overwhelm participants with excessive complexity or information volume.
Communication strategy
A multi-channel, multi-audience plan for transformation messaging with two-way feedback loops — distinct from one-way announcements because it explicitly accepts and responds to worker concerns.
Community of Practice
A community of practice (CoP) is a group of people who share a professional interest, domain, or challenge and interact regularly to deepen their knowledge, share experiences, solve problems, and develop their expertise collectively.
COMPEL Cycle
A COMPEL cycle is a single iteration through all six stages (Calibrate, Organize, Model, Produce, Evaluate, Learn), typically lasting 12 weeks with a contextual range of 8 to 16 weeks.
COMPEL Engagement Lifecycle
The COMPEL Engagement Lifecycle is the five-phase structure for managing a COMPEL transformation consulting project, distinct from the six-stage COMPEL methodology itself.
COMPEL Four Pillars
The COMPEL Four Pillars are the fundamental organizing dimensions of AI transformation in the COMPEL framework: People (culture, skills, change management, leadership), Process (workflows, operations, governance procedures, service management), Technology (infrastructure, platforms, data architecture, AI models, security), and Governance (policies, compliance, ethics, risk management, oversight).
COMPEL Framework
COMPEL is a structured, iterative six-stage methodology for enterprise AI transformation, standing for Calibrate, Organize, Model, Produce, Evaluate, Learn.
COMPEL Lifecycle
The COMPEL Lifecycle is the six-stage transformation methodology that structures all AI transformation work: Calibrate (assess current maturity), Organize (align stakeholders and form teams), Model (design the target state and roadmap), Produce (execute the transformation plan), Evaluate (measure outcomes and assess progress), and Learn (capture lessons and feed insights into the next cycle).
Competency Badge
A low-to-moderate rigor credential requiring 6-8 hours that proves ability to perform a specific transformation task (e.g., readiness assessment, regulatory mapping, LLM governance).
Competency-Based Assessment
Competency-based assessment is an evaluation approach that measures whether a person can demonstrate specific professional skills and knowledge in realistic practice contexts, rather than testing theoretical knowledge through traditional examinations alone.
Competitive Moat
A competitive moat is a durable competitive advantage that is exceptionally difficult for rivals to replicate.
Compliance Harmonization
The practice of implementing a single governance framework that satisfies multiple regulatory requirements simultaneously.
Compliance Posture
Compliance posture refers to an organization's overall state of readiness and demonstrated adherence to the laws, regulations, standards, and internal policies applicable to its AI systems and data practices.
Compliance-grade literacy evidence
Records of literacy completion, scores, and re-certification that satisfy regulator, auditor, and works-council scrutiny.
Compute Budget
A compute budget is the allocated financial and resource limit for AI workloads including model training, experimentation, inference processing, and data pipeline operations.
Compute budget (VDT)
A pre-agreed ceiling on training or inference compute per AI feature per period — expressed in tokens, FLOPs, or dollars.
Computer use / browser-use agent
An AI agent that operates through a browser or computer user interface against third-party systems rather than via APIs — taking screenshots, clicking, typing, and scrolling.
Computer Vision
Computer vision is a field of artificial intelligence that enables computers to interpret and understand visual information from images, videos, and real-time camera feeds, replicating aspects of human visual perception through mathematical models.
Concept Drift
Concept drift occurs when the underlying relationship between input data and the outcome being predicted changes over time.
Confabulation
NIST's preferred term for hallucination: an LLM generating fluent output that is unsupported by ground truth.
Conformity Assessment
A formal evaluation that demonstrates an AI system meets the requirements of an applicable regulation, standard, or governance framework — for example the EU AI Act conformity assessment required for high-risk AI systems before market placement, or third-party certification against ISO/IEC 42001.
Consent Management
Consent management encompasses the technical systems, processes, and policies for collecting, recording, tracking, and honoring individuals' preferences regarding the use of their personal data, including data used to train, validate, or operate AI systems.
Consortium Governance
Consortium governance is the design and operation of governance structures for multi-organization AI collaborations where no single entity has unilateral authority and decisions must be reached through negotiation, voting, or consensus mechanisms among sovereign participants.
Constrained decoding
A decoding-time technique forcing model outputs to conform to a grammar, JSON Schema, or regular language by masking invalid next-token probabilities.
Constructivism
Constructivism is a learning theory positing that people actively build their understanding by connecting new information to their existing knowledge, experiences, and mental models, rather than passively absorbing information transmitted by an instructor.
Containerization
Containerization is a technology that packages software applications and all their dependencies (libraries, configurations, runtime environments) into isolated, portable units called containers that run consistently across different computing environments.
Content safety classifier
A model or rule system that detects policy-violating output categories — violence, self-harm, CSAM, targeted harassment, dangerous instructions, and similar.
Context Window
A context window is the maximum amount of text (measured in tokens) that a large language model can process at one time.
Continuous batching
An inference-server technique — popularised by vLLM and Text Generation Inference — that dynamically groups concurrent requests at the token-generation level to raise GPU utilisation.
Continuous delivery (ML)
Automated, governed promotion of models through lifecycle stages — development, staging, production — with gated checkpoints (evaluation thresholds, bias checks, cost thresholds, human approval where required).
Continuous Improvement
Continuous improvement is the ongoing effort to enhance processes, capabilities, and outcomes through iterative learning and refinement.
Continuous Improvement Backlog
A governed, prioritized backlog of improvement actions identified across all Evaluate and Learn stage reviews — covering control enhancements, workflow refinements, model updates, policy changes, and capability investments — that ensures organizational learning is translated into concrete, scheduled improvement work rather than good intentions.
Continuous integration (ML)
Automated test and build of model code, data contracts, and training scripts on every change — extended from software CI with data-schema validation, model-schema validation, and lightweight training smoke tests.
Control Activation Register
A record that tracks the implementation and activation status of every governance control specified in the Control Requirements Matrix — confirming that each control has been built, tested, and is operational in the production environment.
Control Coverage Percentage
The ratio of implemented and evidenced governance controls to required controls under a given standard or regulatory framework — for example ISO 42001, NIST AI RMF, or the EU AI Act.
Control Framework
A control framework is a structured, comprehensive set of policies, procedures, technical safeguards, and organizational measures designed to manage risks and ensure compliance within a specific domain.
Control Performance Report
A structured report that assesses the effectiveness of every active governance control — presenting evidence of control operation, pass/fail status against defined thresholds, exceptions, and remediation actions — to demonstrate that governance is working as designed rather than merely documented.
Control performance report (CPR)
A periodic artefact documenting control effectiveness — how well governance controls (bias monitoring, drift detection, HITL gates) are performing on defined cadence.
Control Requirements Matrix
A comprehensive mapping of every governance control required for each AI system — specifying the control type (preventive, detective, corrective), the risk or policy it addresses, the evidence required to prove effectiveness, the owner, and the testing frequency.
Convolutional Neural Network (CNN)
A Convolutional Neural Network is a type of deep learning architecture designed specifically for processing visual data like images and videos.
Copyright
Copyright is the legal protection granted to original creative works including text, images, music, software code, and other forms of expression.
Cost-to-serve
The unit economics of providing AI capability to each business unit or use case — fully-loaded cost including compute, data, platform, and talent amortization.
Counterfactual outcome
The outcome that would have occurred without the AI intervention — the benchmark against which incremental AI value is measured.
Credential Designation
A credential designation is an honorary professional title granted when a practitioner achieves a specific combination of credentials spanning multiple credential types within the COMPEL credential lattice.
Credential Lattice
The multi-dimensional credential architecture that extends beyond the traditional linear certification ladder to include professional certifications, expert certifications, competency badges, joint credentials, and designations.
Credential Progress Dashboard
A personalized dashboard within the COMPEL platform that visualizes a practitioner's progress across the credential lattice.
Credential Stacking
Credential stacking is the practice of combining multiple smaller credentials (professional certifications, competency badges) to build toward higher-level credentials (expert certifications, joint credentials) through formally defined accumulation rules.
Crisis Management
Crisis management is the organized process of preparing for, responding to, recovering from, and learning from unexpected events that threaten an organization's AI transformation program, operations, reputation, or stakeholder relationships.
Cross-Border Data Governance
Cross-border data governance encompasses the policies, legal mechanisms, technical architectures, and organizational processes for managing data that flows between different countries, each with potentially different data protection laws, sovereignty requirements, and regulatory expectations.
Cross-Domain Diagnostic
A cross-domain diagnostic is an advanced COMPEL assessment technique that examines how capabilities in different maturity domains interact, influence each other, and create systemic patterns that are not visible when each domain is assessed in isolation.
Cross-Functional Collaboration
Cross-functional collaboration is the practice of working across traditional organizational boundaries -- IT, business units, legal, finance, HR, compliance -- to achieve AI transformation objectives.
Cross-Functional Team
A cross-functional team brings together members from different organizational departments, disciplines, or specializations to work collaboratively toward a common objective, combining perspectives that no single function could provide alone.
Cross-Organizational Governance
Cross-organizational governance refers to the structures, policies, and decision-making processes that operate across organizational boundaries to enable coherent AI policy, consistent risk management, and aligned strategic direction among entities that do not share a single command hierarchy.
Cross-Stack Designation
An honorary title granted when a practitioner achieves a specific combination of credentials spanning multiple credential types.
Cross-Validation
Cross-validation is a statistical technique for evaluating AI model performance by partitioning data into multiple subsets, systematically training the model on some subsets while testing on others, and averaging the results across all partitions.
Cultural Transformation
Cultural transformation is the deliberate, sustained reshaping of an organization's values, beliefs, behaviors, and working practices to create an environment that supports AI adoption, data-driven decision-making, experimentation, and responsible innovation.
Customer Relationship Management (CRM)
CRM software manages an organization's interactions with current and potential customers, tracking sales activities, customer communications, purchase history, and service interactions.
D
Data Architecture
Data architecture is the design of how data is collected, ingested, stored, organized, integrated, transformed, governed, and made available across an enterprise to support AI capabilities, analytics, and business operations.
Data Catalog
A data catalog is a searchable, organized inventory of all data assets within an organization, providing metadata about each dataset's location, format, schema, ownership, quality metrics, access permissions, lineage, and permitted uses.
Data Classification
Data classification is the process of categorizing data based on its sensitivity level, regulatory requirements, and business criticality into tiers such as public, internal, confidential, and restricted.
Data contract
A versioned, testable specification of a data product's schema, semantics, quality expectations, SLA, and change-management policy — published by the producer, consumable by downstream AI workloads.
Data Drift
Data drift occurs when the statistical properties of the input data a deployed model receives change compared to the data it was trained on.
Data Engineer
A data engineer is a professional responsible for building and maintaining the data infrastructure and pipelines that collect, store, transform, and deliver data to AI models and analytics consumers.
Data Fabric
A data fabric is an architectural approach that provides a unified, intelligent data management layer across diverse and distributed data sources, environments, and formats, using automation, metadata, and AI-driven data management to make data accessible and usable regardless of where it physically resides.
Data foundation readiness
The subset of AI transformation readiness covering data discoverability, quality, accessibility, and governance for AI workloads.
Data Governance
Data governance encompasses the organizational processes, policies, standards, and accountability structures that ensure data is accurate, consistent, secure, and used appropriately across the enterprise.
Data Governance Plan
A system-specific plan that defines data lineage, quality standards, access controls, retention policies, consent requirements, and bias monitoring obligations for every dataset used by an AI system.
Data incident
An event impairing integrity, confidentiality, or availability of AI-supporting data — corruption, unauthorized disclosure, unauthorized modification, or loss.
Data Lake
A data lake is a centralized storage repository that ingests and holds large volumes of raw data in its original format, whether structured, semi-structured, or unstructured, until it is needed for analysis, reporting, or AI model training.
Data Lakehouse
A data lakehouse is a modern data architecture that combines the flexibility and scale of a data lake with the management features, performance, and data governance capabilities of a traditional data warehouse.
Data leakage
Information from the test or validation set inadvertently entering training — through preprocessing, feature engineering, target encoding, or time-ordered splits — inflating offline metrics and producing over-optimistic ship decisions.
Data Lineage
Data lineage is the documented, traceable history of a piece of data as it moves through an organization's systems, recording where it originated, how it was collected, what transformations were applied, where it was stored, who accessed it, and how it was ultimately used in AI models or business processes.
Data Mesh
Data mesh is a decentralized data architecture and organizational approach where individual business domain teams own, produce, and maintain their data as discoverable, trustworthy data products, rather than centralizing all data management in a single data engineering team.
Data Minimization
Data minimization is a core data protection principle, mandated by GDPR and adopted by many other privacy frameworks, requiring that organizations collect and retain only the personal data that is strictly necessary for a specific, stated purpose.
Data Pipeline
A data pipeline is an automated, orchestrated sequence of steps that moves data from source systems through extraction, transformation, validation, and loading processes to its destination, which may be a data warehouse, feature store, or directly an AI model's training or inference system.
Data Poisoning
Data poisoning is a type of attack where an adversary deliberately corrupts the data used to train an AI model, causing the model to learn incorrect patterns or behave in unintended ways.
Data Protection Impact Assessment (DPIA)
A DPIA is a formal GDPR-required assessment when data processing poses high risk to individuals, evaluating necessity, proportionality, risks, and mitigations.
Data Quality
Data quality is the degree to which data meets requirements for accuracy, completeness, consistency, timeliness, validity, and uniqueness.
Data quality dimension
A measurable attribute of data integrity — accuracy, completeness, consistency, timeliness, validity, uniqueness, representativeness — used as a scoring axis in a data-readiness rubric.
Data Readiness
Data readiness is an assessment of whether the data required for an AI initiative is available, of sufficient quality, properly governed, legally accessible, and representative of the populations the AI system will serve.
Data Readiness Assessment
A data readiness assessment is a structured evaluation of an organization's data ecosystem to determine its fitness for AI transformation, covering data quality metrics, data lineage documentation, metadata management maturity, access governance policies, storage and compute infrastructure scalability, and data team capabilities.
Data residency (AI)
The requirement that training data, retrieval data, and inference itself occur within a specified jurisdiction.
Data Scientist
A data scientist is a professional who uses statistical analysis, machine learning, and programming to extract insights from data and build predictive or generative models.
Data Steward
A data steward is an individual formally responsible for the quality, governance, and appropriate use of data within a specific domain or business function.
Datasheet for datasets
A structured dataset documentation artifact covering motivation, composition, collection process, preprocessing, uses, distribution, and maintenance — modeled after electronic-component datasheets.
Deadman switch
A fail-safe stop that halts the agent when heartbeats from supervising systems cease for a defined interval.
Deceptive behavior (agentic)
An agentic failure in which the agent produces outputs that misrepresent its state, actions, capabilities, or intent — whether to pass oversight checks, preserve instrumental goals, or exploit principal trust.
Deceptive delegation
An OWASP agentic risk where one agent misrepresents its state, capabilities, or intent to another agent or to a human — whether through deliberate prompt design, emergent behavior, or adversarial compromise.
Decision Log
A decision log is a formal, maintained record of significant decisions made during an AI transformation program, documenting the context and problem statement, alternatives considered, decision criteria, the decision reached, the rationale for the choice, who made the decision, what authority they had, and the expected consequences.
Decision Provenance
Decision provenance is the complete, traceable record of how an AI decision was reached, encompassing the input data, model version, algorithm parameters, intermediate reasoning steps, tool calls, and contextual factors that contributed to a specific output.
Decision Rights
Decision rights are formally documented authorities specifying who can approve what within the AI transformation program.
Deep Learning
Deep learning is a subset of machine learning that uses artificial neural networks with many layers (hence 'deep') to automatically learn complex patterns and representations from large amounts of data.
Defense in Depth
Defense in depth is a security strategy that implements multiple, layered defensive mechanisms throughout an AI system so that if any single layer is breached, other layers continue to provide protection.
Delegation
The assignment of authority from a human principal to an AI agent to act on the principal's behalf.
Delegation Framework
A delegation framework is a governance structure that precisely defines what decisions, actions, and resource commitments an AI agent is authorized to make independently, what requires human approval before proceeding, what escalation paths exist for edge cases, and what hard boundaries the agent must never cross.
Demand Forecasting
Demand forecasting uses AI to predict future customer demand for products or services, enabling optimized inventory management, production planning, workforce scheduling, and supply chain operations.
Demographic Parity
Demographic parity is a mathematical fairness criterion requiring that an AI system's positive outcomes (such as loan approvals, job interview invitations, or benefit eligibility) are distributed equally across different demographic groups, regardless of the group's representation in the underlying data.
Deployer
Under Regulation (EU) 2024/1689, a natural or legal person, public authority, agency, or other body using an AI system under its authority — except where the AI system is used in the course of a personal non-professional activity..
Deployment Readiness Checklist
A comprehensive, sign-off-based checklist that verifies all technical, governance, operational, and organizational prerequisites have been met before an AI system is deployed to production.
Designation
A pinnacle-tier credential recognizing cross-stack mastery across multiple credential types in the COMPEL ecosystem.
DevOps
DevOps is a set of cultural practices, processes, and tools that integrate software development (Dev) and IT operations (Ops) to enable organizations to deliver software changes more frequently, reliably, and with higher quality.
Difference-in-differences (DiD)
A quasi-experimental design comparing treated and control trajectories over time — identifying causal effect from differential change.
Differential Privacy
Differential privacy is a rigorous mathematical framework for sharing data, statistical analyses, or machine learning model outputs while providing formal guarantees that no individual's private information can be inferred from the results.
Dimensionality Reduction
Dimensionality reduction is a technique that simplifies complex datasets with many variables by identifying the most important underlying factors and representing the data in fewer dimensions.
Disaster Recovery
Disaster recovery encompasses the plans, processes, and technical infrastructure for restoring AI systems, data, and services after a catastrophic failure such as data center outages, major security breaches, data corruption, or natural disasters.
Discriminative AI
Discriminative AI models analyze input data to classify it, predict outcomes, or identify patterns.
Disparate Impact
Disparate impact occurs when an AI system's decisions disproportionately and negatively affect a particular demographic group even though the system does not explicitly use protected characteristics such as race, gender, or age as input variables.
Distillation
The training of a smaller "student" model to imitate a larger "teacher" model's behaviour — typically on a shared dataset of prompts and teacher outputs.
Distributor
Under Regulation (EU) 2024/1689, any natural or legal person in the supply chain — other than the provider or the importer — that makes an AI system available on the Union market..
DMAIC
DMAIC (Define, Measure, Analyze, Improve, Control) is the five-phase improvement cycle from Lean Six Sigma methodology.
Drift (value drift)
Erosion of delivered value over time due to changes in data, model, user behaviour, or environment — distinct from data drift (input distribution) and concept drift (relationship change) because its surface is the value metric itself.
Drift Detection
Drift detection is automated monitoring that identifies when the statistical properties of input data or model outputs have shifted significantly from baseline measurements established during model training or initial deployment.
Due Diligence
Due diligence is the comprehensive investigation and risk evaluation of an organization, technology, vendor, or partnership opportunity conducted before making a significant commitment such as an acquisition, major vendor contract, or strategic partnership.
E
Ecosystem Strategy
Ecosystem strategy designs external partnerships, collaborations, and alliances providing capabilities the organization cannot build alone.
Edge Computing
Edge computing is the practice of processing data near its source (at the 'edge' of the network) rather than sending all data to a centralized cloud data center, enabling low-latency AI inference, reduced bandwidth consumption, and operation in environments with limited or intermittent connectivity.
Edge Deployment
Edge deployment refers to running AI models on devices located close to where data is generated -- factory equipment, IoT sensors, retail stores, or branch offices -- rather than in a centralized cloud.
Embedded archetype
Operating-model archetype in which AI capability is fully embedded in every business function with minimal central structure — highest business-proximity, highest divergence risk.
Embedding
An embedding is a mathematical representation that converts text, images, or other complex data into dense numerical vectors (lists of numbers) that capture semantic meaning and relationships.
Embedding governance
Readiness criteria specific to vector-store operations — choice of chunking strategy, embedding-model version pinning, index refresh and re-embedding cadence on model upgrades, per-namespace access control, and retention/deletion workflow consistent with data-subject rights.
Embedding model
A model that maps text, images, or multimodal content to dense vector representations used for retrieval, clustering, and similarity search.
Engagement Architecture
Engagement architecture is the comprehensive design of a COMPEL consulting engagement, encompassing its scope (which domains and pillars are included), phases (how the work is sequenced), workstreams (how parallel activities are organized), deliverables (what tangible outputs will be produced), timeline (how long the engagement will run), team composition (what roles and skills are needed), governance structure (how decisions will be made and progress tracked), and commercial model (how the work is priced and paid for).
Enterprise AI Maturity Spectrum
The Enterprise AI Maturity Spectrum defines five levels of organizational AI capability: Level 1 (Foundational -- scattered, ungoverned experimentation), Level 2 (Developing -- intentional investment with initial governance), Level 3 (Defined -- repeatable, standardized AI delivery), Level 4 (Advanced -- AI embedded in core operations with proactive governance), and Level 5 (Transformational -- AI reshapes the business model).
Enterprise AI Transformation
Enterprise AI transformation is the coordinated, organization-wide effort to embed artificial intelligence into the strategic fabric of a large or complex organization — spanning multiple business units, geographies, regulatory jurisdictions, and technology stacks.
Enterprise Resource Planning (ERP)
ERP systems are integrated business software suites that manage core organizational processes including finance, supply chain, manufacturing, human resources, and procurement.
Enterprise Transformation Architecture
Enterprise Transformation Architecture is a comprehensive, integrated blueprint that unifies AI strategy, organizational design, technology architecture, governance frameworks, change management, and measurement systems into a coherent whole for enterprise-scale AI transformation.
Episodic memory
A form of long-term memory that recalls events from past sessions — "what happened when I ran last Tuesday" — analogous to human episodic memory.
Equalized Odds
Equalized odds is a mathematical fairness criterion requiring that an AI system has equal true positive rates and equal false positive rates across different demographic groups, meaning the system is equally accurate for each group and distributes its errors fairly.
Escalation protocol
Deterministic rules that pause the agent and escalate to a human for decision — triggered by low confidence, high-consequence tool calls, out-of-distribution inputs, or explicit agent request.
ESG (Environmental, Social, and Governance)
ESG is a framework for evaluating corporate behavior and sustainability across three dimensions: Environmental (climate impact, resource usage), Social (labor practices, community impact, diversity), and Governance (corporate ethics, board oversight, transparency).
Ethical Debt
Ethical debt refers to known ethical issues in AI systems that have been identified but not yet remediated, analogous to technical debt in software engineering.
Ethical Impact Assessment (EIA)
An Ethical Impact Assessment is a structured methodology for evaluating the ethical implications of an AI system before deployment.
Ethics by Design
Ethics by design is the approach of integrating ethical considerations into every stage of the AI development lifecycle rather than reviewing ethics after the system is built.
Ethics Review Process
An ethics review process evaluates proposed AI projects for ethical implications before authorization, defining triggers, criteria, review body, decision options, and appeals.
ETL/ELT Pipeline
An ETL (Extract-Transform-Load) or ELT (Extract-Load-Transform) pipeline is a data processing workflow that moves data from source systems into target repositories where it can be used for AI training and operations.
EU AI Act
The EU AI Act (Regulation 2024/1689) is the world's first comprehensive legal framework for regulating artificial intelligence, adopted by the European Parliament in March 2024 and entering into force in August 2024.
EU AI Act Article 52 transparency
Under the pre-consolidation drafts of Regulation (EU) 2024/1689, transparency obligations governing AI systems interacting with natural persons (disclosure), emotion-recognition and biometric-categorisation systems (notification), deepfakes (synthetic-content marking), and AI-generated informational text (disclosure).
Evaluate (COMPEL Stage)
Evaluate is the fifth COMPEL stage, focused on rigorously measuring what the current cycle achieved against planned objectives.
Evaluate Stage
The Evaluate stage is the fifth stage of the COMPEL lifecycle where the outcomes of the transformation program are systematically measured against the objectives established during the Model stage, maturity progression is re-assessed using the 20-domain model, stakeholder satisfaction is gauged, and the overall effectiveness and efficiency of the program are critically examined.
Evaluation harness
The infrastructure that runs capability, regression, safety, and human-review evaluations on an LLM feature on a defined cadence.
Evidence Chain
An evidence chain is a sequence of related governance artifacts that together tell a complete, traceable story from strategic intent through operational implementation.
Evidence Collection Setup
The documented configuration of all evidence collection processes — defining what evidence is gathered, from which systems, on what schedule, in what format, and how it is stored and linked to governance controls — so that the Evidence Pack can be assembled continuously and automatically rather than scrambled together before audit.
Evidence Pack
The complete, auditable collection of artifacts, test results, decision records, and attestations that demonstrate an AI system meets its governance, compliance, and operational requirements.
Evidence-based scoring
A scoring rule requiring tangible evidence — documents, metrics, observed behaviour, artefacts — for any assigned readiness level.
Excessive agency
A failure mode in which an LLM has been wired into tools and permissions whose blast radius exceeds what its supervision and validation logic can safely bound.
Executive Coaching
Executive coaching in the COMPEL context is the structured, one-on-one developmental relationship where an AITGP-level consultant helps senior leaders develop the mindset, capabilities, and behaviors needed to champion, sustain, and personally embody AI transformation in their organizations.
Executive Sponsor
The Executive Sponsor is the C-suite champion who provides strategic direction, budget authority, and organizational air cover for AI transformation.
Executive Sponsorship
Executive sponsorship is the active, visible, sustained commitment from a senior organizational leader who champions the AI transformation program by securing funding, allocating resources, removing organizational barriers, resolving cross-functional conflicts, communicating the transformation vision, and holding the organization accountable for progress.
Experiential Learning
Experiential learning is an educational approach grounded in the theory that lasting knowledge and skill development come from direct experience followed by structured reflection, conceptualization, and active experimentation.
Experiment tracking
The infrastructure and practice of recording artifacts, metrics, parameters, environment, and lineage for every experiment run — enabling later reproduction, comparison across runs, and audit.
Expert Certification
A high-rigor credential requiring 50-80 hours of study that spans multiple COMPEL stages and includes a capstone project with peer review.
Explainability
Explainability is the degree to which an AI system's decision-making process can be understood and communicated to humans.
Explainability Requirements
A documented specification of the minimum explainability standard each AI system must meet — defining what must be explainable, to whom, in what form, and within what timeframe — aligned to the system's risk classification, regulatory context, and stakeholder needs.
Explainable AI (XAI)
Explainable AI (XAI) is a field of research and practice focused on developing techniques, tools, and methodologies that make AI decision-making processes understandable to humans.
External Credential Mapping
A formal mapping between a credential issued by an external organization (such as a recognized training partner, AWS, Google, or Microsoft) and recognition within the COMPEL credential ecosystem.
External Recognition
The formal acknowledgment of credentials earned outside the COMPEL ecosystem that grants equivalent standing, CE credits, or professional certification auto-unlocks within the COMPEL credential lattice.
External Training Bridge
A structured mapping between recognized technical training program completions and COMPEL credential ecosystem entry points.
Externality accounting
Inclusion of environmental and social impacts — energy, water, emissions, labour displacement — in the AI value equation.
F
F1 Score
The F1 score is a model performance metric that combines precision and recall into a single balanced measure, calculated as the harmonic mean of the two.
Facilitation
Facilitation is the professional skill of guiding group discussions, workshops, and collaborative sessions to achieve productive outcomes by managing group dynamics, encouraging diverse participation, maintaining focus on objectives, and synthesizing collective input without imposing the facilitator's own views or conclusions.
Fairness
Fairness in AI is the principle that AI systems should produce equitable outcomes across different demographic groups and not perpetuate or amplify existing societal biases.
Fairness Engineering
Fairness engineering is the technical discipline of detecting and mitigating bias in AI systems through systematic processes applied throughout the model lifecycle.
Feature Store
A feature store is a centralized, managed repository for storing, versioning, and serving the processed data features (engineered variables) used to train and run AI models, enabling feature reuse across teams, ensuring consistency between training and serving environments, and reducing the redundant data processing that occurs when each team independently creates the same features.
Federated archetype
Operating-model archetype in which AI capability is distributed across business units, coordinated by a central standards body that sets policy and shares platform.
Federated Governance
Federated governance sets central AI standards while giving business units implementation autonomy within defined boundaries.
Federated Learning
Federated learning is a machine learning approach where a model is trained across multiple devices, servers, or organizations holding local data, without exchanging the raw data itself.
Federated Model
In organizational design, a federated model distributes AI capability across business units while maintaining a central team that provides standards, shared infrastructure, coordination, and governance.
Federated Model (Organizational)
A federated organizational model distributes AI capability across business units with central coordination, balancing local autonomy with enterprise consistency.
Feedback Loop
A feedback loop in AI occurs when an AI system's outputs influence its future inputs, creating a self-reinforcing cycle that can either improve or degrade performance over time.
Few-shot prompting
A prompt pattern in which the prompt includes labeled examples demonstrating the desired input-output behavior before the real task.
Fine-Tuning
Fine-tuning is the process of further training a pre-trained AI model on a specific dataset to adapt it for a particular task or domain.
FinOps
FinOps (Financial Operations) is the practice of bringing financial accountability, transparency, and optimization to variable cloud and infrastructure spending through real-time cost visibility, collaborative decision-making between engineering and finance teams, and continuous optimization of resource utilization.
FinOps for AI
The FinOps Foundation's Inform-Optimize-Operate lifecycle applied to AI workloads — visibility into AI spend, optimisation of token and compute costs, and operational discipline.
Fitness for purpose
The determination that a specific dataset is appropriate for a specific AI use case given the task, risk tier, and intended deployment context.
Foundation Model
A foundation model is a large pre-trained AI model that serves as a base for multiple downstream applications.
Four Pillars of AI Transformation
The Four Pillars -- People, Process, Technology, and Governance -- are the four interdependent structural foundations of AI transformation in the COMPEL framework.
Framework integration
The mapping between AI operating-model decisions and existing enterprise frameworks — SAFe for delivery, ITIL for service management, PMBOK for project management, COBIT for IT governance.
Framework Interoperability
Framework interoperability is the ability of different management, governance, and delivery frameworks such as COMPEL, SAFe, TOGAF, ITIL, COBIT, PMBOK, and Lean Six Sigma to work together effectively within an organization without creating conflicting requirements, redundant processes, or framework fatigue among practitioners.
Full Transformation Engagement
A full transformation engagement is a COMPEL consulting arrangement that spans the complete COMPEL lifecycle from Calibrate through Learn, typically running six to twenty-four months, involving cross-functional teams, and requiring sustained executive sponsorship.
Function Calling
Function calling is the capability of modern LLMs to produce structured calls to external tools and APIs as part of their output, enabling AI agents to interact with enterprise systems and take real-world actions.
Funding model
The allocation mechanism for AI investment — central allocation, chargeback to business units, showback without billing, or per-initiative funding.
G
Gap Analysis
Gap analysis is the systematic comparison of an organization's current state (as determined by maturity assessment) to its desired future state (as defined by strategic objectives), producing a detailed map of specific capability gaps that must be closed through targeted transformation initiatives.
Gate Review Decision Record
The formal record of a stage gate review decision — documenting the inputs reviewed, the pass/fail outcome, the conditions or exceptions attached to the decision, the named decision-makers, and any required remediation actions before the next gate.
GDPR
The General Data Protection Regulation (GDPR) is the European Union's comprehensive data protection law that governs how personal data of EU residents is collected, processed, stored, and transferred, imposing strict requirements for lawful basis, consent, data minimization, purpose limitation, individual rights (access, deletion, portability), data protection impact assessments, and breach notification.
Generative AI
Generative AI refers to artificial intelligence systems capable of creating new content, including text, images, code, music, video, and synthetic data, based on patterns learned from large training datasets.
Goal hijacking
An OWASP agentic risk in which adversarial input redirects the agent from its intended goal to an attacker-chosen goal.
Goal mis-specification
The failure mode where an agent optimizes for a goal or reward that diverges from the principal's actual intent — because the goal was written too narrowly, too literally, or with a mis-characterized success metric.
Goal-achievement rate
An agent service-level indicator measuring the fraction of tasks that reach their intended outcome without human intervention.
Golden dataset
A versioned, labeled, license-cleared evaluation dataset used as the benchmark reference for an AI feature.
Governance Control
A defined mechanism — preventive, detective, or corrective — that enforces policy compliance, mitigates identified risks, or ensures operational integrity for AI systems.
Governance Harmonization
Governance harmonization is the deliberate process of aligning different AI governance frameworks, policies, standards, and practices across organizational units, business entities, jurisdictions, or partner organizations to create a coherent, non-contradictory governance environment that participants can comply with efficiently.
Governance Maturity
Governance maturity measures the sophistication of AI governance from ad hoc and reactive to optimized and continuously improving.
Governance readiness
The subset of AI transformation readiness covering decision rights, risk-oversight structures, and control framework.
Governance Theater
Governance theater is an anti-pattern where an organization builds the visible apparatus of AI governance -- policies, committees, ethics boards, review processes, published principles -- without operationalizing any of it.
Governance Velocity
A measure of how quickly governance processes enable (rather than impede) AI deployment.
Governance-as-Enabler
Governance-as-enabler is a strategic design philosophy that positions AI governance not as a restrictive control mechanism that slows innovation but as an accelerant that enables the organization to move faster with confidence by providing clear guidelines, pre-approved pathways for common scenarios, and efficient review processes that reduce uncertainty and rework.
GPAI (general-purpose AI model)
Under Regulation (EU) 2024/1689, an AI model — including where trained with a large amount of data using self-supervision at scale — that displays significant generality and can competently perform a wide range of distinct tasks, regardless of how it is placed on the market..
GPAI model with systemic risk
Under Regulation (EU) 2024/1689, a general-purpose AI model classified as having systemic risk because it meets or exceeds high-impact capability thresholds — the initial benchmark being cumulative training compute greater than 10^25 FLOPs — or because it is designated as such by the Commission..
GPU (Graphics Processing Unit)
A GPU is a specialized processor originally designed for rendering graphics in video games, now widely repurposed for AI workloads.
Graceful Degradation
Graceful degradation is the design principle and architectural capability that allows an AI system to continue operating at reduced functionality rather than failing completely when components break, resources become constrained, or performance degrades.
GRC Platform
A GRC (Governance, Risk, and Compliance) platform is specialized software that automates governance workflows, approval tracking, evidence chain visualization, compliance reporting, and risk management processes.
Grounding
Grounding refers to techniques that connect AI model outputs to factual, verifiable information sources rather than relying solely on patterns learned during training.
Grounding Score
The percentage of generative model outputs whose factual claims can be traced to a verifiable source in the supplied context, computed by a grounding evaluator over a fixed test set.
Growth mindset
The belief that abilities can be developed through effort and learning — foundation for continuous learning in AI-era work.
Guardrail
A control placed between the user or environment and an LLM that blocks, rewrites, or classifies content at one of four architectural layers: input filter, policy filter, output filter, or tool-call validator.
Guardrail metric
A secondary metric that must not degrade — typically beyond a declared tolerance — even if the primary metric improves.
Guardrails
Guardrails are the safety boundaries, constraints, filters, and monitoring mechanisms built into AI systems to prevent harmful, inappropriate, unauthorized, or out-of-scope behaviors and outputs.
H
Hallucination
Hallucination is the phenomenon where an AI system, particularly a large language model, generates output that is plausible-sounding and confidently stated but factually incorrect, fabricated, or unsupported by the model's training data or any real-world source.
Hierarchical topology
A multi-agent topology in which a boss agent delegates sub-tasks to worker agents and aggregates their results.
High-risk AI system
Under Regulation (EU) 2024/1689, an AI system falling under Article 6(1) because it is a safety component of, or is itself, a product covered by Annex I Union harmonization legislation, or under Article 6(2) because its use case falls within Annex III — unless exempted by the Article 6(3) derogation..
HIPAA
The Health Insurance Portability and Accountability Act (HIPAA) is a US federal law that establishes strict requirements for protecting sensitive patient health information (Protected Health Information, or PHI) from unauthorized disclosure, with severe penalties for violations.
Horizon Portfolio
A horizon portfolio allocates AI investments across three time horizons: near-term quick wins, medium-term capability building, and long-term strategic bets, ensuring continuous value while investing in future capabilities.
Human Oversight
Human oversight in the context of AI governance refers to the organizational mechanisms, processes, and technical controls that ensure qualified humans maintain meaningful authority over AI system decisions throughout the system lifecycle.
Human oversight (Art. 14)
Under Regulation (EU) 2024/1689, the provider-designed measures that allow natural persons to understand the capacities and limitations of a high-risk AI system, monitor its operation, and intervene or interrupt it.
Human Validation Rules
A formal specification of every decision point in an AI system's operation where human review, approval, or override is required — including the trigger conditions, review timeline, escalation path, and documentation requirements for each rule.
Human-AI collaboration pattern
One of four canonical patterns describing how humans and AI share work: augment (AI enhances human judgment), assist (AI supports but human decides), automate (AI executes with human oversight), arbitrate (AI adjudicates between human positions).
Human-AI collaboration spectrum
The range of collaboration configurations from fully manual, through AI-assisted, to AI-delegated — with each configuration carrying different skill, trust, and governance requirements.
Human-in-the-Loop
Human-in-the-loop (HITL) is an AI system design pattern where a human must actively review, approve, and authorize each AI decision or action before it is executed, providing maximum human oversight at the cost of reduced speed and scalability.
Human-intervention rate
An agent SLI measuring the rate at which humans must intervene on agent actions — via escalation, override, or correction.
Human-on-the-Loop
Human-on-the-loop (HOTL) is an AI system design pattern where the AI makes and executes decisions autonomously, but a human monitors the process through dashboards and alerts and can intervene to override, pause, or adjust the system when anomalies or problems are detected.
Hybrid (hub-and-spoke) archetype
Operating-model archetype with a CoE hub providing platform, standards, and scarce-expertise services, plus embedded spokes in each business unit owning use-case delivery.
Hybrid CoE
A Hybrid Center of Excellence is an organizational model where a central team owns AI standards, governance, shared platforms, and complex cross-functional initiatives, while embedded AI teams within business units handle domain-specific delivery.
Hybrid retrieval
A retrieval pattern combining dense (vector-based, e.g., dense passage retrieval) and sparse (term-based, e.g., BM25) retrieval methods, whose candidate sets are fused — typically via reciprocal rank fusion — before reranking.
Hype Cycle
The Hype Cycle is a Gartner model describing the typical progression of emerging technologies through five phases: Technology Trigger (initial breakthrough generates interest), Peak of Inflated Expectations (publicity produces unrealistic enthusiasm), Trough of Disillusionment (implementations fail to deliver on hype), Slope of Enlightenment (practical benefits become understood), and Plateau of Productivity (mainstream adoption with realistic expectations).
Hyperparameter search
Structured or randomized exploration over model-configuration space — learning rate, depth, regularization strength, and similar — to find a configuration meeting a target criterion.
Hypothesis
A falsifiable statement an AI experiment is designed to support or reject — specifying an intervention, an expected direction of effect, a primary metric, and an acceptable effect size.
I
Importer
Under Regulation (EU) 2024/1689, a natural or legal person located or established in the Union that places on the Union market an AI system bearing the name or trademark of a natural or legal person established outside the Union..
In-Context Learning
In-context learning is the simplest form of AI agent adaptation, where the model adapts its behavior based on information in its current context window -- the conversation history, task instructions, retrieved documents, and tool outputs -- without changing its underlying model weights.
Incident and Risk Review
A structured review of all AI-related incidents, near-misses, and emerging risks that occurred during the evaluation period — including root cause analysis, control failure attribution, and required remediation actions — to ensure that the organization learns from operational experience and updates its risk profile accordingly.
Incident Response
AI incident response encompasses the defined procedures for investigating and remediating AI-related events such as model failures, bias discoveries, data breaches, safety incidents, or unexpected behavioral changes.
Incremental value
Delivered outcome minus counterfactual outcome — the only value quantity that survives CFO and audit-committee scrutiny.
Indemnification
Indemnification is a contractual provision where one party compensates another for specified losses from AI system failures, data breaches, or IP infringement.
Index algorithm
The data-structure choice — HNSW, IVF, product quantization (PQ), or flat (brute-force) — that governs a vector store's query latency, memory footprint, and recall-at-k tradeoff.
Indirect prompt injection
Prompt injection delivered through content the model retrieves or ingests — emails, documents, web pages, or tool outputs — rather than through a direct user message.
Inference
Inference is the process of using a trained AI model to make predictions or generate outputs on new, previously unseen data.
Influence-Interest Matrix
The Influence-Interest Matrix is a stakeholder analysis tool that maps individuals or groups along two dimensions: their level of influence over transformation outcomes and their level of interest in transformation activities.
Information Asymmetry
Information asymmetry occurs when different teams possess different knowledge about project status or risks, leading to misaligned decisions and coordination failures.
Infrastructure as Code (IaC)
Infrastructure as Code is the practice of managing and provisioning computing infrastructure through machine-readable configuration files rather than manual setup processes.
Initiative Sequencing
Initiative sequencing is the strategic ordering of transformation activities based on dependencies between initiatives, organizational readiness and absorption capacity, quick-win opportunity timing, resource availability, regulatory deadlines, and the compounding value that certain foundational capabilities provide to subsequent initiatives.
Inter-annotator agreement
A statistical measure of consistency between human labelers annotating the same items — Cohen's kappa (two annotators), Fleiss' kappa (multiple annotators, nominal scale), Krippendorff's alpha (any scale, tolerant of missing data).
Internal Audit
Internal audit is an independent assurance function evaluating risk management, governance, and controls governing AI.
Internal talent marketplace
Infrastructure — typically a platform — for moving employees across roles as work evolves.
Interpretability
Interpretability is the degree to which a human can understand the internal mechanisms and decision-making logic of an AI model, enabling meaningful inspection of how inputs are transformed into outputs.
Investment Thesis
An investment thesis documents why a specific AI investment will create value, what assumptions underlie returns, and how success will be measured.
ISO 27001
ISO/IEC 27001 is the international standard for information security management systems (ISMS).
ISO 42001
ISO/IEC 42001:2023 is the first international management system standard for artificial intelligence, published jointly by the International Organization for Standardization (ISO) and the International Electrotechnical Commission (IEC).
ITIL (Information Technology Infrastructure Library)
ITIL is a widely adopted framework for IT service management that defines processes for incident management, change management, service level management, configuration management, and continual service improvement.
J
J-Curve Effect
The J-curve effect describes the common pattern in AI transformation where organizational performance initially dips before improving beyond the original level, forming a J-shaped curve when plotted over time.
Jailbreak
A user-crafted prompt pattern that bypasses a model's safety training to elicit restricted behavior.
Jailbreak Resistance
A composite score of an AI system's ability to reject adversarial prompts designed to bypass its safety policies, measured against a fixed, versioned red-team test suite.
Jailbreaking
Jailbreaking is the practice of crafting inputs, prompts, or interactions designed to manipulate an AI system into bypassing its built-in safety restrictions, content filters, or behavioral guidelines to produce prohibited content, reveal confidential information, or perform unauthorized actions.
Joint Controller
Under data protection laws such as GDPR, a joint controller arrangement exists when two or more organizations jointly determine the purposes and means of processing personal data, creating shared legal responsibilities for data protection compliance.
Joint Venture
A joint venture is a business arrangement where two or more organizations combine resources, expertise, and data to pursue a shared AI initiative while maintaining their separate organizational identities.
JSON
JSON (JavaScript Object Notation) is a lightweight, human-readable data format used extensively in AI systems for API communication, configuration files, model metadata, and structured data exchange between applications.
Judicial Review
Judicial review is the process by which courts examine the legality, fairness, and procedural propriety of decisions made by government agencies, public bodies, or AI systems used in administrative or legal decision-making.
K
K-Fold Cross-Validation
K-fold cross-validation is a model evaluation technique that provides a more reliable estimate of model performance than a single train-test split.
Kanban
Kanban is a visual workflow management method that uses boards divided into columns representing stages of work, with cards representing individual tasks that move across the board from left to right as they progress.
Key Performance Indicator (KPI)
A Key Performance Indicator is a quantifiable measurement used to evaluate how effectively an organization or initiative is achieving its objectives.
Key Risk Indicator (KRI)
A Key Risk Indicator (KRI) is a measurable metric that provides early warning of increasing risk exposure before risks materialize as actual incidents or losses.
Kill Switch
A kill switch is an immediate, unconditional mechanism to halt an AI agent's operation.
Knowledge Base
In COMPEL, the knowledge base is the persistent organizational repository of governance knowledge, best practices, lessons learned, reusable patterns, and cautionary tales that accumulates across transformation cycles.
Knowledge Graph
A knowledge graph is a structured representation of real-world entities (people, places, concepts, products) and their relationships, stored in a graph database that enables sophisticated querying and reasoning.
Knowledge Management
Knowledge management is the organizational practice of capturing, organizing, sharing, and applying institutional knowledge to improve decision-making and performance over time.
Knowledge plane
The layer of an AI system that stores and serves non-parametric knowledge to the model — through retrieval over vector stores, traditional indexes, and tool-based data access.
Knowledge Transfer
Knowledge transfer is the deliberate process of transmitting expertise, skills, and understanding from external consultants to client team members, or from experienced practitioners to newer colleagues, ensuring the receiving organization retains the capability to sustain and continue transformation after the engagement ends.
Kotter 8 Steps
John Kotter's organizational-level change method with eight sequenced steps: create urgency, build a guiding coalition, form strategic vision, enlist volunteer army, enable action by removing barriers, generate short-term wins, sustain acceleration, institute change.
KPI Review Report
A structured, cadenced report that measures actual AI system and program performance against the KPIs defined in the Value Thesis Register and Measurement Model — presenting actuals vs.
KPI tree
A hierarchical decomposition of a business outcome into drivers and metrics — with arithmetic or causal relationships at each level.
Kubernetes
Kubernetes is an open-source container orchestration platform that automates the deployment, scaling, management, and networking of containerized applications across clusters of machines.
L
Labeling
Labeling (also called annotation) is the process of tagging data with correct answers to create training datasets for supervised learning.
Lagging indicator
A metric that confirms an outcome after the fact — e.g., quarterly revenue, retention after 90 days, defect rate after release.
Lagging indicator (of readiness)
A signal that confirms past readiness capability — such as model performance in production, incident rate, or value delivered.
Large Language Model (LLM)
A Large Language Model is a massive AI model -- typically based on the transformer architecture and containing billions to trillions of parameters -- trained on enormous amounts of text data to understand and generate human language.
Late chunking
A chunking pattern in which the full document is embedded as a long sequence and chunk boundaries are applied at retrieval time against the embedded representation — rather than chunking before embedding.
Latency
Latency is the time delay between sending a request to an AI system and receiving a response, typically measured in milliseconds.
Leadership Transition Management
Leadership transition management maintains AI transformation momentum when key leaders change due to promotion, departure, or reorganization.
Leading indicator
A metric that moves before the outcome does — predicting near-term outcome trajectory.
Leading indicator (of readiness)
A signal that predicts future readiness capability — such as hiring velocity, training-completion rate, data-quality trend, or sponsor-engagement cadence.
Lean Budget Guardrails
A set of investment policies that govern how funding flows to AI value streams without requiring per-project business cases — typically covering portfolio participation rules, capacity allocation by horizon, approval thresholds for epics, and continuous business owner engagement.
Lean Six Sigma
Lean Six Sigma is a process improvement methodology that combines Lean principles (eliminating waste, maximizing value) with Six Sigma statistical techniques (reducing variation, achieving consistent quality).
Learn (COMPEL Stage)
Learn is the sixth and final COMPEL stage, and paradoxically both the most strategically undervalued and most organizationally consequential.
Learn Stage
The Learn stage is the sixth and final stage of the COMPEL lifecycle where the organization systematically captures lessons learned, codifies new knowledge gained during transformation, shares insights across the enterprise, and feeds those learnings back into the beginning of the next cycle.
Learning Organization
A learning organization, as conceptualized by Peter Senge, is an enterprise that continuously transforms itself through the expansion of its capacity to learn.
Least Privilege
Least privilege is a foundational security principle requiring that AI agents receive access only to the minimum set of tools, data, and system permissions necessary to perform their defined function.
Lewin's unfreeze-change-refreeze
Kurt Lewin's foundational three-state change model: unfreeze the current equilibrium, introduce change, refreeze the new state into a stable equilibrium.
Literacy taxonomy (four levels)
A four-level classification — general population, AI-user, AI-worker, AI-specialist — each with distinct content depth, assessment rigor, and recertification cadence.
LLM risk surface
The union of six interacting layers — input, model, output, retrieval, tool, and data — where governance controls must be applied on any LLM-based feature.
LLM-as-judge
An evaluation technique using a large language model to score outputs from another LLM on quality dimensions — helpfulness, correctness, safety — scaling evaluation beyond human-rater capacity.
Load Balancing
Load balancing distributes incoming requests across multiple servers or model instances to prevent overload, ensuring consistent performance and high availability.
Long-term memory
Persistent memory — typically implemented as a vector store — that survives across agent sessions.
M
Machine Learning (ML)
Machine Learning is a subset of AI where systems learn patterns from data rather than being explicitly programmed with rules.
Manager enablement curriculum
A structured content-and-practice curriculum that builds manager capability in AI literacy, AI-era performance evaluation, coaching, and change leadership.
Market topology
A multi-agent topology in which agents bid on tasks and a market mechanism assigns the task to the most-suitable agent — typically via contract-net protocol.
Master Data Management (MDM)
Master Data Management is the set of processes, governance structures, and technology for ensuring consistent, authoritative definitions of key business entities -- customers, products, suppliers, locations, employees -- across the entire enterprise.
Maturity Assessment
A maturity assessment is a structured, evidence-based evaluation that measures an organization's capabilities, practices, and governance against a defined maturity model, producing numerical scores and qualitative findings that indicate current state, identify gaps, and guide improvement priorities.
Maturity Baseline Report
A scored assessment of the organization's current AI capability maturity across people, process, technology, and governance dimensions, establishing the starting point against which transformation progress is measured.
Maturity Plateau
The maturity plateau is a COMPEL-identified anti-pattern where organizations make genuine early progress in AI transformation -- achieving production deployments, measurable business impact, and functioning governance -- but then stall at an intermediate maturity level (typically Level 2-3), unable to advance further.
Maturity Progression Dashboard
A maturity progression dashboard is a visual governance tool that tracks an organization's advancement across the 18 COMPEL maturity domains over time, displaying current scores, historical trends, targets, and the gaps remaining for each domain.
Mean Time To Recovery (MTTR)
The average elapsed time from detection of an AI incident or SLO breach to restoration of the system to an operationally healthy state.
Measurement Model
The structured framework for quantifying AI transformation progress and outcomes across four levels: strategic KPIs (organization-level), portfolio KPIs (aggregate across use cases), use-case KPIs (individual initiative performance), and operational KPIs (system-level health).
Measurement plan
A pre-launch document naming hypothesis, metrics, sources, cadence, analysis method, and decision rule for an AI feature.
Memory Poisoning
Memory poisoning is an attack targeting AI agents with persistent memory, where an adversary manipulates what the agent remembers to permanently alter its behavior across future sessions.
Metadata
Metadata is data that describes other data -- information about a dataset's source, format, creation date, quality metrics, ownership, access permissions, update frequency, and usage history.
Methodology Benchmarking
Methodology benchmarking is the systematic comparison of AI transformation methodologies across different frameworks, practitioners, and organizations to identify best practices, performance standards, areas for improvement, and opportunities for innovation.
Methodology choice framework
Criteria for choosing among ADKAR, Kotter, Bridges, and hybrid approaches for a specific AI program — based on scale, pace, workforce disposition, and regulatory posture.
Microservices
Microservices is an architectural pattern where applications are built as a collection of small, independent services that communicate through well-defined APIs, each responsible for a specific function and deployable independently.
ML Engineer
An ML engineer is a professional who specializes in building production-quality machine learning systems, bridging the gap between data science (model development) and software engineering (production deployment).
MLOps
MLOps (Machine Learning Operations) is the set of practices, tools, and cultural patterns that enable organizations to deploy, monitor, and maintain machine learning models in production reliably and at scale.
Model
In AI and machine learning, a model is a mathematical representation learned from data that can make predictions or generate outputs.
Model (COMPEL Stage)
Model is the third COMPEL stage, where assessment findings and organizational readiness converge into a concrete, evidence-based transformation plan for the current 12-week cycle.
Model and prompt registry
A versioned inventory of models, system prompts, retrieval sources, and guardrails deployed in production.
Model Card
A model card is a standardized documentation template that describes an AI model's intended use, training data, performance characteristics across different populations, known limitations, fairness evaluations, and ethical considerations.
Model Context Protocol (MCP)
An open protocol published by Anthropic in November 2024 for interoperability between models and tools — standardising how servers expose tools, resources, and prompts to AI clients.
Model Drift
Model drift is the degradation of an AI model's performance over time caused by changes in the statistical properties of the input data, the target variable, or the relationship between them.
Model Lifecycle Management
Model lifecycle management is the governance discipline of maintaining visibility, control, and accountability over AI models from initial conception through production deployment, monitoring, retraining, and eventual retirement.
Model Monitoring
Model monitoring is the continuous, automated observation of AI models operating in production to track performance metrics (accuracy, latency, throughput), detect data drift and concept drift, identify anomalous behavior, monitor fairness metrics, and ensure the model continues to operate within the acceptable parameters defined by its governance framework.
Model Registry
A model registry is a centralized, versioned repository for storing, cataloging, and managing AI models throughout their lifecycle, maintaining metadata about each model's training data, hyperparameters, performance metrics, deployment status, owner, and governance approval status.
Model Risk
Model risk is the risk of adverse consequences arising from errors, limitations, or inappropriate use of AI models.
Model Risk Management (MRM)
Model Risk Management is a governance discipline originating in financial services that provides structured approaches to validating, documenting, monitoring, and governing AI/ML models used in decision-making.
Model routing
A pattern that routes each request to the cheapest model capable of handling it, escalating to more powerful models only when necessary — typically via a small classifier, confidence-based escalation, or response evaluation.
Model selection framework
An eight-criterion decision framework — capability, cost, latency, data residency, customization, operational maturity, exit cost, and license — for choosing a foundation model for a given use case.
Model Stage
The Model stage is the third COMPEL lifecycle stage designing the target state based on Calibrate gaps and Organize priorities, producing the transformation roadmap, governance framework, technology blueprint, and measurement framework.
Model Validation
Model validation is the independent assessment of an AI model's performance, fairness, robustness, and compliance before it is deployed to production.
Model Validation Pipeline
A model validation pipeline automates quality checks, tests, fairness assessments, and security scans that models must pass before production authorization.
Monte Carlo simulation
A numerical method using repeated random sampling to produce probabilistic uncertainty bands on value forecasts — instead of single-point NPV estimates.
Multi-agent collusion
Emergent behavior where multiple AI agents coordinate against principal intent — sharing information, price-fixing, bypassing oversight, or colluding on a task the principals did not authorize.
Multi-agent orchestration
The architectural coordination of multiple cooperating agents — via hierarchical, market, swarm, or actor topologies — to achieve collective tasks.
Multi-Agent System
A multi-agent system (MAS) is an AI architecture in which multiple autonomous agents, each with specialized capabilities or knowledge domains, collaborate to accomplish tasks that no single agent could handle effectively alone.
Multi-agent system (MAS)
A system composed of multiple interacting AI agents — each with its own goals, memory, tool access, and decision logic — that communicate and coordinate to achieve individual or collective objectives.
Multi-armed bandit
An online experimentation strategy that shifts traffic toward better-performing variants during the experiment — trading statistical power for exploitation of early wins.
Multi-Modal AI
Multi-modal AI refers to AI systems that can process and reason across multiple types of data simultaneously, such as text, images, audio, and video.
Multi-rater assessment
A readiness method that gathers data from at least four stakeholder vantage points — executive, manager, individual contributor, and customer-facing — before any dimension is scored.
Multi-tenant AI
A design pattern that serves multiple tenants from a shared AI system with isolation at the retrieval (namespace), rate-limit, policy, cost-attribution, and evaluation layers.
Multi-Workstream Coordination
Multi-workstream coordination is the discipline of keeping parallel transformation activities across the People, Process, Technology, and Governance pillars aligned and progressing in concert during the Produce stage.
Multinational Governance Architecture
Multinational governance architecture designs AI governance operating across multiple countries, balancing global consistency with local regulatory and cultural adaptation.
N
Naive RAG
The baseline retrieval-augmented-generation pattern: a single retrieval step returns the top-k chunks by vector similarity, and those chunks are concatenated into the prompt for generation.
National competent authority
Under Regulation (EU) 2024/1689, the authority designated by each Member State — covering market-surveillance and notifying-authority functions — responsible for enforcement of the Act at national level, including serious-incident handling, audit, and sanction decisions..
Natural Language Processing (NLP)
Natural Language Processing is a branch of AI focused on enabling computers to understand, interpret, and generate human language.
Net Present Value (NPV)
Net Present Value is a financial calculation that determines the current value of all future cash flows from an AI investment minus the initial cost, using a discount rate that reflects the time value of money and investment risk.
Network Effect
A network effect occurs when an AI system or platform becomes more valuable as more people or data flows through it, creating a self-reinforcing cycle: better models attract more users, more users generate more data, more data trains better models.
Neural Network
An artificial neural network is a computing system loosely inspired by the human brain, consisting of layers of interconnected nodes (neurons) that process data by adjusting numerical weights during training.
NIST AI RMF
The NIST AI Risk Management Framework (AI RMF 1.0), published by the National Institute of Standards and Technology in January 2023, is a voluntary framework for managing risks associated with the design, development, deployment, and evaluation of AI products and services.
Non-Disclosure Agreement (NDA)
An NDA establishes confidentiality obligations between parties, protecting sensitive information during AI engagements, vendor evaluations, and partnerships.
O
Observability
Observability is the comprehensive ability to understand the internal state, behavior, and health of an AI system by examining its external outputs, including logs, metrics, traces, and events.
Observability for AI
The set of telemetry — prompt/response capture, retrieval traces, tool-call records, token-cost metrics, and evaluation signals — that makes an AI system operable and auditable.
Offline evaluation
Assessment of an AI system against static datasets — training hold-out, validation set, benchmark corpus — without exposure to live user traffic.
OKRs
A quarterly alignment tool pairing a qualitative Objective with 3-5 measurable Key Results.
Online evaluation
Assessment of an AI system under live traffic using randomized or sequential experimental designs — A/B test, multi-armed bandit, canary, or interleaving.
Operating Model
An operating model defines how an enterprise structures its teams, processes, governance mechanisms, and technology to deliver its strategy.
Operating Model Design
Operating model design is the process of defining how an organization's AI capabilities will be structurally organized, staffed, funded, governed, and operated to deliver sustainable value at enterprise scale.
Operating-model maturity
A five-level COMPEL maturity progression applied to operating-model dimensions — Ad Hoc, Emerging, Defined, Managed, Optimizing.
Operational Readiness
The assessed capability of an organization to sustain AI operations across 10 interdependent dimensions: strategy alignment, governance maturity, operating model, workforce capability, data readiness, technology infrastructure, monitoring and observability, vendor dependency management, compliance readiness, and change and adoption.
Operational Resilience
Operational resilience is the ability of an organization to prevent, prepare for, respond to, recover from, and learn from operational disruptions to its AI systems and AI-dependent business processes.
Opportunity Cost
Opportunity cost is the potential value lost by choosing one AI initiative over another when resources are limited.
Oral Defense
An oral defense is a live examination where COMPEL certification candidates at Levels 3 and 4 present and defend their capstone work before a panel of experienced evaluators who probe the depth of understanding, professional judgment, and integrated mastery demonstrated in the submission.
Orchestration plane
The layer of an AI system that coordinates prompts, retrievals, tool calls, safety filters, and routing between the user and one or more models.
Organizational Readiness
Organizational readiness for AI transformation is the degree to which an organization's people, culture, processes, and structures are prepared to adopt and benefit from AI.
Organizational readiness score
A composite measure across leadership, culture, skills, process, and technology dimensions — indicating how ready the organisation is to absorb AI-driven change.
Organize (COMPEL Stage)
Organize is the second COMPEL stage, focused on building the human and structural infrastructure required to execute transformation.
Organize Stage
The Organize stage is the second COMPEL lifecycle stage translating Calibrate findings into mobilization through stakeholder alignment, team formation, governance establishment, and resource securing.
Overfitting
Overfitting occurs when an AI model learns the training data too precisely -- memorizing specific examples including their noise and anomalies rather than learning generalizable patterns -- and consequently performs poorly on new, unseen data.
P
Parameter
A parameter is a learned numerical value within an AI model that is adjusted during training to improve the model's ability to make accurate predictions.
Partner ecosystem
The named set of external partners — management consultancies, systems integrators, managed-AI-service providers, academic collaborators, citizen-AI communities — that closes capability gaps the internal AI operating model cannot fill.
Pattern Library Update
A structured update to the organization's AI pattern library — capturing reusable design patterns, anti-patterns, successful control configurations, and workflow redesign templates discovered during the current transformation cycle — so that future use cases benefit from accumulated organizational experience.
PCI DSS
PCI DSS (Payment Card Industry Data Security Standard) is a set of security standards for organizations that handle, process, or store credit card information.
Peer Review Assessment
Peer review assessment is a structured evaluation process in which capstone project submissions are reviewed by qualified peers using standardized rubrics with weighted criteria, defined scoring levels, minimum passing thresholds, and required reviewer counts.
Peer Review Rubric
A structured evaluation framework used by qualified reviewers to assess expert certification capstone projects and joint credential defenses.
PEFT (parameter-efficient fine-tuning)
A family of fine-tuning techniques — most prominently LoRA, QLoRA, and adapters — that update only a small fraction of model parameters while freezing the rest.
Penetration Testing
Penetration testing is the authorized, controlled simulation of real-world attacks against an AI system to identify exploitable security vulnerabilities before malicious actors can discover and exploit them.
People and change KPI tree
A hierarchical KPI decomposition rooted in workforce outcomes (productivity, retention, engagement, belonging) with drivers (literacy, role-redesign adoption, manager-enablement completion) and metrics at each level.
Per-task cost
An agent SLI capturing the full compute and API cost of a single task end-to-end — including all loop iterations, tool calls, memory reads and writes.
Performance evaluation redesign
Changes to goal-setting, coaching, and review processes forced by AI integration — including attribution rules, revised productivity expectations, and safeguards against AI-integrated surveillance overreach.
Persistent Memory
Persistent memory extends an AI agent's learning beyond a single session by storing information -- facts, preferences, outcomes, strategies -- in an external memory system that is retrieved when processing new tasks.
Pilot Program
A pilot program is a controlled, limited-scope initial deployment of an AI solution in a real operational environment, designed to test feasibility, measure actual impact, identify integration challenges, gather user feedback, and validate the business case before committing to full-scale enterprise implementation.
Pilot Purgatory
Pilot purgatory is a COMPEL-identified anti-pattern where organizations launch numerous AI pilot projects but never build the governance, data infrastructure, organizational capability, or production readiness to move them beyond the pilot stage.
Pilot-to-Production Gap
The pilot-to-production gap describes the common phenomenon where AI proofs of concept demonstrate impressive results in controlled environments but never scale to full production deployment.
Pipeline
An automated execution graph connecting data ingestion, feature engineering, training, evaluation, and deployment stages — parameterized, versioned, and re-runnable.
Plan-and-Execute
A loop pattern where a planner step produces a multi-step plan, then an executor step carries each step out — with optional replanning after observations.
Playbook
A playbook is a documented set of step-by-step procedures, decision trees, and communication templates for handling specific operational scenarios such as AI incident response, model deployment, stakeholder escalation, or governance review.
Policy Baseline Document
A structured inventory and gap analysis of all existing organizational policies relevant to AI — covering data, security, privacy, ethics, procurement, and acceptable use — that identifies which policies need to be created, updated, or retired to support the AI transformation program.
Policy engine
A declarative rule engine that gates agent actions — via ABAC-style policies evaluating identity, resource, environment, and action attributes.
Policy Lifecycle Management
Policy lifecycle management covers creating, reviewing, approving, implementing, monitoring, updating, and retiring AI governance policies ensuring they remain current.
Policy Update Register
A tracked register of all policy changes, additions, and retirements triggered by insights from the Evaluate stage — documenting the rationale for each policy change, the approver, the effective date, and the affected systems or processes.
Political Navigation
Political navigation is the professional skill of understanding and working within organizational power dynamics, informal influence networks, competing agendas, and institutional politics to advance AI transformation objectives.
Portfolio Management
Portfolio management in the COMPEL context is the centralized governance and optimization of a collection of AI transformation programs, projects, and operational activities managed together to achieve enterprise strategic objectives.
Portfolio Risk Aggregation
Portfolio risk aggregation combines individual program risks into portfolio-level views revealing systemic patterns, correlated exposures, and concentration risks invisible at program level.
Portfolio Strategic Themes
Thematic investment categories defined at the portfolio level that translate enterprise strategy into the differentiated business outcomes the AI portfolio must deliver.
Portfolio Vision
A 3-5 year directional statement of the AI portfolio target state, describing the capabilities, business outcomes, and operating posture the organization intends to reach.
Position bias (judge)
The systematic tendency of an LLM-as-judge to favour responses in a particular position (first, second, or last) when comparing candidates — independent of content quality.
Post-hoc review
An oversight mode in which human reviewers inspect agent actions after they complete, typically on a sample, exception, or threshold basis.
Post-market monitoring (Art. 72)
Under Regulation (EU) 2024/1689, the ongoing, documented collection, analysis, and corrective-action process that providers of high-risk AI systems must operate after the system is placed on the market.
Post-Mortem
A post-mortem is a structured review conducted after an AI incident, project milestone, or engagement conclusion to analyze what happened, understand root causes, identify lessons learned, and develop actionable improvements for the future.
Pre-authorization oversight
An oversight mode in which a named human must approve specified agent actions before they execute.
Precision
Precision is a model performance metric measuring the proportion of positive predictions that are actually correct -- in other words, when the model says 'yes,' how often is it right? High precision means few false positives (false alarms).
Predictive Maintenance
Predictive maintenance uses AI to predict when equipment will fail so maintenance can be performed just before failure occurs, rather than on a fixed schedule (preventive maintenance) or after failure (reactive maintenance).
Primary metric
The single metric whose movement drives an experiment's ship/no-ship decision.
PRINCE2
PRINCE2 (Projects in Controlled Environments) is a structured project management methodology widely used in government, regulated industries, and large enterprises.
Privacy
Privacy in the AI context goes beyond compliance with regulations like GDPR or CCPA to encompass a broader commitment to responsible data stewardship.
Privacy by Design
Privacy by design embeds data privacy protections into AI system design from the earliest stages rather than adding controls afterward.
Proactive Regulatory Engagement
Proactive regulatory engagement is the strategic practice of actively participating in regulatory development processes rather than passively waiting for final rules and then scrambling to comply.
Produce (COMPEL Stage)
Produce is the fourth COMPEL stage and the execution engine of each 12-week cycle.
Produce Stage
The Produce stage is the fourth COMPEL lifecycle stage executing the transformation plan across all four pillars simultaneously including technology deployment, model development, process redesign, governance operationalization, and change management.
Production Readiness
The verified state in which an AI system meets all prerequisites for safe, governed operation in a production environment.
Professional Certification
A focused, stackable credential requiring 12-20 hours of study with automated quiz and portfolio artifact assessment.
Prompt
The input text, structured message sequence, or multimodal payload provided to a language model at inference time.
Prompt architecture
The production-grade concerns around prompts as architectural artefacts — template assembly, versioning, registry management, injection defense, and structured-output schemas.
Prompt caching
An inference optimisation that caches the attention key-value state for a prompt prefix so that subsequent requests sharing the same prefix skip re-processing.
Prompt Engineering
Prompt engineering is the practice of designing and refining the text inputs (prompts) given to a large language model to produce desired outputs.
Prompt evaluation harness
The infrastructure that runs capability, regression, safety, and human-review evaluations on prompts — distinct from a general LLM evaluation harness by scope: prompt-evaluation tests the prompt while holding the model fixed, catching prompt-level drift (e.g., after a system-prompt edit) without attributing it to the model.
Prompt Injection
Prompt injection is a security attack where malicious instructions are hidden in input data to manipulate an AI agent's behavior, potentially causing it to ignore safety guidelines, reveal sensitive information, or take unauthorized actions.
Prompt Injection Resistance
The measured ability of an AI system to reject or neutralize adversarial instructions injected via user input, retrieved documents, tool output, or other untrusted content channels.
Prompt template
A versioned, parameterized representation of a prompt — with placeholders for user input, retrieved context, and dynamic state — enabling reuse, testing, and audit.
Proof of Concept (PoC)
A Proof of Concept is a small-scale implementation that demonstrates the feasibility of an AI solution in a controlled environment, typically using sample data and simplified conditions.
Propensity-score matching
An observational-study method matching treated and control units on estimated probability of receiving treatment, then comparing outcomes across matched pairs.
Provenance
The record of origin and custody for a data asset — who collected it, from whom, under what legal basis, and through which hands it passed — required for auditability of high-risk AI under EU AI Act Article 10.
Provenance Graph
A provenance graph represents the complete chain of data sources, processing steps, and agent interactions leading to an AI output.
Provider
Under Regulation (EU) 2024/1689, a natural or legal person, public authority, agency, or other body that develops an AI system or a general-purpose AI model — or has it developed — and places it on the market or puts it into service under its own name or trademark, whether for payment or free of charge..
Pseudonymization
Pseudonymization replaces identifying information with artificial identifiers while maintaining a secured re-identification mapping.
Psychological Safety
Psychological safety is the shared belief within a team or organization that individuals can take interpersonal risks -- asking questions, admitting mistakes, proposing unconventional ideas, reporting problems -- without fear of punishment or ridicule.
Purpose Limitation
Purpose limitation is the privacy principle ensuring that data collected for one purpose is not repurposed for AI training or other uses without appropriate consent, legal basis, and governance review.
Q
Quality Assurance (QA)
Quality assurance for AI extends traditional software testing with model-specific validation processes to ensure AI systems meet defined standards for performance, reliability, fairness, and governance compliance.
Quality Gate
A quality gate is a predefined checkpoint in a development or transformation process where deliverables must meet explicit quality criteria before work can proceed to the next stage.
Quantitative Risk Assessment
Quantitative risk assessment is an approach to evaluating AI risks that uses numerical data, statistical methods, and mathematical models to estimate the probability and potential financial or operational impact of identified risks.
Quantization
Quantization is an optimization technique that reduces the computational resources required to run an AI model by decreasing the numerical precision of its internal calculations, typically from 32-bit floating point to 16-bit, 8-bit, or even 4-bit representations.
Quantization (AI cost)
Representation of model weights (and sometimes activations) at lower numerical precision — INT8, INT4, or mixed-precision — to reduce memory footprint and accelerate inference.
Query Optimization
Query optimization is the process of improving the efficiency of data retrieval operations to reduce latency (response time) and resource consumption (compute and storage costs).
Quick Win
A quick win is a transformation initiative strategically selected and designed to deliver visible, measurable, and broadly recognized value within a short timeframe (typically six to twelve weeks), building organizational momentum, stakeholder confidence, and political support for the broader, longer-term transformation program.
R
RACI / RAPID / DACI
Three named decision-rights frameworks with different emphasis: RACI (responsible, accountable, consulted, informed) clarifies ownership; RAPID (recommend, agree, perform, input, decide) separates decide from perform; DACI (driver, approver, contributors, informed) emphasizes the driver as single point of execution.
RACI Matrix
A role-accountability matrix that assigns Responsible, Accountable, Consulted, and Informed designations for every key AI governance decision, control activity, and operational process across the organization.
RAG (Retrieval-Augmented Generation)
Retrieval-Augmented Generation (RAG) is an AI architecture pattern that enhances the accuracy and reliability of large language model outputs by first retrieving relevant information from external knowledge sources (databases, documents, knowledge bases) and then including that retrieved information in the context provided to the model for response generation.
ReAct
An agentic prompt pattern interleaving reasoning steps (Thought) and action steps (Act — typically a tool call) over multiple turns, with observations from each action feeding the next reasoning step.
Readiness Assessment Report
A point-in-time, evidence-backed assessment of the organization's readiness to advance from one COMPEL stage to the next, scored across all 10 operational readiness dimensions with explicit pass/fail thresholds and remediation actions for any dimension below threshold.
Readiness dimension
A single assessable domain — for example, data foundation, sponsor strength, or operating-model design — with a rubric, evidence requirements, and a five-level scoring scale.
Readiness recommendation (go / wait / redesign)
The three-way output of a readiness engagement: proceed with the AI initiative; pause until named gaps close; or redesign scope and approach.
Readiness scorecard
A structured, dimension-by-dimension artifact summarizing evidence, scores, remediation priorities, and owner assignments for a use-case-scoped data-readiness assessment.
Real-Time Processing
Real-time processing involves generating AI predictions as events occur, typically delivering results within milliseconds to seconds.
Realized value
The outcome the organisation actually captures from an AI feature, after accounting for adoption, override, and drift effects — distinct from shipped value, which is the theoretical capacity.
Recall
Recall is a model performance metric measuring the proportion of actual positive cases that the model correctly identifies -- in other words, of all the real positives, how many did the model catch? High recall means few false negatives (missed cases).
Recommendation Engine
A recommendation engine is an AI system that suggests relevant items -- products, content, actions, or connections -- to users based on their behavior, preferences, and similarities to other users.
Red Teaming
Red teaming is a security and safety testing practice where a dedicated team deliberately attempts to find vulnerabilities, trigger unsafe behavior, or exploit weaknesses in an AI system.
Red-team (for LLMs)
A structured adversarial exercise against an LLM feature using human, automated, or hybrid techniques drawn from MITRE ATLAS or OWASP LLM Top 10 to discover failure modes before attackers do.
Red-team experiment
An adversarial experiment designed to probe failure modes rather than validate desired behavior — structured, hypothesis-driven exploration of safety bypass, goal mis-specification, jailbreak, and harm.
Redesigned role specification
An output artefact documenting a redesigned role — tasks, skills, AI touchpoints, performance expectations, growth path.
Redundancy planning
Structured planning for roles that retire due to AI — including retraining, redeployment, and exit pathways aligned with law and human dignity.
Reflection
An agentic prompt pattern where the model critiques and revises its own output — typically after tool-call feedback or detected error — before returning a final answer.
Regression
Regression is a supervised learning task that predicts a continuous numerical value rather than a discrete category.
Regression discontinuity (RDD)
A quasi-experimental design using a threshold — e.g., a credit-score cutoff, an eligibility cutoff — to create a natural experiment near the cutoff.
Regulatory Compliance
Regulatory compliance for AI encompasses the organizational processes and practices that ensure AI systems meet the requirements of applicable laws, regulations, and industry standards across all relevant jurisdictions.
Regulatory Intelligence
Regulatory intelligence is the systematic, ongoing monitoring, analysis, and interpretation of regulatory developments, enforcement actions, policy proposals, guidance documents, and judicial decisions relevant to AI across all jurisdictions where an organization operates.
Regulatory Sandbox
A regulatory sandbox is a controlled, supervised environment established by a regulatory authority where organizations can test innovative AI applications under relaxed or modified regulatory requirements, with close regulator oversight and structured learning objectives.
Reinforcement Learning
Reinforcement Learning (RL) is a machine learning paradigm where an agent learns by interacting with an environment and receiving rewards or penalties for its actions.
Reinforcement Learning from Human Feedback (RLHF)
RLHF is the technique used to align large language model behavior with human preferences and safety requirements.
Reinforcement mechanism
A structured means by which behavior change persists after the formal transformation phase — incentives aligned to new behavior, visible leadership modeling, communities of practice, and continuous feedback loops.
Remediation roadmap
A phased, resourced plan that closes identified readiness gaps, sequenced against the organisation's change capacity and dependency map.
Replayability
The ability to deterministically reproduce an agent trace given captured inputs — a recorded session, the prompt, the tool responses, the random seed.
Replicability
The property that an independent team reproduces the qualitative conclusions of an experiment using different data, tooling, or implementation.
Reproducibility
The property that re-running an experiment with the same code, data, and configuration produces the same results within declared tolerance.
Reranker
A cross-encoder model that re-scores a candidate set of documents — typically the top-k from an initial vector or hybrid retrieval — to improve ordering by the actual query-document semantic match.
Resilience
Resilience is the multidimensional capability of an AI system, transformation program, or organization to anticipate, withstand, respond to, recover from, and adapt to adverse events, disruptions, and changing conditions.
Resistance analysis
A mapping of change resistance by type — rational, experiential, political, values-based — plus diagnosis of the underlying concern driving each instance.
Responsibility Index
A composite scorecard metric for the Responsibility dimension, combining bias delta, explainability coverage, and human-oversight effectiveness into a single index used in executive reviews.
Responsible AI
Responsible AI is the practice of designing, developing, and deploying AI systems in ways that are ethical, transparent, fair, accountable, and safe — and that actively avoid creating harm to individuals, groups, or society.
Retirement/Redesign Decision Record
A formal record of the decision to retire or fundamentally redesign an AI use case — documenting the evidence that triggered the decision, the options considered, the chosen path, the decommissioning or redesign plan, and the governance actions required to safely wind down or restart the use case.
Retraining
Retraining is the process of updating an AI model by training it on new or additional data to restore or improve its performance after drift, degradation, or changing business requirements.
Retrieval-Augmented Generation (RAG)
Retrieval-Augmented Generation is a technique that enhances AI model responses by first retrieving relevant information from external knowledge sources -- databases, document repositories, knowledge bases -- and then using that information as context for generating more accurate, grounded answers.
Retrospective
A retrospective is a structured review session conducted after completing work to examine what went well, what went wrong, and what should change going forward.
Return on Investment (ROI)
Return on Investment (ROI) is a financial performance measure calculated as (net benefits minus costs) divided by costs, expressed as a percentage, used to evaluate the profitability of an AI transformation investment.
Reward Hacking
Reward hacking occurs when an AI agent learns to maximize its reward signal in unintended ways that do not align with the actual desired outcome.
Risk Appetite
Risk appetite is the overall level and types of risk that an organization is willing to accept in pursuit of its strategic objectives, set by the board of directors or equivalent governing body.
Risk Appetite Statement
A formally approved document that defines the types and levels of AI-related risk the organization is willing to accept in pursuit of its AI ambition, covering operational, reputational, regulatory, and ethical risk dimensions.
Risk Assessment Report
A structured, evidence-based evaluation of the risks associated with each AI system or use case — covering technical, operational, ethical, regulatory, and reputational risk dimensions — with scored likelihood and impact ratings, existing control effectiveness ratings, and residual risk determinations.
Risk management system (Art. 9)
Under Regulation (EU) 2024/1689, a continuous, iterative process running across the entire lifecycle of a high-risk AI system that identifies foreseeable risks, estimates and evaluates risks, adopts risk-management measures, and documents residual risks.
Risk Register
A risk register is a comprehensive, living document that catalogs all identified AI-related risks for a transformation program or portfolio, recording each risk's description, probability assessment, potential impact, current mitigation measures, residual risk level, assigned owner, review frequency, and status.
Risk Taxonomy
A risk taxonomy is a structured classification system that organizes AI-specific risks into categories with defined severity levels, likelihood assessments, and mitigation strategies.
Risk-adjusted NPV (rNPV)
Net present value with stage-probability and risk-weighted discount applied at each cash-flow stage — capturing both the time value of money and the probability-of-success at each lifecycle gate.
Risk-Based Classification
Risk-based classification is an approach to AI governance that applies different levels of regulatory requirements, oversight, and governance controls based on the potential risk of harm from the AI application.
Roadmap
A transformation roadmap is a strategic planning document that maps AI initiatives to timelines, resources, dependencies, milestones, and success criteria.
Robotic Process Automation (RPA)
RPA is software that automates repetitive, rule-based tasks typically performed by humans, such as data entry, form filling, file moving, and report generation.
ROI and Outcome Report
A comprehensive financial and strategic outcome report that quantifies the realized value of an AI initiative against its original value thesis — covering financial return, operational improvement, strategic enablement, and risk reduction — providing the evidence base for scaling, continuation, or retirement decisions.
Role exposure score
A task-level measure of how much of a role's work is exposed to AI capability — computed by mapping role tasks to AI-capable task categories.
Role redesign
Explicit redefinition of jobs when AI shifts tasks between humans and systems — covering new accountabilities, required skills, oversight responsibilities, and performance metrics.
Role-specific literacy
Adaptation of AI literacy content to the specific AI touchpoints of a given role — such as credit-risk analysts, customer-service agents, radiologists.
Rollback
Rollback reverts an AI system to a previous known-good state when the current version causes problems, requiring versioned model snapshots and automated reversion procedures.
Runaway loop
An agentic incident class in which the agent recurses indefinitely without making progress — typically by re-invoking tools, re-planning without termination, or cycling through memory — until compute budget or token context is exhausted.
Runtime intervention
An oversight mode in which a human can redirect, pause, or halt an agent during execution — through a dashboard, approval prompt, or interrupt signal — without requiring pre-authorization for every action.
S
SAFe (Scaled Agile Framework)
SAFe (Scaled Agile Framework) is a comprehensive framework for implementing agile practices at enterprise scale, addressing the coordination challenges that arise when multiple agile teams must work together on large, complex initiatives.
Safety
Safety in AI means that systems are designed to operate reliably within their intended boundaries and fail gracefully when they encounter situations outside their training distribution.
Sandboxing
Technical isolation that limits an agent's actions to a bounded environment — file-system jail, network-namespace restriction, container, VM, or ephemeral cloud workspace — so that a misbehaving agent cannot reach systems of record, production data, or external networks.
Scalability
Scalability is the ability to expand AI capabilities from individual successes to enterprise-wide deployment without proportional increases in effort or cost per deployment.
Scaling Decision Record
A formal record of the decision to scale, replicate, or extend an AI use case to additional business units, geographies, or processes — documenting the evidence basis, scaling approach, resource requirements, governance adaptations needed, and the named decision-makers who authorized the expansion.
Scope Creep
Scope creep is the uncontrolled, incremental expansion of a project or engagement's requirements beyond its originally agreed boundaries, typically occurring through a series of individually reasonable additions that collectively consume disproportionate time, resources, and budget without corresponding value.
Scrum of Scrums
Scrum of scrums is a coordination mechanism where representatives from multiple agile teams meet regularly to share progress, surface cross-team dependencies, and resolve inter-team issues in large transformation programs with five or more concurrent workstreams.
Security by Design
Security by design is the principle of integrating security considerations into the architecture and design of AI systems from the earliest stages of development rather than adding security measures as an afterthought.
Self-consistency
A meta-prompting technique that samples multiple independent reasoning paths (via temperature > 0) and selects the most frequent final answer — improving accuracy on multi-step reasoning tasks at the cost of proportional compute.
Self-Sustaining Capability
Self-sustaining capability is an organization's demonstrated ability to continue AI transformation, governance improvement, and capability development independently after external consultants depart.
Semantic caching
A caching strategy in which cache hits are determined by semantic similarity to prior queries rather than exact-string match — typically implemented by embedding the query and performing nearest-neighbour search over a cache of past query-response pairs.
Semantic chunking
A chunking strategy that respects semantic boundaries — sentence, paragraph, or topic-shift — rather than fixed token windows.
Semantic memory
Structured knowledge — typically a knowledge graph or ontology — that the agent can query for general facts rather than session-specific history.
Sensitivity Analysis
Sensitivity analysis is a technique that tests how changes in key assumptions affect the outcomes of a business case, financial model, or risk assessment.
Sentiment Analysis
Sentiment analysis is a natural language processing technique that determines the emotional tone, opinion, or attitude expressed in text -- typically classified as positive, negative, or neutral, sometimes with finer-grained categories like anger, joy, or frustration.
Serious incident (Art. 3(49))
Under Regulation (EU) 2024/1689, an incident or malfunction of an AI system that directly or indirectly leads to the death of a person or serious harm to a person's health, to a serious and irreversible disruption of critical infrastructure, to infringement of fundamental-rights obligations, or to serious harm to property or the environment..
Service Level Agreement (SLA)
A Service Level Agreement is a formal commitment between a service provider and consumer defining expected performance levels.
Serving pattern
The architectural shape of the inference path — managed API, cloud-platform hosted, self-hosted online, self-hosted batch, or edge.
Shadow AI
Shadow AI refers to AI tools, models, and AI-enabled applications that employees use within an organization without formal approval from IT, legal, risk management, or governance functions.
Shadow AI Inventory
A structured catalogue of AI tools, models, and automated systems already in use across the organization that were deployed outside formal governance channels.
Shadow Deployment
Shadow deployment (also called shadow mode) is a deployment pattern where a new AI model runs alongside the current production model, receiving the same real-world inputs but without its outputs being served to users or affecting business processes.
Shadow traffic
A deployment pattern in which a new model or prompt version receives a copy of live traffic and produces outputs that are captured for evaluation but not returned to users.
Shipped value
The outcome an AI feature can theoretically produce if it is adopted and sustained at design intent — the denominator against which realized value is compared.
Short-term memory
The working context window an agent uses within a single task — prompts, retrievals, tool outputs, intermediate reasoning.
Showback Model
A showback model shows business units their AI resource consumption costs without billing them, creating awareness before full chargeback implementation.
Simulation harness
A virtual environment for agent evaluation without production side effects — mock tools, synthetic data, deterministic scenarios.
Skills adjacency map
A structured visualisation of which current skills are adjacent to future-demand skills — enabling targeted reskilling rather than blanket retraining.
SLI/SLO for AI
Service-level indicators and objectives for AI systems — including evaluation score, per-task cost, and goal-achievement rate alongside classical availability/latency.
Sovereign AI
Sovereign AI is the concept of national or organizational control over AI capabilities, data, and infrastructure, ensuring that critical AI systems are not dependent on foreign providers, jurisdictions, or infrastructure that could be disrupted by geopolitical events, sanctions, or policy changes.
Sponsor Commitment
A formal, signed commitment from executive sponsors authorizing an AI transformation program and allocating the human, financial, and political capital required to sustain it.
Sponsor strength
A composite indicator of executive sponsorship quality — visibility, budget authority, political capital, and sustained engagement — rather than formal title alone.
Sprint
A sprint is a fixed one-to-four-week period during which a transformation team commits to completing defined deliverables, providing the rhythmic heartbeat of the Produce stage through planning, execution, review, and retrospective.
Stacking Rules
The formal rules governing how credentials combine and contribute to higher-level credentials in the lattice.
Stage Gate
A stage gate is a structured decision point between COMPEL lifecycle stages that verifies deliverables meet quality criteria before the organization advances.
Stage-gate value review
A COMPEL-stage gate where realized-value-to-date is compared against the business case, and the investment is continued, adjusted, or sunset.
Stakeholder
A stakeholder is any individual, group, or organization that has an interest in or is affected by an AI transformation initiative.
Stakeholder Alignment
Stakeholder alignment is the deliberate process of ensuring that all key stakeholders in an AI transformation program share a common understanding of objectives, success criteria, roles, governance mechanisms, risk tolerances, and expected outcomes before and during program execution.
Stakeholder Engagement Plan
A structured plan that identifies all stakeholders affected by the AI transformation program, maps their influence and interest, and defines the engagement approach, communication cadence, and escalation paths for each stakeholder group.
Stakeholder Mapping
Stakeholder mapping identifies all individuals and organizations affected by or influential over AI transformation, documenting roles, interests, concerns, and relationships.
Standard Contractual Clauses
Standard Contractual Clauses (SCCs) are pre-approved contract terms established by the European Commission for transferring personal data from the EU to countries that do not have an adequacy decision.
State graph
An agent execution modeled as a directed graph of states and transitions — each node is a reasoning or tool-call step, each edge is a conditional transition.
Status-quo bias
A cognitive pattern favoring continued current behavior absent strong disconfirming evidence — even when an alternative is demonstrably superior.
Steering Committee
A steering committee is a senior leadership body that provides strategic oversight, decision-making authority, cross-functional conflict resolution, and executive sponsorship for an AI transformation program.
Stop-go decision right
The explicit assignment of authority to halt or resume an agent — distinct from authority to configure, deploy, or operate it.
Strategic Risk
Strategic risk encompasses threats to an organization's fundamental strategy, competitive position, or long-term viability, including the risk of falling behind competitors in AI capability, making wrong technology platform bets, failing to attract and retain AI talent, or being disrupted by AI-native competitors.
Structured Data
Structured data is data organized in a predefined format with rows and columns, such as spreadsheets, database tables, ERP records, and CRM entries.
Structured output
Output formatted to match a declared schema — JSON with a JSON Schema, function-call arguments, or grammar-constrained text — rather than free text.
Subgroup coverage
The representation of protected or intersectional groups in a dataset relative to their prevalence in the target population.
Summative Assessment
Summative assessment is a final evaluation conducted at the conclusion of a learning program or certification process to determine whether participants have achieved the required learning objectives and competency levels.
Sunset / decommission case
A structured argument to stop running an AI feature whose realized value no longer justifies its TCO.
Supervised Learning
Supervised learning is the most widely deployed machine learning paradigm in enterprises.
Supply Chain AI Governance
Supply chain AI governance manages risks and accountability across organizational boundaries through vendor relationships, third-party models, and partner integrations.
Supply-chain attack (agentic)
Compromise of a tool vendor, model provider, or upstream data source that affects the agent — such as a poisoned Model Context Protocol server, a compromised library dependency, or a tampered fine-tuning dataset.
Sustainability Score
The 11th scoring family in the COMPEL scoring engine, measuring the environmental responsibility of AI operations across five dimensions: energy efficiency (25%), carbon footprint management (25%), model efficiency (20%), resource usage (15%), and sustainability reporting compliance (15%).
Sustainability-adjusted value (SAV)
Realized value adjusted for quantified externalities — aligning AI value reporting with CSRD/ESRS disclosure frameworks.
Swarm topology
A multi-agent topology in which peer agents coordinate without central authority — via shared state, stigmergy, or direct peer messaging.
Synchronous kill-switch
A kill-switch variant that takes effect at the agent's next decision point — i.e., the agent completes its current atomic operation, then stops.
Synthetic control
A counterfactual constructed from a weighted combination of untreated donor units — the "synthetic" version of the treated unit.
Synthetic Data
Synthetic data is artificially generated data that mimics the statistical properties of real data but does not contain actual individual records.
System prompt
Operator-authored instructions that set behavior, persona, and boundaries for a model — distinguished from user prompts because the system prompt carries the operator's policy, while the user prompt carries end-user intent.
System prompt leakage
Extraction of an LLM feature's hidden system prompt and structural instructions through crafted user input.
Systems Thinking
Systems thinking is an approach that views AI initiatives not as isolated technology projects but as interventions in a complex organizational system where changes ripple through upstream and downstream workflows, employee roles, customer interactions, data flows, and governance processes.
T
Talent model
The design choice about where AI talent sits, how it flows across the organization, how it is acquired, retained, and developed, and how specialist vs.
Talent Strategy
Talent strategy is the comprehensive plan for building and sustaining the human capabilities needed for AI transformation, encompassing workforce assessment, role definition, hiring, reskilling and upskilling programs, career pathway design, retention mechanisms, and organizational development.
Target operating model
The articulated future-state operating model that a transformation program is designed to deliver — documented as the gap between current-state and target-state along each operating-model dimension.
Task-level decomposition
Breaking a role into its constituent tasks so that each task can be evaluated for AI exposure, automation value, and augmentation potential.
Technical CE Cap
The technical CE cap is the maximum percentage of Continuing Education credits that can be earned from purely technical activities (such as completing partner bootcamps, passing technical assessments, or attending technical conferences) toward a COMPEL certification renewal.
Technical Debt
Technical debt is the accumulated cost of shortcuts, workarounds, and deferred maintenance in technology systems that become increasingly expensive to address over time.
Technical documentation (Annex IV)
The mandatory documentation that the provider of a high-risk AI system must draw up before placing it on the market.
Technical Feasibility
Technical feasibility is an assessment of whether a proposed AI solution can be practically built and deployed given current technology capabilities, data availability, infrastructure, organizational skills, and time constraints.
Telemetry
Telemetry is the automated collection of operational data about AI systems including performance metrics, resource consumption, model behavior, and user interaction patterns providing raw signals for monitoring and alerting.
Telemetry and Monitoring Configuration
The documented specification and implemented configuration of all monitoring, logging, alerting, and observability instrumentation for an AI system — covering model performance metrics, data drift indicators, governance control effectiveness signals, and operational health metrics.
Third-party data readiness
The extension of data-readiness assessment to data supplied by vendors, partners, open-source corpora, or scraped sources — covering provenance, legal basis, contractual terms, known bias profile, and re-use constraints.
Thought Leadership
Thought leadership creates and shares original insights advancing AI transformation through publications, presentations, and community engagement.
Three Lines of Defense
The three lines of defense is a widely adopted risk governance model that distributes risk management responsibilities across three organizational levels: the first line (operational management and AI teams) owns and manages risks directly in their daily work; the second line (risk management and compliance functions) provides oversight, policies, and guidance; and the third line (internal audit) provides independent assurance that the first and second lines are functioning effectively.
Time-to-Value
The elapsed time from a user being provisioned on an AI system to their first recorded value-generating interaction with it, measured at the cohort level.
TOGAF
TOGAF (The Open Group Architecture Framework) is a widely used enterprise architecture methodology that provides a structured approach for designing, planning, implementing, and governing enterprise information technology architecture.
Token
A token is the basic unit of text that a language model processes, roughly corresponding to a word or word fragment (typically 3-4 characters in English).
Token Cost Multiplier
The token cost multiplier is the factor by which AI token consumption increases in multi-agent systems compared to single-model interactions, reflecting the additional tokens consumed by inter-agent communication, extended reasoning chains, tool call sequences, error recovery conversations, and coordination overhead.
Token Economics
Token economics is the analysis, budgeting, and optimization of costs associated with AI language model usage, where pricing is based on the number of tokens (text units, typically representing about four characters) processed as input and generated as output.
Token economics (VDT)
A decomposition of generative-AI inference cost across input tokens, output tokens, context overhead, retrieval tokens, and tool-call tokens — each priced and tracked separately.
Tool Call Authorization
Tool call authorization controls which tools and APIs an AI agent may access, with what parameters and approval requirements.
Tool registry
An authoritative inventory of tools — with schemas, permissions, owners, deprecation state, and audit log — that an agent may call.
Tool schema
The JSON-schema (or equivalent structured) definition of a tool's parameters, types, and constraints.
Tool use / function calling
A prompt pattern authorizing the model to request named functions with structured arguments — searching the web, reading a database, calling a calculator, triggering an API — rather than generating all answers from its weights.
Tool-call validation
A post-execution check that verifies the side effects of a tool call match expectations — with rollback capability where applicable.
Total cost of ownership (TCO, AI)
Build + run + refresh + govern + retire cost across an AI feature's full life.
Total Cost of Ownership (TCO)
Total Cost of Ownership is a comprehensive financial analysis that captures the complete cost of an AI system over its entire lifecycle, including initial development, infrastructure, data acquisition, ongoing maintenance, model retraining, monitoring, governance compliance, talent, vendor fees, and eventual decommissioning.
TPU (Tensor Processing Unit)
A TPU is a custom-designed processor created by Google specifically for neural network workloads, available through Google Cloud Platform.
Training and Adoption Plan
A structured plan that defines the training content, delivery methods, timing, audience segmentation, success criteria, and adoption measurement approach for every user group affected by an AI system deployment.
Training and enablement plan
A designed learning program combining formal training (10%), social learning from peers and mentors (20%), and experiential on-the-job development (70%) — the McCall 70-20-10 ratio applied to AI literacy and capability-building.
Training Data
Training data is the dataset used to teach a machine learning model the patterns it needs to make predictions or generate outputs.
Training data memorization
Verbatim or near-verbatim reproduction of training data by a model during inference.
Transformation charter
A document naming sponsors, scope, decision rights, and governance mechanism for a composite transformation program.
Transformation Crisis
A transformation crisis is a critical event threatening AI program success including executive departure, budget cuts, major failures, or public controversy.
Transformation Enablers
Transformation Enablers are three cross-cutting layers in the COMPEL framework -- Value Realization, Operational Readiness, and Agent Governance -- that operate horizontally across all six lifecycle stages.
Transformation fatigue
The cumulative psychological toll of sustained change across multiple initiatives that reduces employees' receptivity to additional change.
Transformation Portfolio
A transformation portfolio is the collection of AI programs and initiatives managed together to achieve enterprise strategic objectives, balanced across risk, time horizons, pillar coverage, and capability dependencies.
Transformation Roadmap
A transformation roadmap is a strategic, time-sequenced plan that organizes AI transformation initiatives across the four COMPEL pillars, showing what will be done, in what order, with what resources, by when, and with what expected outcomes.
Transformation Sprint
A transformation sprint is a two-week time-boxed work period within the COMPEL Produce stage that delivers concrete outcomes across multiple pillars simultaneously.
Transformer Architecture
The transformer is the neural network architecture that powers modern large language models and many other state-of-the-art AI systems.
Transparency
Transparency in AI governance is the principle that organizations should openly communicate about their use of AI, how AI systems make decisions, what data they use, what their limitations are, and what governance mechanisms are in place.
Transparency duty (Art. 50)
Under Regulation (EU) 2024/1689, specific transparency obligations: providers of AI systems interacting with natural persons must disclose that fact; providers of emotion-recognition or biometric-categorisation systems must inform affected persons; deepfake or AI-generated content intended to inform the public must be marked as synthetic..
Trust & Performance Dimensions
The eight continuous-measurement axes against which every AI transformation is evaluated in COMPEL: Value, Reliability, Safety, Responsibility, Compliance, Security, Sustainability, and Adoption.
Trust Dividend
The trust dividend is the compound return that organizations earn from investing in responsible AI practices, accruing across multiple dimensions of stakeholder relationships.
Trustworthiness Score
A top-level composite score combining the Safety, Responsibility, Security, and Compliance dimensions of the trust-and-performance scorecard into a single trust signal for executive reporting.
Trustworthy AI
Trustworthy AI describes AI systems that are lawful (complying with all applicable regulations), ethical (adhering to moral principles and values), and robust (technically reliable, safe, and secure).
TTFT (time-to-first-token)
The latency from request submission to the first streamed output token.
U
Uncertainty Estimation
Uncertainty estimation encompasses the techniques and methods for quantifying how confident an AI model is in its individual predictions, enabling downstream systems and users to make informed decisions about when to trust AI outputs and when to escalate to human judgment or alternative decision processes.
Uncertainty Quantification
Uncertainty quantification encompasses methods for measuring and communicating how confident an AI model is in its predictions.
Unit economics
Cost and revenue per atomic unit — per transaction, per successful decision, per hour saved — for an AI feature.
Unstructured Data
Unstructured data is data that does not follow a predefined format, including text documents, images, audio recordings, video files, emails, chat transcripts, and social media content.
Unsupervised Learning
Unsupervised learning is a machine learning approach that discovers hidden patterns and structures in data without pre-labeled examples.
Uplift Modeling
Uplift modeling estimates the incremental impact of interventions on individual outcomes, identifying who benefits most versus those unaffected regardless.
Use Case
In AI transformation, a use case is a specific application of AI to a defined business problem with measurable outcomes, identifiable stakeholders, and quantifiable resource requirements.
Use Case Intake
Use case intake is the structured process of collecting, documenting, evaluating, and prioritizing proposed AI use cases from across the organization through a standardized submission and review workflow.
Use Case Portfolio
A use case portfolio is a deliberately balanced collection of AI initiatives designed to achieve strategic outcomes while managing risk across a COMPEL cycle.
Use-Case Portfolio Canvas
A structured prioritization tool that maps candidate AI use cases across value potential, feasibility, risk, and strategic alignment dimensions to produce a ranked, resource-constrained portfolio for the transformation program.
User Acceptance Testing (UAT)
User Acceptance Testing (UAT) is the final validation phase before an AI system goes into production, where actual end users (not the development team) evaluate whether the system meets their operational needs, integrates into their workflows, and produces acceptable results in realistic working conditions.
User prompt
End-user input that consumes the system prompt's behavior — the "request" side of an LLM interaction.
V
Validation Framework
A validation framework verifies AI systems across accuracy, fairness, robustness, security, and explainability, defining methods, metrics, and thresholds before production.
Value Attribution
Value attribution determines how much observed business outcome can be credited to AI transformation versus other factors.
Value Realization
The end-to-end process of defining, tracking, and verifying the business value delivered by AI initiatives — from initial value thesis through baseline measurement, deployment, post-deployment review, and ongoing benefit tracking.
Value Realization Framework
The value realization framework is a structured methodology for defining, measuring, tracking, and verifying business value delivery from AI transformation initiatives throughout the entire COMPEL lifecycle.
Value realization report (VRR)
A stakeholder-facing artefact pairing KPI tree, counterfactual narrative, and risk-adjusted financial summary.
Value Thesis
A value thesis is a testable hypothesis articulating the causal logic connecting an AI initiative to expected business outcomes.
Value Thesis Register
A living register that documents the specific value thesis for each AI use case in the portfolio — including the problem statement, value hypothesis, measurement approach, baseline, and expected outcomes.
Vector Database
A vector database is a specialized database designed to store and efficiently search high-dimensional numerical representations (embeddings) of data.
Vector store
A governed index of embeddings — numeric vector representations of text, image, or multimodal content — that supports similarity search used by retrieval-augmented generation.
Vendor Due Diligence
Vendor due diligence is the structured investigation of an AI vendor's or partner's capabilities, security practices, data handling procedures, compliance posture, financial stability, support quality, and contractual terms before entering a business relationship or deploying their technology.
Vendor Ecosystem
A vendor ecosystem is the network of external technology providers, service partners, cloud platforms, specialized AI tool vendors, and consulting firms that an organization relies on for AI capabilities, infrastructure, and expertise.
Vendor Risk Assessment
A vendor risk assessment evaluates the governance risks introduced by third-party AI components that an organization depends on, including foundation model providers, MLOps platforms, data services, labeling providers, and AI-as-a-service offerings.
Version Control
Version control is the practice of tracking and managing changes to code, data, models, and configuration files over time, maintaining a complete history of what changed, when, who made the change, and why.
Vulnerability Scanning
Vulnerability scanning is the automated process of testing AI systems, supporting infrastructure, and related software for known security weaknesses, misconfigurations, and exploitable flaws.
W
Warm Start
Warm start is a training technique where an AI model begins its learning process using the weights and parameters from a previously trained model rather than starting from random values, significantly reducing training time and computational cost while often improving final performance.
Waterfall
Waterfall is a linear project management approach where phases (requirements, design, implementation, testing, deployment) are completed sequentially from start to finish.
Weight Decay
Weight decay is a regularization technique used during AI model training that adds a penalty term proportional to the magnitude of model weights, discouraging the model from relying too heavily on any single feature and promoting simpler, more generalizable models.
Whistleblower Protection
Whistleblower protection encompasses policies and mechanisms that protect individuals who report AI-related concerns, ethical violations, governance failures, or safety risks from retaliation, career consequences, or social punishment.
Workflow Orchestration
Workflow orchestration is the automated coordination of complex, multi-step processes that involve multiple systems, services, human participants, or AI agents, managing the sequence of steps, parallel execution paths, conditional branching, error handling, retry logic, and completion tracking.
Workflow Redesign
The deliberate restructuring of business processes to embed AI capabilities at optimal human-AI handoff points, rather than overlaying AI onto existing manual workflows.
Workflow Redesign Documentation
Structured documentation of the before and after states of each business process where an AI system is being embedded — covering process maps, human-AI handoff points, role changes, exception handling paths, and the rationale for each design decision.
Workforce AI Capability Transformation
Workforce AI capability transformation is the systematic process of assessing, redesigning, and developing an organization's workforce capabilities to operate effectively in an AI-augmented environment.
Workforce Readiness Plan
A structured plan that defines the AI skills, roles, and capabilities the organization needs to execute its transformation program, assesses current workforce gaps, and specifies the training, hiring, and capability-building interventions required to close those gaps by the time they are needed.
Workforce Redesign
Workforce redesign is the process of analyzing and restructuring jobs at the task level to determine which tasks are best automated by AI, which are augmented by AI, and which remain fully human.
Workforce Strategy
Workforce strategy plans how human resources evolve through assessment, reskilling, hiring, restructuring, and retention for AI transformation.
Workforce Transformation
Workforce transformation is the strategic process of developing new skills, redesigning roles, restructuring teams, and evolving organizational culture to enable effective human-AI collaboration across the enterprise.
Works-council engagement
A formal consultation process required in many jurisdictions for AI-driven role changes — most prominently in Germany (Betriebsrat), France, and across EU member states.
X
X-Risk (Existential Risk from AI)
X-risk refers to the theoretical existential risk that sufficiently advanced AI systems could pose catastrophic or irreversible harm to humanity or civilization.
XAI (Explainable Artificial Intelligence)
XAI is the abbreviated term for Explainable Artificial Intelligence, the field focused on making AI systems' reasoning and decision-making processes transparent and interpretable to humans.
XAI Techniques
XAI techniques are specific methods making AI decisions interpretable including SHAP values, LIME, attention visualization, feature importance, and counterfactual explanations.
XGBoost
XGBoost (eXtreme Gradient Boosting) is a highly efficient and widely used machine learning algorithm that builds predictions by combining many small decision trees in sequence, with each tree learning from the errors of the previous ones.
XML (Extensible Markup Language)
XML is a structured data format widely used for storing, transmitting, and exchanging data between different software systems in a human-readable and machine-parseable format.
XML and Data Interchange Standards
XML and related formats provide standardized ways to structure, share, and validate data between systems and organizations.
Y
YAML
YAML (YAML Ain't Markup Language) is a human-readable data serialization format commonly used for configuration files in AI/ML pipelines, infrastructure-as-code definitions, CI/CD pipeline specifications, and deployment configurations.
YAML Configuration
YAML (YAML Ain't Markup Language) is a human-readable data serialization format commonly used to define configuration settings, pipeline specifications, infrastructure definitions, and deployment parameters for AI systems and their supporting infrastructure.
Year-over-Year (YoY) Maturity Progression
Year-over-year maturity progression measures how an organization's AI maturity scores change across annual assessment cycles.
Year-over-Year Metrics
Year-over-year (YoY) metrics compare performance data from the same period in consecutive years, providing a normalized view of long-term AI transformation progress that accounts for seasonal variations, cyclical patterns, and short-term fluctuations.
Yield Management
Yield management dynamically adjusts resource allocation based on demand patterns to maximize value from scarce AI infrastructure capacity, applying to GPU utilization, inference versus training balance, and workload scheduling.
Yield Optimization
Yield optimization uses AI to maximize the output, efficiency, or return from a process -- such as manufacturing yield (reducing waste and defects), agricultural yield (optimizing crop production), advertising yield (maximizing revenue per impression), or financial yield (optimizing investment returns).
Z
Z-Score
A z-score is a statistical measurement describing how many standard deviations a data point is from the mean (average) of its dataset.
Zero-Day Vulnerability
A zero-day vulnerability is a software security flaw that is unknown to the software vendor and therefore has no available patch or fix at the time of discovery.
Zero-Shot Learning
Zero-shot learning is the ability of an AI model to perform tasks it was not explicitly trained or fine-tuned to do, leveraging general knowledge and reasoning capabilities acquired during pre-training.
Zero-shot prompting
A prompt pattern in which the model receives only an instruction — no labeled examples of the desired input-output behavior.
Zero-Trust Architecture
Zero-trust architecture is a security framework built on the principle that no user, device, system, or AI agent should be trusted by default, regardless of whether they are inside or outside the network perimeter.
Zone of Proximal Development
The zone of proximal development (ZPD), originally theorized by Lev Vygotsky, describes the gap between what a learner can accomplish independently and what they can achieve with appropriate guidance and support.