We use cookies to improve your experience and analyse site traffic.
Organisations operating multiple high-risk AI systems require portfolio-level governance orchestration. Under the EU AI Act, each system's governance pipeline must respond to changes in shared components, and evidence records must remain cleanly separated across concurrent deployments.
Organisations operating multiple high-risk AI systems face a coordination challenge that single-system guidance does not address.
Organisations operating multiple high-risk AI systems face a coordination challenge that single-system guidance does not address. Each system maintains its own governance pipeline, its own gate thresholds, and its own aisdp. Changes to shared infrastructure, shared data sources, shared model components, or shared governance policies can affect multiple systems simultaneously. Without portfolio-level orchestration, a change to one shared component could alter the compliance posture of several systems while only one undergoes re-evaluation.
The governance pipeline at portfolio level must handle three coordination patterns: shared dependency propagation, concurrent deployment coordination, and portfolio-wide compliance visibility. These patterns ensure that the obligations under Governance Pipeline: CI/CD for Regulatory Compliance are discharged consistently across the full portfolio rather than system by system.
When a shared component changes, every system that depends on that component must re-evaluate its governance gates.
When a shared component changes, every system that depends on that component must re-evaluate its governance gates. Shared components include common feature libraries, shared embedding models, and shared data sources. The AI Governance Lead maintains a dependency map linking shared components to the systems that consume them.
When a shared component change is detected, the governance pipeline triggers re-evaluation for all dependent systems automatically. Without this propagation mechanism, a change to a shared embedding model could alter the fairness profile of five systems while only one is re-evaluated. The dependency map is the single source of truth for determining which systems require re-assessment following any shared component update.
When multiple systems deploy changes simultaneously, the governance pipeline must ensure that each system's deployment is evaluated independently and that evidence records are not commingled.
When multiple systems deploy changes simultaneously, the governance pipeline must ensure that each system's deployment is evaluated independently and that evidence records are not commingled. The artefact registry uses a composite key combining the system identifier with the pipeline execution identifier to maintain separation.
This separation is essential for traceability. A competent authority reviewing the evidence trail for one system must not encounter artefacts generated by another system's pipeline execution. The composite key approach ensures that even during periods of high deployment activity across the portfolio, each system's governance evidence remains cleanly partitioned within the artefact registry.
The AI Governance Lead requires visibility across all systems in the portfolio.
The AI Governance Lead requires visibility across all systems in the portfolio. The dashboard aggregates information about which systems have open non-conformities, which are approaching certification expiry, which have pending governance gate failures, and which have documentation currency gaps.
The governance pipeline's audit persistence layer feeds this portfolio dashboard. It serves as both an operational tool for the AI Governance Lead and a reporting input for executive oversight. The dashboard transforms raw pipeline telemetry into actionable compliance intelligence, enabling the governance team to prioritise remediation effort across the portfolio based on risk severity and regulatory deadline proximity.
The dependency map should be maintained as a machine-readable configuration file, version-controlled alongside the GOVERNANCE PIPELINE definitions.
The dependency map should be maintained as a machine-readable configuration file, version-controlled alongside the governance pipeline definitions. The map is structured as a directed graph where nodes represent systems and shared components, and edges represent dependency relationships. When a shared component's pipeline executes successfully, the orchestration layer reads the dependency map and triggers the governance pipeline for each dependent system.
Tools such as Backstage, Spotify's open-source service catalogue, or a custom registry built on a graph database provide the dependency tracking infrastructure. The Technical SME maintains the map on an ongoing basis. The AI Governance Lead reviews it quarterly to confirm that all dependencies are captured and that no system operates outside the portfolio governance framework. This review cadence ensures that new systems or new shared components are promptly integrated into the dependency graph.
The governance pipeline itself must be monitored, because a pipeline that silently fails, skips a gate due to misconfiguration, or stops producing evidence artefacts creates a compliance gap that may not be detected until the next CONFORMITY ASSESSMENT.
The governance pipeline itself must be monitored, because a pipeline that silently fails, skips a gate due to misconfiguration, or stops producing evidence artefacts creates a compliance gap that may not be detected until the next conformity assessment. The engineering team monitors four categories of pipeline health.
First, gate execution rate: every pipeline execution should produce a record for every applicable gate, and a missing gate record indicates a pipeline misconfiguration or failure. Second, gate failure rate over time: a rising failure rate for a specific gate may indicate a systemic issue such as data quality degradation, fairness drift, or threshold miscalibration rather than isolated incidents. Third, evidence generation completeness: every pipeline execution should deposit a defined set of artefacts in the governance artefact registry, and missing artefacts trigger an alert. Fourth, approval latency: the time between a pipeline reaching the deployment authorisation gate and the approver's decision, where excessive latency indicates a governance bottleneck that may create pressure to bypass the gate.
The AI Governance Lead receives alerts for gate execution failures, evidence generation gaps, approval latency exceeding the defined threshold, and any manual override of a governance control. The alerting mechanism is separate from engineering monitoring alerts and is configured to reach the governance team directly, ensuring that Monitoring and Anomaly Detection concerns are surfaced to the appropriate audience.
The governance pipeline itself is a compliance artefact that must be documented within the AISDP.
The governance pipeline itself is a compliance artefact that must be documented within the AISDP. This documentation feeds primarily into AISDP Module 10 covering version control and change management, and Module 2 covering model selection and architecture, with cross-references to every module that receives evidence from the pipeline.
The Technical SME documents five areas. Pipeline architecture covers the orchestration tool, the policy enforcement engine, the evidence storage backend, the AISDP synchronisation mechanism, and the audit persistence layer. Gate definitions specify, for each governance gate, the trigger conditions, evaluation criteria, pass and fail thresholds, evidence produced, and approval authority, all traceable to the regulatory requirements they implement. Policy rules comprise the complete set of policy-as-code rules enforced by the policy engine, version-controlled and referenced by commit hash, with the Legal and Regulatory Advisor confirming they correctly implement regulatory requirements. Change classification criteria cover the classification rules, the substantial modification indicators, and the approval authority matrix, confirmed by the Legal and Regulatory Advisor against the Article 3(23) definition. Evidence lifecycle specifies the retention period, immutability mechanism, access control policy, and disaster recovery provisions for the governance artefact registry.
A competent authority reviewing the AISDP should be able to understand, from the governance pipeline documentation alone, how the organisation ensures that every change to the system is evaluated, evidenced, approved, and recorded. The documentation must demonstrate that the pipeline is not merely designed to enforce compliance but is operationally effective, monitored for failures, and maintained as a living system alongside the AI system it governs. The obligations under Article 11, Article 9, Article 17, and Article 12 are simultaneously discharged through this pipeline infrastructure, and the documentation must make that discharge visible and verifiable.
Tools such as Backstage (Spotify's open-source service catalogue) or a custom registry built on a graph database provide the dependency tracking infrastructure. The dependency map should be maintained as a machine-readable configuration file, version-controlled alongside governance pipeline definitions.
The Technical SME maintains the dependency map on an ongoing basis, while the AI Governance Lead reviews it quarterly to confirm that all dependencies are captured and no system operates outside the portfolio governance framework.
The AI Governance Lead should receive alerts for gate execution failures, evidence generation gaps, approval latency exceeding defined thresholds, and any manual override of a governance control. These alerts should be separate from engineering monitoring and reach the governance team directly.
Four categories: gate execution rate, gate failure rate trends, evidence generation completeness, and approval latency at the deployment authorisation gate.
Document pipeline architecture, gate definitions, policy-as-code rules, change classification criteria, and evidence lifecycle provisions, all traceable to regulatory requirements.