top of page

Strategic Realignment of European AI Governance: Analysis of AI Omnibus Proposal and Impact on EU AI Act

  • Writer: Nelson Advisors
    Nelson Advisors
  • 4 hours ago
  • 15 min read
Strategic Realignment of European AI Governance: Analysis of AI Omnibus Proposal and Impact on EU AI Act
Strategic Realignment of European AI Governance: Analysis of AI Omnibus Proposal and Impact on EU AI Act

The European Union's approach to digital sovereignty and the regulation of emerging technologies has reached a critical juncture, characterised by a fundamental shift from a purely precautionary regulatory stance to one that emphasises industrial competitiveness and operational feasibility. This evolution is most visibly manifested in the "AI Omnibus" proposal, formally unveiled by the European Commission on November 19, 2025, as a pivotal component of the broader Digital Simplification Package.


The proposal seeks to address a burgeoning crisis in the implementation of the EU AI Act (Regulation (EU) 2024/1689), which, despite entering into force on August 1, 2024, has faced significant institutional and technical hurdles that threaten its viability as a global standard. Central to this legislative intervention is a "stop-the-clock" mechanism designed to link the application of high-risk AI obligations to the actual availability of technical standards and guidance, effectively pushing back major compliance deadlines by 12 to 16 months.


This analysis explores the socio-economic drivers, legal mechanisms, institutional shifts and stakeholder contestations that define this historic recalibration of the European digital rulebook.


The Socio-Economic Catalyst: Competitiveness and the Draghi Mandate


The genesis of the AI Omnibus cannot be understood in isolation from the broader economic anxieties currently permeating the European corridors of power. Throughout late 2024 and 2025, a series of strategic reports, most notably the analysis by former European Central Bank President Mario Draghi on the future of European competitiveness—warned that the complexity, fragmentation, and cumulative burden of the EU’s digital acquis were stifling innovation and deterring investment. Draghi’s report highlighted a critical disconnect between the EU’s ambition to be a global rule-setter and the practical capacity of its domestic industry to absorb and comply with these rules without losing ground to global competitors in the United States and China.


The Commission's Digital Simplification Package, which includes the AI Omnibus, is a direct response to these concerns, aiming to streamline the regulatory landscape across data, cybersecurity and artificial intelligence. The overarching objective is to reduce administrative costs for businesses, estimated to save up to €5 Billion by 2029, while ensuring that the EU’s high standards for fundamental rights and safety are not compromised. This tension between "simplification" and "deregulation" forms the core of the current legislative debate, as policymakers attempt to craft a framework that is both "innovation-friendly" and "trustworthy".


The Institutional Architecture of the Digital Omnibus Package

The simplification agenda is structured as a dual legislative track, designed to provide immediate relief while paving the way for a more comprehensive "Digital Fitness Check" in 2027. This architectural choice allows the Commission to isolate the urgent timing issues of the AI Act from the more extensive task of harmonizing the EU’s broader data economy rules.

Legislative Instrument

Primary Focus

Key Acts Amended

Digital Omnibus Regulation

Streamlining data and cybersecurity rules

GDPR, ePrivacy Directive, Data Act, NIS2 Directive, CER Directive

Digital Omnibus on AI Regulation

Targeted adjustments to the AI Act

Regulation (EU) 2024/1689 (AI Act), Regulation (EU) 2018/1139 (Civil Aviation)

This bifurcated approach reflects the unique status of the AI Act as a "living" regulation that requires rapid technical support in the form of harmonised standards and delegated acts. The Digital Omnibus on AI is essentially a corrective measure to ensure that the AI Act’s "high-risk" regime does not collapse under the weight of its own deadlines before the necessary technical infrastructure is in place.


The Temporal Pivot: Deconstructing the "Stop-the-Clock" Mechanism


The most impactful element of the AI Omnibus is the radical overhaul of the implementation timeline for high-risk AI systems. Under the original 2024 text of the AI Act, the majority of obligations for high-risk systems were scheduled to become applicable on August 2, 2026. However, as that date approached, it became increasingly clear that neither the regulators nor the industry were ready.


The "stop-the-clock" mechanism introduced in the Omnibus proposal fundamentally alters the logic of compliance.Instead of an arbitrary fixed date, the application of Chapter III requirements is now linked to a Commission decision confirming that "compliance support tools", specifically harmonised technical standards, common specifications, or formal guidelines, are officially available. This shift acknowledges that without technical standards, companies face immense legal uncertainty, as they cannot verify if their risk management or data governance systems meet the "essential requirements" of the law.


Revised Timelines for High-Risk Systems

The proposal establishes a staggered application window that distinguishes between standalone AI applications (Annex III) and AI embedded as safety components in regulated products (Annex I).

High-Risk Category

Trigger for Application

Proposed Deadline (Long-stop Date)

Original AI Act Deadline

Annex III Systems (e.g., Biometrics, Education, Law Enforcement)

6 months after Commission readiness decision

2 December 2027

2 August 2026

Annex I Systems (e.g., Medical Devices, Industrial Machinery)

12 months after Commission readiness decision

2 August 2028

2 August 2027

The implications of this temporal realignment are profound. For Annex I systems, the extension effectively grants a 12-month reprieve, while Annex III systems receive an additional 16 months of preparation time. This "breathing room" is intended to allow for a more robust and high-quality implementation, avoiding the "tick-box" compliance exercises that often result from rushed deadlines.


However, critics argue that this delay leaves individuals exposed to the risks of unregulated high-risk AI systems for a significantly longer period, potentially undermining the protective intent of the original legislation.


Technical Readiness and the Standardisation Crisis


The necessity of the AI Omnibus is primarily driven by a "standardization crisis" within the European technical infrastructure. The AI Act relies on "harmonised standards" to provide the technical detail missing from its high-level legal principles. These standards are developed by European Standardisation Organisations (ESOs), specifically CEN and CENELEC, under a formal mandate from the European Commission.


As of early 2026, the Joint Technical Committee 21 (JTC 21), responsible for AI standards, has faced significant delays.The process of reaching consensus among hundreds of volunteer experts from diverse national and commercial backgrounds has proven more complex than anticipated.


The Six-Step Standardisation Process and Current Bottlenecks

The complexity of the European standardisation model contributes directly to the implementation delays addressed by the Omnibus.

Stage

Process Description

Status for AI Act Standards

1. Request

Commission issues formal standardisation request

Completed in late 2023

2. Drafting

Technical experts in JTC 21 draft the specifications

Ongoing; significant delays in risk management and data quality

3. Enquiry

Public review and voting by national stakeholders

Initial enquiries faced heavy negative feedback, triggering process resets

4. Formal Vote

National bodies formally approve the final text

Delayed; many standards not expected until late 2026

5. Publication

ESOs publish the approved standard

Pending

6. Citation

Commission cites standards in the Official Journal

Triggers the compliance clock under the AI Omnibus

The IAPP reported that the Commission missed its own February 2, 2026, deadline to provide critical guidance on the classification of high-risk systems under Article 6. Furthermore, CEN-CENELEC officials have signaled that a complete suite of standards will likely not be ready before December 2026 at the earliest. Without these technical blueprints, the "high-risk" obligations of the AI Act are essentially unenforceable in a way that provides legal certainty for businesses.


Redefining Regulatory Scope: The Inclusion of Small Mid-Caps (SMCs)


A secondary but significant objective of the AI Omnibus is the expansion of regulatory relief to a new category of economic operators: the "Small Mid-Cap" (SMC) company. The original AI Act recognised that SMEs and startups faced disproportionate compliance costs and provided them with certain privileges, such as simplified Quality Management Systems (QMS) and lower fines. The Omnibus recognises that these challenges also affect larger, but still relatively modest, firms that form the backbone of the European industrial "Mittelstand".


Comparative Definitions of SME and SMC under the Omnibus


The proposal introduces formal definitions for SMCs, aligning them with existing EU economic classifications while granting them access to the AI Act's "innovation enablers".

Entity Category

Maximum Employee Count

Maximum Annual Turnover

Relief Measures under AI Omnibus

SME

< 250

< €50 Million

Simplified documentation, lower fines, priority sandbox access

Small Mid-Cap (SMC)

< 750

< €150 Million

Facilitated procedures, simplified QMS, proportional penalty calculation

This expansion has been welcomed by industry associations as a pragmatic move to support European scaling. However, consumer groups such as BEUC have criticised this change, arguing that it undermines the "risk-based" logic of the AI Act. Their concern is that an AI system’s risk to fundamental rights is determined by its application (eg. credit scoring or biometric identification), not by the size of the company deploying it. By granting relief to firms with up to 750 employees, critics argue that the EU is exempting significant market players from the full rigors of safety testing.


Algorithmic Fairness vs. Privacy: The Bias Mitigation Paradox


One of the most technically challenging aspects of AI development is the mitigation of algorithmic bias. To detect and correct bias, developers often need to "see" the very sensitive data (eg, race, religion, health status) that they are trying not to discriminate against. The GDPR, however, generally prohibits the processing of such "special categories" of personal data.


The AI Omnibus seeks to resolve this paradox by introducing a new Article 4a to the AI Act. This provision creates a specific legal basis for providers and deployers to process special category data for the sole purpose of bias detection and correction. Crucially, the Omnibus proposes to lower the threshold for this processing from "strictly necessary" to "necessary," while broadening its scope beyond high-risk systems to cover all AI systems.


Safeguards for Bias-Related Data Processing


While the Omnibus facilitates this data use, it maintains a layer of protection designed to prevent mission creep.


  • Necessity Requirement: The developer must demonstrate that bias detection cannot be performed using non-sensitive or synthetic data.


  • Technical Minimisation: Use of state-of-the-art security measures, such as differential privacy or pseudonymisation, is required to prevent the identification of individuals.


  • Purpose Limitation: The data collected under Article 4a cannot be reused for other purposes, such as model training for performance or marketing.


The pharmaceutical and MedTech sectors have identified this amendment as particularly vital, as it provides a clearer legal framework for ensuring that medical AI models perform equally well across diverse patient populations. Conversely, the EDPB and EDPS have expressed "significant concerns," warning that this could lead to the normalization of large-scale sensitive data collection under the guise of fairness.


The Governance Evolution: Strengthening the AI Office


The AI Omnibus signifies a major institutional shift toward centralized enforcement, primarily by expanding the mandate of the European AI Office. Originally envisioned as a coordinating body, the AI Office is increasingly taking on the characteristics of a primary "market surveillance authority" for the most advanced AI systems.


This centralization is intended to address the "fragmentation" of enforcement, where different national authorities might interpret the AI Act in divergent ways, creating obstacles for cross-border operations. The Omnibus grants the AI Office exclusive competence over AI systems based on General-Purpose AI (GPAI) models in cases where the same entity provides both the model and the system. Furthermore, the AI Office will oversee AI systems integrated into Very Large Online Platforms (VLOPs), aligning its work with the Digital Services Act (DSA).


New Tools for Innovation and Oversight


The proposal also introduces new operational tools for the AI Office to foster a more "pro-innovation" environment while maintaining oversight.


  • EU-Level AI Regulatory Sandboxes: While the original Act mandated national sandboxes, the Omnibus allows the AI Office to establish EU-wide sandboxes for GPAI-based systems. This provides a single point of entry for developers operating across multiple Member States.


  • Real-World Testing Agreements: The scope for testing high-risk systems in real-world conditions—outside of laboratory environments—is expanded, allowing sectors like transport and healthcare to validate AI performance in actual operational settings.


  • Streamlined Notified Body Procedures: To address the shortage of conformity assessment bodies, the Omnibus introduces a "single application" process for designation, allowing these bodies to operate across the EU with less administrative repetition.


The centralisation of power in the AI Office is a point of contention with Member States, who are traditionally protective of their national market surveillance prerogatives. The final negotiations will likely center on finding a balance between the efficiency of Brussels-led enforcement and the importance of national expertise and proximity to local markets.


The Reversal of AI Literacy Obligations


One of the most notable "simplification" measures in the Omnibus is the revision of Article 4, which deals with AI literacy. In the original AI Act, providers and deployers were legally mandated to take measures to ensure that their staff attained a "sufficient level of AI literacy". This was seen by industry as a vague and potentially expensive open-ended obligation.


The Omnibus proposes to "re-direct" this responsibility. The mandatory requirement for companies is replaced by a duty for the European Commission and Member States to "encourage" and support AI literacy through training opportunities and informational resources.


While this provides immediate relief for HR and compliance departments, civil society groups warn that it weakens the first line of defence against AI harms: the human operator. The EDPB has noted that without literate staff, the "human-in-the-loop" requirement of the AI Act becomes effectively hollow.


Stakeholder Contestation: Convergence and Conflict


The AI Omnibus has polarised the European digital community, with the dividing lines drawn between those who prioritise economic "productivity" and those who prioritise "rights-based" protection.


Industry Perspectives: The Quest for Legal Certainty


Industry associations, including DigitalEurope, CCIA Europe, and the Technology Industries of Finland (TIF), have been the primary advocates for a delay. Their argument is that the "compliance cliff" of August 2026 is a threat to the EU’s industrial stability. They have broadly welcomed the Omnibus but remain wary of the "dual-trigger" mechanism.


TIF and other groups have recommended that the timeline-related amendments should be "fast-tracked" and separated from the broader, more controversial substantive changes to the AI Act. They fear that if the entire Omnibus package becomes bogged down in political negotiations, the August 2026 deadline will arrive before the extension is legally finalized, leaving businesses in a state of maximum uncertainty.


Civil Society and Regulators: The Warning of Fundamental Rights Rollback

Consumer advocates (BEUC), digital rights NGOs (Access Now, Amnesty Tech), and the EU’s data protection supervisors (EDPB/EDPS) have been sharply critical of several aspects of the proposal. They view the Omnibus not as a "targeted simplification" but as a "reopening" of the fragile political compromise reached during the AI Act's trilogues.

Stakeholder Concern

Argument against Omnibus Changes

Deletion of Registration

Eliminates public transparency and hinders collective redress

SMC Extensions

Exempts significant actors from safety duties based on size rather than risk

Bias Data Processing

Normalizes sensitive data collection and undermines data minimization

Moving Deadlines

Leaves consumers unprotected from high-risk AI for 12-16 additional months

The EDPB and EDPS issued a "Joint Opinion 1/2026" expressing "sincere concerns" about the potential impact on fundamental rights. They argued that while administrative simplification is welcome, it must not lead to a dilution of the core protections that make the EU AI Act a global beacon for ethical technology.


The Parliamentary Scrutiny: The Kokalari-McNamara Draft Report


The legislative fate of the AI Omnibus currently rests with the European Parliament, where the lead committees (IMCO and LIBE) published their draft report on February 5, 2026. The rapporteurs, Arba Kokalari (EPP) and Michael McNamara (Renew), have signaled that while they support the goal of simplification, they disagree with the Commission’s method of implementation.


The Parliament’s draft report proposes a "philosophy of legal certainty" over "discretionary flexibility". The most significant amendment is the replacement of the Commission's flexible "stop-the-clock" mechanism with fixed application dates.

Feature

Commission Proposal (Nov 19, 2025)

Parliament Draft Report (Feb 5, 2026)

Annex III Compliance Date

Linked to Commission readiness decision

Fixed: 2 December 2027

Annex I Compliance Date

Linked to Commission readiness decision

Fixed: 2 August 2028

AI Literacy

Non-binding encouragement

Reinstated as binding obligation

Article 6 Guidance

Flexible delivery

Stricter deadlines for Commission guidelines

This move toward fixed dates is intended to give businesses a clear "North Star" for their compliance programs, removing the uncertainty of waiting for a Commission decision that could be triggered at any moment. Furthermore, the Parliament’s draft report seeks to reinstate the mandatory AI literacy requirement, reflecting the concerns of the EDPB and civil society.


Broader Implications: The GDPR and Cybersecurity Harmonisation


While the AI-specific amendments attract the most headlines, the broader Digital Omnibus package proposes significant changes to the GDPR and the EU’s cybersecurity reporting framework. These changes are intended to address "compliance fatigue" and the proliferation of redundant reporting obligations.


The 96-Hour Rule and Single Entry Point


The proposal addresses the grueling 72-hour reporting window for data breaches under the GDPR, which often forces security teams to file incomplete reports just to meet the deadline.


  • Extended Reporting Window: The Omnibus proposes extending the breach notification deadline from 72 to 96 hours (4 days).


  • Single Entry Point (SEP): Managed by ENISA, a new centralized portal would allow companies to "report once, share many". An incident report submitted to the SEP would automatically satisfy notification requirements under the GDPR, NIS2, DORA, and the CER Directive.


  • Revised Definition of Personal Data: The Omnibus proposes a "relative" definition of personal data. Information would not be considered personal data for a specific entity if that entity has no "reasonable way" to identify the individual, even if another entity could. This aims to provide relief for entities handling pseudonymized data for research or AI training.


These measures are designed to "filter out the noise," allowing security and forensic teams to focus on high-impact threats rather than administrative paperwork. However, data protection authorities warn that these changes could erode the high level of individual protection that the GDPR was designed to provide.


Sectoral Case Study: Life Sciences and MedTech


The intersection of the AI Act and the Medical Device Regulation (MDR) is one of the most complex areas of digital law.For pharmaceutical and MedTech companies, the AI Omnibus offers both tactical relief and strategic clarity.


Integration with Existing Conformity Assessments


The Omnibus confirms that for AI-enabled medical devices, the AI Act’s requirements should be applied within the existing conformity assessment procedures of the MDR and IVDR. This prevents the "dual certification" problem, where a company would have to go to one body for medical safety and another for AI safety.


The proposal also introduces a "grandfathering" clause for legacy systems. If at least one unit of an AI system has been lawfully placed on the market before the relevant compliance date, additional units of that same model can continue to be sold without a new assessment, provided the design remains unchanged. This provides essential stability for long-cycle industrial and medical products.


The Impact of Real-World Testing


For Life Sciences, the expansion of real-world testing opportunities is perhaps the most significant operational change. By allowing AI models to be validated in clinically relevant settings before full market deployment, the Omnibus facilitates a more iterative and safety-conscious development process. This is expected to be a major driver for the adoption of AI in personalized medicine and surgical robotics.


The Path Forward: Negotiations and Global Impact


The AI Omnibus has entered a period of "heightened uncertainty" as the European Parliament, the Council, and the Commission prepare for trilogue negotiations in late spring 2026. The pressure to finalize the file by August 2, 2026, is immense, as failure to do so could result in a "legal vacuum" where the original high-risk rules kick in without any support infrastructure.


Strategic Outlook: The "Brussels Effect" in Flux


The AI Omnibus represents the EU’s attempt to manage the "Brussels Effect"—its ability to set global standards—by making those standards more "practical" and "innovation-friendly". If the EU succeeds in streamlining its AI Act without losing its ethical core, it could solidify its position as the global model for technology governance. However, if the Omnibus is perceived as a significant retreat from safety and rights, it may embolden other jurisdictions to pursue even more deregulatory approaches, potentially leading to a "race to the bottom" in global AI safety.


Future Scenario

Likely Outcome

Impact on European Industry

Smooth Adoption (July 2026)

Omnibus passed with fixed deadlines and restored literacy duties

High legal certainty; manageable transition periods

Negotiation Deadlock (Aug 2026)

Original high-risk rules apply without the extension

High legal risk; potential "compliance cliff" and investment pause

Divergent National Implementation

Member States implement their own temporary rules or pauses

Market fragmentation; high compliance costs for cross-border firms

Nuanced Conclusions and Actionable Analysis


The AI Omnibus proposal is not merely an administrative delay; it is a fundamental recalibration of the European Union's digital strategy. By acknowledging the technical and institutional unreadiness for the original AI Act timelines, the Commission has chosen a path of operational realism that prioritizes the long-term success of the regulation over short-term political posturing.


For professional peers in the regulatory and compliance domains, several key takeaways emerge from this analysis:


The transition from the original "fixed" deadlines to the proposed "flexible" or "long-stop" dates creates a period of strategic ambiguity. Organizations should not treat the extension as a "permission slip" to delay their AI governance programs. Instead, the additional 12-16 months should be used to move beyond "check-box" compliance and toward "safety engineering" by design.


The reinforcement of the AI Office and the introduction of EU-level sandboxes signal a move away from fragmented national oversight and toward a more centralized, expert-led enforcement model. Firms should engage proactively with the AI Office and monitor the development of the Transparency Code of Practice, as these will likely become the primary mechanisms for day-to-day compliance.


The proposed amendments to the GDPR and the introduction of Article 4a in the AI Act represent a critical "safety valve" for the AI industry. The ability to process sensitive data for bias mitigation is a technical necessity that has finally found a legal home, but its use will be subject to intense scrutiny from data protection supervisors. Developers must document their "balancing tests" and necessity assessments with extreme rigor to withstand future audits.


As the EU moves toward final adoption in 2026, the AI Omnibus will be remembered as the moment the "Brussels Effect" met the "Draghi Realism". Whether this produces a more competitive and innovative Europe—or a more vulnerable one—will depend on the final text’s ability to preserve the fragile equilibrium between technological power and human rights.For now, the "stop-the-clock" proposal remains the most vital insurance policy for the future of European artificial intelligence.


Nelson Advisors > European MedTech and HealthTech Investment Banking

 

Nelson Advisors specialise in Mergers and Acquisitions, Partnerships and Investments for Digital Health, HealthTech, Health IT, Consumer HealthTech, Healthcare Cybersecurity, Healthcare AI companies. www.nelsonadvisors.co.uk


Nelson Advisors regularly publish Thought Leadership articles covering market insights, trends, analysis & predictions @ https://www.healthcare.digital 

 

Nelson Advisors publish Europe’s leading HealthTech and MedTech M&A Newsletter every week, subscribe today! https://lnkd.in/e5hTp_xb 

 

Nelson Advisors pride ourselves on our DNA as ‘Founders advising Founders.’ We partner with entrepreneurs, boards and investors to maximise shareholder value and investment returns. www.nelsonadvisors.co.uk



Nelson Advisors LLP

 

Hale House, 76-78 Portland Place, Marylebone, London, W1B 1NT




Nelson Advisors specialise in Mergers and Acquisitions, Partnerships and Investments for Digital Health, HealthTech, Health IT, Consumer HealthTech, Healthcare Cybersecurity, Healthcare AI companies. www.nelsonadvisors.co.uk


Comments


Commenting on this post isn't available anymore. Contact the site owner for more info.
bottom of page