X Close Search

How can we assist?

Demo Request

EU AI Act Implementation: Five Critical Steps Boards Must Take in 2025

Healthcare boards must prepare for the EU AI Act by mapping AI systems, enhancing security, and ensuring compliance by 2025.

Healthcare organizations must act now to meet the EU AI Act's 2025 compliance deadlines. This regulation, effective as of July 2024, classifies AI systems by risk and mandates strict oversight for high-risk applications, especially in healthcare. Non-compliance could result in fines of up to €35 million or 7% of annual turnover.

Key Actions Boards Must Take:

  1. Map and Rate AI Systems: Create an inventory of all AI tools, assess their risks, and prioritize high-risk systems like diagnostic tools and patient care AI.
  2. Strengthen Security: Implement AI-specific cybersecurity measures using standards like ISO/IEC 42001:2023 and NIST AI RMF.
  3. Develop AI Expertise: Train board members on AI basics, regulatory compliance, and security frameworks.
  4. Ensure Documentation: Maintain detailed records, including system specifications, risk assessments, and compliance actions.
  5. Meet Compliance Deadlines: Discontinue prohibited AI by January 2025 and ensure all high-risk systems comply by July 2025.

Quick Overview:

Focus Area Action Deadline
AI System Inventory Catalog and classify all AI systems Q2 2025
Security Controls Implement automated monitoring and risk tools Q3 2025
Board Governance Train and establish oversight processes Q4 2025

By following these steps, healthcare boards can ensure compliance, protect patient data, and avoid costly penalties.

EU AI Act

EU AI Act: Healthcare Requirements

This section outlines the healthcare-specific requirements under the EU AI Act, focusing on patient safety and compliance for AI systems in medical settings.

The EU AI Act takes a risk-based approach to regulate healthcare AI, aiming to protect patients while encouraging advancements in medical technology.

Risk Categories in the EU AI Act

The regulation classifies AI systems into three risk levels:

Risk Level Description Healthcare Examples
Unacceptable Risk AI practices banned due to severe threats to safety or rights AI systems manipulating treatment decisions subliminally
High Risk Systems with significant health impacts, requiring strict oversight Diagnostic tools and AI controlling medical devices
General Purpose AI Flexible systems with compliance needs depending on their specific use case Language models for documentation and administrative AI tools

AI systems directly affecting patient care are almost always categorized as high risk, requiring strict adherence to compliance standards. These classifications inform the Act’s compliance timelines.

2025 Compliance Deadlines

Following the Act’s publication in July 2024, healthcare organizations must meet specific deadlines:

  • January 2025: Discontinue use of prohibited AI systems
  • July 2025: Ensure compliance for high-risk systems
  • July 2025: Complete technical documentation and register applicable systems

This regulation will impact many AI-enabled devices, particularly in radiology, where about 75% of such devices are in use [1].

Healthcare-Specific Rules

Healthcare boards are responsible for implementing robust measures to comply with the Act. These include:

  • Risk Management Systems: Ongoing monitoring of AI performance in clinical environments
  • Data Quality Controls: Strict oversight of training and validation datasets
  • Technical Documentation: Comprehensive records detailing system design and deployment
  • Human Oversight: Clear protocols for professional supervision of AI systems
  • Quality Management: Policies to ensure consistent and reliable AI performance

The Act also provides exemptions for medical uses of facial recognition and emotion detection technologies, acknowledging their diagnostic value. However, these systems must still meet documentation and safety requirements [1].

Step 1: Map and Rate AI Systems

Boards need to catalog AI systems to meet EU AI Act requirements. This inventory is the starting point for accurate risk evaluations and tailored compliance strategies.

Create an AI System Inventory

List all current and planned AI systems, including:

  • Clinical diagnostic tools and medical imaging AI
  • Patient monitoring systems
  • Administrative automation tools
  • Resource management AI systems
  • Clinical decision support systems

Automated tools can help identify and document AI systems within your organization’s tech infrastructure.

Evaluate Risk Levels

Consider these factors when assessing risk:

Risk Factor Assessment Criteria Documentation Requirements
Patient Impact Direct effect on clinical decisions Clinical validation records
System Autonomy Degree of required human oversight Supervision protocols and controls
Data Sensitivity Type and scope of patient data handled Privacy impact assessments
Clinical Use Case Specific medical applications Documentation of use cases

Focus on high-risk systems that directly affect patient care, and ensure these assessments are completed quickly. These evaluations will guide the use of automated tools in the next steps.

Tools for Risk Assessment

Censinet RiskOps™, aligned with NIST AIRMF 1.0, offers:

  1. Automated Risk Scoring
    Standardizes risk evaluations across departments for consistency.
  2. Centralized Documentation
    Keeps all system assessments, specifications, controls, and validation results in one place.
  3. Dynamic Risk Monitoring
    Tracks changes in risk profiles, enabling early issue detection and timely responses.

In August 2024, Censinet introduced an enterprise assessment for the NIST Artificial Intelligence Risk Management Framework 1.0 (NIST AIRMF). This tool helps healthcare organizations identify, manage, and reduce risks tied to AI technologies, ensuring they are implemented and used safely, securely, and ethically across the organization.

sbb-itb-535baee

Step 2: Improve Security and Risk Controls

Healthcare boards must now focus on strengthening cybersecurity measures to comply with the EU AI Act. After mapping and rating AI systems, the next step is to enhance defenses with stricter security protocols.

AI Security Standards

Incorporate AI-specific security measures into current frameworks. Key standards to align with include:

Standard Key Components Focus Areas
ISO/IEC 42001:2023 AI Management Systems Risk assessments and governance structures
NIST AI RMF Risk Management Framework Security controls and monitoring processes

These standards provide a solid foundation for managing AI security effectively.

"The AI arms race has accelerated rapidly in the last year. It appears this Act is trying to establish some guardrails that intend to eliminate or at least reduce the risk of harm to both businesses and private citizens." [3]

Using automated tools can further strengthen these frameworks and streamline compliance efforts.

Automated Risk Management

Censinet offers tools that simplify compliance processes, including:

  • TPRM AI™:
  • ERM AI™:
    • Aligns with the NIST AI Risk Management Framework
    • Automates compliance monitoring
    • Integrates workflows for risk management

Real-World Example: Renown Health

Renown Health

In February 2025, Renown Health adopted the Censinet platform to automate IEEE UL 2933 compliance screening. Under the guidance of CISO Chuck Podesta, this initiative showcased how automation ensures high standards while improving vendor evaluations. Benefits included:

  • Automated compliance checks
  • Seamless integration of risk workflows
  • Improved efficiency in addressing cybersecurity risks

Healthcare boards should embrace these automated solutions to maintain consistent risk management practices and meet EU AI Act requirements effectively.

Step 3: Develop Board AI Expertise

Once risk assessments and system inventories are in place, boards need to strengthen their ability to oversee AI systems effectively. This includes gaining the knowledge required to ensure compliance with the EU AI Act. A good starting point is improving board training programs to cover key aspects of AI oversight.

Board AI Training

Structured training programs should focus on three main areas:

Training Focus Key Components Results
Technical Basics AI system architecture, machine learning models, data governance Clear understanding of AI capabilities and limitations
Regulatory Compliance EU AI Act requirements, risk categories, reporting obligations Ability to evaluate compliance needs effectively
Security Framework Cybersecurity risks, incident response, resilience measures Improved risk management and response strategies

These training sessions should emphasize real-world applications, particularly in healthcare settings. Topics might include how AI systems interact with patient data, integrate into clinical workflows, and adhere to security protocols.

AI Review Process

Training alone isn’t enough. Boards must adopt systematic oversight practices to ensure ongoing compliance and minimize risks. Consider these three key steps:

1. Monthly AI System Audits

Regularly evaluate AI systems for compliance with the EU AI Act. This includes checking system performance, assessing risk levels, and reviewing security measures.

2. Quarterly Risk Assessments

Analyze the impact of AI deployments on patient care, data security, and operational efficiency. Align these assessments with the NIST AI Risk Management Framework for consistency.

3. Annual Tabletop Exercises

Simulate AI-related cybersecurity incidents to prepare board members for potential real-world challenges. These exercises help identify gaps in readiness and improve response strategies.

Compliance Records

Maintaining detailed documentation is critical to meeting EU AI Act requirements. Key records to track include:

Documentation Type Required Elements Update Frequency
AI System Inventory System specifications, risk ratings, deployment scope Monthly
Training Records Certification details, completion dates, assessment results Quarterly
Incident Reports Security events, response actions, remediation steps As needed
Risk Assessments Evaluation results, mitigation strategies, implementation timelines Quarterly

"The spectrum of effective attacks against ML is wide, rapidly evolving, and covers all phases of the ML lifecycle." - NIST [4]

Automated tools can simplify record-keeping, ensuring accuracy and readiness for audits. This approach allows organizations to stay compliant and manage AI-related risks efficiently.

Conclusion: AI Compliance Action Plan

Summary of Board Actions

Boards need to act promptly to meet EU AI Act requirements while upholding strong cybersecurity measures. The timeline highlights three key focus areas that demand attention:

Focus Area Required Actions Timeline
System Assessment Conduct a full inventory of AI systems and classify risks Q2 2025
Security Controls Introduce automated tools for risk monitoring and management Q3 2025
Governance Develop board oversight processes and proper documentation Q4 2025

These steps serve as the foundation for a clear and efficient compliance strategy.

Implementation Guide

To ensure smooth implementation, boards should use automated assessment tools, strengthen documentation efforts, and integrate these processes with current cybersecurity systems.

Key priorities for implementation:

  • Use automated tools to speed up risk assessments while meeting IEEE UL 2933 standards.
  • Build a thorough documentation process to track AI-related decisions.
  • Align governance measures with existing risk management frameworks.

Success hinges on staying ahead with risk assessments and automated monitoring. By adopting advanced risk management platforms, healthcare organizations can achieve compliance while safeguarding patient data and ensuring safety.

FAQs

What risks do healthcare organizations face if they don’t comply with the EU AI Act by 2025?

Healthcare organizations that fail to comply with the EU AI Act by 2025 could face severe financial and legal consequences. Penalties for non-compliance can reach $37 million or 7% of global annual revenue, whichever is higher, for using prohibited AI systems. For violations involving general-purpose AI models, fines may go up to $16 million or 3% of global annual revenue, whichever is greater.

Additionally, providing false, incomplete, or misleading information to regulators can result in fines as high as $8 million or 1% of global annual revenue. Beyond financial penalties, non-compliance may harm an organization’s reputation, disrupt operations, and erode trust with patients and partners. It's critical for healthcare boards to prioritize compliance to avoid these risks.

What steps should healthcare boards take to prioritize and manage high-risk AI systems under the EU AI Act?

Healthcare boards can effectively prioritize and manage high-risk AI systems by focusing on data governance, transparency, and human oversight. Start by ensuring data quality, integrity, and compliance with privacy and security standards. Use high-quality, unbiased, and representative training data to minimize risks and improve AI performance.

Document all data processes thoroughly, including sources and how data influences AI decision-making, to maintain transparency and accountability. Incorporate human oversight during the data training process to identify and eliminate biases, ensuring the AI system aligns with its intended purpose. By implementing these practices, boards can proactively address compliance requirements and mitigate risks associated with high-risk AI systems.

What training and knowledge do board members need to effectively oversee AI systems and ensure compliance with regulations?

To effectively oversee AI systems and ensure compliance, board members should focus on gaining a solid understanding of AI-related risks such as data breaches, algorithm bias, and privacy vulnerabilities. They must also familiarize themselves with cybersecurity best practices and the reporting requirements outlined in the EU AI Act and other relevant regulations.

Additionally, board members should seek training on AI governance frameworks and risk management strategies tailored to the healthcare sector. This can include learning how to assess third-party AI vendors, monitor system performance, and ensure compliance with data protection standards. Staying informed through workshops, certifications, or industry-specific seminars can greatly enhance their ability to oversee AI initiatives effectively.

Related posts

Censinet Risk Assessment Request Graphic

Censinet RiskOps™ Demo Request

Do you want to revolutionize the way your healthcare organization manages third-party and enterprise risk while also saving time, money, and increasing data security? It’s time for RiskOps.

Schedule Demo

Sign-up for the Censinet Newsletter!

Hear from the Censinet team on industry news, events, content, and 
engage with our thought leaders every month.

Terms of Use | Privacy Policy | Security Statement | Crafted on the Narrow Land