• There are no suggestions because the search field is empty.

Impact of the AI Act on the MedTech and Life Sciences Industry

Author Avatar
Pieter Smits, Project Manager at QbD Group
Software Solutions & Services
In Vitro Diagnostics
Medical Devices
Explore how the AI Act impacts MedTech and Life Sciences, and how the QbD Group can guide you through its complexities.
Impact of the AI Act on the MedTech and Life Sciences Industry | QbD Group
4:24

The AI Act was formally adopted by the full European Parliament on 13 March, and we have reviewed the current text to form our initial impressions of the law’s impact on the MedTech industry.

The AI Act will be crucial for any AI-containing product or device used in our industry. It represents the world’s first regulatory framework for AI and aims to ensure trustworthy AI in Europe by mandating that AI systems respect fundamental rights, safety, and ethical principles, and by addressing the risks posed by AI models.

Failure to comply with the AI Act could result in fines of up to €35,000,000 or up to 7% of a company’s global annual turnover, whichever is higher.

In this blog post, we will explore:

  • the different AI categories mentioned in the AI Act,
  • outline their requirements,
and discuss the implications of the AI Act for the MedTech industry.
 

Prohibited AI Systems

The AI Act outlines a number of prohibited AI practices. These include AI models that:

  • use behavioral manipulation techniques.
  • exploit vulnerabilities based on demographics or biometrics.
  • employ facial recognition databases obtained through untargeted scraping, among others.
These prohibitions are designed to protect against manipulative or exploitative practices and to uphold ethical standards in AI usage.
 

General Purpose AI

For general-purpose AI models that can perform a wide range of tasks competently, and for AI systems designed to directly interact with people, the AI Act recognizes and addresses emerging challenges such as deepfakes and emotion recognition technologies.

The AI Act enforces transparency requirements, especially concerning the disclosure of content that has been artificially generated or manipulated.

Provisions for watermarking are intended to increase transparency, with exceptions allowed for certain use cases that are legally authorized or deemed artistic or satirical in nature.

In our industry, relevant examples might include AI models that aid in the creation of high-quality records, such as those for change control or root cause analysis.

 

High-Risk AI Systems

Criteria for high-risk AI systems are designed to mitigate potential harms to health, safety, or fundamental rights.

These criteria include:

  • strict compliance with documentation and registration requirements,
  • adherence to robust data governance processes,
  • and a significant focus on cybersecurity, underscoring the importance of thorough risk assessment and compliance.
 While manufacturers of In Vitro Diagnostics (IVD) and Medical Devices (MD) are familiar with such requirements, these additional requirements represent new challenges for certain equipment and software suppliers to the life sciences industry.

Suppliers impacted by the AI Act must now navigate these requirements in addition to the Good Manufacturing Practice (GMP) standards demanded by their customers.

This heightened emphasis on data governance and cybersecurity aligns with the latest developments in Software as a Medical Device (SaMD) guidelines and regulations.

Examples falling under this category include AI-containing MDs or IVDs.

 

Human Oversight and Employer Obligations

Robust human oversight emerges as another cornerstone of the AI Act, requiring competent personnel with the necessary training and authority.

In particular, employers deploying high-risk AI systems will be required to inform workers of their integration, promoting transparency and accountability in the workplace.

 

Impact on the MedTech industry

While the AI Act does not impact the entire MedTech and Life Sciences industry, software developers and device manufacturers planning to integrate AI models into their products — whether as industry support tools or as software as a medical device — should consider several requirements during the design and development phases. Given the recent surge in AI-enabled software, the influence of the AI Act will be a pivotal factor in the market introduction of these products.

In response to these developments, organizations should adopt a proactive stance by first defining how the AI Act may affect their products. This approach will equip them with a clear strategy for navigating the regulatory landscape once the AI Act is enforced. 

 

 

 

Stay ahead in life sciences

Keeping up with the fast-paced life sciences industry doesn’t have to be overwhelming.

Our newsletter delivers the latest insights, industry updates, and expert content directly to your inbox, helping you stay informed and make smarter decisions.

Circles-banner-short

Discover more expert content

preview_image
Whitepaper

How to keep computerized systems in the operational phase

Ensure compliance and efficiency with best practices for maintaining computerized systems in the operational phase. Download our expert whitepaper now!
preview_image
Webinar

From Requirements to Code: a unified MDSW development cycle that covers all requirements

Watch our webinar on demand to master medical device software development. Learn about IEC standards, cybersecurity, AI integration, and FDA expectations.
preview_image
Whitepaper

Mobile health on the rise: exploring the regulatory landscape for reimbursement

This whitepaper will help you navigate the maze of the DTx regulatory environment, highlighting several important countries and regulations.
preview_image
Whitepaper

GAMP 5 Software Validation Approach for GMP, GCP and GLP regulations

Learn how to comply with GMP, GCP, and GLP regulations using the GAMP 5 Software Validation Approach. Download the whitepaper for more insights.
preview_image
Blog

Establishing and Maintaining the Right Level of Clinical Evidence under the EU IVDR

We’re proud to highlight a new publication by Pieter Bogaert, PhD—senior...
preview_image
Blog

Post-Market Surveillance (PMS) and Post-Market Clinical Follow-Up (PMCF) under the MDR: Ensuring safety and performance

With the introduction of the Medical Device Regulation (MDR), the...
preview_image
Blog

What Makes Usability Testing Crucial for Near-Patient and Self-Testing Devices under IVDR?

It shouldn’t be a surprise that today, “Near-Patient Testing (NPT)” and...
preview_image
Blog

When does Annex XIV apply in Performance Studies, and what key documentation is needed for compliance?

In the European regulatory landscape, conducting performance studies for in...