Summary and assessment: Using Artificial Intelligence & Machine Learning in the Development of Drug & Biological Products: Discussion Paper and Request for Feedback (FDA)

The FDA discussion paper begins with a lengthy section covering potential use cases for AI/ML in drug development. There is some overlap here with a 2023 publication by CDER and CBER staffers1, although it goes much further, considering potential use cases as well as those already seen by FDA.  

It follows with a series of questions on which it seeks stakeholder feedback. These fall under three broad headers: 

  1. Human-led governance, accountability, and transparency 
  2. Quality, reliability, and representativeness of data 
  3. Model development, performance, monitoring, and validation 

Since the text poses questions, but does not offer an explicit viewpoint, one must infer what might be in the minds of the authors. However, we see many similar themes to those raised by EMA.

Human-led governance, accountability, and transparency

FDA’s first question is interesting: Where is more regulatory clarity needed when it comes to AI in drug development? To which one is tempted to say ‘everywhere!’. However, this may speak to a desire to draw on existing guidances where they exist, something which EMA also highlights. Questions in this section relate to definitions for transparency, and the extent to which ‘how the AI system operates’ can be understood in detail. This leans on the concept of explainability. It also asks about human oversight of functioning AI systems, and how this can be made ‘meaningful’, by which it would seem they mean, simply, how do we know it works as intended? Traceability and auditability are also raised as aspects of technology validation that are ordinarily required from technology systems deployed in clinical development. Additionally, FDA asks about pre-specification, presumably regarding AI systems that may still be being trained in-use. 

Quality, reliability, and representativeness of data

In this section, FDA ask how we ensure data used to train AI algorithms are fit-for-purpose. This indicates that an issue uppermost in their mind is the potential for biased training data to lead to biases in the AI system performance characteristics.  In terms of AI system performance, FDA also asks about practices deployed to enhance reproducibility and replicability – perhaps in part reflecting on known behaviors of generative AI models. 

They also ask what AI system developers are doing to ensure the integrity of their training data pipeline, manage issues such as missing data, and ensure the data are accurate, consistent, and complete. Like EMA, data privacy and data security are areas highlighted where FDA wants to understand how these twin issues are managed. 

Model development, performance, monitoring, and validation

This final set of questions seeks input that may help establish a shared view on relevant best practices for AI developers. These relate to system documentation, algorithm development, validation, model evaluation, model accuracy and model explainability. There is also an interesting reference to use of open-source AI, asking when its use might be appropriate. 

When can we expect to see regulatory guidance?

With so many questions to be answered, it seems unlikely that a formal regulatory guidance will be published very quickly.  However, given the complexity to be addressed, this is perhaps to be welcomed – even if the one thing sponsors and other participants in clinical development want more than anything else is regulatory clarity. 

For more on this topic, read our full blog: Regulatory acceptability of AI: Current perspectives

References

1. Liu, Q et al.  Landscape analysis of the application of artificial intelligence and machine learning in regulatory submissions for drug development from 2016 to 2021.  Clin Pharm & Therapeutics; 113(4): April 2023. 

Return to Insights Center

Related Insights

Playbook

Are you using real-world evidence?

Feb 1, 2023

Podcast

RBQM Podcast Series | Episode 3: Staying within the Guardrails: How to Push the Boundaries in a Highly Regulated Industry

Jun 16, 2022

Blog

Leveraging the draft FDA Guidance on PBPK for your drug development program

Feb 24, 2021

Blog

AI Milestones: FDA’s ISTAND program accepts AI-based assessment tool for depression

Mar 19, 2024

Blog

Summary and assessment of EMA’s reflection paper on the use of artificial intelligence (AI) in the medicinal product lifecycle

Mar 7, 2024

Blog

Regulatory acceptability of AI: Current perspectives

Mar 7, 2024

Blog

Studying multiple versions of a cellular or gene therapy product in an early-phase clinical trial

Nov 19, 2021

Blog

CNS Summit Recap: The Future is Collaborative

Nov 22, 2021

Video

Creating EU-CTR compliant and patient-friendly lay language summaries (LLS)

Jan 26, 2022

Article

New endpoints for early-stage cancer are gaining regulatory traction

Jan 28, 2022

Article

How biotechs can strengthen their value story with advanced analytics

Feb 15, 2022

Video

On-demand webinar: An expert guide to EU-CTR

Mar 10, 2022