The intersection of AI and pharmaceutical regulation is rapidly evolving, with potential implications for IP strategy. The recently issued draft guidance on the Considerations for the Use of AI To Support Regulatory Decision-Making for Drug and Biological Products (FDA-2024-D-4689) is the FDA's first attempt to standardize how AI-generated data and models will be evaluated in the context of drug safety and effectiveness. New AI applications in the drug regulatory process are emerging every day, from personalised medicine to clinical trial design and manufacturing quality control. As innovators make more and more use of these technologies, they will face the challenge of conforming to the regulatory transparency requirements whilst protecting their IP.
Use of AI to support regulatory decision making
The focus of the FDA guidance is the application of AI to processes directly supporting regulatory decision-making about the safety, effectiveness, or quality of drugs. As such, the guidance explicitly excludes pre-clinical drug discovery activities that occur prior to any interaction with regulatory authorities. However, whilst much of the public interest for the use of AI in pharma has focused on drug discovery tools such as AlphaFold, AI companies are already moving beyond drug discovery to develop solutions for the whole drug development process.
Kat assisted drug development |
The FDA identifies numerous potential applications for AI throughout drug development. These include computer models for pharmacokinetic, pharmacodynamic, and toxicological studies, predictive modelling for dose determination and the integration of data from different sources such as clinical trials, genetic databases, and social media to improve understanding of a disease.
Current AI applications in drug development already demonstrate the regulatory complexities that the FDA guidance seeks to address, particularly in the realm of personalised medicine. For Moderna's cancer vaccine mRNA-415/V940, for example, AI algorithms analyse patient tissue samples to identify tumour-specific neoantigens for creating patient-unique mRNA vaccines. This approach raises questions for the regulatory authorities, such as how the safety and efficacy of the vaccine can be assessed when each patient receives a potentially unique vaccine sequence.
The FDA identifies several key challenges associated with AI use in drug development. First, the FDA guidance points out that the variability in quality, size, and representativeness of training datasets may introduce bias and raise questions about the reliability of AI-driven results. Additionally, the complexity of the underlying AI models can make it difficult to understand how these systems reach their conclusions. This lack of interpretability necessitates what the FDA terms "methodological transparency", namely detailed disclosure of the methods and processes used to develop AI models in regulatory submissions.
Here we see a parallel with patent law, specifically with the requirement in patent law for the invention to be sufficiently disclosed (Europe) or described and enabled (US). The patent law on the sufficiency of AI inventions is still developing. The question thus remains whether the regulatory requirements relating to the use of AI in pharmaceutical development will be more or less stringent than the patentability requirements for the AI itself.
To satisfy the requirement for methodological transparency, the FDA guidance places particular emphasis on disclosure of the quality and characteristics of the training data used to develop an AI model. To ensure that the data is as representative of the target patient population or manufacturing process involved, the guidance recommends describing data collection, processing, annotation, storage, and control procedures, along with rationales for dataset selection and methods for establishing labels or annotations. The guidance also suggests including descriptions of model inputs and outputs, architecture, features, selection processes, and training hyperparameters. For models using pre-trained components, the guidance indicates that the datasets used for pre-training should be specified, and there should be a description of how the models were obtained or developed. In this Kat's view, the level of disclosure required by the guidance is therefore currently far greater than is necessitated for obtaining a patent relating to an AI model.
The credibility assessment approach
At the heart of the FDA guidance is our old friend "credibility". Credibility (and its synonym plausibility) is of course a loaded term in European patent law, applied to the evidence standard for therapeutic inventions and the patentability of software inventions (IPKat), see for example, Case Law of the Boards of Appeal I-D.9.2.8. The FDA Guidance also chooses "credibility" to define the evidence standard for AI in the context of the drug regulatory process.
FDA guidance provides a 7-step risk-based credibility assessment process for AI in the drug regulatory process:
Step 1: What is the question of interest to be addressed by the AI model?
Step 2: What is the "context of use" for the AI model? The 2nd step involves defining the "context of use" (COU) for the AI model. The COU is supposed to define the AI system's role and scope, describing in detail what will be modelled and how model outputs will be used.
Step 3: What is the risk of the AI model? This step requires an assessment of the risk of using the AI model, which combines two factors: model influence and decision consequence. Model influence represents the contribution of AI-derived evidence relative to other contributing evidence used to inform the question of interest. In other words, what is the benefit of using AI over more traditional approaches? Decision consequence is supposed to describe the significance of an adverse outcome resulting from an incorrect decision by the AI. For these two steps, the guidance provides a risk matrix showing how combinations of model influence and decision consequence determine overall model risk. For instance, if an AI system is the sole determinant for a high-stakes decision (high model influence) and errors could result in serious patient harm (high decision consequence), the model carries high risk requiring stringent credibility assessment activities.
Step 4: Develop a plan to establish the credibility of AI model output within the COU. Step 4 involves developing a "credibility assessment plan" tailored to the model risk and COU. For high-risk models, this plan should include comprehensive documentation of model architecture, training data characteristics, development processes, and evaluation methods. The guidance provides detailed recommendations for describing models, training data, model development processes, and evaluation methodologies. It is clear from the details and complexity of this process that innovators developing an AI model should engage with the FDA as early as possible to ensure they are documenting the correct information and mitigating relevant risks appropriately.
Steps 5-7 involve executing the credibility assessment plan, documenting the results of the credibility assessment plan and establishing whether the AI is adequate for the COU.
When model credibility proves insufficient for the intended context of use, the guidance suggests some actions that may be taken to remedy the situation. These include reducing the dependency on the AI model by submitting alternative evidence, conducting more stringent credibility assessment activities or improving the AI model performance with additional training data.
Impact on IP strategy for AI
One of the most critical decisions for AI companies and companies using AI is whether to disclose and protect their models via patent filings, or to keep the details of the model a trade secret. What is clear from even this brief outline of the FDA 78-step process, is that clear documentation and description of what, when and where AI was used in regulatory submissions will be critical. It is unclear from the guidance how much of this information may be made public. If full or even partial public disclosure of the intricate details of an AI model used in the regulatory process is required by the FDA, then this may significantly impact the IP strategy of the AI provider, and necessitate a patent-focused, as opposed to a trade secret-focused, strategy
Furthermore, even if the full information provided to the FDA about an AI model is not made publicly available, there are still potential IP implications. In many cases, an innovator company may not itself have all the details about all the AI models to the level of detail stipulated by the FDA guidance. In these circumstances, the AI provider may have to provide more information about their product and processes to the pharma company than they are comfortable with from a trade secrets perspective. Additionally, there are some aspects of the disclosure required by the guidance that simply cannot be protected by patents, including the detailed information about the training data.
Response from the industry: What about foundation models?
Over a hundred comments on the guidance have been submitted by interested parties, including big pharma and AI biotech companies. There are a number of recurring themes in the comments including the need for more comprehensive examples of how the 7-step credibility assessment process will work. There are also calls for greater harmonisation with international standards, particularly the OECD's AI definition and the International Medical Device Regulators Forum's Good Machine Learning Practice principles, to ensure global consistency in regulatory approaches.
A number of technical challenges were also highlighted by commentators including the unique considerations required for foundation models and agentic AI platforms. Foundation models are large-scale models, such as ChatGPT, are trained on vast datasets. Foundational models can then be fine-tuned for numerous downstream applications, including those in the drug regulatory process such as patient stratification, biomarker discovery, and clinical trial design. LLMs, for example, may well be used to write clinical trial protocols and regulatory submissions. Foundation models thus operate as general-purpose platforms that may not fit into context-of-use categorisation. However, the draft guidance focuses on context-specific credibility assessments. Commenters such as the AI biotech Owkin wondered whether there should therefore be a dual assessment structure, evaluating both the foundation model itself for data governance and privacy compliance and each specific application derived from it.
Final thoughts
It is clear from the draft guidance that, as the use and complexity of AI models in the pharmaceutical industry inexorably grows, the FDA is likely to face an ever increasing burden of reviewing the safety and efficacy of AI models. Innovators will also face an increasing burden of documenting and reviewing the AI models they use throughout the regulatory process, including those provided by third parties. Layered on top of this is the complexity of the IP strategy that will be required to ensure alignment between the regulatory process and maintaining a competitive IP position. A key lesson that this Kat has learnt throughout her career is reinforced by the emergence and challenges of AI in the drug regulatory process. To be effective, IP cannot exist in a silo, but must be integrated and aligned with the drug development and R&D teams and the entire regulatory process. This latest FDA guidance highlights that alignment between the IP and regulatory strategy for AI models and AI generated data continues to be absolutely crucial.
On the theme of AI, don't forget to secure your place at CIPA congress: IP & AI this year!
Further reading
- USPTO call for comments: Impact of AI on patentability (May 2024)
- OpenAI's large language model (LLM) patents (Aug 2024)
- New USPTO Guidance on the use of AI in precision medicine (Aug 2024)
- When is the inventor of an AI model also an inventor of the model's output? A closer look at the USPTO Guidance for AI-Assisted Inventions (July 2024)
- IP strategy for AI-assisted drug discovery (Oct 2024)
- AlphaFold: From Nobel Prize to drug-discovery gold mine? (Nov 2024)
- Insilico Medicine: Lessons in IP strategy from a front-runner in AI-drug discovery (Feb 2025)
- Too broad, too early? AI platform for cell analysis found to lack technical character and sufficiency (T 0660/22, Cell analysis/NIKON) (Aug 2025)
No comments:
All comments must be moderated by a member of the IPKat team before they appear on the blog. Comments will not be allowed if the contravene the IPKat policy that readers' comments should not be obscene or defamatory; they should not consist of ad hominem attacks on members of the blog team or other comment-posters and they should make a constructive contribution to the discussion of the post on which they purport to comment.
It is also the IPKat policy that comments should not be made completely anonymously, and users should use a consistent name or pseudonym (which should not itself be defamatory or obscene, or that of another real person), either in the "identity" field, or at the beginning of the comment. Current practice is to, however, allow a limited number of comments that contravene this policy, provided that the comment has a high degree of relevance and the comment chain does not become too difficult to follow.
Learn more here: http://ipkitten.blogspot.com/p/want-to-complain.html