FDA and OpenAI Explore AI Integration for Accelerating Drug Evaluations

8 Sources

Share

The FDA is in talks with OpenAI to implement AI in drug evaluation processes, aiming to speed up approvals. This move raises questions about data security and the reliability of AI in critical healthcare decisions.

News article

FDA and OpenAI Collaborate on AI-Assisted Drug Evaluations

The U.S. Food and Drug Administration (FDA) is reportedly in discussions with OpenAI to integrate artificial intelligence into its drug evaluation processes. This collaboration aims to accelerate the notoriously lengthy drug approval timeline, which can often exceed a decade

1

2

.

Project cderGPT and Key Players

At the center of these talks is a project called cderGPT, likely referring to an AI tool for the Center for Drug Evaluation and Research (CDER), which regulates over-the-counter and prescription drugs in the U.S.

1

2

. The discussions have involved a small team from OpenAI, FDA officials, and associates from Elon Musk's Department of Government Efficiency

2

.

Jeremy Walsh, the FDA's first-ever AI officer, has been leading these discussions. Additionally, Peter Bowman-Davis, an undergraduate student on leave from Yale serving as the acting chief AI officer at the Department of Health and Human Services, has been involved in conversations about the FDA's AI ambitions

2

.

Potential Benefits and Timeline

FDA Commissioner Marty Makary has expressed enthusiasm for modernizing the drug approval process with AI, stating that the agency has already completed its first AI-assisted scientific review for a product

2

. The FDA aims to achieve full AI integration across all its centers by June 30, an ambitious timeline that has raised some concerns

5

.

Concerns and Challenges

While the integration of AI in drug evaluation processes holds promise, it also raises several questions:

  1. Data Security: There are concerns about securing the vast amount of proprietary company data involved in the drug approval process

    4

    .

  2. AI Reliability: Given the critical nature of drug approvals, the potential for AI hallucinations and errors is a significant concern

    5

    .

  3. Training Data: Questions remain about which models are being used to train the AI and what inputs are being provided for specialized fine-tuning

    4

    .

  4. Regulatory Framework: There is a need for policy guidance on acceptable AI model performance and the types of data used for training

    2

    .

Broader Context of AI in Government

The FDA's AI initiative is part of a larger trend of AI adoption in federal agencies during the Trump administration. This includes OpenAI's ChatGPT Gov, designed to process sensitive government information, and efforts to use AI in other departments such as the General Services Administration and Social Security

5

.

Expert Opinions

Rafael Rosengarten, CEO of Genialis and a board member of the Alliance for AI in Healthcare, supports automating certain tasks in the drug review process but emphasizes the need for careful implementation

2

. Former FDA Commissioner Robert Califf noted that the agency has been experimenting with AI for several years and sees broader opportunities beyond final reviews

2

5

.

As the FDA moves forward with its AI integration plans, balancing innovation with potential risks will be crucial. The outcome of this initiative could set a precedent for AI use in critical government functions and healthcare decision-making processes.

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo