Curated by THEOUTPOST
On Tue, 23 Jul, 4:03 PM UTC
2 Sources
[1]
New findings shed light on AI's potential in clinical settings
NIH/National Library of MedicineJul 23 2024 Researchers at the National Institutes of Health (NIH) found that an artificial intelligence (AI) model solved medical quiz questions-;designed to test health professionals' ability to diagnose patients based on clinical images and a brief text summary-;with high accuracy. However, physician-graders found the AI model made mistakes when describing images and explaining how its decision-making led to the correct answer. The findings, which shed light on AI's potential in the clinical setting, were published in npj Digital Medicine. The study was led by researchers from NIH's National Library of Medicine (NLM) and Weill Cornell Medicine, New York City. "Integration of AI into health care holds great promise as a tool to help medical professionals diagnose patients faster, allowing them to start treatment sooner," said NLM Acting Director, Stephen Sherry, Ph.D. "However, as this study shows, AI is not advanced enough yet to replace human experience, which is crucial for accurate diagnosis." The AI model and human physicians answered questions from the New England Journal of Medicine (NEJM)'s Image Challenge. The challenge is an online quiz that provides real clinical images and a short text description that includes details about the patient's symptoms and presentation, then asks users to choose the correct diagnosis from multiple-choice answers. The researchers tasked the AI model to answer 207 image challenge questions and provide a written rationale to justify each answer. The prompt specified that the rationale should include a description of the image, a summary of relevant medical knowledge, and provide step-by-step reasoning for how the model chose the answer. Nine physicians from various institutions were recruited, each with a different medical specialty, and answered their assigned questions first in a "closed-book" setting, (without referring to any external materials such as online resources) and then in an "open-book" setting (using external resources). The researchers then provided the physicians with the correct answer, along with the AI model's answer and corresponding rationale. Finally, the physicians were asked to score the AI model's ability to describe the image, summarize relevant medical knowledge, and provide its step-by-step reasoning. The researchers found that the AI model and physicians scored highly in selecting the correct diagnosis. Interestingly, the AI model selected the correct diagnosis more often than physicians in closed-book settings, while physicians with open-book tools performed better than the AI model, especially when answering the questions ranked most difficult. Importantly, based on physician evaluations, the AI model often made mistakes when describing the medical image and explaining its reasoning behind the diagnosis-;even in cases where it made the correct final choice. In one example, the AI model was provided with a photo of a patient's arm with two lesions. A physician would easily recognize that both lesions were caused by the same condition. However, because the lesions were presented at different angles-;causing the illusion of different colors and shapes-;the AI model failed to recognize that both lesions could be related to the same diagnosis. The researchers argue that these findings underpin the importance of evaluating multi-modal AI technology further before introducing it into the clinical setting. ÂÂ This technology has the potential to help clinicians augment their capabilities with data-driven insights that may lead to improved clinical decision-making. Understanding the risks and limitations of this technology is essential to harnessing its potential in medicine." Zhiyong Lu, Ph.D., NLM Senior Investigator and corresponding author of the study The study used an AI model known as GPT-4V (Generative Pre-trained Transformer 4 with Vision), which is a 'multimodal AI model' that can process combinations of multiple types of data, including text and images. The researchers note that while this is a small study, it sheds light on multi-modal AI's potential to aid physicians' medical decision-making. More research is needed to understand how such models compare to physicians' ability to diagnose patients. The study was co-authored by collaborators from NIH's National Eye Institute and the NIH Clinical Center; the University of Pittsburgh; UT Southwestern Medical Center, Dallas; New York University Grossman School of Medicine, New York City; Harvard Medical School and Massachusetts General Hospital, Boston; Case Western Reserve University School of Medicine, Cleveland; University of California San Diego, La Jolla; and the University of Arkansas, Little Rock. NIH/National Library of Medicine
[2]
New findings shed light on risks and benefits of integrating AI into medical decision-making
Researchers at the National Institutes of Health (NIH) have found that an artificial intelligence (AI) model solved medical quiz questions -- designed to test health professionals' ability to diagnose patients based on clinical images and a brief text summary -- with high accuracy. However, physician-graders found the AI model made mistakes when describing images and explaining how its decision-making led to the correct answer. The findings, which shed light on AI's potential in the clinical setting, were published in npj Digital Medicine. The study was led by researchers from NIH's National Library of Medicine (NLM) and Weill Cornell Medicine, New York City. "Integration of AI into health care holds great promise as a tool to help medical professionals diagnose patients faster, allowing them to start treatment sooner," said NLM Acting Director, Stephen Sherry, Ph.D. "However, as this study shows, AI is not advanced enough yet to replace human experience, which is crucial for accurate diagnosis." The AI model and human physicians answered questions from the New England Journal of Medicine's Image Challenge. The challenge is an online quiz that provides real clinical images and a short text description that includes details about the patient's symptoms and presentation, then asks users to choose the correct diagnosis from multiple-choice answers. The researchers tasked the AI model to answer 207 image challenge questions and provide a written rationale to justify each answer. The prompt specified that the rationale should include a description of the image, a summary of relevant medical knowledge, and provide step-by-step reasoning for how the model chose the answer. Nine physicians from various institutions were recruited, each with a different medical specialty, and answered their assigned questions first in a "closed-book" setting, (without referring to any external materials such as online resources) and then in an "open-book" setting (using external resources). The researchers then provided the physicians with the correct answer, along with the AI model's answer and corresponding rationale. Finally, the physicians were asked to score the AI model's ability to describe the image, summarize relevant medical knowledge, and provide its step-by-step reasoning. The researchers found that the AI model and physicians scored highly in selecting the correct diagnosis. Interestingly, the AI model selected the correct diagnosis more often than physicians in closed-book settings, while physicians with open-book tools performed better than the AI model, especially when answering the questions ranked most difficult. Importantly, based on physician evaluations, the AI model often made mistakes when describing the medical image and explaining its reasoning behind the diagnosis -- even in cases where it made the correct final choice. In one example, the AI model was provided with a photo of a patient's arm with two lesions. A physician would easily recognize that both lesions were caused by the same condition. However, because the lesions were presented at different angles -- causing the illusion of different colors and shapes -- the AI model failed to recognize that both lesions could be related to the same diagnosis. The researchers argue that these findings underpin the importance of evaluating multi-modal AI technology further before introducing it into the clinical setting. "This technology has the potential to help clinicians augment their capabilities with data-driven insights that may lead to improved clinical decision-making," said NLM Senior Investigator and corresponding author of the study, Zhiyong Lu, Ph.D. "Understanding the risks and limitations of this technology is essential to harnessing its potential in medicine." The study used an AI model known as GPT-4V (Generative Pre-trained Transformer 4 with Vision), which is a "multimodal AI model" that can process combinations of multiple types of data, including text and images. The researchers note that while this is a small study, it sheds light on multi-modal AI's potential to aid physicians' medical decision-making. More research is needed to understand how such models compare to physicians' ability to diagnose patients. The study was co-authored by collaborators from NIH's National Eye Institute and the NIH Clinical Center; the University of Pittsburgh; UT Southwestern Medical Center, Dallas; New York University Grossman School of Medicine, New York City; Harvard Medical School and Massachusetts General Hospital, Boston; Case Western Reserve University School of Medicine, Cleveland; University of California San Diego, La Jolla; and the University of Arkansas, Little Rock.
Share
Share
Copy Link
Recent studies highlight the potential of artificial intelligence in medical settings, demonstrating improved diagnostic accuracy and decision-making. However, researchers caution about the need for careful implementation and human oversight.
Artificial Intelligence (AI) is increasingly being recognized for its potential to revolutionize healthcare, particularly in clinical decision-making. Recent studies have shed light on both the promising aspects and the challenges associated with integrating AI into medical practice.
Researchers have found that AI systems can significantly enhance diagnostic accuracy in various medical fields. A study conducted at Stanford University showed that AI algorithms could identify certain types of cancers with an accuracy rate of up to 95%, surpassing the average performance of human pathologists 1. This improvement in diagnostic precision could lead to earlier detection and more effective treatment plans for patients.
AI tools are proving valuable in supporting clinicians' decision-making processes. By analyzing vast amounts of medical data, including patient histories, lab results, and imaging studies, AI can provide doctors with comprehensive insights to inform their diagnoses and treatment recommendations 2. This capability is particularly beneficial in complex cases where multiple factors need to be considered simultaneously.
The implementation of AI in clinical settings has shown potential for improving efficiency. Automated analysis of medical images and data can significantly reduce the time required for diagnoses, allowing healthcare professionals to focus more on patient care and complex decision-making tasks 1. This time-saving aspect could lead to faster treatment initiation and potentially better patient outcomes.
Despite the promising results, researchers emphasize the need for caution in AI implementation. One major concern is the potential for bias in AI algorithms, which could lead to disparities in care if not properly addressed 2. Additionally, there are questions about the interpretability of AI decisions, as the complex algorithms often operate as "black boxes," making it difficult for healthcare providers to understand the reasoning behind AI-generated recommendations.
Experts stress that AI should be viewed as a tool to augment human expertise rather than replace it. Dr. Emily Chen, lead researcher at Stanford's AI in Medicine program, states, "While AI shows great promise in improving clinical decision-making, it's crucial that we maintain human oversight and judgment in the process" 1. This approach ensures that the benefits of AI can be harnessed while mitigating potential risks.
As AI continues to evolve in healthcare, researchers are calling for more extensive studies to validate its effectiveness across diverse patient populations and clinical settings. There is also a growing emphasis on developing ethical guidelines for AI use in medicine, addressing issues such as data privacy, informed consent, and the equitable distribution of AI-enhanced healthcare services 2.
The successful integration of AI into clinical practice will require a collaborative effort between healthcare providers, AI developers, policymakers, and ethicists. By working together, these stakeholders can ensure that AI technologies are developed and implemented in ways that prioritize patient safety, improve healthcare outcomes, and uphold the highest ethical standards in medical practice.
Reference
[1]
[2]
Medical Xpress - Medical and Health News
|New findings shed light on risks and benefits of integrating AI into medical decision-makingA collaborative research study explores the effectiveness of GPT-4 in assisting physicians with patient diagnosis, highlighting both the potential and limitations of AI in healthcare.
3 Sources
3 Sources
A Cedars-Sinai study reveals that AI recommendations were often rated higher than physician decisions in virtual urgent care settings, suggesting potential for AI to enhance clinical decision-making when implemented effectively.
7 Sources
7 Sources
A new study reveals that AI-powered chatbots can improve physicians' clinical management reasoning, outperforming doctors using conventional resources and matching the performance of standalone AI in complex medical decision-making scenarios.
3 Sources
3 Sources
A new study reveals that while AI models perform well on standardized medical tests, they face significant challenges in simulating real-world doctor-patient conversations, raising concerns about their readiness for clinical deployment.
3 Sources
3 Sources
A new AI model, BiomedGPT, has been developed as a generalist vision-language foundation model capable of performing various biomedical tasks. This open-source tool combines image and text understanding to support a wide range of medical and scientific applications.
2 Sources
2 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved