A Survey of Large Language Models in Medicine: Progress, Application, and Challenge
Project Overview
The document explores the integration of generative AI, specifically large language models (LLMs), in the field of education, particularly within medical training and healthcare. It outlines various applications of LLMs that enhance educational outcomes, such as improving diagnostic accuracy, aiding in clinical decision-making, streamlining administrative tasks like clinical coding and report generation, and enriching the training of medical professionals. The text emphasizes the crucial need for interdisciplinary approaches, including bilingual education programs and in-house development strategies, to effectively train bilingual healthcare professionals while safeguarding patient data. The potential benefits of LLMs in medical education are acknowledged, alongside the challenges posed by ethical concerns, the necessity for rigorous evaluation, and the establishment of regulatory frameworks. Overall, the document underscores a collaborative relationship between AI and healthcare education, highlighting both the transformative impact of generative AI on learning and the ongoing challenges that must be navigated to harness its full potential in medical practice.
Key Applications
AI-Enhanced Healthcare Support
Context: Assisting healthcare professionals and students through content generation, personalized learning, patient interaction, and psychological support by leveraging language models for various educational and clinical applications.
Implementation: Utilizes fine-tuned LLMs (such as ChatGPT and RoBERTa) for generating educational materials, automating medical coding, summarizing clinical reports, facilitating patient interactions, and providing psychological support through conversational interfaces. The implementations involve dynamic prompts, knowledge graphs, and training on clinical notes.
Outcomes: Improved educational experiences, enhanced patient care, and accurate clinical documentation. Achieved metrics include 30.72 ROUGE-L for diagnosis summarization, 0.926 AUC in ICD code prediction, and 47.93 ROUGE-L in report summarization.
Challenges: Potential for hallucinations and biases in AI outputs, need for expert review for clinical relevance, risks of misinformation, and lack of emotional understanding in conversational agents.
AI-Enhanced Medical Education Tools
Context: Providing comprehensive training and resources for healthcare professionals and students to improve learning outcomes through the integration of AI technologies into medical education.
Implementation: Employs LLMs to deliver personalized educational content, assessments, and resources, enhancing the understanding of both medical and AI concepts.
Outcomes: Results in better-informed healthcare professionals and improved integration of AI technology in healthcare practice.
Challenges: Ensuring the accuracy and reliability of AI-generated content in medical training and the need for effective frameworks for interdisciplinary collaboration.
Implementation Barriers
Technical
Hallucination of LLMs leading to inaccurate information generation, and challenges in developing effective frameworks for sharing data between rural clinics and AI systems.
Proposed Solutions: Implement training-time correction, generation-time correction, retrieval-augmented correction, and research into interdisciplinary frameworks to facilitate data sharing.
Evaluation
Lack of comprehensive benchmarks and metrics to evaluate LLM performance in medical contexts.
Proposed Solutions: Develop domain-specific benchmarks that assess trustworthiness, helpfulness, and explainability.
Data Limitations
Limited access to high-quality medical datasets for training LLMs.
Proposed Solutions: Generate high-quality synthetic datasets and use smaller open-sourced datasets for fine-tuning.
Knowledge Adaptation
Challenges in updating LLMs with new medical knowledge efficiently.
Proposed Solutions: Utilize model editing and retrieval-augmented generation to incorporate new knowledge.
Ethical
Concerns regarding the ethical use of LLMs in medical settings, including accountability and safety.
Proposed Solutions: Establish guidelines for ethical practices and ensure transparency in AI decision-making.
Regulatory
Complex regulatory landscape for the integration of LLMs in healthcare.
Proposed Solutions: Develop adaptive and robust regulatory frameworks to ensure safety and ethical standards.
Educational Barrier
The need for bilingual professionals who are versed in both medicine and AI technology.
Proposed Solutions: Creation of bilingual education programs that train healthcare professionals in both fields.
Privacy Barrier
Concerns about patient data protection when integrating LLM technology in healthcare.
Proposed Solutions: Implementing effective in-house development methods to safeguard patient data.
Project Team
Hongjian Zhou
Researcher
Fenglin Liu
Researcher
Boyang Gu
Researcher
Xinyu Zou
Researcher
Jinfa Huang
Researcher
Jinge Wu
Researcher
Yiru Li
Researcher
Sam S. Chen
Researcher
Peilin Zhou
Researcher
Junling Liu
Researcher
Yining Hua
Researcher
Chengfeng Mao
Researcher
Chenyu You
Researcher
Xian Wu
Researcher
Yefeng Zheng
Researcher
Lei Clifton
Researcher
Zheng Li
Researcher
Jiebo Luo
Researcher
David A. Clifton
Researcher
Contact Information
For information about the paper, please contact the authors.
Authors: Hongjian Zhou, Fenglin Liu, Boyang Gu, Xinyu Zou, Jinfa Huang, Jinge Wu, Yiru Li, Sam S. Chen, Peilin Zhou, Junling Liu, Yining Hua, Chengfeng Mao, Chenyu You, Xian Wu, Yefeng Zheng, Lei Clifton, Zheng Li, Jiebo Luo, David A. Clifton
Source Publication: View Original PaperLink opens in a new window
Project Contact: Dr. Jianhua Yang
LLM Model Version: gpt-4o-mini-2024-07-18
Analysis Provider: Openai