Skip to main content Skip to navigation

PALLM: Evaluating and Enhancing PALLiative Care Conversations with Large Language Models

Project Overview

The document explores the transformative role of generative AI, particularly large language models (LLMs) like GPT-4, in the field of education, with a focus on enhancing communication skills and learning outcomes. It highlights the application of LLMs in evaluating communication metrics such as empathy and understanding, particularly in sensitive contexts like palliative care. Findings indicate that LLMs surpass traditional methods in assessing clinical conversations, thereby providing a powerful tool for personalized feedback and fostering better interactions. Furthermore, the study illustrates the potential of LLMs to fine-tune educational practices and generate synthetic data, ultimately improving both teaching and learning experiences. By integrating these advanced AI capabilities, educational institutions can leverage technology to enhance student engagement and learning efficacy, paving the way for a more responsive and effective educational environment. The outcomes suggest that adopting generative AI in educational settings can lead to significant advancements in communication, understanding, and overall educational effectiveness.

Key Applications

Evaluation of palliative care communication using LLMs

Context: Healthcare providers in palliative care settings, specifically during patient-provider conversations.

Implementation: The study utilized simulated clinical scripts crafted by healthcare professionals to evaluate LLMs like GPT-4 and LLaMA2 for assessing communication metrics.

Outcomes: LLMs demonstrated over 90% accuracy in identifying communication metrics, significantly improving evaluation performance compared to traditional methods.

Challenges: Challenges include the need for sensitive clinical data, potential biases in model outputs, and the scarcity of real-world clinical dialogue datasets.

Implementation Barriers

Data Privacy

The requirement to upload sensitive clinical data to proprietary models raises significant privacy concerns.

Proposed Solutions: Developing in-house LLMs to ensure sensitive data remains within secure, institution-controlled environments.

Data Availability

There is a notable absence of publicly available datasets containing clinical communication, which hinders model training and evaluation.

Proposed Solutions: Generating synthetic datasets using advanced LLMs like GPT-4 to create realistic clinical dialogue scenarios for training.

Model Limitations

LLMs may struggle to fully grasp the nuances of clinical communication, especially in complex emotional contexts. This includes challenges in understanding and reasoning capabilities.

Proposed Solutions: Refining prompting techniques and training models on more diverse datasets to improve understanding and reasoning capabilities.

Project Team

Zhiyuan Wang

Researcher

Fangxu Yuan

Researcher

Virginia LeBaron

Researcher

Tabor Flickinger

Researcher

Laura E. Barnes

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Zhiyuan Wang, Fangxu Yuan, Virginia LeBaron, Tabor Flickinger, Laura E. Barnes

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies