Skip to main content Skip to navigation

Contextual Candor: Enhancing LLM Trustworthiness Through Hierarchical Unanswerability Detection

Project Overview

The document explores the application of generative AI, specifically through Reinforced Unanswerability Learning (RUL), to improve the reliability and trustworthiness of large language models (LLMs) in educational contexts. It highlights the pervasive challenge of LLMs producing hallucinated or factually incorrect answers, which can undermine user trust and learning outcomes. By introducing a hybrid training approach that incorporates unanswerability detection into the generative processes of LLMs, the method significantly enhances the model's ability to recognize unanswerable questions and generate suitable refusal responses. This advancement not only increases the accuracy of AI interactions but also enriches user experiences, making AI tools in education more dependable. The findings suggest that by fostering better interactions and trust, generative AI can play a pivotal role in educational settings, providing learners and educators with more effective and reliable AI-assisted resources. Overall, the integration of RUL represents a promising direction for enhancing the efficacy of AI in education, contributing to improved learning experiences and outcomes.

Key Applications

Reinforced Unanswerability Learning (RUL)

Context: Conversational AI systems for information retrieval and customer service, applicable to various user demographics seeking reliable information.

Implementation: RUL integrates unanswerability detection into the training of LLMs with a multi-stage learning strategy: supervised fine-tuning followed by reinforcement learning with human feedback.

Outcomes: Higher accuracy in detecting unanswerable questions and generating appropriate refusal responses, leading to improved user trust and satisfaction.

Challenges: The complexity of implementing hierarchical unanswerability detection and the need for extensive annotated datasets.

Implementation Barriers

Technical Barrier

LLMs may still attempt to answer unanswerable questions, leading to hallucinated responses.

Proposed Solutions: Implementing a hybrid training approach like RUL that integrates unanswerability detection directly into the LLM's generative process.

Data Barrier

The requirement for large, annotated datasets that include examples of unanswerable questions and appropriate refusal responses.

Proposed Solutions: Developing comprehensive datasets, such as the Enhanced-CAsT-Answerability (ECA), to facilitate supervised fine-tuning.

Project Team

Steven Robinson

Researcher

Antonio Carlos Rivera

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Steven Robinson, Antonio Carlos Rivera

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies