Skip to main content Skip to navigation

Distractor Generation in Multiple-Choice Tasks: A Survey of Methods, Datasets, and Evaluation

Project Overview

The document explores the transformative role of generative AI in education, particularly in generating distractors for multiple-choice and fill-in-the-blank questions, marking a shift from traditional assessment methods to advanced AI techniques like deep learning and pre-trained language models. It underscores the importance of effective distractors in evaluating student knowledge while addressing the challenges of maintaining their quality and relevance. Furthermore, it examines various AI models and methodologies that contribute to this distractor generation, paving the way for future research in the field. Additionally, the document highlights the development and analysis of multiple-choice question datasets, which are critical for generative AI applications in education. These datasets, covering a wide range of subjects, serve vital functions in tasks such as distractor generation and reading comprehension, showcasing state-of-the-art performance metrics and enhancing the efficacy of AI-driven educational tools. Overall, the findings indicate that generative AI has significant potential to improve assessment quality and educational outcomes through innovative applications and robust datasets.

Key Applications

Generative AI for Assessment and Question Generation

Context: Utilized in educational assessments across various domains including Science, English, Math, and Medicine for creating multiple-choice questions and evaluating comprehension in students.

Implementation: Involves the use of neural networks, pre-trained models, and various learning-based methods to generate plausible distractors and educational datasets. Datasets such as CLOTH, RACE, and SciQ are created from educational materials and textbooks, which may include blogs and multi-modal content.

Outcomes: Enhances the quality of assessment tools, leading to fairer evaluations of student understanding and improved readability and comprehension assessment.

Challenges: Includes generating distractors that are both plausible and misleading, ensuring diversity, avoiding answer revealing issues, and addressing data availability and licensing requirements for certain datasets.

Implementation Barriers

Technical Barrier

Challenges in generating high-quality and diverse distractors that do not reveal the correct answer.

Proposed Solutions: Integration of reinforcement learning and few-shot examples to improve the quality of generated distractors.

Resource Barrier

Limited availability of low-resource datasets for training models on distractor generation.

Proposed Solutions: Efforts to build non-English and multi-modal datasets to facilitate better training and evaluation.

Evaluation Barrier

Current automatic metrics for evaluating distractor quality have significant limitations.

Proposed Solutions: Development of new quality metrics and automation of existing evaluation rubrics.

Access Barrier

Some datasets are publicly accessible while others require payment or specific requests for access.

Proposed Solutions: Encouraging open access initiatives for educational datasets and collaboration with creators to broaden accessibility.

Project Team

Elaf Alhazmi

Researcher

Quan Z. Sheng

Researcher

Wei Emma Zhang

Researcher

Munazza Zaib

Researcher

Ahoud Alhazmi

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Elaf Alhazmi, Quan Z. Sheng, Wei Emma Zhang, Munazza Zaib, Ahoud Alhazmi

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies