Skip to main content Skip to navigation

Automated Reading Passage Generation with OpenAI's Large Language Model

Project Overview

The document explores the application of OpenAI's GPT-3 in generating automated reading passages for educational assessments, particularly within the framework of the Progress in International Reading Literacy Study (PIRLS). It highlights the potential of generative AI to enhance the efficiency of item development for reading comprehension tests by producing passages that can closely resemble the quality and appropriateness of those crafted by humans. While the research shows promising results, indicating that AI-generated texts can effectively meet educational standards, it also acknowledges certain limitations related to coherence and originality. Therefore, the document stresses the importance of incorporating human oversight in the process to ensure that the generated content aligns with educational objectives and maintains suitability for learners. Overall, the findings suggest that generative AI holds significant promise for improving educational assessments, provided that adequate measures are taken to address its limitations.

Key Applications

Automated Reading Passage Generation using OpenAI's GPT-3

Context: Used for generating reading comprehension passages for fourth graders in the context of the PIRLS assessment.

Implementation: AI-generated passages were created using prompts based on existing reading passages, evaluated by human experts for coherence and appropriateness.

Outcomes: Generated passages were found to be adequate for the target audience, showing promise in matching the text difficulty and engagement levels of original PIRLS passages.

Challenges: Some AI-generated passages were found to be less coherent and more distracting than human-generated texts, with challenges in determining the main topic.

Implementation Barriers

Quality and Technical Limitations

AI-generated passages may lack coherence and structured organization, making it difficult for readers to identify main topics and engage with the content. Additionally, errors in technical or specialized content may occur in AI outputs, necessitating human review for accuracy.

Proposed Solutions: Incorporating a human-in-the-loop evaluation process to assess and refine AI-generated content can mitigate issues of coherence and organization. Furthermore, human editors should review AI-generated passages to correct grammatical and factual errors before usage.

Project Team

Ummugul Bezirhan

Researcher

Matthias von Davier

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Ummugul Bezirhan, Matthias von Davier

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies