Skip to main content Skip to navigation

Teaching Astronomy with Large Language Models

Project Overview

The document explores the integration of Large Language Models (LLMs) in undergraduate astronomy education through a specialized tutoring system known as AstroTutor. It underscores the structured incorporation of AI in the learning process, which not only enhances educational outcomes but also fosters AI literacy and preserves critical thinking skills among students. Key applications of the LLMs include delivering personalized feedback that promotes student independence and facilitating LLM-assisted assessments that tackle conventional grading issues. The findings suggest that these AI tools can significantly improve the learning experience, while also highlighting the necessity of thorough documentation and critical reflection in their application. Overall, the study illustrates the transformative potential of generative AI in educational settings, particularly in providing tailored support and addressing longstanding challenges in assessment and feedback.

Key Applications

LLM-based Assessment and Grading System

Context: Assessment and grading in final-year undergraduate astronomy courses, focusing on student understanding, skills evaluation, and homework assignments in astrostatistics and machine learning.

Implementation: A comprehensive assessment system was developed using various LLM models, including AstroTutor principles for individualized assessments and automated grading. The system facilitated Socratic-style interviews and provided immediate feedback, maintaining grading consistency and accuracy.

Outcomes: ['Students showed improved AI literacy and developed effective prompting strategies.', 'LLM assessments demonstrated strong consistency with human evaluations and provided detailed feedback on student misconceptions.', 'LLM grading showed strong correlation with human grading, offering detailed feedback and improving grading efficiency.']

Challenges: ['Initial dependency on LLMs and difficulty in balancing prompt specificity.', 'Concerns regarding ethical implications and privacy in automated assessments.', 'LLMs exhibited systematic bias towards stricter grading compared to human evaluators.']

Implementation Barriers

Technical Barrier

Students struggled with prompt engineering, leading to ineffective interactions with LLMs. Limited familiarity with modern IDEs and computational tools among students.

Proposed Solutions: Educators should provide training on effective prompt crafting and contextualizing queries, as well as comprehensive training on Integrated Development Environments (IDEs) and API usage.

Ethical Barrier

Concerns about academic integrity and privacy in utilizing LLMs for assessments.

Proposed Solutions: Implement strict guidelines for privacy and ensure that human oversight remains in grading processes.

Dependence Barrier

Initial student reliance on LLMs for answers rather than using them as learning tools.

Proposed Solutions: Encourage reflective documentation of AI usage to foster critical thinking and independence.

Project Team

["Yuan-Sen Ting", "Teaghan O"Briain"]

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: ["Yuan-Sen Ting", "Teaghan O"Briain"]

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies