Skip to main content Skip to navigation

LMCD: Language Models are Zeroshot Cognitive Diagnosis Learners

Project Overview

The document introduces a groundbreaking framework known as LMCD (Language Models as Zeroshot Cognitive Diagnosis Learners), which leverages large language models (LLMs) to advance cognitive diagnosis in educational contexts. By addressing the cold-start problem that traditional cognitive diagnosis models face due to limited interaction data, LMCD enhances the assessment of student understanding. The framework operates in two distinct phases: Knowledge Diffusion, which generates enriched content related to exercises and knowledge concepts to strengthen semantic connections, and Semantic-Cognitive Fusion, which combines this enriched content with students' cognitive states to develop comprehensive profiles of their learning. The implementation of LMCD has yielded notable improvements in performance, particularly in exercise-cold and cross-domain cold-start situations, outperforming existing methodologies. Overall, this innovative use of generative AI in education demonstrates its potential to personalize learning experiences and improve cognitive diagnostic capabilities, ultimately leading to better educational outcomes.

Key Applications

LMCD (Language Models as Zeroshot Cognitive Diagnosis Learners)

Context: The framework is designed for online education environments targeting students in various educational stages, particularly where mathematical exercises are involved.

Implementation: LMCD was implemented by leveraging LLMs to generate enriched content and integrate student cognitive states into cognitive diagnosis models.

Outcomes: The framework significantly outperformed state-of-the-art methods in cold-start scenarios, improving the accuracy of cognitive diagnosis and personalized learning.

Challenges: The main challenges include diagnosing new students without prior data and the computational demands of the LLM-based approach.

Implementation Barriers

Technical Barrier

The LLM-based approach involves substantial computational resources, making it less suitable for time-sensitive applications.

Proposed Solutions: Potential solutions include developing smaller, more efficient models as LLM technology advances.

Data Barrier

The method struggles with diagnosing new students, as it requires existing student data to create embedding representations.

Proposed Solutions: One potential workaround is substituting new students with trained students having similar response patterns.

Project Team

Yu He

Researcher

Zihan Yao

Researcher

Chentao Song

Researcher

Tianyu Qi

Researcher

Jun Liu

Researcher

Ming Li

Researcher

Qing Huang

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Yu He, Zihan Yao, Chentao Song, Tianyu Qi, Jun Liu, Ming Li, Qing Huang

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies