Skip to main content Skip to navigation

Not a Swiss Army Knife: Academics' Perceptions of Trade-Offs Around Generative Artificial Intelligence Use

Project Overview

The document explores the transformative role of generative AI (Gen AI) in education, emphasizing its potential to democratize access to knowledge and function as a valuable research assistant for academic professionals. However, it also addresses significant challenges such as trust concerns, threats to academic integrity, potential biases in AI-generated content, and the detrimental effects on critical thinking skills among students. The findings underscore the necessity for well-defined implementation strategies, ethical frameworks, and educational policies that can harness the advantages of Gen AI while mitigating associated risks. By addressing these issues, the integration of generative AI into educational settings can be optimized to enhance learning outcomes and foster a more equitable academic environment.

Key Applications

AI as a Writing and Learning Assistant

Context: Higher education environment for faculty and students, focused on academic research, writing, and democratization of knowledge through AI tools that enhance accessibility and engagement.

Implementation: Utilization of AI technologies like ChatGPT to assist in writing tasks (proofreading, citation formatting, overcoming writer's block) while also encouraging critique of AI-generated content and adapting curricula to integrate AI tools. Additionally, AI is used to provide access to learning resources and support for diverse socioeconomic groups.

Outcomes: Improved efficiency in writing and research tasks, enhanced brainstorming capabilities, increased engagement with knowledge resources among underrepresented groups, and development of critical thinking skills.

Challenges: Loss of personal voice in writing, overreliance on AI leading to diminished critical thinking, need for faculty training on AI use, potential for academic dishonesty, and the risk of exacerbating social inequalities due to unequal access to technology.

Implementation Barriers

Ethical

Concerns regarding biases in AI outputs, academic integrity, and the need for transparency in AI training data

Proposed Solutions: Development of ethical guidelines for AI use in education and increased transparency in AI algorithms

Technical

Inaccuracies and misinformation generated by AI models

Proposed Solutions: Implementation of fact-checking mechanisms and transparency in AI algorithms

Societal

Digital divide affecting access to AI technologies among different socioeconomic groups

Proposed Solutions: Promoting AI literacy and ensuring equitable access to technology in education

Project Team

Afsaneh Razi

Researcher

Layla Bouzoubaa

Researcher

Aria Pessianzadeh

Researcher

John S. Seberger

Researcher

Rezvaneh Rezapour

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Afsaneh Razi, Layla Bouzoubaa, Aria Pessianzadeh, John S. Seberger, Rezvaneh Rezapour

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies