Skip to main content Skip to navigation

Generative AI Systems: A Systems-based Perspective on Generative AI

Project Overview

Generative AI (GenAI), particularly through Large Language Models (LLMs), has revolutionized interactions with technology by enhancing natural language processing capabilities, thus paving the way for innovative applications in education. These systems are capable of handling diverse data types—text, images, audio, and video—facilitating multimodal learning experiences that cater to various learning styles. The document explores the intricate architecture and components of GenAI systems, highlighting their potential to personalize learning, provide instant feedback, and support educators in curriculum development. However, it also addresses the challenges inherent in training these complex systems, such as data quality, bias, and the need for robust evaluation methods. By adopting a systems-based approach, the paper underscores the importance of thoughtful design and implementation to maximize the benefits of GenAI in educational settings. The findings suggest that when effectively integrated, GenAI can enhance student engagement, improve learning outcomes, and transform traditional pedagogical methods, ultimately better preparing learners for a rapidly evolving digital landscape.

Key Applications

Generative AI for Information Retrieval and Content Generation

Context: Applications in health-related fields, automatic speech recognition, and image generation, enhancing educational settings by providing accurate information, transcribing spoken content, and creating visual content.

Implementation: Utilizes advanced AI technologies, including Retrieval-Augmented Generation (RAG) with two LLMs (encoder and decoder) for information retrieval, speech-to-text processing using encoder-decoder transformer models for audio signals, and Large Vision Models (LVMs) like Stable Diffusion for generating images from textual prompts.

Outcomes: Improved accuracy and reliability in generating responses based on factual data, state-of-the-art performance in speech recognition tasks, and the ability to create high-quality images from text prompts, thereby enhancing creative workflows and educational content.

Challenges: Complexity in ensuring the relevance and accuracy of retrieved data, resource-intensive model training, performance issues with large model sizes, and potential biases in image generation.

Implementation Barriers

Technical Barrier

The size and complexity of GenAI systems make end-to-end training infeasible with current hardware.

Proposed Solutions: Utilizing pre-trained models and freezing certain components during training.

Operational Barrier

Fine-tuning large models is resource-intensive and may not be quick or cost-effective.

Proposed Solutions: Implementing Parameter-Efficient Fine-Tuning methods (PEFT) to reduce the overhead.

Project Team

Jakub M. Tomczak

Researcher

Contact Information

For information about the paper, please contact the authors.

Authors: Jakub M. Tomczak

Source Publication: View Original PaperLink opens in a new window

Project Contact: Dr. Jianhua Yang

LLM Model Version: gpt-4o-mini-2024-07-18

Analysis Provider: Openai

Let us know you agree to cookies