Model Card for ML-GT/Llama-3.3-70B-Instruct-Edubuddy
EduBuddy is a Socratic AI Teaching Assistant designed for Georgia Tech’s CS 4641/7641 Machine Learning course.
It integrates fine-tuning and retrieval-augmented generation (RAG) to replicate TA-style pedagogical communication — guiding students through reasoning rather than providing direct answers.
Model Details
Model Description
EduBuddy fine-tunes Llama-3.3-70B-Instruct to emulate the Socratic teaching style observed in 8,197 authentic student–TA conversations from five years of Georgia Tech ML coursework.
It uses a dual-RAG architecture to retrieve:
- Course-specific lecture and homework materials.
 - Structured learning subgoals for scaffolding conceptual understanding.
 
EduBuddy’s responses are contextually grounded, Socratic in nature, and aligned with the structure of the ML course’s assignments and lectures.
- Developed by: Georgia Tech ML TA Team
 - Model type: Instruction-tuned large language model (LLM) fine-tuned for pedagogical dialogue
 - Language(s): English
 - License: No license (research-only use)
 - Finetuned from: 
meta-llama/Llama-3.3-70B-Instruct - Intended users: Instructors, researchers, and educational technologists exploring Socratic AI tutoring
 - Primary domain: University-level Machine Learning education (Georgia Tech CS 4641/7641)
 
Model Sources
- Paper: EduBuddy: A Socratic AI Teaching Assistant combining Fine-Tuning and Retrieval-Augmented Generation for ML Education
 - Repository: [Not released publicly]
 - Demo: None (research prototype only)
 
Uses
Direct Use
- To assist ML students by asking guiding questions and offering step-by-step reasoning in conceptual and coding queries.
 - To emulate TA responses for course-related Q&A forums or tutoring systems.
 
Downstream Use
- Research on Socratic dialogue systems, AI pedagogy, or educational LLM design.
 - Fine-tuning base models for similar instructional contexts.
 
Out-of-Scope Use
- Providing direct answers or code solutions.
 - Use outside educational contexts (e.g., general tutoring, grading automation, or production deployment).
 
Bias, Risks, and Limitations
- Course specificity: Designed for Georgia Tech ML course materials; may perform poorly on unrelated topics.
 - Data artifacts: Training data includes informal language, emojis, and stylistic variation from real forums.
 - Scale dependency: Smaller models (e.g., 8B) underperform significantly; pedagogical reasoning emerges reliably only at large scale (70B+).
 - RAG noise sensitivity: Retrieval documents can introduce irrelevant context when not aligned with queries.
 
Recommendations
Users should treat EduBuddy as a pedagogical research prototype, not a general-purpose tutor.
Before adaptation to new domains, retraining on course-specific dialogues and materials is recommended.