Pedagogical goal: Students can translate technical AI concepts to non-technical audiences.
This is the culmination. Everything prior was in service of this.
Time: 4–5 hours (recommend full week to practice)
Before Phase 4: Orientation Activity (1 hour)
Purpose: Show students why this matters.
Activity:
- Invite a non-technical colleague (or manager, or peer) to class
- Have them ask a question: “Why do LLMs hallucinate? Should we be worried?”
- Have a student attempt to answer (no prep)
- Debrief: “What was hard? What assumptions did we make?”
- Frame: “Phase 4 teaches you to answer this well.”
Resource 9–11 — Teaching Strategy
Session 1: Understand the Architecture (60 min)
- Show Resource 9 (Illustrated Transformer)
- Goal: students understand attention (not every detail, but the concept)
- Activity: “Annotate this attention visualization. Which words is the model paying attention to? Why?”
Session 2: Understand Failures (45 min)
- Read Resource 10 (hallucinations)
- Discuss: “Why can’t we just train it better to not hallucinate?”
- Activity: “Brainstorm three ways to detect hallucinations”
Session 3: The Explanation Skill (60 min)
- Read Resource 11 (interpretability tools)
- Discuss: “We have tools to explain decisions. Are they good enough?”
- Activity: “Choose one misconception from Phase 4 Capstone (below). Write a 2-min explanation that addresses it.”
Phase 4 Capstone Activities (In Depth)
Activity 1: Misconception Buster
Misconception 1: “LLMs understand language”
- Your reframe: “LLMs predict language. Understanding would mean grasping meaning; prediction means finding statistical patterns.”
- Student task: Write a 1-min explanation of this distinction for a CEO.
Misconception 2: “More training data = no hallucinations”
- Your reframe: “Hallucinations aren’t a data quality problem; they’re architectural. Hallucination happens because the model is optimized to generate plausible text, not factual text.”
- Student task: Explain to a customer why we can’t promise “no hallucinations.”
Misconception 3: “Attention visualization shows what the model thought”
- Your reframe: “Attention shows what words mattered, not why they mattered or how they influenced the decision.”
- Student task: Show an attention visualization and explain what it does and doesn’t tell us.
Misconception 4: “Explainability = safety”
- Your reframe: “Explaining a bad decision doesn’t make it good. Transparency and defense mechanisms (monitoring, guardrails) matter more than explanation.”
- Student task: Propose a defense strategy for a company deploying LLMs, combining explanation + monitoring + guardrails.
Activity 2: Role-Play Scenarios
Scenario A: The Skeptical Executive
- Student role-plays explaining hallucinations to a CEO who’s read about ChatGPT errors
- Class observes and gives feedback: “What was clear? What was confusing?”
Scenario B: The Non-Technical Customer
- Student explains why the LLM gave a wrong answer, what we did about it, and what the customer should do
- Class rates clarity and reassurance level
Scenario C: The Peer Who Disagrees
- Student defends a position (e.g., “we should use LLMs here despite hallucination risk”)
- Peer argues the opposite
- Class evaluates quality of reasoning
Phase 4 Student Talk-Back (Assessment — The Capstone)
Final project: “Choose a non-technical audience (CEO, customer, journalist, board member). Write or record a 3-minute explanation addressing this question: ‘What’s an LLM, where does it work, where does it fail, and how do you know?’”
Grade on:
- No jargon (or jargon explained)
- Accurate (doesn’t overstate capability)
- Complete (answers all four parts)
- Persuasive (audience feels informed, not scared)
Rubric:
- Excellent: Clear analogies, acknowledges limitations, proposes solutions (monitoring/guardrails)
- Good: Clear analogies, mentions limitations, but shallow on solutions
- Adequate: Some jargon leakage, mentions hallucinations but doesn’t explain why
- Below target: Heavy jargon, overstates capability or dismisses risks