Day 3 Overview: Understanding AI Under the Hood
Focus: How AI actually works: from words to predictions
Learning Objectives
By the end of Day 3, you will:
- Understand how computers turn words into mathematical coordinates (embeddings)
- Recognize why "Large" matters in LLMs (scale and attention mechanisms)
- Explain why AI hallucinates fake papers and non-existent data
- Navigate the difference between truth and probability in AI outputs
Why This Session Matters
As engineers, you trust tools you understand. You've been using AI, but to trust it responsibly, you need to look under the hood.
Core Insight
AI doesn't "know" things. It predicts patterns in mathematical space.
Understanding this changes how you use it—from blind trust to informed application.
Session Structure
Part 1: NLM Foundations
How Words Become Coordinates
- Tokenization: Breaking text into unit cells
- Embeddings: The mathematical map of meaning
- Semantic space: Words as points in multidimensional space
- Why "Steel" and "Iron" are close neighbors
Key Analogy: Just like materials occupy positions in property space (density, strength, conductivity), words occupy positions in semantic space.
Part 2: The "Large" in LLM
Scaling Up & Understanding Context
- The context problem: "Lead" the metal vs "lead" investigator
- Attention mechanism: Dynamic weighting system
- Why "Large" = billions of parameters
- From pattern matching to probability engines
Key Analogy: Attention is like a dynamic weighting system in composite design—adjusting importance based on surrounding elements.
Part 3: Truth vs. Probability
Why Models Hallucinate & How to Ground Them
- The "Stochastic Parrot" problem
- Why AI fabricates plausible-sounding citations
- The difference between pattern completion and fact retrieval
- Grounding techniques: RAG, temperature control, chain-of-thought
Key Principle: AI prioritizes plausibility over veracity—your job is to ground it in reality.
What You'll Gain Today
Technical Understanding
- Demystification: See through the "AI magic" to understand the mechanism
- Critical evaluation: Know when to trust vs verify AI outputs
- Better prompts: Understand why certain instructions work better than others
Practical Skills
- Hallucination detection: Spot fabricated data and citations
- Grounding techniques: Force AI to stick to actual sources
- Temperature control: Adjust creativity vs certainty
- RAG implementation: Ground AI in your documents
Engineering Perspective
Throughout this session, we'll use materials science thinking to understand AI:
| Materials Concept | AI Equivalent |
|---|---|
| Property space (density, strength) | Semantic space (word meanings) |
| Phase diagrams | Probability distributions |
| Crystal lattice unit cells | Tokens |
| Composite weighting | Attention mechanism |
| Quality control | Verification protocols |
You already have the mental models,we're just mapping them to AI.
Success Criteria
You're ready for Day 4 when you can:
- Explain to a colleague how AI turns words into coordinates
- Describe why context matters using the attention mechanism
- Predict when AI is likely to hallucinate
- Apply at least 2 grounding techniques to prevent fabrication
- Recognize the difference between plausibility and accuracy
Let's Begin!
Ready to understand what's really happening when you type a prompt?
Next: NLM Foundations: How words become coordinates →