Lectures
You can download the lectures here. We will try to upload lectures prior to their corresponding classes.
-
Lecture 1 - Introduction to Responsible AI
tl;dr: Overview of AI history, example models (NLP/LLM, Computer Vision, Robotics), and core challenges in responsible AI: explainability, uncertainty, and potential harms
[slides]
Topics Covered:
- What is AI: Agents, perceptions, actions, and environment
- AI Timeline: From 1900s to 2025+ (AI winters, transformers, LLMs, AGI)
- Example AI Models:
- Natural Language Processing (LLM): Transformer, code generation
- Computer Vision: CLIP, image captioning, medical segmentation
- Robotics: Vision-Language-Action models (RT-2)
- AI Responsibility Challenges:
- Explainability: Understanding model decisions (from linear models to LLMs), Chain-of-Thought limitations, mechanistic interpretability
- Uncertainty: Aleatoric vs. epistemic uncertainty, impact on LLM performance, robotics navigation, self-driving vehicles
- AI Harms: Safety (hallucination, jailbreaking, prompt injection, data poisoning), privacy (data leakage, memorization), fairness
-
Lecture 2 - Transformer and Vision Models
tl;dr: Deep dive into Transformer architecture and its applications in computer vision, including ViT, CLIP, SAM, and multi-modal models
[slides]
Topics Covered:
- Transformer Architecture: Positional embedding, Multi-head attention, Feed-forward layers, Residual connections, Layer normalization
- Vision Transformer (ViT): Applying transformers to images, patch embedding, scaling laws
- ViT Variants: MAE (self-supervised pre-training), Swin Transformer (multi-scale patches), DeiT (distillation)
- CLIP: Contrastive language-image pre-training, zero-shot classification
- SAM: Segment Anything Model, promptable segmentation
- LLaVA: Visual instruction tuning, multi-modal reasoning
- Responsibility Issues: Uncertainty estimation, out-of-distribution detection, fairness, hallucination
-
Lecture 3 - Model-Agnostic Explanations and SHAP
tl;dr: Post-hoc explanation methods that work on any AI model by approximating complex logic with simpler, local models or using game theory to calculate feature importance
[slides]
Topics Covered:
- Local vs. Global Explanations
- LIME (Local Interpretable Model-agnostic Explanations)
- Shapley Values & SHAP
- Properties of SHAP
-
Lecture 4 - Uncertainty in AI
tl;dr: How to understand why LLMs produce certain outputs—covering token-level attributions, perturbation and counterfactual explanations, reasoning structures (CoT, ToT), and mechanistic interpretability inside transformers.
[slides]
Topics Covered:
- Gradient-based explanations
- Perturbation & counterfactual explanations
- Chain-of-Thought, Tree / Graph of Thoughts, and adaptive reasoning
-
Lecture 5 - AI Safety and Robustness
tl;dr: The vulnerability of AI systems to malicious attacks and environmental noise.
[slides]
Topics Covered:
- Adversarial Examples
- Attack Methods
- Defensive Strategies
-
Lecture 6 - Backdoor Attacks in AI
tl;dr: Backdoor Attacks, a form of data poisoning where a model is trained to behave normally on clean data but perform specific malicious actions when a hidden trigger is present.
[slides]
Topics Covered:
- Backdoor Attack Mechanism
- Attack vs. Adversarial Examples
- Poisoning Strategies
-
Lecture 7 - Algorithmic Fairness
tl;dr: This lecture explores the growing tension between AI performance and data privacy, highlighting how models can unintentionally memorize and leak sensitive training information
[slides]
Topics Covered:
- Data Leakage Risks
- Membership Inference Attacks
- Reconstruction Attacks
-
Lecture 8 - Privacy in AI
tl;dr: Understanding the broader environmental and social implications of AI
[slides]
Topics Covered:
- Environmental impact of AI
- Social and economic implications
- Sustainable AI practices
-
Lecture 9 - AI Governance and Intellectual Property
tl;dr: This lecture examines the legal and regulatory frameworks governing AI, focusing on how governments manage AI risks and how copyright laws apply to AI-generated content
[slides]
Topics Covered:
- AI Regulation Frameworks
- Risk Categorization
- Intellectual Property
-
Lecture 10 - Reinforcement Learning and Safety
tl;dr: This lecture introduces a more sophisticated approach to Reinforcement Learning (RL) that models the entire probability distribution of future rewards rather than just their average
[slides] [distributed RL]
Topics Covered:
- From Mean to Distribution
- Aleatoric vs. Epistemic Uncertainty
-
Lecture 11 - LLM-based Agentic Systems
tl;dr: This lecture explores the transition from static Large Language Models (LLMs) to autonomous agents that can reason, use tools, and interact with the world
[slides]
Topics Covered:
- From LLM to Agent
- Agentic Architectures
- Multi-Agent Systems
-
Lecture 12 - LLM-based Agentic Systems
tl;dr: This lecture explores the transition from static Large Language Models (LLMs) to autonomous agents that can reason, use tools, and interact with the world
[slides]
Topics Covered:
- From LLM to Agent
- Agentic Architectures
- Multi-Agent Systems
