You can download the lectures here. We will try to upload lectures prior to their corresponding classes.

  • Lecture 1 - Introduction to Responsible AI
    tl;dr: Overview of AI history, example models (NLP/LLM, Computer Vision, Robotics), and core challenges in responsible AI: explainability, uncertainty, and potential harms
    [slides]

    Topics Covered:

    • What is AI: Agents, perceptions, actions, and environment
    • AI Timeline: From 1900s to 2025+ (AI winters, transformers, LLMs, AGI)
    • Example AI Models:
      • Natural Language Processing (LLM): Transformer, code generation
      • Computer Vision: CLIP, image captioning, medical segmentation
      • Robotics: Vision-Language-Action models (RT-2)
    • AI Responsibility Challenges:
      • Explainability: Understanding model decisions (from linear models to LLMs), Chain-of-Thought limitations, mechanistic interpretability
      • Uncertainty: Aleatoric vs. epistemic uncertainty, impact on LLM performance, robotics navigation, self-driving vehicles
      • AI Harms: Safety (hallucination, jailbreaking, prompt injection, data poisoning), privacy (data leakage, memorization), fairness
  • Lecture 2 - Transformer and Vision Models
    tl;dr: Deep dive into Transformer architecture and its applications in computer vision, including ViT, CLIP, SAM, and multi-modal models
    [slides]

    Topics Covered:

    • Transformer Architecture: Positional embedding, Multi-head attention, Feed-forward layers, Residual connections, Layer normalization
    • Vision Transformer (ViT): Applying transformers to images, patch embedding, scaling laws
    • ViT Variants: MAE (self-supervised pre-training), Swin Transformer (multi-scale patches), DeiT (distillation)
    • CLIP: Contrastive language-image pre-training, zero-shot classification
    • SAM: Segment Anything Model, promptable segmentation
    • LLaVA: Visual instruction tuning, multi-modal reasoning
    • Responsibility Issues: Uncertainty estimation, out-of-distribution detection, fairness, hallucination
  • Lecture 3 - Model-Agnostic Explanations and SHAP
    tl;dr: Post-hoc explanation methods that work on any AI model by approximating complex logic with simpler, local models or using game theory to calculate feature importance
    [slides]

    Topics Covered:

    • Local vs. Global Explanations
    • LIME (Local Interpretable Model-agnostic Explanations)
    • Shapley Values & SHAP
    • Properties of SHAP
  • Lecture 4 - Uncertainty in AI
    tl;dr: How to understand why LLMs produce certain outputs—covering token-level attributions, perturbation and counterfactual explanations, reasoning structures (CoT, ToT), and mechanistic interpretability inside transformers.
    [slides]

    Topics Covered:

    • Gradient-based explanations
    • Perturbation & counterfactual explanations
    • Chain-of-Thought, Tree / Graph of Thoughts, and adaptive reasoning
  • Lecture 5 - AI Safety and Robustness
    tl;dr: The vulnerability of AI systems to malicious attacks and environmental noise.
    [slides]

    Topics Covered:

    • Adversarial Examples
    • Attack Methods
    • Defensive Strategies
  • Lecture 6 - Backdoor Attacks in AI
    tl;dr: Backdoor Attacks, a form of data poisoning where a model is trained to behave normally on clean data but perform specific malicious actions when a hidden trigger is present.
    [slides]

    Topics Covered:

    • Backdoor Attack Mechanism
    • Attack vs. Adversarial Examples
    • Poisoning Strategies
  • Lecture 7 - Algorithmic Fairness
    tl;dr: This lecture explores the growing tension between AI performance and data privacy, highlighting how models can unintentionally memorize and leak sensitive training information
    [slides]

    Topics Covered:

    • Data Leakage Risks
    • Membership Inference Attacks
    • Reconstruction Attacks
  • Lecture 8 - Privacy in AI
    tl;dr: Understanding the broader environmental and social implications of AI
    [slides]

    Topics Covered:

    • Environmental impact of AI
    • Social and economic implications
    • Sustainable AI practices
  • Lecture 9 - AI Governance and Intellectual Property
    tl;dr: This lecture examines the legal and regulatory frameworks governing AI, focusing on how governments manage AI risks and how copyright laws apply to AI-generated content
    [slides]

    Topics Covered:

    • AI Regulation Frameworks
    • Risk Categorization
    • Intellectual Property
  • Lecture 10 - Reinforcement Learning and Safety
    tl;dr: This lecture introduces a more sophisticated approach to Reinforcement Learning (RL) that models the entire probability distribution of future rewards rather than just their average
    [slides] [distributed RL]

    Topics Covered:

    • From Mean to Distribution
    • Aleatoric vs. Epistemic Uncertainty
  • Lecture 11 - LLM-based Agentic Systems
    tl;dr: This lecture explores the transition from static Large Language Models (LLMs) to autonomous agents that can reason, use tools, and interact with the world
    [slides]

    Topics Covered:

    • From LLM to Agent
    • Agentic Architectures
    • Multi-Agent Systems
  • Lecture 12 - LLM-based Agentic Systems
    tl;dr: This lecture explores the transition from static Large Language Models (LLMs) to autonomous agents that can reason, use tools, and interact with the world
    [slides]

    Topics Covered:

    • From LLM to Agent
    • Agentic Architectures
    • Multi-Agent Systems