|
Tue, Aug 26 |
Lec 1. Introduction
What: demos, formulation
Why: opportunities, risks
How: syllabus, logistics
|
The Social Dilemma
|
lab1
|
|
Language Generation (Week 1-7)
|
Symbolic Representation
|
Thu, Aug 28 |
Lec 2. Logic and Grammars
Logic: propositional logic, first-order logic
Grammar: lexicon, syntax, Eliza
|
Eliza
|
hw1
(sol)
|
|
Statistical Representation
|
Tue, Sep 2 |
Lec 3. N-Gram Language Models
Probability: language modeling
Statistics: MLE, MAP
N-Gram: smoothing, perplexity
|
Sasha Rush:
LLM In 5 Formulas
|
lab2
|
survey |
Neural Representation
|
Thu, Sep 4 |
Lec 4. Linear Models
Task: regression, classification
Data: (X, Y-continuous/discrete)
Model: linear layer, sigmoid/softmax
Loss: MSE/cross entropy
Optimization: analytical/SGD
|
3Blue1Brown:
Essence of Linear Algebra,
Essence of Calculus
|
hw2 (sol)
|
|
Tue, Sep 9 |
Lec 5. Multi-Layer Perceptrons
Model: Activation, Dropout layer
Optimization: Backpropagation
|
|
lab3
|
|
Thu, Sep 11 |
Lec 6. Word Embeddings
Counting: Word-context matrix
Learning: NLM, Word2Vec (Skip-gram, CBOW)
Vector Properties: Analogies
|
Papers:
NLM,
Word2Vec
|
hw3
(sol)
|
|
Tue, Sep 16 |
Lec 7. Recurrent Neural Networks
RNN: Hidden state & sequential modeling
LSTM: Gates & cell state for long memory
NLP applications: encoder, decoder
|
Papers:
RNN,
LSTM
|
lab4
|
|
Thu, Sep 18 |
Lec 8. Attention Module
Machine Translation: Seq2Seq
Attention Mechanism: Query, Key, Value
|
Papers:
Seq2Seq,
Attention
|
hw4
(sol)
|
team info due |
Tue, Sep 23 |
Lec 9. Transformer Model
Encoder: MHA, FFN,LayerNorm, Residual Connection
Decoder: Masked Attn, Cross Attn
Embedding: BPE, Sinuositional Encoding
|
Attention is All You Need!
Paper:
Transformer
Code:
Annotated Transformer
|
|
|
Thu, Sep 25 |
Lec 10. GPT 1-3
GPT-1: Finetuning
GPT-2: Zero-shot
GPT-3: Few-shot (In-context Learning)
|
Scaling is all you need!
Papers:
GPT-1,
GPT-2,
GPT-3
|
lab5
|
problem statement due (Fri) |
Tue, Sep 30 |
Lec 11. InstructGPT
Supervised Fine-Tuning
Reinforcement Learning from Human Feedback
|
Alignment is all you need!
Paper:
InstructGPT
|
|
|
Thu, Oct 2 |
Lec 12. Reinforcement Learning Basics
RL Framework: state, action, reward
Value-based RL: Bellman Equation (Q-learning)
Policy-based RL: ε-greedy
|
|
hw5
|
|
Tue, Oct 7 |
Lec 13. ChatGPT
RLHF: PPO, DPO
DeepSeekR1: GRPO
|
|
|
|
Thu, Oct 9 |
Mid-term Exam
|
topics,
practice
(sol)
|
|
|
Tue, Oct 14 |
No Class (Happy Fall Break) |
|
|
|
Image Generation (Week 8-12)
|
Thu, Oct 16 |
Lec 14. Image Basics and Filtering
Digital Representation: 2D arrays, grayscale, RGB
channels Convolution Operation: Mathematical
definition & properties Classic Filters:
Gaussian blur, Laplacian edge detection
|
|
|
|
Probability and Statistics
|
Tue, Oct 21 |
Lec 15. Statistical Image Modeling
MRFs & CRFs: Spatial dependencies & energy
functions Patch-Based Synthesis: Efros & Leung,
PatchMatch Early Deep Models: RBMs, Deep Belief
Networks
|
|
|
|
Deep Learning
|
Thu, Oct 23 |
Lec 16. CNN Models for Image Generation
CNN Architecture: Convolutional, pooling, activation
layers Encoder-Decoder: U-Net, skip
connections, transposed convs Advanced CNNs:
ResNet, DenseNet, attention mechanisms
|
|
|
Game Theory
|
Thu, Oct 30 |
Lec 17. Generative Adversarial Networks (GANs)
GAN Framework: Generator vs. discriminator game Architecture
Evolution: DCGAN, Progressive Growing, StyleGAN Challenges:
Mode collapse, training instability, solutions
|
|
|
Information Theory
|
Tue, Oct 28 |
Lec 16. Variational Autoencoders (VAEs)
VAE Architecture: Encoder q(z|x), decoder p(x|z),
prior p(z) ELBO Objective: Reconstruction + KL
divergence Training: Reparameterization trick,
KL vanishing
|
|
|
Dynamical Systems
|
Tue, Nov 4 |
Lec 18. Neural ODEs
Autoregressive Principle: Pixel-by-pixel
generation PixelRNN/PixelCNN: Sequential
modeling with LSTM/conv Image Transformers:
Self-attention over image patches
|
|
|
Thu, Nov 6 |
Lec 19. Normalizing Flows
Change of Variables: Exact likelihood computation Flow
Architectures: RealNVP, Glow, continuous flows Properties:
Invertibility, differentiability, composition
|
|
|
Stochastic Process
|
Tue, Nov 11 |
Lec 20. Stochastic Differential Equations
Forward Process: Gradual noise addition to images Reverse
Process: Learning to denoise step by step DDPM:
Denoising diffusion probabilistic models
|
|
|
Thu, Nov 13 |
Lec 21. Diffusion Models
Text-to-Image: DALL·E 2, Midjourney, Imagen Advanced
Guidance: Classifier & classifier-free guidance Latent
Diffusion: Stable Diffusion, faster training
|
|
|
Guest Lectures (Week 13-14)
|
Tue, Nov 18 |
Lec 22. Image Style Transfer
Siyu Huang (Clemson University)
|
|
|
Thu, Nov 20 |
Lec 23. Molecule Generation for Drug Discovery
Wengong Jin (Northeastern University)
|
|
|
Tue, Nov 25 |
Lec 24. Biomedical 3D Generation
Jiancheng Yang (Aalto University)
|
|
|
|
|
Thu, Nov 27 |
No Class (Happy Thanksgiving) |
|
|
|
Finals (Week 15-16)
|
Tue, Dec 2 |
Project Presentations I
|
|
|
|
Thu, Dec 4 |
Project Presentations II
|
|
|
|
Thu, Dec. 11 |
|
|
|
report/code due |