Ch.00
Advanced DL: Large Models and Generative AI Paradigm
Advanced Deep Learning (Ch.00) is the entry point that connects “why models got so large” with “how generative AI systems actually work.” We go beyond learning representations from data: how large Transformers build contextual understanding, predict the next token, and then how we align, control, and deploy those models for real users.
An advanced roadmap toward large generative models
This roadmap gradually fills from Ch01 onward, showing how each chapter contributes to the full system.
What you will learn in Ch01~Ch26
- Ch.01Transformer 1: Self-Attention and Parallelization
- Ch.02Transformer: Positional Encoding and Feed-Forward
- Ch.03Transformer Lineage: Encoder (BERT) vs Decoder (GPT)
- Ch.04Attention Optimization: FlashAttention and Sparse Attention
- Ch.05Vision Transformer (ViT) and Image Patches
- Ch.06Self-Supervised Learning
- Ch.07Prompt Engineering and In-Context Learning
- Ch.08PEFT 1: PEFT and LoRA
- Ch.09PEFT 2: QLoRA and Quantization Tuning
- Ch.10Alignment and RLHF
- Ch.11DPO: Alignment without Reinforcement Learning
- Ch.12RAG: Hallucination Control Architecture
- Ch.13LLM Agents and Tool Use
- Ch.14GNN and Message Passing
- Ch.15XAI in Deep Learning: Grad-CAM
- Ch.16Autoencoder and Unsupervised Dimensionality Reduction
- Ch.17VAE: Probability-Based Generative Space
- Ch.18GAN Basics
- Ch.19Conditional GAN (cGAN) and Applications
- Ch.20Diffusion Model 1: Forward and Reverse Process
- Ch.21Diffusion Model 2: Latent Diffusion
- Ch.22Vision-Language Model and CLIP
- Ch.23Speech-to-Text and Audio Processing
- Ch.24Model Compression and Knowledge Distillation
- Ch.25Inference Optimization and Deployment
- Ch.26Advanced DL Summary: AI Architecture and Future
What is Advanced DL? (Generative AI system view)
Foundation models / LLMs are trained with the objective of predicting the next token. In other words, they maximize , learning language flow and patterns that go beyond simple grammar.
A practical way to understand generative AI is to split it into stages: pretraining (broad knowledge), instruction / SFT (follow user intent), and alignment (preference, safety, and reduced hallucinations).
The backbone is mostly Transformers. Self-attention creates “token-to-token” context, and feed-forward + normalization layers refine it so the model stays consistent even with long contexts.
Bigger models can improve capability, but they also make training less stable and dramatically increase cost. Advanced DL therefore focuses on more than accuracy: training stability, efficiency (compute/memory), and reproducibility.
In the real world, generative AI is judged by trust: truthfulness, safety, and reliability. Achieving that requires alignment, evaluation, and control mechanisms.
Finally, deployment constraints (latency, cost, server limits) matter. So advanced DL continues from training to inference optimization, compression, and serving strategies.
In production, systems usually follow a pipeline like `text/image -> tokenization -> context window -> Transformer -> decoding (greedy/beam/sample)`. Decoding strategy and prompt design strongly affect output quality.
Alignment and control can be done in multiple ways. For example, RLHF / DPO uses preferences to improve the model, and RAG retrieves external knowledge to ground answers.
From a product perspective, tool use, caching/batching, and optimization such as quantization or knowledge distillation are part of the whole stack. The same base model can feel very different depending on how you run it.
- TopicNext-token prediction
- Role in an AI systemBuilds general language ability
- Connection concept in this courseprobabilistic generation, representation learning
- TopicInstruction / SFT
- Role in an AI systemMakes responses follow intent
- Connection concept in this coursedata format, fine-tuning
- TopicAlignment
- Role in an AI systemControls preferences, safety, truthfulness
- Connection concept in this coursepreference learning, reward model
- TopicRAG / grounded generation
- Role in an AI systemReduces ungrounded claims
- Connection concept in this courseretrieval, embeddings, context assembly
- TopicInference optimization
- Role in an AI systemCuts latency and cost
- Connection concept in this coursequantization, caching, distillation
| Topic | Role in an AI system | Connection concept in this course |
|---|---|---|
| Next-token prediction | Builds general language ability | probabilistic generation, representation learning |
| Instruction / SFT | Makes responses follow intent | data format, fine-tuning |
| Alignment | Controls preferences, safety, truthfulness | preference learning, reward model |
| RAG / grounded generation | Reduces ungrounded claims | retrieval, embeddings, context assembly |
| Inference optimization | Cuts latency and cost | quantization, caching, distillation |