Chapter 00
Why Basic Math?
Why math is needed to understand deep learning and machine learning, and what math is used.
Math diagram by chapter
Select a chapter to see its diagram below. View the flow of basic math at a glance.
What you learn in Ch01–Ch12
Understanding deep learning and machine learning requires basic math such as functions, exponential and log, limits, derivatives, integrals, and probability and distributions. Ch01–Ch12 cover exactly that. Functions are the basis of input→output; derivatives and gradients are what the model uses to decide where and how much to change parameters when learning; probability and distributions are needed for prediction and uncertainty.
- Ch.01Functions
A function is a rule that assigns one output to each input. Neurons and layers in deep learning are functions.
- Ch.02Exponential and Exponential Functions
Exponentiation is repeated multiplication of the same base; an exponential function fixes the base and uses the exponent as the variable. Used in activation and loss design in deep learning.
- Ch.03Logarithm
A logarithm answers 'how many times we multiply the base to get this number?' It is the inverse of exponentiation and is used with exponentials in loss and probability in deep learning.
- Ch.04Limit and Epsilon-Delta (ε-δ)
A limit describes what happens when we get "arbitrarily close" to some value. Epsilon-delta is the precise way to define that idea and is the basis for derivatives and deep learning.
- Ch.05Continuity
Continuity at a point means the limit exists and equals the function value there. It is the basis for differentiability and for understanding activation and loss functions in deep learning.
- Ch.06Derivative
Differentiation gives the instantaneous rate of change (slope) at a point. The derivative as a function is the basis for gradient descent and backprop in deep learning.
- Ch.07Chain Rule
When you differentiate a function inside another, multiply outer derivative × inner derivative. That's the core of backprop.
- Ch.08Partial Derivative & Gradient
When there are several variables, partial derivative is the derivative w.r.t. one variable with others fixed. The gradient is the vector of those partial derivatives. It's the basis of gradient descent.
- Ch.09Integral
Integration is the inverse of differentiation. It is used for area under a curve, cumulative quantities, and for probability and expectation.
- Ch.10Random Variable & Distribution
A random variable assigns numbers to outcomes of an experiment; a probability distribution summarizes how likely each value is. Used in deep learning for prediction and uncertainty.
- Ch.11Mean & Variance
- Ch.12Uniform & Normal Distribution