Chapter 12
Uniform & Normal Distribution
Uniform distribution spreads probability evenly over an interval; normal distribution is bell-shaped around the mean. Used in AI for initialization, noise, and priors.
Math diagram by chapter
Select a chapter to see its diagram below. View the flow of basic math at a glance.
Uniform & Normal Distribution
Uniform distribution spreads probability evenly over an interval; normal distribution is bell-shaped around the mean. Used in AI for initialization, noise, and priors.
Uniform & Normal distribution
Uniform and normal are two of the most used continuous distributions. Their shape is determined by mean and variance (Ch10–Ch11).
Uniform distribution — Same height over an interval . Density for . Used when outcomes are equally likely (e.g. one face of a die).
Uniform mean is , variance is . The center of the interval is the mean; wider intervals give larger variance.
Normal distribution — Determined by mean and standard deviation . Density . Fits measurement error, height, scores (values cluster near the mean).
Bell curve — Normal is highest at the mean and tapers off on both sides. Symmetric about ; about 68% in , about 95% in .
Why these two? — Uniform is used when we have no prior information (initialization, flat prior). Normal appears for noise/error and via the central limit theorem (averages tend to be normal), so both are central in AI and statistics.
Priors — In Bayesian settings, uniform is a common 'uninformative' prior; normal is used when we have beliefs about mean and variance.
Noise and error — Regression errors, VAE and diffusion noise are often modeled as normal. The math is simple and matches the central limit idea.
Central limit theorem — With many independent trials, the sample mean tends to a normal distribution. So confidence intervals and hypothesis tests rely on normality.
In deep learning and ML — Weight initialization (uniform/normal), dropout and noise (normal), VAE latent space (normal), diffusion (Gaussian) all use these distributions.
Initialization — Weights are drawn from uniform or normal. Too large or biased values hurt training; usually small-variance normal is used.
Noise — VAE samples the latent vector from a normal; diffusion models add and remove Gaussian noise step by step.
Regression — Assuming normal errors makes least squares (OLS) equivalent to maximum likelihood. Prediction intervals use .
Bayesian — Uniform or normal priors are common; after observing data we compute the posterior. Neural network weights can have normal priors.
Math flow — Ch10 random variables and distributions, Ch11 mean and variance, then Ch12 two concrete distributions (uniform and normal). Knowing these helps read 'initialization', 'noise', and 'prior' in AI papers.
Uniform — On , density , mean , variance . Normal — Mean , variance ; interval probabilities from standard normal table or calculator.
Example (uniform). On , mean is , variance , standard deviation .
Example (normal). For mean and standard deviation , about 68% lie in –, about 95% in –.