nbammpuzzEnjMIeF4pY23vVQA9UM1HC4C+f322rBrBDJEqrQ72OCF7TiFc19H0y3GsyFv6HAMcxWDf0bGPN0VjeW8/657Hxgzi1YqLlhZe4WSALqFqt1mbbD/0iEyvddzKVtC04YVYPmSZhPZxbw13ixt782qoUYnq4wsiRcCE9Z3OCe39wLJru4DaDumkcl2PGhgNOK0HCBQBB6ejkluECyOxAvUwwcP5Sy1cZaG3SzJFLqEyfb2LXs1CHvF2RdfQ1ZStlA5AOwK2cIqXOhWheHUF4WretkwftAw1JD96YDP0jyWLxC5Bg77B7EDqy3LQ9xgIRWXefgYBJDRznIDqAQ6RSwBofCFVwrHNdNWir8ftVv1yNfg6i5zAEuhmLZdlWWvmXbAw4V3t01eBog12Il77uHVoKlGYz4XhEmAiCDAIz15pKipYeMgQpz7UsIQXoZDGYQBmM1sKepTo2Sdc9kba48XQlnQ/Gp8DEx+ooGAYt8GQyKPyQSBwBlE3K0jq1BKTjnR+ZQcv+NimFXbKmVCN+7w1xd6rJ0oUzX+tMT2SLyJkY+Ustklf59fpf3GW9oOe6KiVsW5vB2IrAnOJRkha7l8ErBMphlJWDmI67tbggVpbVaYKVATjgXDFilzzy0E49Ix+TUVmy+KZzsYnwFfgrOwqFJ+knxiTSLOiSnAJiLce8wv2GY57aNnIYO/dOceMDRxo7hUTzQaeGcpt9Rq0MPonCf9ZNlhLj/ZsfFmXeccDVBrXA6oYaI4UEgnojGXCaNsiDPTVgSUafICBGe+U/CWmFP+xb7+avNncgnjNfdNkbDYSiCJQKMp4g8cTmnekg4OK+QD1Kywbk+J

Probability & Distributions

Model uncertainty with probability and understand how random variables behave.

Probability Basics

P(A) = favorable outcomes / total outcomes
0 ≤ P(A) ≤ 1
P(A ∪ B) = P(A) + P(B) − P(A ∩ B)
P(A') = 1 − P(A)

Probability quantifies uncertainty. These rules come from set theory — unions and intersections. Counting techniques like combinatorics (permutations and combinations) are essential for computing probabilities in finite sample spaces.

Conditional Probability & Bayes' Theorem

P(A|B) = P(A ∩ B)/P(B)
Independent events: P(A ∩ B) = P(A)·P(B)
Bayes: P(A|B) = P(B|A)·P(A)/P(B)
mepl6ReeJGhZpNj07GmWsAN2/ISu2vl98UIXJYro+1SAbCrBv2LuWWaPKdKnJuUj702RvCp31E2bxesf2n8hjjpBbdckKqO+hDvaiekxVs+ap9jSDAGKSj1pioSc3qrda3zMOawYdvpGlRaaFtyzyIRCMIWwaUcm96i+Cw2/HGyz4bzy8UNrI5/JAgWYZLBLOVTrhUvRLijQ3vSXhCjCRr3c9Gm0QORwdHJi2Cs/BgQTxtyxmK1sQKOrBMTAUYOCe+I/6YTfmLbBPJOCQFe8kMTEYVIWXaVayAzZYWCotI8qhXQXsjBKEppFVMJ5dbG5R95YACgOWZqRl1TxSfuAOOPdXpCN+jdyUFPMkpg3sZjNBwg3zTxpSNVmomuwrdU6cNw98LDzXMFtZYWYrlQRgtaVCKa5ZmxnpkxGHnSAh2CpxqWDz7hqh2YOxzD1OP8Ztc+YOKxtMRdzh4J4MSGAy+BjQn7BA1JWKIZfMJ3v8bgC+aXwnlcLrOsDRr1tGkTGpwimUZYKvJI0t0+ROlKgLg6N+7HWR3CqgfWHPJ8TsvFTWDu7SVOiytU+9EERomb7nWzHxIFX6nlg/jAnMZHmr8R2NY1M6bxUwwPZ6HWhL8KevhJIWRVN5Gk0bSSBx+I/HVauFePG6iHERNq3waiFE2I4RIHYpVdoX4YrzO1NByXqSUS7Rl75xAobJqAr0ag99MrTJVZZadDk1Jmc6zJZl5CKbQZ32cnIjCuYsgwxkqLVczVgHVDqf6zNpKizr6LjDtXKNsBbUrmdbzaQKIPsl24OGRxSdkKklFaX47W8hw/66cJb/ngyQxlbhtRQl3vl9sLd8dkYK9vq7jxp/QotZD

Bayes' Theorem is the backbone of Bayesian inference and machine learning. It lets us update beliefs with new evidence. Total probability connects this to partition: P(B) = Σ P(B|Aᵢ)·P(Aᵢ).

Example: Disease Testing

Disease prevalence: 1%. Test sensitivity: 99%. False positive rate: 5%.

P(Disease | Positive) = (0.99×0.01)/(0.99×0.01 + 0.05×0.99) ≈ 0.0099/0.0594 ≈ 16.7%

Even with a good test, a positive result is only 16.7% likely to be a true positive when the disease is rare!

Discrete Distributions

W8OjytBfQqN6mt6OMqOtLnpyGO9uAGkLH9mgcgZE96yoDp9D257OcBI8nHJVG/n9QOaT17wxhZv9Z1bLGIKB2QaaVAn8viGR6fAUKfPO7Ymk9rEYuJmUnV7b3rUgPf7HwlDlVmNn2AXRNT0Z2dDJzIl0JufqlX9frNYLIVgplfiXbQjySRbURvUkHGn6UD0nmHYVVsOnPLxu+tztO+BbN+jSnHgi+kBZ0sRdeTV2YZj6c1QMXB7CKJimYYrEeXBa8mmOQoEul8hjwTdfRlLIszTidat5olX8AhF4VNfw7I9YQENRHHtbaA+n+CIYL0vRN1hc4eZaBcqGmaj8fgKYGxIN8rM1IV2K2rOpGFxtJSZtWQVEdZ9Tba7BRRTHtduiIMHu//x5Cmmzus2LLXKjMF2Xb8f0++FL+ExZrWJ1CUZ5CRE5OCjQ345enedOtmGY+RUJDxhTYNAhEA4x+Rt3FVqs/XXuJUKXGk3z4dhi+0LNij3VGcOgInv02/doaHaAhQxORcIiTrEdw7F0Vrw2V0iP6DMGmKjhw+Buxl3yJu7xwYjDtnIqndO0uYJSpJVD0XehTjL6DOPnD36+CdPn7F4ihqfkNa8JxEKaGqyrpXL/w+bsw2Wwq6ojAMqu3wrK4RPvBMwJIyuRSfOaoPaWhUmSDLVox8UFgeBlw+6LR244x5nMqmx9mdXwsaNyd6HEwMFM9f3wSeAOAr+VdSPW294x0r1RtkNKDm6hv+YWrQLR8leQhlIoxl06w+pLRpQL3jHSHVGyS4t3hC4Ac+Tr85rQtR8GO6sSClScPFMUyvzdAueWsRUDMrmrVk0JItYh29j0osK3kfqu6Db7KS0+7N
  • Bernoulli: Single trial, p = success. E[X] = p, Var = p(1−p)
  • Binomial(n, p): P(X = k) = C(n,k)·pᵏ(1−p)ⁿ⁻ᵏ — uses polynomial expansion
  • Poisson(λ): P(X = k) = e⁻λ·λᵏ/k! — models rare events per interval
  • Geometric(p): P(X = k) = (1−p)ᵏ⁻¹·p — trials until first success. Connects to exponential functions

Continuous Distributions

Normal: f(x) = (1/σ√(2π))·e^(−(x−μ)²/(2σ²))
Z-score: z = (x − μ)/σ
Standard normal: μ = 0, σ = 1

For continuous distributions, probabilities are areas under the curve — you need integration. The normal (Gaussian) distribution is the most important, governing everything from measurement error to stock prices.

  • Uniform(a, b): f(x) = 1/(b−a) — constant density
  • Exponential(λ): f(x) = λe⁻λˣ — time between events
  • t-distribution: Used in hypothesis testing with small samples
  • Chi-squared: Sum of squared standard normals — used in goodness-of-fit tests
9Pskpmc3VnlnoiUblTsxcPkS3iGPfM7Pa1DpdeKYGLkvIrMq1SfvKBf3i/vgdQzq0r4ApIaBiWWscuvdB3NqjBiZ2pj6hz/QoD+eFz4ru6OO8bEmvpzjGBH5yAJC2IAnTYdPiAUmVeRXHJX87NGpf5mvOYPZkZJl06oMkHQbNFuazWEeoh1rd/dtzJXbVSdbRlMmv2/vpKGajvZWHTV8CyWSeRD73a7BX2mpPeHstcgJJsIumVOeqtDJcot20nJyrxPMonNGwUjM64nbzYOfxmaiGscUPT8Bjh/cHc42v3AaMV/TbNTkoneu1tn4BcmySC/6wmX5NMZ8Md7NOb8QXjm2h5smAaQqrygzaPPln6vq7MnLIRjKDFTvPQ91YBFuLpA5kywIqN/LDi7ZUGvVfjTgUVXElbQAZjY0OHA3Ra+vZMVgsq+0tk5ymLvO5qIYQ3PT8rJ6/5rlX/yBhkvUta0GBbHvCMMDxMkFM3lB3y106+tMDAuhVgx52JhYDedTWl2f3mXx5CnKrnGlYDrmDqq0DFINIPldUg0A4yfS9lSrmJhirXQSx+tqF9sUlL1Lr7Wkjl5wU+PmxJfOzO8Urn2P9fbM62Eot8EVmFLgZYEBrOMF+Pt3ZnA/WOuUU4A9XjJlQEyZHw9qO1ZIG7gAbjkhHvyyCqEmAsr8KBKX1ZskNOl4iy8ov3hgxZo69HiINdNynO4OYcHBnnOclnqtea208fBP2V9kEMXlJ2WM6cNcY27E9UD11wUGXnerGvMAGUbA6Pp+lkkpreUeVuWfq+r8iJoBfV0gZ72Rzsx3orc3WvqmtLBrrBLtQzzB17b1w+wDcV87/QHQIpsscHKa36

Central Limit Theorem

S/7B+h84dHNufnfIrO3W8kG35dbJCgjJxj+F5siet5XTXpI+aGsktdnkinwaAzINUTF6MxVuB/bE9jiUgUOoG7AViL5Oi9aWalMCjTtMjwuWY0RHOfCGrFteYvA8DMZEjuxGTrD94ZavZ84+ukGdwMAHhJqbg0HjkSv6BkQmwmST3/UZf7od+mFm0DUiGbdtiiPhgLoctv2tPnsxvWQ0JAr8PKbnkzDtufpDjch9t1w50x0ppbGOuv20bI1TICwqHjkdDtMNF2iqLeAydk556umNjfZSTGrIxrDelitwF0wbt1Exe2px8/qBTWM5Nwzcv78iuHXK+KmWcLZyCrjKedTfoRB84iJH0VOuPzAz3vrg72x12zUAD749cRwdLtWoI9Qbx9QyUKnPY/M7vS/X/53SuaKD3/DKUSHONE0eZ+u6D1j8gkLlv72PQIpzs1PMvCcuAjxbTv0/pP8FYtSJ3SNuvs8JNXNkX+hB+T1sl0tWsDoDvBZyHS7VFZK1SamctK9/Bs2d551R297vfjrk7uZ1JlycXGL5WWsroB1nJPsf99XiRSSRmR5/p5vh8Pzf+rUyprKv65m1zOCVcrK/iBAjmIMDp4bQupOCk4kaCY6tEysC/dMj3vzCK/QR798XLVn1pxQXUrOdR/9SNUDZbgFRpsrzRCcXbmCUUQm7V8hfOjFeR8w+r9rmK9+/U5JnzvKu4QNOYkN5Qq9CEwpjACPHGoXJ2hsRvwsPBIKzSaeIfy2KqNpTYE84uk6Q8zTvJ5xDid+87C0kkscwhzCScr9H0jKvAAzrpx5CGSU1gibjNJthNn6V5zzw8C8nJytvUudyTPeADnt3YeOdS4/slz
Central Limit Theorem: Regardless of the population distribution, the sample mean X̄ approaches a normal distribution N(μ, σ²/n) as n → ∞. This is why the normal distribution is so important, and why statistical inference works. The convergence concept mirrors limits in calculus.