In This Lesson Probability Basics Conditional Probability & Bayes' Theorem Discrete Distributions Continuous Distributions LH7s3hYlaPJBFrZJWAH9fFn3GBg2cobmWLYubdj5smEZO2o1uoIBua84pZ1OAO+PDrrKEzV+OXPSUdV/Vo0e4cjcz/XhodfhEFV7sodJDapRgAc6IEZou66FylCadV5x1OxyjzBtl4q3rknMKDuYzV5GoGB7PgOYfqqsbXGhEC9mChH3BkeF8Nf/GQZ2DieBz3GlHKqWpjHNbyYo/f/leS1jMxTLds5weqfuhIYZENHFVIi4iNJIBD5L2dTD4Cam/nMFGKHdaGt+SUaWqYVDEdfkdzUKHhQYvUlNfd2vbU4IvP8LgHqJdsSZw+sG33h+DrylfNVDvlH77t7d7f1TXZdTeO84Yunyce5KJsiup7+FDJx6KrJWuw1oI94MFQQLrDtKnxHvHn4itPK3Wo8LAOo2oB/SyKrL2Cnn61buXI345D84LSMMNrtWKZR8nS0i6coj0cZjWQkJVF+kCcvqqo9YfuJ7nzCTB/1UjjZy+YjQfld6rrldQ+k8noI1vwbwZIdOC4xq3uUe+ICns4IX/Ba50mdlxkGiAbE6Nwy4etSBVLWru3EgcPweL5gMQMctfVwLtIwRKGg+R9XZVvz6QbD4iX/rnSY6Qb+BGMpgWGq8VeT+RiykqKneOfJKalqhr1NQ7xYe6xl9rtdU0A+hj3YmS5aYn7URjR7FP8ZolrcfCBq0iGqvjYZhdge5rIVNYUwf7Fhm+/kZSVm277AH1lXwn9LOQsm/ygC3FxvvSmlnqLvc3phwhkwZ2ZZhILbLIIbZTMrIZ1brJISSq1RQVrUuLPlwEBn1z7okKKydhNluGYcQDq1nc4duxRvw9Br6suDwIAAb3HAYuq+8ei9aDg/88EM8HP Central Limit Theorem UhzT/HWaO2cpB8Jl7/4X2hL9AEeO/5HTFxR6UKqufLlWCV2YPKxO+SBZXfb4ntK1EqjUdZ994PgbIHj/pgdpTADR9awoKnKMES232msVhrVOrRomCBrqeLRz8rOR5509W761dCJNk/5Dq34hH6FOHZE8iEOguP3d8wsauSn8TelmJOSvz2hse/TYAHBaVgZY5pUHDIJ+vIO4LyqlwF+BklbzsZWrAeixzCmMI5nN7xgGVbSDjeP9TF1BEmZQwtxYN9rJ+9Lb4V2P6Woqg1h2l9annoJ3FWRsTzTaWcU6CrSS8C5W4pDL71ApPqHU7P3DkM657nnGTZZIVAbJeCpFrknlvIVKSA+URA9wTWg7AQWGO7iER6Br/oRrFetW5G/yy52HtbDR/4eTuV9KDa5LrekTgXRYxS6qeOWqTrqSDUgL3zTXuvuJQxBhsAa7niKJewdDSiHp3imwhOonMD+yBD4Eyq3mrPWUa5uej9CRXiOtrYRgfMMxC7AIQuelHZLAZkW+9xUOaySMle0VJB80TIsjCd361S0RRSnod3AfbTTRijQ8V8ejeZ0A/JreEzR3xX4Syc5rKuEycdIPCELqaKhFBffEGFgE2VjugamzY66FMKABIUDOOpF2kCdmXSZvXz/jsFpVXfS2K+P7Xd08PkJfD0JjlC9IN2PRM4HILC7i8ZXv5kGNAsxgOmdpBkJ/hQdLE6pCUrLXTN/zD3UjfXMgqYk3yKYYoe3YS3/0xrPHUgljmzegmNeKwcGoEFi/imlmTSZl9nhCfZe67nv4J7NybT7xJlKH7Id09X9oX/I2+UF6jfyn/UtCGN6026dzyHk8cZX2fI4pm3Jc891OxhJY7x0BKr FpmiaaMKkmho9h4Mi7AaQu6FUB6DS0b9GYp3eKNCEgp4NVzW+yClfwCX1DmpGKVlhpJFWgj3bjwe6HGFbO3wPnGlq5+qId6VnDCM11hhePH6/RQn7H5oqJabxMesVYYjikxt6ubyAsCMfICMClqDBq0+MSXvhgxiroa9oPIqXv3BGF1WdAlC7IbyM7mJVaaEd0ekskb7lxZSyp4Xd847UrJ1NtX7dB0CJ6XUia0fRDdtPcnLO4S1FmQErtNOvSLr92r7ToNtLuH6LLL4IfGMTabO7S1pMPKltL9pTX2upMp4whIo16TUD9fBTU2ZXotGoMs6gyM7foHjb2lBd+5F10jzqpmXpN8GqI2X4oxUi61nWQ4sdWTfBjncXM59TD/VT3wBSSyrIvqyxXMOhgx+YOwBfI32Clxw58NOBBz4kNKcc5GIOpcXdCWpLizYBzOSF84kTJMZWsixQmGbv0fCOhuRCFVyAcEhEVsPb1exUtkLVWpRagSZTPFCShO1/mxyMrcQwLqAX9CASxt/kmenN13WiIXQ6Ia0i4HMcf/1fmljIMmT8on7Ix1amGHkPKdtEd3rfOiQDGKcqYtivd4WXJ7Uv3tPwOcDFGcve3a8mbujdL+ywy3Nlco0GRHltCkwV+/k0vuiaTAAz1izdZh5bXBUZ9j46QsrUb9ybx2LQ+/Gzmkto8HGoEQxbUdysngsBTaZMdHwvma2S5lnxGVbzD0XO4LGW8TkGsNWJtj812UojnCk6Isew2+Fz/6JZ8hmYIOFHSfZ+ZXH6eizYpmlkF4SkInZEvFyp52sh2RAevrBlNBUxAnRYwQtPhf/cPAsiOAfkryIysqOxOSLN0vCx5i7VNGVLK 5kWXVUnMas9490xROYdx//gpVpfrJ6m/NJpS+Csf2D99QOEPrFGEyFb5xNXLqmoN7hyoZBzoV5X+KpPr+XSj/VX9ONMTiOAN+POM1jwF3zMsgJd/PtedNlTplLBtBvn815rsP/XhRni9WgCtun7MCGyU2mRjjF+EE64C18a0CO54WsHvbL8lDV9cXhJfwg98HfxTbxlrGZYK90XQvvQaiaWeLLmtGRD0sHYpsRlSN1t0sWfMszk6C2v+yjtIMQAOf3zhM6y1+Wq/+JJKFgzE+zvLZPjjizJmsKV4pflE7vmqbcH8tYZjG3vl9DBK9awvyF17kKKLeZY1j9kA0UPb/uxBRQfdvpdIwULY9csjlDu/XNtiezt5z2SUC/qyCu7pWVA1RiAIKqbZW7dq1+Sl9KC1+jACW682lSAKloD7fJHjf1VhqjJh0NyJ5gA9f6eNeKuPGQ7lazKCD1nlMhT2yYi7FhSPrYXNsgv2vkYJdHv9kwTFxipP376tQZkbgdJy8kH1dH9Q7gdFyRSVucprd0MHDOySiqEQquVpXww9EDH3lxVuF5jJoMNFkzLwAH8MfVzbYmKK9+9X3z/1YGB+1ajbpANMKsIIVlpDKqwPo+vsuBnKGHRj91kvh8q5qLoYo6fra257oYkbZJX913MRsNNWCXvBHWMz/dJqV+RCNehVziaUUw2gCMvIlLV++isXxePK06p9i71dp9eOd2fqgPK2VzavIsrzKa5x4nayl/S/kxtkkvXvKo8c6s4Aq/nfqduFxmzIqVwgv0Y/UiDFWPcyrqjQAm65DCdqnS4KJtdYJdW3dpNYvEFmrkgmiId2fqWhY8snijPzvnga8AJGoRZv5QOWcm Probability Basics
P(A) = favorable outcomes / total outcomes
0 ≤ P(A) ≤ 1
P(A ∪ B) = P(A) + P(B) − P(A ∩ B)
P(A') = 1 − P(A)
Probability quantifies uncertainty. These rules come from set theory — unions and intersections. Counting techniques like combinatorics (permutations and combinations) are essential for computing probabilities in finite sample spaces.
Conditional Probability & Bayes' Theorem VlupIQzSf8w/BRXKygkNiXt9ToW4CyE+5CdmPcJ8UcnKjp4ZYbzU37k1+QLAjTX3Q9jfdTg2nu2vGUME46gq/fOgnFr6QK6IeuRyD9lVI+3zRXnqBaP9HGX0XGOKAYrl0o0JlUw59DE/1QwwXn9nod00kaISIG/vGpETcFvSE8Tkpr7sG9PMbCEtVyHEAd4vMrlc+lvQWhygoqir0xsugl+sBmnKBgwvELz1QjWt1H4H4KRNr+oyAyyYoHdS+mofvRnHGop5hryLPzwgglT7zpXuibo1Q1CAoS+KbcRFWo9kmw6gVfjUc8KbSRbpB9vV9MCsCkjHqUXHFwAUNEQVyFr8iIgN7Q9O7MaGlm5sPRNyXYFHcgznfgGi7tiuwqjpLsWK26eU5PTY+wZCq3trhA3yOBTv2NoDrZmBTOLylY6Ez/G9ipRAIScDtbyP7Oofaz/WbdKSwX9bVeqdBlqmngS57Alysw/VSRc6YEGQ/NLpLiERur9mA21U0sJJg3f8L+XJykXeLnSsJUXxM6rafn8eENiDcnrGYXVtNMz29uWPcqG6TGc3jSD6fwaKcrxkTFncUHZL9Yo02BromLSFn0es8XJuzlZD0hbCx4O3/q0eHIfLVmxt4pGhVNkj15kIhbcnMY3q1B78ZLqdemypXarH58a18djOqrj74AQHZH6iQ/37Pv0IW93k4n1P2tp/unkDdjtKapVJUxb1C8S66f6etCIPRcCMXaHTZ4mJtAzqwuZvTUQ2nr67HxlZXJ3/8rv/lYrhxPsgLXUm2uOjg0tDcQ5X4D4eU96f+LM8SLKfpenEPwrrNUOY24M62V3eKf8dtA9qbg5xQTLNY0brgOjVC2X1fd
P(A|B) = P(A ∩ B)/P(B)
Independent events: P(A ∩ B) = P(A)·P(B)
Bayes: P(A|B) = P(B|A)·P(A)/P(B)
Bayes' Theorem is the backbone of Bayesian inference and machine learning. It lets us update beliefs with new evidence. Total probability connects this to partition: P(B) = Σ P(B|Aᵢ)·P(Aᵢ).
ECyStNcO9WcTqUtBKyV3eHtp5w+ScsIl//RLh7yGNIeo3K/CAnCGghq2/bGBgM3WsvYvZFzsLZrP79IzAtFZE7LuAcxN0fqSw8WOrMxA8TA5MyxOWAbGlH/+IHgRvuSsyVEY0ue5ACabrYYgsHcKHmgvIvarnEYizapy8ZNebYgiR3nDhHOZoqkcKkOxhUXuYzJyc6h9cfi3spFXN4K56+hGQmpXpSv1ArNWdCk2OF0RgkjIIzV9kSu7B793wkc8vwB8IFezLS+QiKgGsKt8FPLKldZgFEw996kSraEtJdAOAThxIva/gQM4+d+YLsDJ1p0tMRmwoFod+rqTa248EQoNDAlQexOwMBQb3uTPkeRWU703i/Xl4f4pavzIM9+z1o4+zRSP/4OKJ9I0+FZ3i/YRTqYvkYXi8JnQHD3r6iqfH8h9claTKirjjmIcFOrqKlVHzuMYpH2QchKR6om2aHpAtJre4LwPQe6T98hma77+pw3WiK4MA9HIwojeOE0QF8rt5mLcLvzk50rowcKijRjE8k2ouE1/q+bG8Wfcru9Mdx2dC3t1x5sfJ4b3xoltlJYX3pXVgE2puwNCXNdu7dME9eepeluRnUkezfyoI3kGD/lcq8OQEwJ3fADna2TY+UuELtxvNvCPHk0Ptyfp1bgP4G3DCr9eRQ2JyaNyakaJJgkkELU6XtYiSBwxY5g89Gp25O6y8JdQUgyTg3lCQPrY+759MULqpS322rs+V16m4yWOiblHBiFz7SCn9QOjdoePp5/Q/Ox85gmEKbeywX6kQxrUzJSQtqefmOEDKsoV4d9m7sJjH3lIEZvvFWiGmdt2eXqdJgUYiTLSQaRlP3PSLWZlMv Example: Disease Testing H3eZW6PYZ6SA3/SLOe0grSLnW4PHAjXysxg2OZMt/FGfLwBuNaTnWqjFBdl2fDMukRAaPvj/UYpKvMaS9k1yPApcxTXbg4cisfYxz5rY1y4OnaH9m+kRfNBSch+ssTYXK3xPCK++RL3Tqlqt04r8FJyXO/Bz2wAYfW+KRm6131YA/76hTJCX05R7FzG5RrFfPLVbKcSmi3QrvVcJ5lmSNFuLT3H9bilpYNv2rblP6W0mntwXSivghCpjqGPrNNzaxcRBsxulv3+8LEtnPzcXxo+VLvDbCAADjKf1v5O5USn3PIloXINvVQQCCHyJRjjZEOqxg6vas8dCKBWuYWNPEekBuElA/yaaFt70AtMJOjXQd904fAf3RGqDhaVXdxC3QTJ4xwVTgZlvtRdiVOHuH5L4q5Ev7ksGUtH9qoTbEBX/VpCXpV15ZxyHDA9WvdcO52sq69wxXg+AHKtmN1MEQzPbTEseLGTivyzH/jLZRKQPd6Icxx9XCauLmiAOuU7vD0gMc8V8wzBlFNt4ntw0Ldwprgxc8+y9/m+Nc6FVMRpPaENkqrqaZpq5g2/pa2OUqanRTC/Uv/5RHBo1fLEwzgCctVkqru1el9v3ibYHHXbhAYH+eKPF2jkg89NkOovjRmlHZKbTI5qYOo4Vqf4ndeh6swFrMJawEkeLt6e5zOrSC6o9rJZVigfvBjk7qKluiqfO6lz14LVhZhomYj4ixmrPoEy6vPW+tnSpytj7/8qcQPK+WA+AT4RMLSjJOOcvDo6PZHckP4N8i6OZyB4oqDNFUmBzD9Ql8YKZ9PjhDQKcTqxJWCxHgLeiVSu50CdfB22YR1oV+DWZp5P07ovwIi4aBRk+p Disease prevalence: 1%. Test sensitivity: 99%. False positive rate: 5%.
P(Disease | Positive) = (0.99×0.01)/(0.99×0.01 + 0.05×0.99) ≈ 0.0099/0.0594 ≈ 16.7%
Even with a good test, a positive result is only 16.7% likely to be a true positive when the disease is rare!
Discrete Distributions Bernoulli: Single trial, p = success. E[X] = p, Var = p(1−p) Binomial(n, p): P(X = k) = C(n,k)·pᵏ(1−p)ⁿ⁻ᵏ — uses polynomial expansion Poisson(λ): P(X = k) = e⁻λ·λᵏ/k! — models rare events per interval Geometric(p): P(X = k) = (1−p)ᵏ⁻¹·p — trials until first success. Connects to exponential functions Continuous Distributions
Normal: f(x) = (1/σ√(2π))·e^(−(x−μ)²/(2σ²))
Z-score: z = (x − μ)/σ
Standard normal: μ = 0, σ = 1
For continuous distributions, probabilities are areas under the curve — you need integration. The normal (Gaussian) distribution is the most important, governing everything from measurement error to stock prices.
Uniform(a, b): f(x) = 1/(b−a) — constant density Exponential(λ): f(x) = λe⁻λˣ — time between events t-distribution: Used in hypothesis testing with small samples Chi-squared: Sum of squared standard normals — used in goodness-of-fit tests Central Limit Theorem Central Limit Theorem: Regardless of the population distribution, the sample mean X̄ approaches a normal distribution N(μ, σ²/n) as n → ∞. This is why the normal distribution is so important, and why
statistical inference works. The convergence concept mirrors
limits in calculus .