Post

Learn Exam P - Probability

Exam P Study Guide

Learn Exam P - Probability

A. Discrete Probability

1. Fundamentals of Probability

1) Fundamentals of Probability

For any two events \(A\) and \(B\):

\[P(A \cup B) = P(A) + P(B) - P(A \cap B)\]

2) Complements

The complement of event \(A\), written as \(A'\), is the event that \(A\) does not occur.

\[P(A') = 1 - P(A)\]

3) Venn Diagrams

If \(A\) and \(B\) are mutually exclusive, then they cannot occur at the same time:

\[P(A \cap B) = 0\]

So:

\[P(A \cup B) = P(A) + P(B) - 0\]

or simply:

\[P(A \cup B) = P(A) + P(B)\]

4) De Morgan’s Laws

The complement of \(A \cup B\) means neither \(A\) nor \(B\) occurs:

\[(A \cup B)' = A' \cap B'\]

The complement of \(A \cap B\) means at least one of \(A\) or \(B\) does not occur:

\[(A \cap B)' = A' \cup B'\]

5) Inclusion-Exclusion

For three events \(A\), \(B\), and \(C\):

\[P(A \cup B \cup C) = P(A) + P(B) + P(C) - P(A \cap B) - P(A \cap C) - P(B \cap C) + P(A \cap B \cap C)\]

2. Conditional Probability

1) Conditional Probability

\[P(A \mid B) = \frac{P(A \cap B)}{P(B)}\]

So:

\[P(A \cap B) = P(B)P(A \mid B) = P(A)P(B \mid A)\]

2) Independence

For two events \(A\) and \(B\):

\[P(A \cap B) = P(A)P(B \mid A)\]

If \(A\) and \(B\) are independent, then:

\[P(A \cap B) = P(A)P(B)\]

Also, if independent:

\[P(B) = P(B \mid A)\]

3) Sequence of Events

4) Bayes’ Theorem and Law of Total Probability

If \(A_i\) represents a case:

\[P(A_1 \mid B) = \frac{P(B \mid A_1)P(A_1)} {\sum_i P(B \mid A_i)P(A_i)}\]

3. Discrete Moments

1) Mode and Median

A median \(m\) satisfies:

\[F(m) = P(X \le m) \ge \frac{1}{2}\]

3) Mean and Law of Total Expectation

For a discrete random variable \(X\):

\[E(X) = \sum_x xP(X = x)\]

For a function \(g(X)\):

\[E[g(X)] = \sum_x g(x)P(X = x)\]

Linearity of expectation:

\[E(aX + b) = aE(X) + b\]

Law of total expectation:

\[E(X) = \sum_i E(X \mid A_i)P(A_i)\]

5) Survival Approach

For a nonnegative integer-valued random variable \(X\):

\[E(X) = \sum_{x=0}^{\infty} P(X > x) = \sum_{x=1}^{\infty} P(X \ge x)\]

where \(x \ge 0\).

7) Variance

Variance formulas:

\[\operatorname{Var}(X) = E(X^2) - E(X)^2 = E[(X - \mu)^2]\]

Linear transformation:

\[\operatorname{Var}(aX + b) = a^2\operatorname{Var}(X)\]

Coefficient of variation:

\[CV(X) = \frac{SD(X)}{E(X)} = \frac{\sigma}{\mu}\]

Standard deviation:

\[\sigma = SD(X) = \sqrt{\operatorname{Var}(X)}\]

8) Discrete Uniform

For a discrete uniform random variable \(N\) on the integers from \(a\) to \(b\):

\[P(N = n) = \frac{1}{b - a + 1}\]

where \(b - a + 1\) is the number of possible values.

\[E(N) = \frac{a + b}{2}\] \[\operatorname{Var}(N) = \frac{(\text{# possible values})^2 - 1}{12}\]

4. Combinatorics

1) Permutations and Combinations

Permutations:

\[{}_nP_r = \frac{n!}{(n - r)!}\]

where order is important.

Combinations:

\[{}_nC_r = \frac{n!}{r!(n - r)!} = \binom{n}{r} = \binom{n}{n-r}\]

where order is not important.

3) Binomial Distribution

For \(N \sim \operatorname{Binomial}(n, p)\):

\[P(N = k) = \binom{n}{k}p^k(1 - p)^{n-k}\] \[E(N) = np\] \[\operatorname{Var}(N) = np(1 - p)\]

4) Multinomial Distribution

\[P(N_1 = k_1, N_2 = k_2, N_3 = k_3) = \frac{n!}{k_1!k_2!k_3!} p_1^{k_1}p_2^{k_2}p_3^{k_3}\]

5) Hypergeometric Distribution

With replacement, trials are independent, so the distribution is binomial:

\[P(g) = \binom{n}{g} \left(\frac{G}{N}\right)^g \left(\frac{N - G}{N}\right)^{n-g}\]

Without replacement, the distribution is hypergeometric:

\[P(g) = \frac{ \binom{G}{g} \binom{N - G}{n - g} }{ \binom{N}{n} }\]

5. Key Discrete Distributions

1) Geometric Distribution

There are two common versions of the geometric distribution.

At 1: trials until first success

\[P(N = n) = (1 - p)^{n - 1}p\] \[E(N) = \frac{1}{p}\] \[\operatorname{Var}(N) = \frac{1 - p}{p^2}\]

Survival probabilities:

\[P(N > n) = (1 - p)^n\] \[P(N \ge n) = (1 - p)^{n - 1}\]

At 0: failures before first success

\[P(N = n) = (1 - p)^n p\] \[E(N) = \frac{1 - p}{p}\] \[\operatorname{Var}(N) = \frac{1 - p}{p^2}\]

Survival probabilities:

\[P(N \ge n) = (1 - p)^n\] \[P(N > n) = (1 - p)^{n + 1}\]

2) Memoryless Property

For a geometric random variable \(N\):

\[P(N > n + k \mid N > n) = P(N > k) = (1 - p)^k\]

Expectation memoryless property:

\[E(X - s \mid X > s) = E(X)\]

So:

\[E(X \mid X > s) = s + E(X)\]

Variance memoryless property:

\[\operatorname{Var}(X - s \mid X > s) = \operatorname{Var}(X)\]

So:

\[\operatorname{Var}(X \mid X > s) = \operatorname{Var}(X)\]

3) Negative Binomial Distribution

The \(r^{\text{th}}\) success occurs on trial \(n\):

\[P(N = n) = \binom{n - 1}{r - 1} p^r(1 - p)^{n - r}\]

The number of failures \(k\) before the \(r^{\text{th}}\) success:

\[P(X = k) = \binom{r + k - 1}{k} p^r(1 - p)^k\]

Example: for \(\operatorname{NegBin}(p = 0.6, r = 4)\),

\[P(N = 5) = \binom{4}{3}(0.6)^3(0.4)^1(0.6)\]

or:

\[P(X = 1) = \binom{4}{1}(0.4)^1(0.6)^3(0.6)\]
4th A

Mean and variance:

\[E(X) = r\frac{1 - p}{p}\] \[\operatorname{Var}(X) = r\frac{1 - p}{p^2}\]

4) Poisson Distribution

For \(N \sim \operatorname{Poisson}(\lambda)\):

\[P(N = n) = e^{-\lambda}\frac{\lambda^n}{n!}\] \[E(N) = \operatorname{Var}(N) = \lambda\]

5) Sum of Poisson Variables

If \(X_1 \sim \operatorname{Poisson}(\lambda_1)\) and \(X_2 \sim \operatorname{Poisson}(\lambda_2)\), then:

\[X_1 + X_2 \sim \operatorname{Poisson}(\lambda_1 + \lambda_2)\]

So:

\[P[\operatorname{Poisson}(\lambda_1 + \lambda_2) = k] = e^{-(\lambda_1 + \lambda_2)} \frac{(\lambda_1 + \lambda_2)^k}{k!}\]

6. Deductibles and Policy Limits

Total cost/loss:

\[\text{Total cost/loss} = \text{Insurance Payment (covered)} + \text{Company Cost (uncovered)}\] \[E(X) = E(Y) + E(U)\]

Ordinary Deductible

For a deductible \(d\), the insurance payment \(Y\) is:

\[Y = \begin{cases} 0, & X \le d \\ X - d, & X > d \end{cases}\]

The uncovered cost \(U\) is:

\[U = \begin{cases} X, & X < d \\ d, & X \ge d \end{cases}\]

Policy Limit

For a policy limit \(B\), the insurance payment \(Y\) is:

\[Y = \begin{cases} X, & X < B \\ B, & X \ge B \end{cases}\]

Deductible with Policy Limit

For a deductible \(d\) and policy limit \(B\), the insurance payment \(Y\) is:

\[Y = \begin{cases} 0, & X \le d \\ X - d, & d < X < d + B \\ B, & X \ge d + B \end{cases}\]

B. Continuous Probability

0. Calculus Review

Derivatives

\[(x^n)' = nx^{n - 1}\] \[(\ln x)' = \frac{1}{x}\] \[(e^x)' = e^x\] \[[c f(x)]' = c f'(x)\] \[[f(u)]' = f'(u)u'\] \[(uv)' = u'v + uv'\] \[\left(\frac{u}{v}\right)' = \frac{u'v - uv'}{v^2}\]

Integrals

\[\int a \, dx = ax\] \[\int x^n \, dx = \frac{1}{n + 1}x^{n + 1}\] \[\int \frac{1}{x} \, dx = \ln x\] \[\int e^{bx} \, dx = \frac{1}{b}e^{bx}\] \[\int a^x \, dx = \frac{1}{\ln a}a^x\]

Integration by Parts

\[\int u \, dv = uv - \int v \, du\]

Example

Find:

\[\int x^2 e^{2x} \, dx\]
2x 2 0 + + e²ˣ (1/2)e²ˣ (1/4)e²ˣ (1/8)e²ˣ

This tabular integration-by-parts setup gives:

\[\int x^2 e^{2x} \, dx = x^2\left(\frac{1}{2}\right)e^{2x} - 2x\left(\frac{1}{4}\right)e^{2x} + 2\left(\frac{1}{8}\right)e^{2x} - 0\]

So:

\[\int x^2 e^{2x} \, dx = \frac{x^2}{2}e^{2x} - \frac{x}{2}e^{2x} + \frac{1}{4}e^{2x} + C\]

1. Densities and CDFs

2) Densities and CDFs

For a continuous random variable \(X\):

\[P(a < X \le b) = \int_a^b f(x) \, dx = F(b) - F(a)\]

The survival function is:

\[P(X > t) = S(t) = \int_t^\infty f(x) \, dx\]

3) Mixed Distributions

For a mixed distribution with a point mass at \(x\):

\[P(X = x) = F(x) - \lim_{t \uparrow x} F(t)\]

2. Continuous Moments

1) Moments of Continuous Distributions

For a continuous random variable \(X\):

\[E(X) = \int x f(x) \, dx\]

For a function \(g(X)\):

\[E[g(X)] = \int g(x) f(x) \, dx\]

2) Moments of Mixed Distributions

For a mixed random variable \(X\):

\[E(X) = \int x f(x) \, dx + \sum_x x P(X = x)\]

3) Survival Function Approach

For a nonnegative continuous random variable \(X\):

\[E(X) = \int_0^\infty P(X > x) \, dx\]

3. Key Continuous Distributions

1) Continuous Uniform

For \(X \sim \operatorname{Uniform}(a, b)\):

\[f(x) = \frac{1}{b - a}\] \[F(x) = P(X \le x) = \frac{x - a}{b - a}\] \[E(X) = \frac{b + a}{2}\] \[\operatorname{Var}(X) = \frac{(b - a)^2}{12}\]

3) Exponential Random Variables

For \(X \sim \operatorname{Exponential}(\theta)\):

\[f(x) = \frac{1}{\theta}e^{-x/\theta}\] \[F(x) = 1 - e^{-x/\theta}\] \[S(x) = e^{-x/\theta}\] \[E(X) = \theta = \frac{1}{\lambda}\] \[\operatorname{Var}(X) = \theta^2\]

Memoryless property:

\[E(T - s \mid T > s) = \theta\]

So:

\[E(T \mid T > s) = s + \theta\]

4) Gamma, Exponential, and Poisson

For \(X \sim \operatorname{Gamma}(\alpha, \theta)\):

\[f(x) = \frac{x^{\alpha - 1}}{\Gamma(\alpha)\theta^\alpha} e^{-x/\theta}\]

When \(\alpha = 2\), this becomes:

\[f(x) = \frac{x}{\theta^2}e^{-x/\theta}\] \[F(x) = 1 - e^{-x/\theta} - \frac{x}{\theta}e^{-x/\theta}\] \[E(X) = \alpha\theta\] \[\operatorname{Var}(X) = \alpha\theta^2\]

5) Beta and Pareto

4. Normal Approximations

1) Normal Distribution

For \(Z \sim N(0, 1)\):

\[P\left(Z \le \frac{x - \mu}{\sigma}\right) = \Phi(z)\]

where:

\[z = \frac{x - \mu}{\sigma}\]

Symmetry:

\[P(Z > z) = P(Z \le -z)\] \[1 - \Phi(z) = \Phi(-z)\]

Common values:

\[\Phi(z) = 0.95\] \[z = \Phi^{-1}(0.95) = 1.6449\] \[\Phi(z) = 0.75\] \[z = \Phi^{-1}(0.75) = 0.6745\]

2) Linear Interpolation

Example 1:

\[\Phi(0.87) = 0.8078\] \[\Phi(0.88) = 0.8106\]

Interpolating for \(0.81\):

\[\Phi^{-1}(0.81) = 0.87 + \frac{0.81 - 0.8078}{0.8106 - 0.8078} (0.88 - 0.87) = 0.8779\]

Example 2:

\[\Phi(1.31) = 0.9049\] \[\Phi(1.32) = 0.9066\]

Interpolating for \(1.3168\):

\[\Phi(1.3168) = 0.9049 + \frac{1.3168 - 1.31}{1.32 - 1.31} (0.9066 - 0.9049) = 0.906056\]

3) Central Limit Theorem

For a sum \(S\) of \(n\) independent and identically distributed random variables:

\[E(S) = n\mu\] \[\operatorname{Var}(S) = n\sigma^2\]

For the sample mean \(\bar{X}\):

\[E(\bar{X}) = \frac{1}{n}n\mu = \mu\] \[\operatorname{Var}(\bar{X}) = \frac{1}{n^2}n\sigma^2 = \frac{\sigma^2}{n}\]

4) Continuity Correction

When approximating a discrete random variable with a continuous normal distribution:

\[P(X \ge k) \approx P(X > k - 0.5) \quad \text{"At least"}\] \[P(X > k) \approx P(X > k + 0.5) \quad \text{"Greater than"}\] \[P(X \le k) \approx P(X < k + 0.5) \quad \text{"At most"}\] \[P(X < k) \approx P(X < k - 0.5) \quad \text{"Less than"}\]

5) Lognormal Random Variables

5. Deductibles with Continuous

C. Multivariate Probability

1. Joint Distributions

1) Joint Distributions

2) Marginal and Conditional Distributions

2. Joint Moments

1) Means of Multivariate Distributions

For discrete random variables \(X\) and \(Y\):

\[E[g(X, Y)] = \sum_x \sum_y g(x, y)P(X = x, Y = y)\]

For \(E(X)\):

\[E(X) = \sum_x \sum_y xP(X = x, Y = y) = \sum_y \sum_x xP(X = x, Y = y)\]

2) Covariances and Correlations

Variance of a linear combination:

\[\operatorname{Var}(aX + bY) = a^2\operatorname{Var}(X) + b^2\operatorname{Var}(Y) + 2ab\operatorname{Cov}(X, Y)\]

Covariance:

\[\operatorname{Cov}(X, Y) = E(XY) - E(X)E(Y)\]

Covariance and correlation:

\[\operatorname{Cov}(X, Y) = \operatorname{Corr}(X, Y)SD(X)SD(Y)\]

For a sum \(S = X_1 + X_2 + \cdots + X_n\):

\[\operatorname{Var}(S) = n\operatorname{Var}(X_i) + 2\sum_{i < j}\operatorname{Cov}(X_i, X_j)\]

3) Conditional Moments

Law of total expectation:

\[E(X) = E[E(X \mid Y)]\]

Law of total variance:

\[\operatorname{Var}(X) = E[\operatorname{Var}(X \mid Y)] + \operatorname{Var}[E(X \mid Y)]\]

3. Order Statistics

1) Order Statistics

For the maximum order statistic \(Y = \max(X_1, X_2, \dots, X_k)\):

\[P(Y \le y) = P(\text{all } X_i \le y) = [F(y)]^k\]

For the minimum order statistic \(Y = \min(X_1, X_2, \dots, X_k)\):

\[P(Y > y) = P(\text{all } X_i > y) = [1 - F(y)]^k\] \[P(Y \le y) = 1 - P(\text{all } X_i > y) = 1 - [1 - F(y)]^k\]

Law of total probability for \(Y\):

\[P(A) = \int P(A \mid Y = y) f_Y(y) \, dy \quad \text{continuous}\] \[P(A) = \sum_y P(A \mid Y = y)P(Y = y) \quad \text{discrete}\]

2) General Order Stats

This post is licensed under CC BY 4.0 by the author.

Trending Tags