Please refresh the page if equations are not rendered correctly.
---------------------------------------------------------------
For observation of a random phenomenon, the probability of a particular outcome is the proportion of times that outcome would occur in an indefinitely long sequence of like observations, under the same conditions.
- complement 补
- intersection 交 (and)
- union 并 (or)
- A probability of event A, given event B occurred, is called conditional probability. It is denoted by P(A \mid B).
Placing P(B) in the denominator corresponds to conditioning on (i.e., restricting to) the sample points in B, and it requires P(B)>0. We place P(A B) in the numerator to find the fraction of event B that is also in event A (see Figure 2.3).
P(A \mid B)=\frac{P(A B)}{P(B)}
- Random variable: For a random phenomenon, a random variable is a function that assigns a numerical value to each point in the sample space. Random variables can be discrete or continuous. We use
- upper-case letters to represent random variables,
- with lower-case versions for particular possible values.
- Probability distribution: A probability distribution lists the possible outcomes for a random variable and their probabilities.
- Probability density function for continuous random variables (
pdf
) - probability mass function for discrete random variable (
pmf
) - Cumulative distribution function (
cdf
)
The probability P(Y \leq y) that a random variable Y takes value \leq y is called a cumulative probability. The cumulative distribution function is F(y) = P(Y \leq y), for all real numbers y.
By convention, we denote a
pdf
orpmf
by lower-case f and ac d f
by upper-case F.F(y) = P(Y \leq y) = \int_{-\infty}^{y} f(u) d u .
Because we obtain the
cdf
for a continuous random variable by integrating thepdf
, we can obtain thepdf
from thecdf
by differentiating it. - Probability density function for continuous random variables (
- Expected value (mean, expectation) of a discrete random variable:
-
E(Y)=\sum_{y} y f(y),
- For a discrete random variable Y with
pmf
f(y) and E(Y)=\mu, the variance is denoted by \sigma^{2} and defined to be: -
\sigma^{2}=E(Y-\mu)^{2}=\sum_{y}(y-\mu)^{2} f(y),
The variance of a discrete (or continuous) random variable has an alternative formula, resulting from the derivation:
\begin{aligned}
\sigma^{2} & =E(Y-\mu)^{2}=\sum_{y}(y-\mu)^{2} f(y) \\
& =\sum_{y}\left(y^{2}-2 y \mu+\mu^{2}\right) f(y) \\
& = \sum_{y} y^{2} f(y)-2 \mu \sum_{y} y f(y)+\mu^{2} \sum_{y} f(y) \\
& =E\left(Y^{2}\right)-2 \mu E(Y)+\mu^{2} \\
& =E\left(Y^{2}\right)-\mu^{2}
\end{aligned}\operatorname{var}(Y)=\sigma^{2}=E\left(Y^{2}\right)-\mu^{2}
- The standard deviation of a discrete random variable, denoted by \sigma, is the positive square root of the variance.
- Expected value and variability of a continuous random variable for a continuous random variable Y with probability density function f(y),
-
E(Y)=\mu=\int_{y} y f(y) d y , \\
\text { variance } \sigma^{2}=E(Y-\mu)^{2}=\int_{y}(y-\mu)^{2} f(y) d y,and the standard deviation \sigma is the positive square root of the variance.
Bayes’ theorem
For any two events A and B in a sample space with P(A)>0,
P(B \mid A)=\frac{P(A \mid B) P(B)}{P(A \mid B) P(B)+P\left(A \mid B^{c}\right) P\left(B^{c}\right)}
Generally, for two events A and B, the multiplicative law of probability states that:
P(A B)=P(A \mid B) P(B)=P(B \mid A) P(A) .
Sometimes A and B are independent events, in the sense that P(A \mid B)=P(A), that is, whether A occurs does not depend on whether B occurs. In that case, the previous rule simplifies:
P(A B)=P(A) P(B) .
Comments NOTHING