Expectations¶
For details, see Sections 4.1, 8.3, 10.2, and 13.3.
The expected value of a random variable, , is the weighted average of possible against the PMF/PDF:
The expected value is equivalent to the center of mass of the distribution
Long run sample averages converge to the expected value
The expected value of a function of a random variable, , is the weighted average over each , of , weighted by the PMF/PDF
The expected value of a function of multiple random varibales is defined analogously. For example:
The conditional expectation of a random variable is its expected value when sampled from a conditional distribution, for example, if and are jointly distributed, continuous variables, then:
The expected value is distinct from the:
Mode: the most likely outcome, or collection of outcomes that maximize the PMF/PDF.
Median: the “midpoint” value such that .
Rules of Expectations¶
For details, see Sections 4.1, 4.2, 7.1, and 10.2.
Expectations of key distributions:
Constants: .
Indicators: if , then .
Symmetric: If is drawn symmetrically about , then .
Linearity: .
Remember, this rule only works for linear functions. If is a nonlinear function of , then need not equal .
Additivity: for any pair of random variables and , .
Jensen’s Inequality: If is a strictly convex function and is a random variable with nonzero variance, then:
Tail Sums and Integrals:
If is a count valued (integer valued) random variable, then .
If is a continuously distributed, nonnegative random variable, then .
Iterated Expectation: If and are drawn jointly, then:
In particular:
Variance:¶
For details, see Section 4.3.
Given, , the variance and standard deviation of a random variable are:
The standard deviation measures the breadth, spread, or width of the distribution
Properties of Variance:
To compute variances, we often use:
The variance in a random variable is its expected square, minus its squared expectation.
Covariance¶
All definitions and results are available in Sections 11.1 and 13.1.
The covariance between the random variables, and , is defined:
where and The variables and are centered.
It may be expanded as the expected product of the variables minus the product of their expectations:
Covariance Matrices: If is a collection of random variables, then the covariance matrix is the array with entries
Properties of covariance:
The covariance is unchanged by translations (adding constants) to the variables, .
The correlation does depend on the scale of each variable, and .
The sign of the covariance indicates the sign of the association between two variables.
The covariance is zero if and are independent. However, dependent variables may also share a covariance equal to zero.
The covariance between a random variable and itself is the variance, .
The covariance between any random variable and a constant is zero.
Variance of Sums and Sample Averages:
The variance of a sum of random variables is a sum of all the pairwise covariances:
In the special case when :
The variance of a sample average is the average of all the pairwise covariances:
Correlation¶
All definitions and results are available in Sections 11.1 and 11.2.
The correlation between two random variables, and , is defined as the covariance in the standardized variables. It may be computed:
The correlation measures the strength of the association between and .
Properties of correlation:
The correlation is unchanged by translations (adding constants) to the variables, .
The correlation does depend on the scale of each variable, and if and .
The sign of the correlation indicates the sign of the association between two variables.
The correlation is zero if and are independent. However, dependent variables may also be uncorrelated (have correlation equal to 0).
The correlation is between and if and only if is a linear function of .
Correlation Interpretation:
The empirical correlation between a collection of sample pairs is the cosine of the angle between the vectors formed by the centered samples.
The correlation between two variables equals the slope of the best fit line between the standardized variables.