For and any integer , the generalized binomial coefficient is defined as
Specially, if where , then
1.2 Newton Binomial Theorem
Let and with , then
2 Bernoulli Trial and Bernoulli distribution
A Bernoulli trial is a random experiment having 2 possible outcomes commonly labeled as success and failure. If is the indicator of success, then X follows a Bernoulli distribution or if
2.1 Expectation and variation of Bernoulli distribution
Assume that , therefore
3 Binomial Distribution
3.1 Introduction of binomial distribution
Assume that you are testing the toys manufactured by a factory, where the probability that a toy is defective is . In order to decide whether to accept the toys, you randomly sample of them from the whole batch. Let denotes the number of broken toys (). Then follows a binomial distribution with parameters and The pmf of is
Tip
We can consider Binomial distribution as independent Bernoulli trials and the random variable denotes the number of successes in independent Bernoulli trials:
3.2 Expectation and Variation of binomial distribution
We can directly compute the expectation and variation by calculating the first and second moment of .
therefore
2. Let denote the indicator variable of success of the th Bernoulli trial. Then the number of successes can be expressed as . Using the expectation and variation of Bernoulli distribution and linearity of expectation and variation for independent random variables. We get
4 Geometric Distribution
4.1 Pmf of Geometric Distribution
The geometric distribution gives the probability that the first occurrence of success requires independent trials, each with probability . Assume a random variable , then
4.2 Expectation of Geometric Distribution
Here we use an interesting trick to solve where .
4.3 Variation of Geometric Distribution
4.4 Sum of Geometric-distributed random variables
Assume we have random variables satisfying , . From the definition of Geometric distribution we know that it is about the number of trials when the first success occurs. Then indicates the number of trials when the success occurs. Therefore the last trial has succeeded and the remaining trials have successes. Thus We call this distribution Negative Binomial Distribution.
5 Pascal Distribution
In independent Bernoulli trials with success probability , the pascal distribution focuses on the number of failures when the th success happens. If a random variable , then
5.1 Expectation and variation of Pascal distribution
If , then
Proof of expectation and variation using geometric distribution
If we define the geometric distributed random variable as the number of failures before the first success happens in Bernoulli trials (since the original version is defined as the number of Bernoulli trials, we denote this the second version). Then
We can find that starting from the th success to the th success, the number of failure follows the second version of geometric distribution. Therefore satisfies
For this version of geometric distribution, we have
Using the property of linear combination of random variables we get
Direct proof of expectation and variation using generalized binomial coefficient
6 Negative Binomial Distribution
In independent Bernoulli trials with success probability , the negative binomial distribution focuses on the number of trials when the th success happens. If a random variable , then From the above discussion about geometric distribution, we have where , where denotes the number of trials needed after the st success to obtain the th success. ( are independent)
6.1 Expectation and variation of negative binomial distribution
7 Gaussian Distribution
7.1 Pdf of Gaussian Distribution
If a random variable , where is the mean and is the variance, then the pdf of is
7.2 Mgf of Gaussian Distribution
Assume a random variable , then the mgf of is
Proof of the mgf of Gaussian variable
By the definition of expectation we have
7.3 Characteristic function of Gaussian Variables
Since we know the mgf of Gaussian variable is therefore the characteristic function of is equal to
7.4 Moments of standard Gaussian
Assume that the random variable and that , then
First we state an important lemma:
Relation between Moments and derivatives of mgf
Proof
The Taylor's expansion of is Inserting this expansion into the definition of the mgf yields:Therefore, by differentiating term by term, we obtain .
Since we have the mgf of Gaussian variables, moments can be obtained via differentiation. We also provide a direct proof for the moments of the standard Gaussian.
Direct proof of moments of standard Gaussian
If is odd, then it's obvious that =0, so the following process assumes that is even. Then
7.5 An important property of Standard Gaussian and Mills Ratio
A standard Gaussian variable has a good property in terms of derivatives that
7.6 Expectation of absolute value of Gaussian variables
If a random variable ,
Proof
7.7 Rotational Invariance of Gaussian Variables
If a matrix is orthogonal, indicating that , given a random vector , it satisfies This can be proved using linear transformation of multivariate Gaussian variables.
8 Multivariate Gaussian Distribution
8.1 Definition of Standard Normal Random Vector
A real random vector is called a standard normal random vector if all of its components () are independent standard Gaussian variables. We denote it where the mean vector is and the covariance matrix is .
8.2 Definition of Normal random vector
A real random vector is called a normal random vector if there exists a random normal random vector , a matrix and a -dim vector , such that . (Wikipedia) We denote it , where is the mean vector and is the covariance matrix with where is the covariance between and () and is the mean of ().
8.3 Joint pdf of multivariate Gaussian distribution
Assume that , then
8.4 Characteristic function of multivariate Gaussian variables
If , then
8.5 Linear Transformation of multivariate Gaussian variables
If and , then
9 Chi-squared Distribution
9.1 Definition of Chi-squared Distribution
Assume that we have samples from . Then which is called Chi-squared Distribution with degrees of freedom.
9.2 Pdf of Chi-squared Distribution
From the first part, if , then . Then Let denote the primitive function of . Therefore where is because .
Given that if a random variable , then , therefore for a random variable ,
9.5 Variation of Chi-squared Distribution
Given that if a random variable , then , therefore for a random variable ,
9.6 Mgf of Chi-squared Distribution
Given that for a Gamma distributed random variable , the mgf is since is equivalent to , therefore
9.7 Characteristic function of Chi-squared Distribution
Since for , , therefore
9.8 Asymptotic Property of Chi-squared Distribution
9.8.1 LLN
From the definition of Chi-squared Distribution we have If we treat every as an sample from the same distribution with mean . By the law of large number we can derive thus
9.8.2 CLT
Similar to the LLN part, since we treated every as an sample, therefore we can apply Central Limit Theorem to get convergence in distribution:
10 Exponential Distribution
10.1 Pdf of Exponential Distribution
If a random variable , then the pdf of is
10.2 The tail probability of Exponential Distribution
If with , then the tail probability is
10.3 Memoryless Property
Memoryless property of Exponential distributed variable
Since
Changing the expression to we can conclude that and have the same distribution.
10.4 Relation to Gamma Distribution
Exponential Distribution is a special form of Gamma Distribution, using shape-rate version we can find that if :Therefore
11 Poisson Distribution
11.1 Pmf of Poisson Distribution
Assume random variable , the pmf of is given as follows:
11.2 Expectation and variance of Poisson Distribution
First we have an important property of Poisson Distribution:Using this property we can quickly get that therefore
11.3 Mgf of Poisson Distribution
If a random variable where , then the mgf of is
Proof
By the definition of moment generating function, we have
11.4 Reproducibility of Poisson distribution
If are independent variables satisfying , then
Proof: This is proved by using mgf
11.5 Poisson approximation to the Binomial distribution
For Bernoulli trials with probability , if is large enough and is small enough, then the Binomial distribution is approximately equal to Poisson distribution with parameter .
Proof
If a random variable , then
It's trivial that
Which is the pmf of Poisson distribution.
12 Gamma Distribution
12.1 function
The Gamma function is defined as follows:
An integration by parts shows thatGiven that Thus if is a positive integer greater than 1,
Recursive property of Gamma function
For all , the Gamma function satisfies the following recursion:
12.2 Distribution (shape-scale version)
We say that the continuous random variable has a with parameters and if its pdf is we often write that distribution where is the shape parameter and is the scale parameter.
12.3 Distribution (shape-rate version)
We say that the continuous random variable has a with parameters and if its pdf is we often write that distribution where is the shape parameter and is the rate parameter. Throughout this article we use this version of Gamma distribution.
12.4 Expectation of Distribution
We can use the definition of the Gamma function to simplify the computation of the integral:
12.5 Variation of Distribution
Similarly, we can use the definition of the Gamma function to simplify the computation of the integral:Therefore
12.6 Mgf of distribution
12.6.1 Additivity property of the Gamma Distribution
If and are independent, then
This can be proved using mgf of Gamma distribution.
13 Inverse Gamma Distribution
13.1 Pdf of Inverse Gamma Distribution
If a random variable , then
14 Beta Distribution
14.1 Beta Function
Beta function is defined by integral
Association between Beta function and Gamma function
Proof
Apply the variable substitution by letting , , then the above integral is equivalent to thus we have finished the proof.
Another expression of Beta function
We can get another version of Beta function using variable substitution . Then the integral is equivalent to
14.2 Pdf of beta distribution
The beta distribution is a two-parameter distribution with range and pdf
14.3 Expectation of beta distribution
We can use the pdf of beta distribution to get the expectation of a beta distribution random variable easily, as the computation that follows:
14.4 Variation of beta distribution
Similarly, we can use the pdf of beta distribution to derive , using we can get the variation.
14.5 Relation to Chi-squared Distribution
Let and are independent variables and satisfy . Then we have
Proof: Main idea is variable substitution
15 Cauchy Distribution
15.1 Pdf of Cauchy Distribution
A random variable is said to follow a Cauchy Distribution with location parameter and scale parameter if Therefore .
A special property of Cauchy Distribution
The expectation of Cauchy Distribution doesn't exist.
15.2 Median of Cauchy Distribution
15.3 Characteristic function of Cauchy distribution
16 T Distribution
16.1 Definition
Assume that and are independent. Then the statistic
Is said to follow a -distribution with degrees of freedom.
16.2 Pdf of t-distribution
For a random variable , the density function of is given by
This can be proved using variable substitution
17 F Distribution
17.1 Definition
Assume that and are independent. Then the statisticis said to follow a -distribution with parameters and .
17.2 Pdf of F-distribution
From the definition we have , this is an example of the probability density function of the ratio of two Chi-squared distributed variables. The pdf of is given bellow:
Proof: Mainly using substitution of variables
We use the common technique which is variable substitutions to change the variable:
17.3 Property of F-distribution
Proof
Thus we have finished the proof.
17.4 Expectation and Variation of -distribution
If a random variable , then
Proof of expectation
Given , then the pdf of is given by
Therefore
Let , then the above integral is equivalent to
For variation of -distribution we can use similar calculation as its expectation using the second version of Beta function.