This file gives the elementary information on Expectation, Variance and Covariance for the Random Variables.
If X is a random variable with output space H and with density f(x).
Then, E[X]=∫Hxf(x)dx
If X and Y are two random variables with respective outputs spaces H and K, with respective densities f and g, with a joint density f(x,y) and with a finite expectation.
We have E[X+Y]=E[X]+E[Y]
Let consider the function q defined by q(x,y)=x+y
E[q(X,Y)]=∫H∫Kq(x,y)f(x,y)dxdy
E[Z]=∫H∫K(x+y)f(x,y)dxdy
E[Z]=∫H∫Kxf(x,y)dxdy+∫H∫Kyf(x,y)dxdy
E[Z]=∫Hx∫Kf(x,y)dydx+∫Ky∫Hf(x,y)dxdy
E[Z]=∫Hxf(x)dx+∫Ky(y)g(y)dy
E[Z]=E[X]+E[Y]
We also have
E[aX]=aE[X]
If X is a positive random variable i.e P(X≥0)=1 then
E[X]≥0
Where X is a positive random variable, if E[X]=0, then
P(X=0)=1
E[a]=a
Cov(X,Y)=E[(X−E[X])(Y−E[Y])]
Cov(X,Y)=Cov(Y,X)
Sometimes more convenient
Cov(X,Y)=E[XY]−E[X]E[Y]
In particular
Var(X)=Cov(X,X)=E[X2]−E[X]2
Cov(aX+bY,Z)=aCov(X,Z)+bCov(Y,Z)
Cov(X,X)=E[(X−E[X])(X−E[X])]=Var(X)
Note that the variance is always positive (as the expectation of a square of a random variable).
Cov(X,a)=0
Consequence :
Var(a)=0
Reciprocally, if a random variable has a variance equal to 0, then the variable is constant.
Var(n∑i=1λiZi)=n∑i=1n∑j=1λiλjCov(Zi,Zj)
Var(aX)=a2Var(X)
Var(aX+bY)=a2Cov(X,X)+2abCov(X,Y)+b2Cov(Y,Y)
Var(aX−bY)=a2Cov(X,X)−2abCov(X,Y)+b2Cov(Y,Y)
Var(X+a)=Var(X)
When we have a set of random variables Z1,…,Zn.
For each pair (k,l), if we denote ckl=Cov(Zk,Zl)
We can store the ckl’s in a matrix Σ=[c11…c1nc21…c2n⋮⋱⋮cn1…cnn]
Σ is named the covariance matrix of the random vector Z=[Z1⋮Zn]
Note that we can rewrite
Var(n∑i=1λiZi)=λTΣλ
where λ=[λ1⋮λn]
and T designates the transposition
λT=[λ1…λn]
Since a variance is always positive, the variance of any linear combination as to be positive. Therefore, a covariance matrix is always (semi-)positive definite, i.e
For each λ λTΣλ≥0
Let consider two random vectors X=(X1,…,Xn) and Y=(Y1,…,Yp).
We can consider the cross-covariance matrix Cov(X,Y) where element corresponding to the row i and the column j is Cov(Xi,Yj)
If A and B are some matrices (of constants)
Cov(AX,BY)=ACov(X,Y)BT
Suppose that we want to estimate a quantity modeled by a random variable Z0 as a linear combination of known quanties Z1,…,Zn stored in a vector Z=[Z1⋮Zn]
We will denote Z⋆0=n∑i=1λiZi=λTZ this (random) estimator.
We know the covariance matrix of the full vector (Z0,Z1,…,Zn) that we write with blocks for convenience:
[σ20cT0c0C]
where
Compute the variance of the error Z⋆0−Z0
Var(Z⋆0−Z0)=Cov(Z⋆0−Z0,Z⋆0−Z0)
Var(Z⋆0−Z0)=Var(Z0)−2Cov(Z⋆0,Z0)+Var(Z0)
Var(Z⋆0−Z0)=Var(λTZ)−2Cov(λTZ,Z0)+σ20
Var(Z⋆0−Z0)=λTVar(Z)λ−2λTCov(Z,Z0)+σ20
Var(Z⋆0−Z0)=λTCλ−2λTc0+σ20
The covariance is a measure of the link between two variables. However it depends on the scale of each variable. To have a similar measure which is invariant by rescaling, we can use the correlation coefficient:
ρ(X,Y)=Cov(X,Y)√Var(X)Var(Y)
When the correlation coefficient is equal to 1 or −1, we have
Y=aX+b
with
Note that ρ(X,Y) can be equal to 0 even if the variables are strongly linked.
The usual example is a variable X with a pair density (f(−x)=f(x)) and Y=X2:
Cov(X,Y)=Cov(X,X2)=E[X3]−E[X]E[X2]=E[X3]=∫Rx3f(x)dx=0