In bivariate, these constant density is an ellipsoids formed when f(x)= a constant and ei are the basis of that plane, ig?
The solid ellipsoid of x values satisfying (x−μ)⊺∑−1(x−μ)≤χp2(α) has probability 1−α
The p-variate normal density has a maximum value when the squared distance is zero, that is, when X=μ. Thus, is the point of maximum density, or mode, as well as the expected value of X , or mean.
The fact that is the mean of the multivariate normal distribution follows from the symmetry exhibited by the constant-density contours: These contours are centered, or balanced, at μ
The following are true for a random vector X having a multivariate normal distribution:
Linear combinations of the components of X are normally distributed.
All subsets of the components of X have a (multivariate) normal distribution.
Zero covariance implies that the corresponding components are independently distributed.
The conditional distributions of the components are (multivariate) normal
Proposition:
If ∑ is a positive definite, so that ∑−1 exists, then ∑e=λe implies ∑−1e=λ1e so (λ,e) are Eigenvalue-Eigenvector pair for ∑ corresponding to the pair (1/λ,e) for λ−1
also, ∑−1 is positive definite
Linear combination variable of component of X: If X is distributed as Np(μ,Σ), then any linear combination of variables a⊺X is distributed as N(a⊺μ,a⊺∑.a).
Also, if a⊺X is distributed as N(a⊺μ,a⊺∑.a) for every a, then X must be Np(μ,∑)
Matrix tranformation variable: If X is distributed as Np(μ,∑), the q linear combinations Aq×pXp×1 (each equation reprensents 1 row of A) are distributed as Nq(Aμ,A∑.A⊺).
and ∑p×p=[∑q×q11∑(p−q)×q21∑q×(p−q)12∑(p−q)×(p−q)22]
then X1 is distributed as Nq(μ1,∑11)
Dependency between combination elements as variable:
If X1(q1×1) and X2(q2×1) are independent, then Cov(X1,X2)=0, a q1×q2 matrix of zeros
If partition [X1X2] is Nq1+q2([μ1μ2],[∑11∑21∑12∑22]) then X1 and X2 are independent if and only if ∑12=0
If X1 and X2 are independent and are distributed as Nq1(μ1,∑11) and Nq2(μ2,∑22) respectively, then [X1X2] has multivariate normal distribution Nq1+q2([μ1μ2],[∑1100∑22])
Conditional expectation: Let X=[X1X2] be distributed as Np(μ,∑) with μ=[μ1μ2],∑=[∑11∑21∑12∑22] and ∣∑22∣>0. Then the conditional distribution of X1 given that X2=x2, is normal and has:
mean = μ1+∑12∑22−1(x2−μ2)
variance = ∑11−∑12∑22−1∑21
note that the covariance does not depend on the value x2 of the conditioning variable
Let X be distributed as Np(μ,∑) with ∣∑∣>0. Then:
The Np(μ,∑) distribution assigns probability 1−α to the solid ellipsoid {x:(x−μ)⊺∑−1(x−μ)≤χp2(α)} where χp2(α) denotes upper (100α)-th percentile of χp2 distribution
Let X1,...,Xn be mutually independent with Xj distributed as Np(μi,∑), then V1=c1X1+...+cnXn is distributed as Np(∑j=1ncjμj,(∑j=1ncj2)Σ)
Moreover, V1 ad V2=b1X1+...+bnXn are jointly multivariate normal with covariance matrix [(∑j=1ncj2)Σ(b⊺c)Σ(b⊺c)Σ∑j=1nbj2)Σ]
consequently, V1 and V2 are independent if b⊺c=0
Multivariate Normal Likelihood:
Assume that the p×1 vectors X1,...,Xn represent a random sample from a multivariate normal population with mean vector μ and covariance matrix ∑
The joint density function of all observations is product of the marginal normal density:
With that, we can have joint density function L(μ,∑)=(2π)−np/2∣∑∣−n/2×exp{−tr[∑−1(∑j=1n(xj−xˉ)(xj−xˉ)⊺+n(xˉ−μ)(xˉ−μ)⊺)]/2}
Maximum likelihood estimation of μ and ∑:
Lemma: Given a p×p symmetric positive definite matrix B and scalar b>0, it follows that ∣∑∣b1exp{−tr(∑−1B)/2}≤∣B∣b1(2b)pbe−pb for all positive definite p×p matrix ∑, with equality holding only for ∑=2b1B
Proposition: Let X1,...,Xn be a random sample from a normal population with mean μ and covariance ∑.
Then μ^=Xˉ and ∑^=n1∑j=1n(Xj−Xˉ)(Xj−Xˉ)⊺=n−1nS are maximum likehood estimators of μ,∑
Their observed values, xˉ and n1∑j=1n(Xj−Xˉ)(Xj−Xˉ)⊺ are maximum likelihood estimates of μ and ∑
Maximum likelihood estimators possess an invariance property
Let θ^ be the maximum likelihood estimator of θ^ and consider estimating the parameter h(θ) Then the maximum likelihood estimate of h(θ) is h(θ^)
Let X1,...,Xn be a random sample from a multivariate normal population with mean μ and covariance μ, then Xˉ and S=n−11∑j=1n(xj−xˉ)(xj−xˉ)⊺ are sufficient statistics
For the multivariate case, Xˉ has a normal distribution with mean μ and covariance matrix (1/n)∑
The sampling distribution of the sample covariance matrix is called the Wishart distribution, defined as the sum of independent products of multivariate normal random vectors
Wm=Wishart distribution with m df = distribution of ∑j=1mZjZj⊺ where Zj are each independently distributed as Np(0,∑)
denote with Wp(n,∑)
theorem Let X1,...,Xn be a random sample of size n from a p-variate normal distribution with mean μ and covariance matrix ∑, then:
Use squared generalized distances dj2=(x−xˉ)⊺S−1(x−xˉ),j=1,2,...,n
applies for all variables with dimension p≥2
When the parent population is multivariate normal and both n and n−p are greater than 30, each of the squared distances dj2 should behave like a chi-square random variable.
Although these distances are not independent or exactly chi-square distributed, it is helpful to plot them as if they were. The resulting plot is called a chi-square plot.
Steps:
find μ by finding all μi
find S by finding all:
Sii=Var(Xi), rmb sample variance
Sij=Cov(Xi,Xj), rmb sample covariance
Construct chi-square plot:
order dj2 from smallest to largest
Graph the pairs (qc,p(j−0.5)/n,d(j)2)
where qc,p(j−0.5)/n is the 100(j−0.5)/n quantile of the chi-square distribution with p degrees of freedom
qc,p(j−0.5)/n=χp2((n−j+0.5)/n,p)← use chi square (=CHISQ.INV.RT((n-j+0.5)/n,2))