When A is not full rank, it is advisable to choose the minimum-norm solution to the set of solutions. This particular solution is x=A†y
Variants of least-square problem:
Minimum-norm solutions to linear equations:
When equations are under-determined, we may want to single out a solution with minimum Euclidean norm: minx∣∣x∣∣2:Ax=y
Here, A∈Rm×n with m<n, and y∈R(A)
Optimality conditions and full row rank case:
From the fundamental theorem of linear algebra, any candidate x can be written as x=A⊺v+r where Ar=0. Since A⊺v,r are orthogonal we see that ∣∣x∣∣22=∣∣A⊺v∣∣22+∣∣r∣∣22 which proves optimal r=0 and AA⊺v=y
If A is full row rank, AA⊺ is invertible and the unique v is (AA⊺)−1y then x∗=A⊺(AA⊺)−1y
Equality-constrained LS:
A generalization of the basic LS problem allows for the addition of linear equality constraints on the x variable, resulting in the Constrained Least Squareminx∣∣Ax−y∣∣22 subject to Cx=d
This problem can be converted to standard LS by eliminating the equality constraints via a standard procedure
First, suppose the problem is feasible, let xˉ be such that Cxˉ=d
All feasible points (regardless of least square) can be expressed as a form x=x^+Nz where N contains by the column basis for N(C) and z is a new variable
Then we need to find z such that minz∣∣A~z−y~∣∣22 where A~≐AN and y~≐y−Ax~
then replace x=x~+Nz
Weighted LS:
Sometimes the square errors have different weights. The objective function can be written as f0(x)=∣∣W(Ax−y)∣∣22=∣∣Awx−y∣∣22 where
W=diag(w1,...,wm),Aw≐WA,yw=Wy
Then the problem turns to a ordinary ls problem
l2 regularized LS:
minx∣∣Ax−y∣∣22+λ∣∣x∣∣22,λ≥0
∣∣Ax−y∣∣22+λ∣∣x∣∣22=∣∣A~x−y~∣∣22 where A~≐[AλIn] and y~≐[y0n]
λ≥0 is a tradeoff parameter. Interpretation in terms of tradeoff between output tracking accuracy and input effort.
The optimal regulation parameter, x∗(λ):=argminx∣∣Ax−y∣∣22+λ∣∣x∣∣22