Home









                          Least squares




regularized least squares


     We can add a regularization term to an error function in or‐
der  to control over‐fitting, so that the total error function to
be minimized takes the form:



where λ is the regularization coefficient that controls the rela‐
tive importance of the data‐dependent error E_D and the  regular‐
ization term E_W.

The simplest form of regularizer:








     Inconsistent  systems  arise  often in applications.  When a
solution is demanded and none exists, the best one can do  is  to
find an x that makes Ax as close as possible to b.

The general least‐squares problem is to find an x that makes ||b‐
Ax|| as small as possible.

The  adjective  "least‐squares"  arises  from the fact that ||b ‐
Ax|| is the square root of a sum of squares.

The most important aspect of the least‐squares problem is that no
matter what x we select, the vector Ax will necessarily be in the
column space, Col A. So we seek an x that makes  Ax  the  closest
point  in  Col  A  to b. (Of course, if b happens to be in Col A,
then b is Ax for some x, and such an x is a "least‐squares  solu‐
tion."

When we talk about spaces, we talk about points and distances be‐
tween points, and opertations allowed between points. What is the
distance  between  a  space and a point which does not lie in the
space? (projection)












                               ‐2‐




Such an xˆ is a list of weights that will build  bˆ  out  of  the
columns  of  A. Think of multiplying these weights to each corre‐
sponding column of A, and then sum up each row of A to get Axˆ.

What is a column space?

Col A is the set of all linear combinations of the columns of A.




What is a null space?

Nul A is the set of all solutions  of  the  homogeneous  equation
Ax=0.