18 In-Class Assignment: Least Squares Fit (LSF)
Contents
18 In-Class Assignment: Least Squares Fit (LSF)#
0. Final Exam Review Survey#
The two class periods during the week of 4/24-4/28 will be used to review for the final exam. There will not be any PCAs for next week, and there are no graded ICAs next week either, but instructors will likely have some sort of review activity for you to work on in class. To help instructors decide what material to focus on during class next week, please fill out the following survey by this Friday 4/21 https://forms.office.com/r/XfSSM3mQnt.
1. LSF Pre-class Review#
19–LSF_pre-class-assignment.ipynb
2. Finding the best solution in an overdetermined system#
Let \(Ax = y\) be a system of \(m\) linear equations in \(n\) variables. A least squares solution of \(Ax = y\) is an solution \(\hat{x}\) in \(R^n\) such that:
Note we substitute \(y\) for our typical variable \(b\) here because we will use \(b\) later to represent the intercept to a line and we want to try and avoid confusion in notation. It also consistent with the picture above.
In other words, \(\hat{x}\) is a value of \(x\) for which \(Ax\) is as close as possible to \(y\). From previous lectures, we know this to be true if the vector $\(y - A\hat{x}\)\( is orthogonal (perpendicular) to the column space of \)A$.
We also know that the dot product is zero if two vectors are orthogonal. So we have
$\(a \cdot (Ax - y) = 0, \)\(
for all vectors \)a\( in the column spaces of \)A$.
The columns of \(A\) span the column space of \(A\). Denote the columns of \(A\) as $\(A = [a_1, \cdots, a_n].\)\( Then we have \)\(a_1^\top (Ax - y) = 0, \\ a_2^\top(Ax-y)=0\\\vdots \\a_n^\top(Ax-y)=0.\)\( It is the same as taking the transpose of \)A\( and doing a matrix multiply: \)\(A^\top (Ax - y) = 0.\)$
That is:
$\(A^\top Ax = A^\top y\)$
The above equation is called the least squares solution to the original equation \(Ax=y\). The matrix \(A^\top A\) is symmetric and invertable. Then solving for \(\hat{x}\) can be calculated as follows:
The matrix \((A^\top A)^{-1}A^\top\) is also called the left inverse.
Example: A researcher has conducted experiments of a particular Hormone dosage in a population of rats. The table shows the number of fatalities at each dosage level tested. Determine the least squares line and use it to predict the number of rat fatalities at hormone dosage of 22.
Hormone level |
20 |
25 |
30 |
35 |
40 |
45 |
50 |
---|---|---|---|---|---|---|---|
Fatalities |
101 |
115 |
92 |
64 |
60 |
50 |
49 |
%matplotlib inline
import matplotlib.pylab as plt
import numpy as np
import sympy as sym
import time
from answercheck import checkanswer
sym.init_printing(use_unicode=True)
H = [20,25,30,35,40,45,50]
f = [101,115, 92,64,60,50,49]
plt.scatter(H,f)
plt.xlabel('Hormone Level')
plt.ylabel('Fatalities')
We want to determine a linear model that is expressed by the following equation
to approximate the connection between Hormone dosage (\(H\)) and Fatalities \(f\). That is, we want to find \(a\) (y-intercept) and \(b\) (slope) for this line. First we define the variable \( x = \left[ \begin{matrix} a \\ b \end{matrix} \right] \) as the column vector that needs to be solved. Then, we need to convert the overdetermined system of equations $\(a + bH_0 = f_0\)\( \)\(a + bH_1 = f_1\)\( \)\(\vdots\)\( \)\(a + bH_6 = f_6\)\( into matrix form \)\(Ax = y\)$.
✅DO THIS: Rewrite the system of equations in matrix form \(Ax=y\) by defining your numpy
matrices A
and y
using the data from above:
# put your code here
checkanswer.matrix(A,'50278283443c1d502428e602f0d363db');
checkanswer.matrix(y, '37e2ed57a1516fb4a17eb2a3e9e99d2d');
✅ QUESTION: Calculate the square matrix \(A^\top A\) (Call it AtA
) and the modified right hand side vector as \(A^\top y\) (Call it Aty
):
#Put your code here
# AtA=
# Aty=
checkanswer.matrix(AtA,'3f579409e55e9cc409bc727c35e09c74');
checkanswer.matrix(Aty,'1dec77018852144a8b90600161c8bff6');
✅QUESTION: Find the least squares solution by solving \(A^\top Ax=A^\top y\) for \(x\).
# Put your code here
# x=
checkanswer.matrix(x,'3213b988975659ed46496a07542eff33');
✅QUESTION: Given the solution above, define the two scalars y-intercept a
and slope b
.
#put your code here
#a=
#b=
checkanswer.float(a,'87e53cb122536f53f434cdbccf0aca94');
checkanswer.float(b,'f7dfa5a5ed0f65f0e0a87f29efd9cd74');
The following code will plot the original data and the linear model whose coefficients we found by performing a least squares fit.
H = [20,25,30,35,40,45,50]
f = [101,115, 92,64,60,50,49]
plt.scatter(H,f)
H2 = np.linspace(np.min(H), np.max(H))
f2 = a + b*H2
plt.plot(H2, f2)
✅QUESTION: Now, let’s fit a quadratic model instead of a linear model to our data, i.e. $\(a + bH + c H^2 = f\)$
To do this, the vector of unknowns is now \(x = \begin{bmatrix}a\\b\\c\end{bmatrix}\), and the overdetermined system of equations is $\(a + bH_0 + cH_0^2 = f_0\)\( \)\(a + bH_1 + cH_1^2 = f_1\)\( \)\(\vdots\)\( \)\(a + bH_6 + cH_6^2= f_6\)$
Again, put this system of equations into matrix form \(Ax = y\), solve \(A^\top Ax = A^\top y\) to find the least squares solution for \(x\), and then extract the coefficients \(a\), \(b\), and \(c\) for your model.
# put your code here
checkanswer.float(a,'8acc26f9f9da4edfa44529a811ebb8a9')
checkanswer.float(b,'b1e317387458d5471bd7cec94e4959fa')
checkanswer.float(c,'451fd8dc5a793f8483160363b42fa8cd')
✅QUESTION: Now, plot the original data along with the curve for the quadratic model \(f = a+bH+cH^2\) whose coefficients we found by performing the least squares fit.
# put your code here
✅QUESTION: Now, see if you can generalize your code above to fit a degree-\(d\) polynomial model $\(x_0 + x_1H + x_2H^2 + \cdots + x_dH^d = f\)\( for some coefficients \)x_0,x_1,x_2,\ldots,x_d\(, where \)3 \le d \le 6$. Again, you should first put the system of equations into matrix form, then find the least squares solution, and finally make a plot showing the original data and the curve for the polynomial model.
# put your code here
d = 3 # Set d to be 3, 4, 5, or 6
# Form a len(H)x(d+1) matrix A whose entries are A[i,j] = H[i]**j
# Solve Ax = y using least squares
# Plot the original data points along with the curve for the polynomial model
✅QUESTION: The interactive function below allows you to adjust the degree of the polynomial model as well as the limits of the \(x\)-axis in the plot. Play with the interactive function below by adjusting the degree of the least-squares fit approximation and extending the x_min
and x_max
parameters. Do you think that an eighth-order polynomial is a good model for this dataset? Why or why not?
from ipywidgets import interact, fixed
import ipywidgets as widgets
@interact(x=fixed(H), y=fixed(f), degree=widgets.IntSlider(min=1, max=8, step=1, value=8), x_min=widgets.IntSlider(min=min(H)-10, max=min(H), step=1, value=min(H)), x_max=widgets.IntSlider(min=max(H), max=max(H)+10, step=1, value=max(H)))
def graphPolyN(x, y, x_min, x_max, degree):
p = np.polyfit(x, y, degree)
f = np.poly1d(p)
x_pred = np.linspace(x_min, x_max, 1000)
y_pred = f(x_pred)
plt.scatter(x, y, color="red")
plt.plot(x_pred, y_pred)
Put your answer to the above question here
3. Pseudoinverse#
In this class we often talk about solving problems of the form:
Currently we have determined that this problem becomes very nice when the \(n \times n\) matrix \(A\) has an inverse. We can easily multiply each side by the inverse:
Since \(A^{-1}A = I\) the solution for \(x\) is simply:
Now, let us consider a a more general problem where the \(m \times n\) matrix \(A\) is not square, i.e. \(m \neq n\) and its rank \(r\) maybe less than \(m\) and/or \(n\). In this case we want to find a Pseudoinverse (which we denote as \(A^+\)) which acts like an inverse for a non-square matrix. In other words we want to find an \(A^+\) for \(A\) such that:
Assuming we can find the \(n \times m\) matrix \(A^+\), we should then be able to solve for \(x\) as follows:
How do we know there is a Psudoinverse#
Assuming the general case of a \(m\times n\) matrix \(A\) where its rank \(r\) maybe less than \(m\) and/or \(n\) (\(r\leq m\) and \(r\leq n\)). We can conclude the following about the fundamental spaces of \(A\):
The rowspace of \(A\) is in \(R^n\) with dimension \(r\)
The columnspace of \(A\) is in \(R^m\) also with dimension \(r\).
The nullspace of \(A\) is in \(R^n\) with dimension \(n-r\)
The nullspace of \(A^T\) is in \(R^m\) with dimension \(m-r\).
Because the rowspace of \(A\) and the column space \(A\) have the same dimension then \(A\) is a the one-to-one mapping from the row space to the columnspace. In other words:
For any \(x\) in the rowspace, we have that \(Ax\) is one point in the columnspace. If \(x'\) is another point in the row space different from \(x\), we have \(Ax\neq Ax'\) (The mapping is one-to-one).
For any \(y\) in the columnspace, we can find \(x\) in the rowspace such that \(Ax=y\) (The mapping is onto).
The above is not really a proof but hopefully there is sufficient information to convince yourself that this is true.
How to compute pseudoinverse#
We want to find the \(n\times m\) matrix that maps from columnspace to the rowspace of \(A\), and \(x=A^+Ax,\) if \(x\) is in the rowspace.
Let’s apply SVD on \(A\): $\(A= U\Sigma V^\top,\)\( where \)U\( is a \)m\times m\( matrix, \)V^\top\( is a \)n\times n\( matrix, and \)\Sigma\( is a diagonal \)m\times n\( matrix. We can decompose the matrices as \)\(A = \begin{bmatrix}\vdots & \vdots \\ U_1 & U_2 \\ \vdots &\vdots\end{bmatrix} \begin{bmatrix}\Sigma_1 & 0 \\ 0 & 0\end{bmatrix} \begin{bmatrix}\cdots & V_1^\top & \cdots \\ \cdots & V_2^\top &\cdots \end{bmatrix}.\)\( Here \)U_1\( is of \)m\times r\(, \)U_2\( is of \)m\times (m-r)\(, \)\Sigma_1\( is of \)r\times r\(, \)V_1^\top\( is of \)r\times n\(, and \)V_2^\top\( is of \)(n-r)\times n$.
The columnspace of \(U_1\) is the columnspace of \(A\), and columnspace of \(U_2\) is the nullspace of \(A^\top\).
The rowspace of \(V_1\) is the rowspace of \(A\), and rowspae of \(V_2\) is the nullspace of \(A\).
If \(x\) is in the rowspace of \(A\), we have that \(V_2^\top x=0\). We have \(Ax = U_1\Sigma_1 V_1^\top x\).
If we define a matrix \(B=V_1\Sigma_1^{-1}U_1^\top\), we have that \(BAx=V_1\Sigma_1^{-1}U_1^\top U_1\Sigma_1 V_1^\top x=V_1V_1^\top x\). That is \(BAx=x\) is \(x\) is in the rowspace of \(A\).
The matrix \(B\) is the pseudoinverse of matrix \(A\). $\(A^+ = V_1\Sigma_1^{-1}U_1^\top\)\( \)\(A^+ = \begin{bmatrix}\vdots & \vdots \\ V_1 & V_2 \\ \vdots &\vdots\end{bmatrix} \begin{bmatrix}\Sigma_1^{-1} & 0 \\ 0 & 0\end{bmatrix} \begin{bmatrix}\cdots & U_1^\top & \cdots \\ \cdots & U_2^\top &\cdots \end{bmatrix}.\)$
Example 1: Let $\(A=\begin{bmatrix}1 & 0 & 1\\ 1 & 2 & 3\end{bmatrix}\)\( we know that \)r=m=2\( and \)n=3$.
A = np.matrix([[1,0,1],[1,2,3]])
✅TODO: Calculate the pseudoinverse \(A^+\) of \(A\) using the numpy.linalg
function pinv
:
# put your code here
# Apinv=
checkanswer.matrix(Apinv,'7724dc0d898980cb5a80dac73bfa7aa7');
✅DO THIS: Compute \(AA^+\) and \(A^+A\)
#put your code here
✅QUESTION: If \(x\) is in the nullspace of \(A\) what is the effect of \(A^+Ax\)?
Put your answer to the above question here
✅QUESTION: If \(x\) is in the rowspace of \(A\) what is the effect of \(A^+Ax\)?
Put your answer to the above question here
Left inverse is pseudoinverse#
We can compute the left inverse of \(A\) if \(r=n\leq m\). In this case, we may have more rows than columns, and the matrix \(A\) has full column rank.
In this case, the SVD of \(A\) is $\(A = U\Sigma V^\top .\)\( Here \)U\( is of \)m\times n\(, \)\Sigma\( is of \)n\times n\( and nonsingular, \)V^\top\( is of \)n\times n\(. The pseudoinverse of \)A\( is \)\(A^+ = V\Sigma^{-1}U^\top\)$
The left inverse of \(A\) is $\((A^\top A)^{-1}A^\top= (V\Sigma U^\top U\Sigma V^\top )^{-1} V\Sigma U^\top = V(\Sigma \Sigma )^{-1} V^\top V\Sigma U^\top = V\Sigma ^{-1} U^\top =A^+\)$
Example 2: Let $\(A=\begin{bmatrix}1 & 1 \\ 0 & 2 \\ 1 & 3\end{bmatrix}\)\( we know that \)r=n=2\( and \)m=3$.
A = np.matrix([[1,1],[0,2],[1,3]])
A
✅DO THIS: Calculate the pseudoinverse \(A^+\) of \(A\).
#Put your answer here
# pInv_A =
checkanswer.matrix(pInv_A,'fda9d63c9e6cdc7bdb2acc57c4bdf490');
✅DO THIS: Calculate the left inverse of \(A\), and verify that it is the same as \(A^+\).
#Put your anaswer here
# leftInv_A =
checkanswer.matrix(leftInv_A,'fda9d63c9e6cdc7bdb2acc57c4bdf490');
Written by Dr. Dirk Colbry with interactive code David Yonkers, Michigan State University. Some modifications by Dr. Santhosh Karnik, Michigan State University
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.