Please answer the following questions in complete sentences in a clearly prepared manuscript and submit the solution by the due date on Gradescope.
Remember that this is a graduate class. There may be elements of the problem statements that require you to fill in appropriate assumptions. You are also responsible for determining what evidence to include. An answer alone is rarely sufficient, but neither is an overly verbose description required. Use your judgement to focus your discussion on the most interesting pieces. The answer to "should I include 'something' in my solution?" will almost always be: Yes, if you think it helps support your answer.
Please identify anyone, whether or not they are in the class, with whom you discussed your homework. This problem is worth 1 point, but on a multiplicative scale.
Make sure you have included your source-code and prepared your solution according to the most recent Piazza note on homework submissions.
Show that is a convex function. Feel free to use the result proved on the last homework.
Show that the null-space of a matrix is a convex set. (A convex set satisfies the condition that, for every pair of points in the set, any point on the line joining those points is also in the set.)
The Ridge Regression problem is a variant of least squares: This is also known as Tikhonov regularization.
Show that this problem always has a unique solution, for any if , using the theory discussed in class so far.
This property is one aspect of the reason that Ridge Regression is used. It is also a common regularization method that can help avoid overfitting in a regression problem.
Use the SVD of to characterize the solution as a function of .
What is the solution when ?
What is the solution when .
Give the solutions of the least squares problem using the sports teams data when and . (Here, you will not need the constraint that the sum of entries is 1.) Any notable differences from the ranking giving in class?
Suppose that you only want to regularize two components of the solution, say, , so that your optimization problem is Show how to adapt your techniques in this problem to accomplish this goal.
Find the minimizer of where also (that is, we have added one linear constraint) and justify, as concisely and correctly as you can, that you have found a solution. Do the terms local and global minimizer apply? If so, use one of them to describe your solution. Furthermore, evaluate the gradient of at your solution. What is notable about the gradient at your solution in comparison with the solution where have no constraints?
Consider the constrained least squares problem: where , and rank .
Convert this problem into the standard constrained least squares form.
Form the augmented system from the Lagrangian as we did in class.
Manipulate this problem to arrive at the normal equations
for a least-squares problem: .
Discuss any advantages of the systems at intermediate steps.