Assignment 6 CS 750/850 Machine Learning

$30.00

Category: You will Instantly receive a download link for .zip solution file upon Payment || To Order Original Work Click Custom Order?

Description

5/5 - (2 votes)

Problem 1 [33%]
In this problem, we will establish some basic properties of vectors and linear functions.
1. The L2 norm of a vector measures the length (size) of a vector. The norm for a vector x of size n is
defined as:
kxk2 =
vuutXn
i=1
x
2
i
Show that the norm can be expressed as the following quadratic expression:
kxk
2
2 = x
T x
.
2. Let a and x be vectors of size n = 3 and consider the following linear function f(x) = a
T x. Show that
the gradient of f is: ∇xf(x) = a.
3. Let A be a symmetric matrix of size 3 × 3 and consider the following quadratic function f(x) = x
T Ax.
Show that the gradient of f is: ∇xf(x) = 2Ax. A matrix is symmetric if Aij = Aji for all i and j.
Problem 2 [34%]
Hint: You can follow the slides from the March 4th class, or the LAR reference from the class website. See
the class website for some recommended linear algebra references.
You will derive the formula used to compute the solution to ridge regression. The objective in ridge regression
is:
f(β) = ky − Aβk
2
2 + λkβk
2
2
Here, β is the vector of coefficients that we want to optimize, A is the design matrix, y is the target, and λ is
the regularization coefficient. The notation k·k2 represents the Euclidean (or L2) norm.
Our goal is to find β that solves:
min
β
f(β)
Follow the next steps to compute it.
1. Express the ridge regression objective f(β) in terms of linear and quadratic terms. Recall that
kβk
2
2 = β
T β. The results should be similar to the objective function of linear regression.
1
2. Derive the gradient: ∇βf(β) using the linear and quadratic terms above.
3. Since f is convex, its minimal value is attained when
∇βf(β) = 0
Derive the expression for the β that satisfies the inequality above.
4. Implement the algorithm for computing β and use it on a small dataset of your choice. Do not forget
about the intercept.
5. Compare your solution with glmnet (or another standard implementation) using a small example. Are
the results the same? Why yes, or no?
Problem 3 [33%]
Using the MNIST dataset, which we used already in Assignment 2, compare whether boosting, bagging, and
random forests work the best. You may may want to use only a subset of the data.
Optional: Use xgboost (also available for Python) to see whether the results are better than other boosting
methods.
Optional Problem 4 [+10%]
Hint: We did not cover this material in class, but it is important regardless. The slides from March 9th may
be helpful when implementing this.
In this problem, you will implement a gradient descent algorithm for solving a linear regression problem.
Recall that the RSS objective in linear regression is:
f(β) = ky − Aβk
2
2
1. Consider the problem of predicting revenue as a function of spending on TV and Radio advertising.
There are only 4 data points:
Revenue TV Radio
20 3 7
15 4 6
32 6 1
5 1 1
Write down the design matrix of predictors, A, and the response vector, y, for this regression problem. Do
not forget about the intercept, which can be modeled as a predictor with a constant value over all data points.
The matrix A should be 4 × 3 dimensional.
2. Express the RSS objective f(β) = ky − Aβk
2
2
in terms of linear and quadratic terms.
3. Derive the gradient ∇βf(β) using the linear and quadratic terms above.
4. Implement a gradient descent method with a fixed step size in R/Python.
5. Use your implementation of linear regression to solve the simple problem above and on a small dataset
of your choice. Compare the solution with linear regression from R or Python (sklearn). Do not forget
about the intercept.
2