Starting from:

$24.99

CSCI5521 Homework 0 Solution


Problem 1: Linear regression learns a linear function of feature variables X to fit the responses y. In this problem, you will derive the closed-form solution for linear regression formulations.
1. The standard linear regression can be formulated as solving a least square problem
minimize ||Xw − y||2,
w
where X ∈ Rn×m (n ≥ m) represents the feature matrix, y ∈ Rn×1 represents the response vector and w ∈ Rm×1 is the vector variable of the linear coefficients. This is a convex objective function of w. Derive the optimal w by setting the derivative of the function wrt w to zero to minimize the objective function.
2. In practice, a L2-norm regularizer is often introduced with the least squares, calledRidge Regression, to overcome ill-posed problems where the hessian matrix is not positive definite. The objective function of ridge regression is defined as
minimize ||Xw − y||2 + λ||w||2,
w
where λ > 0. This objective function is strictly convex. Derive the solution of the ridge regression problem to find the optimal w.
Problem 2: Consider a coin with probability of heads equal to Pr(H) = p and probability of tails Pr(T) = 1 − p. You toss it 5 times and get outcomes H,H,T,T,H.
1. What is the probability of observing the sequence H,H,T,T,H in five tosses. Also givethe formula for the natural logarithm of this probability. Your formulas should be a function of p.
2. You have a box containing exactly 2 coins, one fair with p = 1/2 and one biased with p = 2/3. You choose one of these two coins at random with equal probability, toss it 5 times and get the outcome H,H,T,T,H.
(a) Give the joint probability that the coin chosen was the fair coin (p = 1/2) and the outcome was H,H,T,T,H.
(b) Give the joint probability that the coin chosen was the biased coin (p = 2/3) and the outcome was H,H,T,T,H.
3. What should the bias p = Pr(H) be to maximize the probability of observing H,H,T,T,H, and what is the corresponding probability of observing H,H,T,T,H (i.e., what is the maximum likelihood estimate for p), assuming p were unknown? Show the derivation. Hint: maximize the log of the function.
Problem 3: Below is the pseudo-code of perceptron algorithm for binary classification,
1 w = w0.
2 Do Iterate until convergence
3 For each sample (xt,rt)
4 If(< w,xt > rt ≤ 0)
5 w = w + rtxt
1. Implement the perceptron algorithm and test it on the provided data. To begin, load the file data1.mat into Python. X ∈ R40×2 is the feature matrix of 40 samples in 2 dimensions and y ∈ R40×1 is the label vector (+1/−1). Initialize w to be vector [1,−1]. Visualize all the samples (use different colors for different classes) and plot the decision boundary defined by the initial w.
Now, run your perceptron algorithm on the given data. How many iterations does it take to converge? Plot the decision boundary defined by the w returned by the perceptron program.
. Once the plot is generated you could use xlim() and ylim() functions in pyplot module of matplotlib package to make sure your axes are in a right range.
When you are done your plots will look like the following figures:

2. In previous question the samples from the two classes are linearly separable. Now let’s look at a linearly non-separable case. Load the file data2.mat into Python and run your perceptron algorithm with w = [1,−1]. Can the perceptron algorithm converge? Explain why. To improve the algorithm, we can introduce a ”soft” linear classifier to tolerate errors. It turns out we can solve the following LP problem:
minimize
w,ξt
subject to rt(wT xt) ≥ 1 − ξt ξt ≥ 0.
Here ξt is the error which needs to be minimized. The function linprog() from optimize module of scipy package can be used for the problem. Now, run the following Python code to solve the LP problem on data2.mat.
import numpy as np from scipy . optimize import linprog
m, n = np. shape (X)
X = np. hstack ((X, np. ones ((m, 1)))) n = n + 1
f = np. append(np. zeros (n) , np. ones (m))
A1 = np. hstack ((X∗np. t i l e (y , (n, 1)).T, np. eye (m))) A2 = np. hstack ((np. zeros ((m,n)) , np. eye (m))) A = −np. vstack ((A1, A2)) b = np. append(−np. ones (m) , np. zeros (m)) x = linprog ( f ,A,b)
w = x [ ’x ’ ] [ 0 : n]
Apply this algorithm to data2.mat, visualize the data and plot the boundary by the w returned by LP.
Submission
• Things to submit:
1. hw0 sol.pdf: a document contains all the derivations of Problem 1&2 and the three plots asked by Problem 3.
2. MyPerceptron.py: a Python function defined as MyPerceptron(X, y, w0) with w and step returned, where X is the feature matrix, y is a label vector and w0 is the initialization of the parameter w. In the output, w is the parameter found by perceptron and step represents the number of steps the algorithm takes to converge. The function should also display the plot of samples and boundary.
Note that only numpy pacakge and functions mentioned above are allowed to use in this assignment.
3. Zip all the files into a single zipped file and name it as your name.
• Submit: All material must be submitted electronically via Canvas. This homework will not be graded but required as a proof of satisfying the prerequisites for taking the class.

More products