Starting from:

$29.99

ISYE6740 Final Exams Solution

If you think a question is unclear or multiple answers are reasonable, please write a brief explanation of your answer, to be safe. Also, show your work if you want wrong answers to have a chance at some credit: it lets us see how much you understood.
You are only required to finish ONE of the Programming Questions, i.e., please pick either Question 8 or Question 9 to work on. If you decide to try both, you will receive BONUS points (please indicate which question you would like to use as BONUS).
Please submit a zip folder, with your final answers (in pdf file format) and code for your Program Question(s).
Name:
GT ID:
GT Account:

Question 1 [10 points]
Question 2 [10 points]
Question 3 [10 points]
Question 4 [10 points]
Question 5 [15 points]
Question 6 [15 points]
Question 7 [10 points]
Question 8 [20 points]
Question 9 [20 points]
1 K-means (10 points)
Given m = 5 data points configuration in Figure 1. Assume K = 2 and use Manhattan distance (a.k.a. the `1 distance: given two 2-dimensional points (x1,y1) and (x2,y2), their distance is |x1 − x2| + |y1 − y2|). Assuming the initialization of centroid as shown, after one iteration of k-means algorithm, answer the following questions.
(a) (3 points) Show the cluster assignment;
(b) (4 points) Show the location of the new center; (c) (3 points) Will it terminate in one step?

Figure 1: Question 1.
2 Clustering [10 pts]
1. (3 points) Explain what is the difference between K-means and spectral clustering?
2. (3 points) What is in common, and what is the main difference between spectral clustering and PCA?
3. (4 points) For the following data (two moons), give one method that will successfully separate the two moons? Explain your rationale.

3 Principal Component Analysis (10 pts)
Suppose we have 4 points in 3-dimensional Euclidean space, namely (1,0,0.5), (6,14,3), (11,28,5.5), and (7,21,3.5).
(a) (3 points) Find the first principal direction. First, write down the data matrixand make sure you fill in numbers specific for this problem. Then explain how to find the first principal direction and what is the optimization problem you need to solve? Find the first principal direction either by calculation in hand, or using program or software report the first principal direction you found.
(b) (2 points) What are the first principle components, for each of the data points?
(c) (2 points) When we reduce the dimensionality from 3 to 1 based on the principaldirection you found in (a), what is the remaining variance in the data?
(d) (3 points) You are given the following 2-D datasets, approximately draw the firstand second principal directional on each plot.

(a) (b)
4 Classification [10 points]
1. (5 points) List all methods below which can be used for classification:
(a) AdaBoost (b) Decision Trees (c) EM and Gaussian Mixture (d) Histogram (e) Knearest neighbors (f) K-means (g) Kernel density estimation (h) Linear Regression (i) Logistic Regression (j) Naive Bayes.
2. (5 points) Which of the decision boundaries below correspond to (a) Random Forest, (b) Decision Tree, (c) SVM. Explain your reasons to fully justify your answers.

5 SVM [15 points]
Suppose we only have four training examples in two dimensions as shown in Fig. The positive samples at x1 = (0,0), x2 = (2,2) and negative samples at x3 = (h,1) and x4 = (0,3).

1. (5 points) For what range of parameter value h > 0 be so that the training points are still linearly separable?
2. (5 points) Does the orientation of the maximum margin decision boundary change as a function of h when the points are separable?
3. (5 points) Explain why only the data points on the “margin” will contribute to the decision boundary?
6 Boosting algorithms [15 points]
In this problem, we test your understanding of AdaBoost algorithm. The figure shows a dataset of 8 points, equally divided among the two classes (positive and negative). The figure also shows a particular choice of decision stump h1 picked by Adaboost in the first iteration.

(a) (6 points) Explain the weights D2(i) for each sample after the first iteration. You can explain by drawing figures similar to what we have in class.
7 Variable section [10 points]
Suppose we have data {xi,yi}, i = 1,...,m, where xi ∈ Rp corresponds to p features.
1. (3 points) Write down the optimization problem we solve with Ridge Regression and Lasso. Make sure you explain your notations: which are the decision variables, and which are data.
2. (3 points) Which of the solution paths below corresponds to Ridge regression and which corresponds to Lasso?

3. (2 points) Explain what’s the difference between Lasso and Ridge regression. We need Lasso for what setting?
4. (2 points) Explain how to tune the regularization parameters for Lasso and Ridge regression (hint: CV).
You are only required to finish ONE of the Programming Questions, i.e., please pick either Question 8 or Question 9 to work on. If you decide to try both, you will receive BONUS points (please indicate which question you would like to use as BONUS).
Please submit a zip folder, with your final answers (in pdf file format) and code for your Program Question(s).
8 PCA for face recognition (20 points)
This question is a simplified illustration of using PCA for face recognition using a subset of data from the famous Yale Face dataset.
Remark: you have to perform downsampling of the image by a factor of 4 to turn them into a lower resolution image before we do anything.
1. (10 points) First, given a set of images for each person, we generate the so-called eigenface using these images. The procedure to obtain eigenface is explained as follows. Given n images of the same person denoted by x1,...,xn. Each image originally is a matrix. We vectorize each image to form the vector xi ∈ Rp. Now form a matrix
X = [x1,...,xn] ∈ Rp×n.
Extract the largest k eigenvector of the data matrix X>X, denoted as u1,u2,...,uk. The eigenfaces correspond to the projections eigenfacei = Xui, i = 1,...,k.
Perform analysis on the Yale face dataset for subject 14 and subject 01, respectively, using all the images EXCEPT for the two images named subject01-test.gif and subject14-test.gif. Plot the top 6 eigenfaces for each subject. When visualizing, you have to reshape the eigenvectors into images with the same dimension as the original images.
What is the interpretation of the top 6 eigenfaces?
2. (10 points) Now we will perform a face recognition task.
For doing face recognition through PCA we proceed as follows. Given the test image subject01-test.gif and subject14-test.gif, we vectorize each image. Take the top eigenfaces of Subject 1 and Subject 14, respectively, project the 2 vectorized test images using the vectorized eigenfaces to obtain scores, respectively. (Hint: use (eigenface1)>(test image) )
Report four scores: (1) projecting test image of Subject 1 using eigenface of Subject 1; (2) projecting test image of Subject 1 using eigenface of Subject 14; (3) projecting test image of Subject 14 using eigenface of Subject 1; (4) projecting test image of Subject 14 using eigenface of Subject 14.
Explain whether or not (and how) can you recognize the faces of the test images using these scores.
You are only required to finish ONE of the Programming Questions, i.e., please pick either Question 8 or Question 9 to work on. If you decide to try both, you will receive BONUS points (please indicate which question you would like to use as BONUS).
Please submit a zip folder, with your final answers (in pdf file format) and code for your Program Question(s).
9 Programming: Bayes and KNN classifier [20 points]
In this programming assignment, you are going to apply the Bayes Classifier to handwritten digits classification problem. Here, we use the binary 0/1 loss for binary classification, i.e., you will calculate the miss-classification rate as a performance metric. To ease your implementation, we selected two categories from USPS dataset in usps-2cls.mat (or usps-
2cls.dat, usps-2cls.csv).
1. (10 points) Your first task is implementing the classifier by assuming the covariance matrices for two classes are a diagonal matrix Σ1, Σ2.
Using slides from “Classification I”, assuming P(y = 1) = P(y = −1) (i.e., the prior distribution for two classes are the same), using Bayes decision rule to write down the decision boundary. (Hint, it should be a quadratic decision boundary.)
Now we will estimate the mean vector and the sample covariance matrices for two classes using the training data (hint: you can use sample mean and sample covariance vector). Report the misclassification rate (error rate) over the training set and over the testing set averaged over the 100 random train/test splits by using different value of splitting ratio p. Explain and compare the performance of each classifier.
After implementing these methods, you should evaluate your algorithm on the given set. Repeat 100 times: split the dataset into two parts randomly, use p portion for training and the other 1−p portion for testing. Let p change from 0.1, 0.2, 0.5, 0.8,
0.9.
Please implement the algorithm from scratch yourself. Make sure to provide code, results (required above) together with necessary explanations to your results.
2. (10 points) Now repeat the classification again using K-nearest neighbors, for K = 5,10,15,30. Repeat 100 times: split the dataset into two parts randomly, use p portion for training and the other 1 − p portion for testing. Let p change from 0.1, 0.2, 0.5, 0.8, 0.9. Report the training error and testing error for each case.

More products