Starting from:

$30

Machine Learning Homework 4 -Solved


I. Pen-and-paper [11v] 
Given the bivariate observations {(
1
2
) , (−11) , (
1
0
)}, 
and the multivariate Gaussian mixture 
𝐮1 = (2
2) , 𝐮2 = (0
0) , 𝚺1 = (2 

1 2
) , 𝚺2 = (2 


2) , 𝜋1 = 0.5, 𝜋2 = 0.5. 
1) [7v] Perform one epoch of the EM clustering algorithm and determine the new parameters. 
Indicate all calculus step by step (you can use a computer, however disclose intermediary steps). 
2) Given the updated parameters computed in previous question: 
a. [1.5v] perform a hard assignment of observations to clusters under a MAP assumption. 
b. [2.5v] compute the silhouette of the larger cluster using the Euclidean distance.
 
II. Programming and critical analysis [9v] 
Recall the pd_speech.arff dataset from earlier homeworks, centered on the Parkinson diagnosis from 
speech features. For the following exercises, normalize the data using sklearn’s MinMaxScaler. 
1) [4.5v] Using sklearn, apply k-means clustering fully unsupervisedly (without targets) on the 
normalized data with 𝑘 = 3 and three different seeds (using random ϵ {0,1,2}). Assess the 
silhouette and purity of the produced solutions. 
2) [1.5v] What is causing the non-determinism? 
3) [1.5v] Using a scatter plot, visualize side-by-side the labeled data using as labels: i) the original 
Parkinson diagnoses, and ii) the previously learned 𝑘 = 3 clusters (random= 0). To this end, select 
the two most informative features as axes and color observations according to their label. For feature 
selection, select the two input variables with highest variance on the MinMax normalized data. 
4) [1.5v] The fraction of variance explained by a principal component is the ratio between the 
variance of that component (i.e., its eigenvalue) and total variance (i.e., sum of all eigenvalues). 
How many principal components are necessary to explain more than 80% of variability? 
Hint: explore the DimReduction notebook to be familiar with PCA in sklearn. 
END

More products