$34.99
100 points
Overview
In this assignment you will explore different approaches to analyzing Markov chains. You will use one data sets for this assignment:
• http://www.cs.utah.edu/˜jeffp/teaching/cs5140/A8/M.csv
1 Finding q∗ (100 points)
We will consider four ways to find q∗ = Mtq0 as t →∞.
Matrix Power: Choose some large enough value t, and create Mt. Then apply q∗ = (Mt)q0. There are two ways to create Mt, first we can just let Mi+1 = Mi ∗ M, repeating this process t − 1 times. Alternatively, (for simplicity assume t is a power of 2), then in log2 t steps create M2i = Mi ∗ Mi.
State Propagation: Iterate qi+1 = M ∗ qi for some large enough number t iterations.
Random Walk: Starting with a fixed state q0 = [0,0,...,1,...,0,0]T where there is only a 1 at the ith entry, and then transition to a new state with only a 1 in the jth entry by choosing a new location proportional to the values in the ith column of M. Iterate this some large number t0 of steps to get state . (This is the burn in period.)
Now make t new step starting at and record the location after each step. Keep track of how many times you have recorded each location and estimate q∗ as the normalized version (recall kq∗k1 = 1) of the vector of these counts.
Eigen-Analysis: Compute LA.eig(M) and take the first eigenvector after it has been L1-normalized.
A (40 points): Run each method (with t = 1024, q0 = [1, 0, 0, ..., 0]T and t0 = 100 when needed) and report the answers.
B (20 points): Rerun the Matrix Power and State Propagation techniques with q0 = [0.1,0.1,...,0.1]T. For what value of t is required to get as close to the true answer as the older initial state?
D (8 points): Is the Markov chain ergodic? Explain why or why not.
E (8 points): Each matrix M row and column represents a node of the graph, label these from 0 to 9 starting from the top and from the left. What nodes can be reached from node 4 in one step, and with what probabilities?
2 BONUS: Taxation (5 points)
Repeat the trials in part 1.A above using taxation β = 0.85 so at each step, with probability 1 − β, any state jumps to a random node. It is useful to see how the outcome changes with respect to the results from Question 1. Recall that this output is the PageRank vector of the graph represented by M.
Briefly explain (no more than 2 sentences) what you needed to do in order to alter the process in question 1 to apply this taxation.