$25
CSE 312: Foundations of Computing II
Assignment #8
Maestro Lazy decides to assign final grades in CSE 312 by ignoring all the work thestudents have done and instead using the following probabilistic method: each student independently will be assigned an A with probability θ, a B with probability 4θ, a C with probability , and an F with probability . When the quarter is over, you discover that only 5 students got an A, 25 got a B, 100 got a C, and 90 got an F. Find the maximum likelihood estimator for the parameter θ that Maestro Lazy used. Give an exact answer as a simplified fraction.
Maestro Lazy decides to assign final grades in CSE 312 by using a more sophisticatedprobabilistic method: each student independently will be assigned an A with probability θ1, a B with probability θ2− θ1, a C with probability , and an F with probability
. When the quarter is over, you discover that only 5 students got an A, 25 got a B, 100 got a C, and 90 got an F. Find the maximum likelihood estimators for the parameters θ1 and θ2 that Maestro Lazy used. Give exact answers as simplified fractions. For this problem, you do not have to demonstrate that your estimators are local maxima, which turns out to be more challenging than just computing second derivatives.
(a) Let x1,x2,...,xn be independent samples from a geometric distribution with unknown parameter p. What is the maximum likelihood estimator for p?
(b) If the samples from the geometric distribution are 5, 4, 3, 2, 9, 5, 3, 9, what is the maximum likelihood estimator for p? Give an exact answer as a simplified fraction.
Let x1,x2,...,xn be independent samples from an exponential distribution with unknown parameter λ. What is the maximum likelihood estimator for λ? (If you do Exercises 3a and 4 correctly, you should see a very natural relationship between the estimators.)
(a) Suppose x1,x2,...,xn are the numbers of independent requests arriving at a web server during each of n successive minutes. Under the assumption that these numbers are independent samples from a Poisson distribution with unknown rate λ per minute, what is the maximum likelihood estimator for λ? (If you do Exercises 4 and 5a correctly, you should see a very natural relationship between the estimators.)Is your estimator unbiased? Justify your answer.
I have played 12273 games against the devilish Schnapsen program Doktor Schnaps and have won 5926 of them. Let’s suppose that each game’s outcome Xi is Ber(p), where p is my probability of winning a single game, and that the outcomes are all independent. The goal of this exercise is to estimate my long-term winning probability p against Doktor Schnaps. Since Doktor Schnaps and I are nearly evenly matched, you can use the estimate Var(Xi) = 0.25 (which would be exact if). Give all your answers to 4 significant digits.What is the maximum likelihood estimate ˆp of p for the 12273 games I have played? You do not need to rederive the formula for this estimator, since we did it in lecture. Just substitute the numbers into the formula to arrive at a numerical estimate.
Approximate the 99% confidence interval for p. Give your answer as the least value of ∆ such that P(ˆp − ∆ ≤ p ≤ pˆ + ∆) ≥ 0. (For consistency for the grader’s sake, when you use the standard normal distribution table, choose the entry that barely makes this inequality true.) Once you have provided ∆, also specify the confidence interval [ˆp − ∆,pˆ+ ∆] to 4 significant digits.
For your interest, Doktor Schnaps actually makes use of these results: the player’s score in the last column of the leaderboard at http://schnapsen.realtype.at/index.php? page=gesamtwertung is the lower boundary ˆp−∆ of the 99% confidence interval. This choice prevents a player who has only played a few lucky games from being ranked too high, since ∆ decreases as the number of games played increases. You can see examples of how this affects ranking in the leaderboard table, where the next-to-last column is pˆ but the table is sorted on the last column ˆp − ∆.
(a) Let x1,x2,...,xn be independent samples from the continuous uniform distribution on [0,θ], where θ is the unknown parameter. What is the maximum likelihood estimator θˆ for θ?
Hint: It might help to roughly sketch out the shape of the likelihood function as a function of θ, that is, θ on the horizontal axis and likelihood L on the vertical axis. Start your sketch by asking what happens to L as θ goes to infinity. Then figure out the shape as you get closer to θ = 0. From your sketch, at what value of θ do you maximize L?
Does your answer to part (a) seem as though it would produce the most accurateestimator for θ? Explain.
In the remaining parts of this exercise, you will quantify your answer to part (b) by computing the bias of the estimator θˆ. Begin by computing the cumulative distribution function for the random variable θˆ. Recall that this is simply the function F(x) = P(θ <ˆ x). Focus first on the interval 0 ≤ x ≤ θ, but when you’re done with that, don’t forget to also define F(x) on the rest of the real numbers.
From your answer to part (c), compute the probability density function f(x) of θˆ.
From your answer to part (d), compute E[θˆ]. Is θˆ an unbiased estimator of θ? Was your intuition in part (b) consistent with your answer to this part?
Starting from the value of E[θˆ] you computed in part (e), determine an unbiased estimator of θ and show that it is unbiased. Although there are other unbiased estimators of θ, you are to derive yours by modifying your answer from part (e).