Starting from:

$30

Stat435 Homework3- Experiments with Turbo Solved

The homework comes with a test data set ”test-data.R” in ”dump” format. For the test data, n = 100, and x1,...,xn are equi-spaced in [0,2π]. The true conditional expectation is f(x) = sin(x), and the error sd is σ = 0.4.

1.     Experiments with Turbo
Turbo is an expansion based smoother that fits 2nd order (linear) splines. It is described in the article Flexible Parsimonious Smoothing and Additive Modeling by J.H. Friedman and B.W. Silverman (Technometrics, Vol. 31, No. 1, 1989, pp 3 – 39.

a)   (10 points) Define basis functions Bi(x) = (x−xi)+,i = 1,...,n−1, and

Bn(x) = 1. Write a function truncated.power.design.matrix(x) that generates the n × n design matrix for this set of basis functions.

b)   (10 points) Install the package ”leaps” and take a look at the documentation. Write a function

regsubsets.fitted.values <- function(X, regsubsets.out, nterm) that computes the fitted values for a model with nterm terms.

c)    (10 points) For the test data produce a plot of residual sum of squares as a function of the number k of basis functions in the model.

d)   (10 points) Plot the GCV score as a function of k. Surprised? Why? Explanation?

e)   (10 points) F&S (pp 9–10) propose to fix this problem by charging 3 degrees of freedom for each of B1,...,Bn−1 entered into the model. Plot this modified GCV score as a function of the number of basis functions in the model. Surprised? Problems with the F&S definition of GCV? (If you have trouble figuring out where the constant term is included in the model, you may charge 3 degrees of freedom for each of B1,...Bn.)

f)     (10 points) Restricting yourself to suitable small values of k, find the ”forward” and ”backward” models with the smallest (modified) GCV scores and plot them.

1

2.     Experiments with order 2 smoothing splines
Training data and basis functions as in (1) above. Define the n × n matrix X by Xij = Bj(xi).

An order 2 smoothing spline is a function of the form

g(x)
=
X

aˆjBj(x),where

=
argmina ky − Xak2 + λaTΩai with h

=
diag(0,1,...,1,0).
The vector of predicted values for the training sample is yˆ = Xaˆ.

(a)    (10 points) Show that

yˆ  = X(XTX + λΩ)−1XTy =    Sλ y.

(b)   (10 points) Read the data in the file test-data.R. Use the glmnet package to plot data and spline for λ = 0,1,10,106. Verify (graphically) that the spline for λ = 106 is very close to the least squares line.

(c)    (20 points) Use the glmnet package to find the optimal value of λ by cross-vaildation. Print out λopt and plot the corresponding spline.

3.     ISLR Section 6.8 Problem 1 (25 points)

4.     ISLR Section 6.8 Problem 4 (25 points)

More products