程序代写 CSCI4390/6390 – Data Mining, Fall 2020, Exam II CSCI4390:100pts, CSCI6390:1

CSCI4390/6390 – Data Mining, Fall 2020, Exam II CSCI4390:100pts, CSCI6390:110pts, Bonus: 10pts
Name: RIN: RCS ID:
You must show all work for full credit; it is not enough to show only final answers. You can use a calculator, but not any other device or software. Use at least two decimals for precision (more when necessary).
X1 X2 Y x1 2 2 1 x2 1 2 0 x3 3 1 1 x4 2 3 0

Copyright By PowCoder代写 加微信 powcoder

Table 1: Regression data
1. (CSCI4390:30 points, CSCI6390:40 points) Consider the points in Table 1. Ignore the re-
sponse variable Y . Define the homogeneous quadratic kernel between any two point as follows: K(xi,xj) = (xTi xj)2
(a) (10 points) Compute the kernel matrix K.
(b) (10 points) Compute the total variance in feature space using only K.
(c) (10 points) Compute the angle between x1 and x3 using the Gaussian kernel, given the gaussian kernel matrix:
 1 0.37 0.14 0.37 K = 0.37 1 0.007 0.14  0.14 0.007 1 0.007
0.37 0.14 0.007 1
(d) (CSCI6390: 10 points) Let the dominant eigenvector of the kernel matrix K be
c1 = (0.00325, 0.085, −0.75, 0.66)T Let the centered homogeneous quadratic kernel be given as:
 0.625 1.125 −0.875 −0.875  K =  1.125 18.625 −11.375 −8.375  −0.875 −11.375 33.625 −21.375
−0.875 −8.375 −21.375 30.625
What is the variance along the first kernel PC, and what fraction of the total variance
is captured by the first kernel PC?
2. (30 points) Consider the data shown in Table 1. Let Y be the response variable, and X1 and X2 the predictor variables. The task is to compute the predicted response Yˆ for a linear regression model using the geometric approach. Answer the following questions:

(a) (15 points) Compute an orthogonal basis for independent variables via the QR factor- ization approach for the linear regression model.
(b) (15 points) Using the orthogonal basis, find the predicted response vector Yˆ.
3. (30 points) Given the data in Table 1, assume that the initial augmented weight vector for
logistic regression is w ̃ = (1, 0.5, 0.5)T . Answer the following questions: (a) (15 points) What is the cross-entropy error for this dataset?
(b) (15 points) Update the weight vector w ̃ by computing the gradient at x1; assume the step size for gradient ascent is η = 1 for maximum likelihood estimation.
z3 o2 Figure 1: Neural network
4. (10 points) Given the neural network in Fig. 1. Let bias values be fixed at 0, and let the weight matrices between input and hidden, and hidden and output layers, be given as:
0.5 1 Wh=􏰀1 1 −1􏰁 Wo=1 2
Assume that hidden layer uses ReLU, whereas output layer uses linear activation. Assume that the loss is given via the squared error. Compute the loss when the input is x = 2 and the true response is y = (5, 9)T :
X1 Y 11 23 44 63
Table 2: Kernel Regression Data
5. (Bonus: 10 points) Assume that a kernel ridge regression model is applied on the dataset in Table 2. Find the bias term and regression coefficients assuming we use an inhomogeneous quadratic kernel (with the kernel constant c = 4), and with mixture coefficient vector c = (1, 0.5, −0.25, 0)T , and regression constant α = 10.

程序代写 CS代考 加微信: powcoder QQ: 1823890830 Email: powcoder@163.com