[SOLVED] 机器学习代写:CS 189 HW01

30 $

File Name: 机器学习代写:CS_189_HW01.zip
File Size: 301.44 KB

SKU: 6514799466 Category: Tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , ,

Or Upload Your Assignment Here:


CS 189 Introduction to Machine Learning Fall 2017

This homework is due Friday, September 1 at 12pm noon. 1 Getting Started

HW1

You may typeset your homework in latex or submit neatly handwritten and scanned solutions. Please make sure to start each question on a new page, as grading (with Gradescope) is much easier that way! Deliverables:

1. Submit a PDF of your writeup to assignment on Gradescope, “HW1 Write-Up” 2. Submit all code needed to reproduce your results, “HW1 Code”.
3. Submit your test set evaluation results, “HW1 Test Set”.

After you’ve submitted your homework, be sure to watch out for the self-grade form.

(a) Before you start your homework, write down your team. Who else did you work with on this homework? List names and email addresses. In case of course events, just describe the group. How did you work on this homework? Any comments about the homework?

(b) Please copy the following statement and sign next to it:

I certify that all solutions are entirely in my words and that I have not looked at another student’s solutions. I have credited all external sources in this write up.

CS 189, Fall 2017, HW1 1

2 “Sample Complexity” of coupon collecting

Roald Dahl’s mathematical version of his book has the following problem. Willy Wonka has hidden 50 different types of cards in his chocolate wrappers, and some of the people to find all 50 types will be allowed to enter his factory and participate in a questionable experiment. Charlie goes to a local store to buy his chocolates, and can afford at most one chocolate a day. The store stocks 20 chocolates of each card type, and every time a chocolate is bought, it is replaced with another chocolate containing an identical card. Charlie picks a chocolate uniformly at random.

  1. (a)  What is the expected number of days it takes Charlie to qualify for Willy Wonka’s challenge if he picks a random chocolate every time he goes to the shop?
  2. (b)  The game has now changed. Instead of having to collect all d card types, at the end of the game Willy Wonka will pick a card uniformly at random and if someone has that card in their collection, they win a smaller prize. If you want your probability of winning to be at least 1 − δw, how many cards should you have in your collection?
  3. (c)  Charlie just decides to buy chocolates for n days. At the end of this, he has some random number of distinct cards. Given that there are d distinct cards and assuming that Charlie buys αd chocolates, what is the probability that Charlie wins a prize?
  4. (d)  What happens as d gets large?
  5. (e)  Now, think about something that might seem completely different. We want to learn a com- pletely unstructured function f on a finite domain of size d. We get training data from ran- dom samples, each of which picks a domain point U uniformly at random and then returns us (U , f (U )) — i.e. the domain point along with the function f ’s evaluation on that domain point. How big of a training set should we collect so that with probability 1 − δ we will be able to successfully estimate the function when presented with a uniformly drawn domain point.

3 The accuracy of learning decision boundaries

This problem exercises your basic probability (e.g. from 70) in the context of understanding why lots of training data helps improve the accuracy of learning things.

For each θ ∈ (1/4, 3/4), define fθ : [0, 1] → {0, 1}, such that 1 if x > θ

fθ (x) = 0 otherwise. The function is plotted in Figure ??.

We draw samples X1,X2,…,Xn uniformly at random and i.i.d. from the interval [0,1]. Our goal is to learn an estimate for θ from n random samples (X1, fθ(X1)),(X2, fθ(X2)),…,(Xn, fθ(Xn)).

Let N0 := #{i : fθ (Xi) = 0} denote the number of samples we obtain that have function evaluation 0.

CS 189, Fall 2017, HW1 2

Figure 1: Plot of function fθ (x) against x.
LetTmin =max({1}∪{Xi|fθ(Xi)=0}). Weknowthatthetrueθ mustbelargerthanTmin.

4

LetTmax =min({3}∪{Xi|fθ(Xi)=1}). Weknowthatthetrueθ mustbesmallerthanTmax. 4

The gap between Tmin and Tmax represents the uncertainty we will have about the true θ given the training data that we have received.

  1. (a)  WhatistheprobabilitythatTmax−θ>εasafunctionofε.
  2. (b)  Whatistheprobabilitythatθ−Tmin>εasafunctionofε.
  3. (c)  Suppose that you would like to have an estimate for θ within an accuracy of 2ε with probability at least 1 − δ . Please bound or estimate how big of an n do you need?
  4. (d)  Let us say that instead of getting random samples (Xi,f(Xi)), we were allowed to choose where to sample the function, but you had to choose all the places you were going to sample in advance. Propose a method to estimate θ. How many samples suffice to achieve an estimate that is ε-close (within an interval of size 2ε) as above? (Hint: You need not use a randomized strategy.)
  5. (e)  Suppose that you could pick where to sample the function adaptively — choosing where to sample the function in response to what the answers were previously. Propose a method to estimate θ . How many samples suffice to achieve an estimate that is ε -close (within an interval of size 2ε) as above?
  6. (f)  Compare the scaling of n with ε and δ in the three sampling approaches above: random, deterministic, and adaptive.

CS 189, Fall 2017, HW1 3

4 Eigenvalue and Eigenvector review

A square matrix A ∈ Rd×d has a (right) eigenvalue λ ∈ C and (right) eigenvector ⃗x ∈ Cd 0 if P⃗x = λ⃗x. Left eigenvalues and eigenvectors are defined analogously — ⃗xT P = λ⃗xT . Since the definition is scale invariant (if ⃗x is an eigenvector, then t⃗x is an eigenvector for any t ̸= 0), we adopt the convention that each eigenvector has norm 1.

(a) Compute the right and left eigenvalues and eigenvectors of the following matrices. 3 2

i)A= 1 3 5 2

ii)B= 2 5

iii) A2 iv) B2 iv) AB v) BA

  1. (b)  Compute the singular value decompositions of the matrices above. In addition, please compute 5 2

    theSVDof:C=2 5 3 2

    13

  2. (c)  Show from the definition of a right eigenvalue that the quantity λ is an eigenvalue with asso-

    ciated eigenvector ⃗x iff for all 1 ≤ i ≤ n, we have
    (λ −Aii)xi = ∑Aijxj.

    j̸=i

  3. (d)  Now for an arbitrary eigenvalue λ of A and its associated eigenvector ⃗x, choose index i such

    that |xi| ≥ |xj| for all j ̸= i. For such an index i, show that |λ −Aii| ≤ ∑|Aij|.

    j̸=i

    You have just proved Gershgorin’s circle theorem, which states that all the eigenvalues of a d × d matrix lie within the union of d disks in the complex plane, where disk i has center Aii, and radius ∑ j̸=i |Ai j |.

5 Fun with least squares

Ordinary least squares for a scalar helps us learn linear predictors for a scalar b given observations ⃗a. This is done by taking our training data points and constructing a tall vector⃗b by stacking up

CS 189, Fall 2017, HW1 4

our training data’s bi and a corresponding tall matrix A by stacking up our training data’s ⃗aTi , and then finding the weights⃗x that minimize ∥A⃗x−⃗b∥2. The resulting weights⃗x can be used to predict b’s by⃗xT⃗a.

Let us think about the components of ⃗a as being a sequence of measurements in time. The first measurement, the second measurement, the third measurement, and so on.

  1. (a)  Suppose that you wanted to construct not one linear predictor for b but a sequence of them. One that just used the first measurement. One that used the first two measurements. All the way up to one that uses all the measurements. How would you do this in the most straightforward manner?
  2. (b)  Someone suggests that maybe the measurements themselves are partially predictable from the previous measurements. They suggest a two part strategy. First we predict the next mea- surement based on the measurements so far. And then we look at the difference (sometimes deemed the “innovation”) between the actual measurement we made and our prediction for it, and just use that to update our prediction for b.

    Give a way to learn (from the training data) these best predictors of the next measurement from the previous measurements as well as learn (from the training data) the weights used to update the prediction for b based on the innovation.

  3. (c)  Is this two-part prediction strategy equivalent to the straightforward approach? Why or why not?

    HINT: Think about what it might mean to orthonormalize the columns of the training matrix A above.

  4. (d)  (BONUS — but worth doing) A student complains to you that the case of learning a linear predictor for⃗b given ⃗a seems too complicated. She suggests just setting up the problem directly as learning a matrix X so that ⃗b ≈ X⃗a. She takes the training data (⃗ai,⃗bi) and stacks them horizontally into fat matrices A and B and computes the hypothetical difference matrix B − X A.

    She would like to minimize an appropriate norm for this and chooses the simple entry-wise 2-norm (also known as the Frobenius norm) squared. So minX ∥B − X A∥2F . Recall that the Frobenius norm of a real matrix (need not be square) L is ∥L∥2F = trace(LT L). Use this and vector calculus to directly solve for the minimizing X.

  5. (e)  Comment on why it makes sense that the computation of the best linear prediction of a vector just turns into a set of best linear predictions of each scalar component of the vector being predicted.

6 System Identification by ordinary least squares regression

Making autonomous vehicles involves using machine learning for many purposes. One of which is learning how the car actually behaves based on data about it.

Make sure to submit the code you write in this problem to “HW1 Code” on Gradescope.

CS 189, Fall 2017, HW1 5

  1. (a)  Consider a noisy time sequence of scalars x[t]. Let x[t +1] = Ax[t]+Bu[t]+w where A,B ∈ R and w is the noise. Fit a line x[t +1] = Ax[t]+Bu[t] to the values of x and u as provided by system identification programming a.mat.
  2. (b)  Consider a noisy time sequence of vectors x[t]. Let x[t +1] = Ax[t]+Bu[t]+w where A,B ∈ R3×3 and w is the noise. Fit a line x[t +1] = Ax[t]+Bu[t] to the values of x and u as provided by the system identification programming b.mat.
  3. (c)  Consider a dynamical system consisting of a cars driving on a straight road. The dynamics (accelerations) for a car i are governed by a linear function of its own position and velocity, as well as the position and velocity of the car i − 1 preceding it. This is called a linear car following model. That is, we may write the dynamical system as follows:

    x ̈i =axi+bx ̇i+cxi−1+dx ̇i−1+w(t)
    where w(t) ∼ N (0,σ). Re-write the dynamical system in matrix form, i.e. in the form x ̇ =

    Ax+Bu+w(t).

  4. (d)  Given data measurements from a dynamical system, we wish to estimate the parameters of the system. This is called system identification. In the car following model example, given vehicle traces consisting of positions, velocities, and accelerations of vehicles and the vehicles preced-

    ˆˆˆ ing them (that is, given samples of x ̈i,xi,x ̇i,xi−1,x ̇i−1, denoted by row vectors x ̈i,xˆi,x ̇i,xˆi−1,x ̇i−1,

    respectively), give the analytical solution to parameters (a,b,c,d) which gives the best linear fit to the data collected from the dynamical system.

  5. (e)  Implement your estimator using data file system identification programming train.mat, whichcontainsadictionarywithseveralusefulvalues:ddotxi, xi, dotxi, xi-1,
    dot x i-1. There are 40,000 data points. This data is a processed, filtered, and sampled form
    of data collected from the I-80 highway here in California; it is available from the Next Gener-

    ation Simulation (NGSIM) dataset, commonly used in traffic simulation. The data is given in units of m/s2,m,m/s,m,m/s, respectively. More precisely, the velocities given are deviations from the speed limit of 29.0576m/s (65mph). Give the resulting parameters below. Describe qualitatively how the dynamics evolve with respect to the state x and input u variables. When are the accelerations positive?

  6. (f)  Use your estimated parameters and data file system identification programming eval.mat to generate a vehicle trace and submit the evaluation results to Gradescope. You will be given
    x i(0), dot x i(0), x i-1(t), dot x i-1(t) and be asked to submit x i(t)
    for 150 time steps. The provided data is given at 15Hz and remember that the given velocities

    are deviations from 29.0576m/s.

    Instructions: Generate a file named submission.txt. Double check that this file is in plaintext. (i.e., cat submission.txt does not give you strange characters at the start of your file). Put only one float x i(t) on each line, where the first line is x i(0). Submit the file to “HW1 Test Set” on Gradescope. You are allowed two submissions per every 24 hours, so start early! If the script catches a formatting error, the submission will not count towards your two daily submissions.

CS 189, Fall 2017, HW1 6

7 Your Own Question

Write your own question, and provide a thorough solution.

Writing your own problems is a very important way to really learn material. The famous “Bloom’s Taxonomy” that lists the levels of learning is: Remember, Understand, Apply, Analyze, Evaluate, and Create. Using what you know to create is the top-level. We rarely ask you any HW questions about the lowest level of straight-up remembering, expecting you to be able to do that yourself. (e.g. make yourself flashcards) But we don’t want the same to be true about the highest level.

As a practical matter, having some practice at trying to create problems helps you study for exams much better than simply counting on solving existing practice problems. This is because thinking about how to create an interesting problem forces you to really look at the material from the perspective of those who are going to create the exams.

Besides, this is fun. If you want to make a boring problem, go ahead. That is your prerogative. But it is more fun to really engage with the material, discover something interesting, and then come up with a problem that walks others down a journey that lets them share your discovery. You don’t have to achieve this every week. But unless you try every week, it probably won’t happen ever.

CS 189, Fall 2017, HW1 7

Reviews

There are no reviews yet.

Only logged in customers who have purchased this product may leave a review.

Shopping Cart
[SOLVED] 机器学习代写:CS 189 HW01
30 $