[Solved] EE559 Homework 10

$25

File Name: EE559_Homework_10.zip
File Size: 160.14 KB

SKU: [Solved] EE559 Homework 10 Category: Tag:
5/5 - (1 vote)

Note: The problems in this assignment are to solved by hand. If you want to use a computer to help with the plots, you may; but you will probably find that unnecessary.

  1. In a 2-class problem with 1 feature, you are given the following data points:

S1 : 3, 2, 0 S2 : 1, 2

  • Give the k-nearest neighbor estimate of p(xS1) with k=3, for all x. Give both algebraic (simplest) form, and a plot. On the plot, show the location (x value) and height of (each) peak.
  • Give the Parzen Windows estimate of p(xS2) with window function:

( ) 0.25, 2 u< 2

u =

0, otherwise

Give both algebraic (simplest) form, and a plot. On the plot, show the x value of all significant points, and label the values of p(xS2) clearly.

  • Estimate prior probabilities P(S1) and P(S2) from frequency of occurrence of the data points.
  • Give an expression for the decision rule for a Bayes minimum error classifier using the density and probability estimates from (a)-(c). You may leave your answer in terms of p(xS1), p(xS2), P(S1), P(S2) , without plugging in for these quantities.
  • Using the estimates you have made above, solve for the decision boundaries and regions of a Bayes minimum error classifier using the density and probability estimates from (a)-(c). Give your answer in 2 forms:
    • Algebraic expressions of the decision rule, in simplest form (using numbers and variable x);
    • A plot showing the decision boundaries and regions.

Tip: you may find it easiest to develop the algebraic solution and plot and the same time.

  • Classify the points: x=5, 0.1, 0.5 using the classifier you developed in (e).
  • Separately, use a discriminative 3-NN classifier to classify the points x=5, 0.1, 0.5. (Hint: if this takes you more than a few steps for each data point,

you are doing more work than necessary.)

Assignment continues on next page

  1. [Comment: this problem is on parameter estimation, which is covered in Lecture 26 on Monday, 4/27.]

In a 1D problem (1 feature), we will estimate parameters for one class. We model the density p(x) as:

px)= ex, x 0

0, otherwise

in which 0 .

You are given a dataset Z : x1,x2,!,xN , which are drawn i.i.d. from p(x).

In this problem, you may use for convenience the notation:

1 N

m!xi .

N i=1

  • Solve for the maximum likelihood (ML) estimate ML , of , in terms of the given data points. Express your result in simplest form.

For parts (b) and (c) below, assume there is a prior for , as follows:

p()= aea, 0

0, otherwise

in which a0 .

  • Solve for the maximum a posteriori (MAP) estimate MAP , of , in terms of the given data points. Express your result in simplest form.
  • Write MAP as a function of ML and given parameters. Find limMAP , in which is the standard deviation of the prior on . What does this limit correspond to in terms of our prior knowledge of ?
a
3. [Extra credit] Comment: this problem is not more difficult than the regular-credit problems above; it is extra credit because the total length Problems 1 and 2 above is already sufficient and reasonable for one homework assignment.
In a 2-class problem with D features, you are to use Fishers Linear Discriminant to find an

Hint: the standard deviation of for the given p() is: =1 .

optimal 1D feature space. You are given that the scatter matrices for each class (calculated from the data for each class) are diagonal:

= 12 2 0 , S2 = 12 0

S1 2 22

0 ! 2 0 ! D2 D

and you are given the sample means for each class:

m1(1) m1(2) m1 = m2(1) , m2 = m2(2) .

! !

m(D 1) m(D2)

  • Find the Fishers Linear Discriminant w . Express in simplest form.
  • Let D= Suppose 12 =422 , and 12 =422 , and:

m1 = 22 , m2 = 12

Plot vectors m1, m2, (m1m2), and w.

  • Interpreting your answer of part (b), which makes more sense for a 1D feature space direction: (m1m2) or w? Justify your answer.

Reviews

There are no reviews yet.

Only logged in customers who have purchased this product may leave a review.

Shopping Cart
[Solved] EE559 Homework 10[Solved] EE559 Homework 10
$25