Q1. Consider the following decision rule for a two-category one-dimensional problem:
Decide 1 if x > ; otherwise decide 2.
- Show that the probability of error for this rule is given by
(1)
- By differentiating, show that a necessary condition to minimize P(error) isthat satisfy p(|1)P(1) = p(|2)P(2)
Q2. Let the conditional densities for a two-category one-dimensional problem be given by the Cauchy distribution
2 (2)
Assuming P(1) = P(2), show that P(1|x) = P(2|x) if x = (a1 +a2)/2, i.e., the minimum error decision boundary is a point midway between the peaks of the two distributions, regardless of b.
Q3. Suppose we have three equi-probable categories in two dimensions with the following underlying distributions:
By explicit calculation of posterior probabilities, classify the point x =
for minimum probability of error.
1
Q4. a. Write a procedure to generate random samples according to a normal distribution N(,) in d dimensions.
- Write a procedure to calculate the discriminant function for a given
normal distribution with = 2I and prior probability P(i).
- Compare the discriminant functions values for two different distributions
) and = 2 dimensions.
Assume the test sample to be and P(1) = 1/3 and P(2) = 2/3.
In a general process, you would be given several samples from two (or more) classes. Counting each class frequency will give the priors. With these samples as d dimensional vectors, you can estimate mean and covariance using MLE or other techniques, which is a part of later lecture. This computed info is sufficient for computing discriminants and thereby classifying the sample into one of the classes.
2

![[Solved] SML Assignment 1](https://assignmentchef.com/wp-content/uploads/2022/08/downloadzip.jpg)

![[Solved] SML Assignment 4-Regression](https://assignmentchef.com/wp-content/uploads/2022/08/downloadzip-1200x1200.jpg)
Reviews
There are no reviews yet.