[SOLVED] algorithm matlab python graph statistic Introduction to Data Mining (Summer 2019) Assignment #2

$25

File Name: algorithm_matlab_python_graph_statistic_Introduction_to_Data_Mining_(Summer_2019)_Assignment_#2.zip
File Size: 894.9 KB

5/5 - (1 vote)

Introduction to Data Mining (Summer 2019) Assignment #2
Introduction to Data Mining: Assignment #2 Summer 2019
Due: May 30th, 23:59:59 CST (UTC +8).
1. A Walk Through Linear Models
In this problem, you will implement a whole bunch of linear classifiers and compare their performance and properties.
We assume our target function is f : [1, 1] [1, 1] {1, +1}. In each run, choose a random line in the plane as your target function f, where one side of the line maps to +1 and the other maps to -1.
Skeleton code and MATLAB/Python functions including mkdata and plotdata are given for your convenience, see the comments in the code for more details. What you need to do is implementing each algorithm and write scripts to play with them. Your algorithm implementations should be able to handle data in arbitrary dimension. For most of the questions below, you should repeat the experiments for 1000 times and take the average. See run.m/run.ipynb for a script example.
Hint: Do not forget to add the bias constant term! (a) Perceptron
Implement Perceptron learning algorithm (in perceptron.m/perceptron.py), then answer the following questions.
(i) What is the training error rate and expected testing error rate (we can generate a large set of testing points to estimate test error), when the size of training set is 10 and 100 respectly?
(ii) What is the average number of iterations before your algorithm converges when the size of training set is 10 and 100 respectively ?
(iii) What will happen if the training data is not linearly separable (Use mkdata(N, noisy) to generate non-linearly separable data) ?
(b) Linear Regression
Implement Linear Regression (in linear_regression.m/linear_regression.py), then an-
swer the following questions.
Note that we use Linear Regression here to classify data points. This technique is
called Linear Regression on indicator response matrix1.
(i) What is the training error rate and expected testing error rate if the size of
training set is 100 ?
(ii) What is the training error rate and expected testing error rate if the training data is noisy and not linearly separable (nTrain = 100) ?
1http://books.google.com/books?id=VRzITwgNV2UC&pg=PA81 1

Introduction to Data Mining (Summer 2019) Assignment #2
(iii) Run Linear Regression on dataset poly_train.mat. What is the training error
rate? What is the testing error rate on poly_test.mat ?
(iv) Do the following data transformation
(1, x1, x2) (1, x1, x2, x1x2, x21, x2).
on dataset poly_train.mat. After this transformation, what is the training error
rate? What is the testing error rate on poly_test.mat ?
(c) Logistic Regression
Implement Logistic Regression (in logistic.m/logistic.py), then answer the following questions.
Remember the form of Logistic Regression
Let us assume
then
p(y|x) = 1 . 1 + eT x
P(y = 1|x;) = h(x), P(y = 0|x;) = 1 h(x), P (y|x; ) = (h (x))y (1 h (x))1y .
Now you can learn your logistic model using MLE. Not like Linear Regression, we have no closed-form solution to solve the MLE problem. So you should use gradient descent to maximize the log-likelihood function iteratively.
Derive the gradient descent update rule, choose a proper learning rate, and repeat the experiment for 100 times to take average.
(i) What is the training error rate and expected testing error rate if the size of training set is 100 ?
(ii) What is the training error rate and expected testing error rate if the training data is noisy and not linearly separable (nTrain = 100) ?
(d) Support Vector Machine
Implement Support Vector Machine (in svm.m/svm.py) without introducing slack vari-
ables , then answer the following questions.
Hint: For MATLAB users, using MATALB built-in function quadprog, you should be able to implement SVM in less than 10 lines of code. For Python users, using scipy.optimize, you should be able to implement in 20 lines of code.
(i) What is the training error rate and expected testing error rate if the size of training set is 30 ? (Use plotdata to plot your learnt model, hope the graph will help you understand SVM as a maximum margin classifier.)
(ii) What is the training error rate and expected testing error rate if the size of training set is 100 ?
(iii) For the case nTrain = 100, what is average number of support vectors in your trained SVM models?
2

Introduction to Data Mining (Summer 2019) Assignment #2
2. Regularization and Cross-Validation
We now learnt the undesirable consequence of overfitting, just as this quotation says: If you torture the data for long enough, it will confess to anything.
Regularization and Cross-Validation2 are the two most common used methods to con- trol overfitting. In this problem, we will use these two techniques together to relieve the overfitting problem.
Cross-Validation is a standard technique for adjusting regularization parameters of learn- ing models. We first split the dataset into training set and validation set. Then we compute the validation error on validation set using different values for regularization parameters. Finally, we pick the parameter with the lowest validation error and use it for training our learning model. To reduce variability, we can run validation multiple rounds using different dataset partitions, and the validation results are averaged over all the rounds.
In the following questions, we will use leave-one-out cross-validation to choose regular- ization parameters. As the name suggests, leave-one-out cross-validation (LOOCV) involves using a single observation from the original sample as the validation data, and the remaining observations as the training data. This is repeated such that each observation in the sample is used once as the validation data.
You may use validation.m/validation.ipynb as a skeleton.
(a) Implement Ridge Regression (in ridge.m/ridge.py), and use LOOCV to tune the reg- ularization parameter .
Train your algorithm on digit_train.mat and test on digit_test.mat.
Hint: Since the dataset is a set of raw images (hand-written digits) without prepro- cessing, you should do feature normalization3 to make the feature have zero-mean and unit-variance. You can visualize the dataset using show_digit.m.
(i) What is the chosen by LOOCV ?
(ii) What is mi=1 i2 with and without regularization (let = 0) ?
(iii) Whats training/testing error rate with and without regularization ?
(b) Implement Logistic Regression with regularization (in logistic_r.m/logistic_r.py), and
use LOOCV to tune the regularization parameter .
Train your algorithm on digit_train.mat and test on digit_test.mat. Then report the training/testing error rate with and without regularization (let = 0). You should also report the chosen by LOOCV.
2https://en.wikipedia.org/wiki/Cross-validation_(statistics) 3https://en.wikipedia.org/wiki/Feature_scaling
3

Introduction to Data Mining (Summer 2019) Assignment #2
3. Bias Variance Trade-off
Let us review the bias-variance decomposition first.
The intuition behind it is straight-forward: if the model is too simple, the learnt function
is biased and does not fit the data. If the model is too complex then it is very sensitive to small changes in the data.
If we were able to sample a dataset D infinite many times, we will learn different g(x) for each time, and get an expected hypothesis g (x). So bias means the difference between the truth and what you expect to learn. It measures how well our model can approximate the truth at best.
However, it is impossible to sample the training dataset multiple time, so variance means the difference between what you learn from a particular dataset and what you expect to learn.
Now please answer the following questions:
(a) True of False:
(i) If a learning algorithm is suffering from high bias, adding more training examples will improve the test error significantly.
(ii) We always prefer models with high variance (over those with high bias) as they will able to better fit the training set.
(iii) A model with more parameters is more prone to overfitting and typically has higher variance.
(iv) Introducing regularization to the model always results in equal or better perfor- mance on the training set.
(v) Using a very large value of regularization parameter cannot hurt the perfor- mance of your hypothesis.
Please submit your homework report to at http://assignment.zjulearning.org:8081 in pdf format, with all your code in a zip archive.
4

Reviews

There are no reviews yet.

Only logged in customers who have purchased this product may leave a review.

Shopping Cart
[SOLVED] algorithm matlab python graph statistic Introduction to Data Mining (Summer 2019) Assignment #2
$25