Starting from:
$35

$29

Assignment 3 Solution

Instructions




Read all instructions in this section thoroughly.




Collaboration: Make certain that you understand the course collaboration policy, described on the course website. You must complete this assignment individually; you are not allowed to collaborate with anyone else. You may discuss the homework to understand the problems and the mathematics behind the various learning algorithms, but you are not allowed to share problem solutions or your code with any other students. You must also not consult code on the internet that is directly related to the programming exercise. We will be using automatic checking software to detect academic dishonesty, so please don't do it.




You are also prohibited from posting any part of your solution to the internet, even after the course is complete. Similarly, please don't post this PDF le or the homework skeleton code to the internet.




Formatting: This assignment consists of two parts: a problem set and program exercises.




For the problem set, you must write up your solutions electronically and submit it as a single PDF document. We will not accept handwritten or paper copies of the homework. Your problem set solutions must use proper mathematical formatting. For this reason, we strongly encourage you to write up your responses using LATEX. (Alternative word processing programs, such as MS Word, produce very poorly formatted mathematics.)




Your solutions to the programming exercises must be implemented in python, following the precise instruc-tions included in Part 2. Portions of the programing exercise may be graded automatically, so it is imperative that your code follows the speci ed API. Several parts of the programming exercise ask you to create plots or describe results; these should be included in the same PDF document that you create for the problem set.




Homework Template and Files to Get You Started: The homework zip le contains the skeleton code and data sets that you will require for this assignment. Please read through the documentation provided in ALL les before starting the assignment.




Citing Your Sources: Any sources of help that you consult while completing this assignment (other students, textbooks, websites, etc.) *MUST* be noted in the your README le. This includes anyone you brie y discussed the homework with. If you received help from the following sources, you do not need to cite it: course instructor, course teaching assistants, course lecture notes, course textbooks or other readings.




Submitting Your Solution: You will be submitting only the following les, which you created or modi ed as part of homework 3:




README




hw3-GTACC.pdf (a PDF of your homework writeup)




boostedDT.py




predictions-BestClassifier.dat predictions-BoostedDT.dat




naiveBayes.py







1



Please follow the naming conventions exactly, and do not submit additional les including the test scripts or data sets. Your PDF writeup of Homework 3 should be named hw3-GTACC.pdf, where \GTACC" is your own Georgia Tech account name (for example, my le would be named \hw3-bboots3.pdf").




Submission Instructions: Please place all of your les into a single folder le named hw3-GTACC and compress the folder as a zip le with the same name. Upload the zip le as an attachment on T-Square. Again, please follow the naming conventions exactly, and do not submit additional les including the test scripts or data sets. For example my le would be named \hw3-bboots3.zip"




You may resubmit multiple times, only your last submission will be counted.

PART I: PROBLEM SET




Your solutions to the problems will be submitted as a single PDF document. Be certain that your problems are well-numbered and that it is clear what your answers are. Additionally, you will be required to duplicate your answers to particular problems in the README le that you will submit.




 
Probability decision boundary (10pts)




Consider a case where we have learned a conditional probability distribution P (y j x). Suppose there are only two classes, and let p0 = P (y = 0 j x) and let p1 = P (y = 1 j x). A loss matrix gives the cost that is incurred for each element of the confusion matrix. (E.g., true positives might cost nothing, but a false positive might cost us $10.) Consider the following loss matrix:






y = 0 (true)
y = 1 (true)






y^ = 0 (predicted)
0
10






y^ = 1 (predicted)
5
0









 
Show that the decision y^ that minimizes the expected loss is equivalent to setting a probability threshold and predicting y^ = 0 if p1 < and y^ = 1 if p1 .




 
What is the threshold for this loss matrix?




 
Double counting the evidence (15pts)




Consider a problem in which the binary class label Y 2 fT; F g and each training example x has 2 binary attributes X1; X2 2 fT; F g.

Let the class prior be p(Y = T ) = 0:5 and p(X1 = T j Y = T ) = 0:8 and p(X2 = T j Y = T ) = 0:5. Likewise,




p(X1 = F j Y = F ) = 0:7 and p(X2 = F j Y = F ) = 0:9. Attribute X1 provides slightly stronger evidence about the class label than X2.




Assume X1 and X2 are truly independent given Y . Write down the naive Bayes decision rule.




 
What is the expected error rate of naive Bayes if it uses only attribute X1? What if it uses only X2?




The expected error rate is the probability that each class generates an observation where the decision



^
; X2) be the predicted class label. Then the
rule is incorrect. If Y is the true class label, let Y (X1
^
; X2)).


expected error rate is p(X1; X2; Y j Y 6= Y (X1


 
Show that if naive Bayes uses both attributes, X1 and X2, the error rate is 0.235, which is better than if using only a single attribute (X1 or X2).




 
Now suppose that we create new attribute X3 that is an exact copy of X2. So for every training example, attributes X2 and X3 have the same value. What is the expected error of naive Bayes now?




 
Brie y explain what is happening with naive Bayes (2 sentences max).




 
Does logistic regression su er from the same problem? Brie y explain why (2 sentences max).


















PART II: PROGRAMMING EXERCISES




 
Challenge: Generalizing to Unseen Data (50 pts)




One of the most di cult aspects of machine learning is that your classi er must generalize well to unseen data. In this exercise, we are supplying you with labeled training data and unlabeled test data. Speci cally, you will not have access to the labels for the test data, which we will use to grade your assignment. You will t the best model that you can to the given data and then use that model to predict labels for the test data. It is these predicted labels that you will submit, and we will grade your submission based on your test accuracy (relative to the best performance you should be able to obtain). Each instance belongs to one of nine classes, named '1' . . . '9'. We will not provide any further information on the data set.




You will submit two sets of predictions { one based on a boosted decision tree classi er (which you will write), and another set of predictions based on whatever machine learning method you like { you are free to choose any classi cation method. We will compute your test accuracy based on your predicted labels for the test data and the true test labels. Note also that we will not be providing any feedback on your predictions or your test accuracy when you submit your assignment, so you must do your best without feedback on your test performance.




Relevant Files in the Homework Skeleton1




boostedDT.py




test boostedDT.py




data/challengeTrainLabeled.dat: labeled training data for the challenge




data/challengeTestUnlabeled.dat: unlabeled testing data for the challenge







1.1 The Boosted Decision Tree Classi er




In class, we mentioned that boosted decision trees have been shown to be one of the best \out-of-the-box" classi ers. (That is, if you know nothing about the data set and can't do parameter tuning, they will likely work quite well.) Boosting allows the decision trees to represent a much more complex decision surface than a single decision tree.




Write a class that implements a boosted decision tree classi er. Your implementation may rely on the decision tree classi er already provided in scikit learn (sklearn.tree.DecisionTreeClassifier), but you must implement the boosting process yourself. (The scikit learn module actually provides boosting as a meta-classi er, but you may not use it in your implementation.) Each decision tree in the ensemble should be limited to a maximum depth as speci ed in the BoostedDT constructor. You can con gure the maximum depth of the tree via the max depth argument to the DecisionTreeClassifier constructor.







Your class must implement the following API:




init (numBoostingIters = 100, maxTreeDepth = 3): the constructor, which takes in the num-ber of boosting iterations (default value: 100) and the maximum depth of the member decision trees (default: 3)




fit(X,y): train the classi er from labeled data (X; y)




predict(X): return an array of n predictions for each of n rows of X







1Bold text indicates les that you will need to complete; you should not need to modify any of the other les.







4



Note that these methods have already been de ned correctly for you in boostedDT.py; be very careful not to change the API. You should con gure your boosted decision tree classi er to be the best \out-of-the-box" classi er you can; you may not modify the constructor to take in additional parameters (e.g., to con gure the individual decision trees).




There is one additional change you need to make to AdaBoost beyond the algorithm described in class. AdaBoost by default only works with binary classes, but in this case, we have a multi-class classi cation problem. One variant of AdaBoost, called AdaBoost-SAMME, easily adapts AdaBoost to multiple classes.

t
2




1






1










Instead of using the equation
=
1
ln
1
in AdaBoost, you should use the AdaBoost-SAMME equation










t


2




























=








ln








+ ln(K


1)
;






























where K is the total number of classes. This will force t 0 as long as the classi er is worse than random guessing (in this case random guessing would be 1=K, so the error rate would need to be greater than




1 1=K). Note that when K = 2, AdaBoost-SAMME reduces to AdaBoost. For further information on SAMME, see http://web.stanford.edu/~hastie/Papers/samme.pdf.




Test your implementation by running test boostedDT.py, which compares your BoostedDT model to a regular decision tree on the iris data with a 50:50 training/testing split. You should see that your BoostedDT model is able to obtain 97.3% accuracy vs the 96% accuracy of regular decision trees. Make certain that your implementation works correctly before moving on to the next part.







Once your boosted decision tree is working, train your BoostedDT on the labeled data available in the le data/challengeTrainLabeled.dat. The class labels are speci ed in the last column of data. You may tune the number of boosting iterations and maximum tree depth however you like. Then, use the trained BoostedDT classi er to predict a label y 2 f1; : : : ; 9g for each unlabeled instance in data/challengeTestUnlabeled.dat. Your implementation should output a comma-separated list of predicted labels, such as




1, 2, 1, 9, 4, 1, 3, 1, 5, 3, 4, 2, 8, 3, 1, 6, 3, ...




Be very careful not to shu e the instances in data/challengeTestUnlabeled.dat; the rst predicted label should correspond to the rst unlabeled instance in the testing data. The number of predictions should match the number of unlabeled test instances.




Copy and paste this comma-separated list into the README le to submit your predictions for grading. Also, record the expected accuracy of your model in the README le. Finally, also save the comma-separated list into a text le named predictions-BoostedDT.dat; this le should have exactly one line of text that contains the list of predictions.




1.2 Training the Best Classi er




Now, train the very best classi er for the challenge data, and use that classi er to output a second vector of predictions for the test instances. You may use any machine learning algorithm you like, and may tune it any way you wish. You may use the method and helper functions built into scikit learn; you do not need to implement the method yourself, but may if you wish. If you don't want to use scikit learn, you may use any other machine learning software you wish. If you can think of a way that the unlabeled data in data/challengeTestUnlabeled.dat would be useful during the training process, you are welcome to let your classi er have access to it during training.







Note that you will not be submitting an implementation of your optimal model, just its predictions.




Once again, use your trained model to output a comma-separated list of predicted labels for the unlabeled instances in data/challengeTestUnlabeled.dat. Again, be careful not to shu e the test instances; the order of the predictions must match the order of the test instances.




Copy and paste this comma-separated list into the README le to submit your predictions for grading. Also, record the expected accuracy of your model in the README le. Finally, also save the comma-separated list into a text le named predictions-BestClassifier.dat; this le should have exactly one line of text that contains the list of predictions.










5



If you believe that your boostedDT classi er (from the previous section) is actually the best set of predictions for this challenge data, then you would submit the boostedDT predictions twice in the README le (and have two identical les of predictions).




Write a brief paragraph (3{4 sentences max) describing the best machine learning classi er you found, its optimal parameter settings (if any), and how you trained the model. Include that paragraph in your PDF writeup, and also in the README.




 
Naive Bayes (25 pts)




In this implementation exercise, you will implement naive Bayes for batch learning.We will then use your naive Bayes implementation in the next assignment in an application to text processing.




Relevant Files in the Homework Skeleton2




naiveBayes.py




test naiveBayes.py







2.1 Implementing Batch Naive Bayes




Implement a multinomial naive Bayes classi er in the NaiveBayes class in naiveBayes.py. Your implemen-tation should support Laplace smoothing. Whether or not to use Laplace smoothing is controlled via an argument to the constructor; Laplace smoothing is enabled by default. Your implementation must follow the API below. (Note that all matrices are actually 2D numpy arrays in the implementation.)




init(useLaplaceSmoothing=True) : constructor




fit(X,Y): method to train the naive Bayes model




predict(X): method to use the trained naive Bayes model for prediction




predictProbs(X): outputs a matrix of predicted posterior class probabilities







The training data for multinomial naive Bayes is speci ed as feature counts: X[i,j] is the number of times feature j occurs in instance i (or you can think of it as that instance i is characterized by a particular real-valued amount of feature j).




Although you we aren't actually dealing with such data sets in this problem, for simple data sets with multi-valued discrete features (e.g., the Tennis play/don't play dataset), in order to use them in this classi er, we must rst convert them to a set of binary features. (e.g, output = fsunny, overcast, rainyg to three features: isSunny?, isOvercast?, isRainy?). The ith instance X[i,:] is then a binary vector for the presence or absence of each feature value.




The predictProbs(X) function takes in a matrix X of n instances and outputs an n K matrix of posterior probabilities. Each row i of the returned matrix represents the posterior probability distribution over the K classes for the ith training instance. (Note that each row of the returned matrix will sum to 1.)




Run test naiveBayes.py to test your implementation. Your naive Bayes should achieve 89% accuracy.




















































2Bold text indicates les or functions that you will need to complete; you should not need to modify any of the other les.




6

More products