Courses:

Machine Learning >> Content Detail



Syllabus



Syllabus

Amazon logo Help support MIT OpenCourseWare by shopping at Amazon.com! MIT OpenCourseWare offers direct links to Amazon.com to purchase the books cited in this course. Click on the Amazon logo to the left of any citation and purchase the book from Amazon.com, and MIT OpenCourseWare will receive up to 10% of all purchases you make. Your support will enable MIT to continue offering open access to MIT courses.

A list of topics covered in the course is presented in the calendar.



Description


This introductory course gives an overview of many concepts, techniques, and algorithms in machine learning, beginning with topics such as classification and linear regression and ending up with more recent topics such as boosting, support vector machines, hidden Markov models, and Bayesian networks. The course will give the student the basic ideas and intuition behind modern machine learning methods as well as a bit more formal understanding of how, why, and when they work. The underlying theme in the course is statistical inference as it provides the foundation for most of the methods covered.



Problem Sets


There will be a total of 5 problem sets, due roughly every two weeks. The content of the problem sets will vary from theoretical questions to more applied problems. You are encouraged to collaborate with other students while solving the problems but you will have to turn in your own solutions. Copying will not be tolerated. If you collaborate, you must indicate all of your collaborators.

Each problem set will be graded by a group of students with the guidance of your TAs. Each problem set will be graded in a single grading session, usually on the first Monday after it is due, starting at 5pm. Every student is required to participate in one grading session. You should sign up for grading by contacting a TA, by email or in person; doing it early increases the chances of getting the preferred grading schedule. Students who do not register for grading by the third week of the course, will be assigned to a problem set by us.

If you drop the class after signing up for a grading session, please be sure to let us know so we can keep track of students available for grading. If you add the class during the term, please remember to sign up for grading.



Exams


There will be two in-class exams, a midterm midway through the term and a final the last day of class.



Project


You are required to complete a class project. The choice of the topic is up to you so long as it clearly pertains to the course material. To ensure that you are on the right track, you will have to submit a one paragraph description of your project a month before the project is due. Similarly to problem sets, you are encouraged to collaborate on the project. We expect a four page write-up about the project, which should clearly and succinctly describe the project goal, methods, and your results. Each group should submit only one copy of the write-up and include all the names of the group members (a two person group will have 6 pages, a three person group will have 8 pages, and so on). The projects will be graded on the basis of your understanding of the overall course material (not based on, e.g., how brilliantly your method works). The scope of the project is about 1-2 problem sets.

The projects are due in Lec #23. Electronic submission is required but we can accept only postscript or pdf documents. The short proposal should be turned in on or before Lec #12.

The projects can be literature reviews, theoretical derivations or analyses, applications of machine learning methods to problems you are interested in, or something else (to be discussed with course staff).



Grading


Your overall grade will be determined roughly as follows:


ACTIVITIESPERCENTAGES
Midterm15%
Problem sets30%
Final25%
Project30%



Text


There are a number of useful texts for this course but each covers only some part of the class material.

Amazon logo Bishop, Christopher. Neural Networks for Pattern Recognition. New York, NY: Oxford University Press, 1995. ISBN: 9780198538646.

Amazon logo Duda, Richard, Peter Hart, and David Stork. Pattern Classification. 2nd ed. New York, NY: Wiley-Interscience, 2000. ISBN: 9780471056690.

Amazon logo Hastie, T., R. Tibshirani, and J. H. Friedman. The Elements of Statistical Learning: Data Mining, Inference and Prediction. New York, NY: Springer, 2001. ISBN: 9780387952840.

Amazon logo MacKay, David. Information Theory, Inference, and Learning Algorithms. Cambridge, UK: Cambridge University Press, 2003. ISBN: 9780521642989. Available on-line here.

Amazon logo Mitchell, Tom. Machine Learning. New York, NY: McGraw-Hill, 1997. ISBN: 9780070428072.

You are responsible for the material covered in lectures (most of which will appear in lecture notes in some form), problem sets, as well as material specifically made available and indicated for this purpose. The weekly recitations/tutorials will be helpful in understanding the material and solving the homework problems.



Recommended Citation


For any use or distribution of these materials, please cite as follows:

Tommi Jaakkola, course materials for 6.867 Machine Learning, Fall 2006. MIT OpenCourseWare (http://ocw.mit.edu/), Massachusetts Institute of Technology. Downloaded on [DD Month YYYY].



Calendar



LEC #TOPICSKEY DATES
1Introduction, linear classification, perceptron update rule
2Perceptron convergence, generalization
3Maximum margin classification
4Classification errors, regularization, logistic regressionProblem set 1 out
5Linear regression, estimator bias and variance, active learning
6Active learning (cont.), non-linear predictions, kernalsProblem set 1 due
7Kernal regression, kernelsProblem set 2 out
8Support vector machine (SVM) and kernels, kernel optimization
9Model selectionProblem set 2 due
10Model selection criteria
Midterm
11Description length, feature selectionProblem set 3 out 3 days before Lec #11
12Combining classifiers, boosting
13Boosting, margin, and complexity

Problem set 3 due

Problem set 4 out

14Margin and generalization, mixture models
15Mixtures and the expectation maximization (EM) algorithm
16EM, regularization, clusteringProblem set 4 due
17Clustering
18Spectral clustering, Markov modelsProblem set 5 out
19Hidden Markov models (HMMs)
20HMMs (cont.)
21Bayesian networks
22Learning Bayesian networksProblem set 5 due
23

Probabilistic inference

Guest lecture on collaborative filtering

Projects due
Final
24Current problems in machine learning, wrap upExams back

 








© 2017 Coursepedia.com, by Higher Ed Media LLC. All Rights Reserved.