Daniel Bauer (office hours Monday 4-5 and Friday 3-4:30 in the SpeechLab, 7LW3 CEPSR (Shapiro) building)
Sara Rosenthal (office hours Tuesday 1.30-2.30, and Thursday from 3-4, in 726 CEPSR)
The Final exam will take place on Tuesday, 12/20/2011 at 4:10pm - 7:00pm in 602 Hamilton Hall. All material from the semester may be covered, excluding the last lecture (on 12/8). The exam is closed book, closed electronic devices (except standard calculators). Students are allowed to bring one letter sized study sheet (double sided).
Nov. 5th: Homework 2 is now posted the homeworks page.
Reminder: the second quiz will be in class, on November 1st. All material up to and including the lecture on 10/27 will be covered. For some example questions on the IBM translation models, see questions 1-4 of this previous quiz.
Sept. 19th: Homework 1 is now posted on the homeworks page.
Reminder: the first quiz will be in class, on October 4th. All material up to and including the class on 9/29 will be covered. For some example questions, see questions 1, 9, 10 of this previous quiz (solutions here), and questions 1 and 4 of this previous quiz (other questions are on topics not yet covered in the class).
Prof. Collins's office hours on September 28th will be held 1-2pm, instead of the usual time of 4-5pm.
Some people are asking for clarification on problem 2 of homework 1. Here is a hint: take a look at slide 20 of the lecture slides on language models.
|9/6||Introduction and Overview
|9/8||Language Modeling||Notes on language modeling
(Required reading: updated version posted Sep 15th, fixing a typo in the definition of missing mass on pages 11 and 12.).
|9/13||Tagging, and Hidden Markov Models||Notes on HMMs
|9/15||Tagging, and Hidden Markov Models (continued)|
|9/20||Parsing, context-free grammars, and probabilistic CFGs||Note on PCFGs (required reading)|
|9/22||Parsing, context-free grammars, and probabilistic CFGs (continued)||9/27||Parsing, context-free grammars, and probabilistic CFGs (continued)|
|9/29||Lexicalized probabilistic CFGs|
|10/6||Guest lecture by Nizar Habash|
|10/11||Lexicalized probabilistic CFGs (continued, see previous slides)|
|10/13||Machine translation part 1||(Note: we didn't cover the final section, on evaluation using BLEU, but I've kept it in the slides in case it's of interest.)|
|10/18||Machine translation part 2||Note on IBM Models 1 and 2 (required reading)|
|10/20||Machine translation part 2 (continued)|
|10/27||Phrase-based translation models||
Note on phrase-based models (required reading: updated on 30th October to include details of learning a phrase-based lexicon from training examples (sections 1 and 2))|
Slides from the tutorial by Philipp Koehn
|11/3||Reordering for statistical MT|
Note on log-linear models (required reading).
|11/15||Log-linear tagging (MEMMs)|
|11/17||Global linear models|
|11/22||Global linear models (continued)|
|11/29||Global linear models part II|
Global linear models part III,|
|12/6||The Brown word-clustering algorithm|
|12/8||The EM algorithm for HMMs|