Office Hours (in computer science TA room unless otherwise stated):
Daniel Mitropolsky (instructor): Mondays 10.30am-12pm
David Wan (TA): Mondays 1pm-2.30pm
Alyssa Hwang (TA): Tuesdays 12-1.30pm
Vu Anh Phung (TA): Wednesdays 3-6pm (NOTE: changed to 12pm-3pm on February 20th only)
Viraj Rai (TA): Thursdays 4pm-5.30pm
Gary Buranasampatanon (TA): Thursdays 1.30-3pm, also Fridays 1-2:30 PM
Past midterms for the class are here: fall 2011, fall 2012, fall 2013, fall 2014, fall 2017, spring 2018.
|Date||Topics||Video Lectures||References||Flipped Classroom Materials|
|Week 1 (January 23rd-29th)||Introduction to NLP,
Language Modeling (Slides: we will cover slides 1-50 inclusive)
|Video lectures in Courseworks: All of Module 1-2; All of Module 3; Sections 4.1, 4.2 in Module 4.||Sections 1.1-1.4.1 inclusive of
Notes on language modeling (required reading).
||Questions (part 1), Solutions (part 1) Questions (part 2), Solutions (part 2)||Week 2 (January 30th-February 4th)||Tagging, and Hidden Markov Models (Slides)||All videos in Module 6 in courseworks: The tagging problem (10:01) to Summary (1:50) inclusive.||Notes on tagging problems, and hidden Markov models (required reading)
|Week 3 (February 6th-11th)||Log-Linear Models (Slides)||All videos in Module 15 on Courseworks.||Notes on Log-Linear Models (required reading)
||Questions, Solutions, Past midterm question|
|Week 4 (February 13th-18th)||Parsing, and Context-free Grammars (Slides)||Courseworks videos:  All of Module 7;  Module 8, sections 8-1 to 8-3 inclusive||Questions, Solutions|
|Week 5 (February 20th-25th)||Probabilistic Context-free Grammars (continued), and lexicalized context-free grammars (Slides part 1) (Slides part 2), (Slides part 3)||Courseworks videos:  Module 8, sections 8-4 to 8-6;  All of Module 9;  All of Module 10||Notes on Probabilistic Context-Free Grammars (required reading)
|Week 6 (February 27th-March 5th)||Log-Linear Models for Tagging, and for history-based parsing (Slides part 1), (Slides part 2).||Modules 16 and 17 in courseworks.||Notes on MEMMs (Log-Linear Tagging Models) (required reading)
||Questions on CRFs, solutions are in section 4 of this note. Additional questions, Solutions|
|Week 7 (March 6th-March 12th)||Feedforward Neural Networks (Slides)||Module 22 videos in Courseworks.
|Notes on Feedforward Neural Networks (required reading)
|Week 8 (March 13th)||Mid-term||Note: no flipped classrooms.|
|Week 9 (March 27th-April 2nd)||Computational Graphs, and Backpropagation (Slides)||Module 23 videos in Courseworks.
|Notes on Computational Graphs, and Backpropagation (required reading)
|Week 10 (April 3rd-9th)||Word Embeddings in Feedforward Networks; Tagging and Dependency Parsing using Feedforward Networks (Slides)||Module 24 videos in Courseworks.
|Week 11 (April 10th-16th)||Recurrent Networks, and LSTMs, for NLP (Slides)||Module 25 in Courseworks.||Questions, Solutions|
|Week 12 (April 17th-23rd)||Recurrent Networks, and attention, for statistical machine translation (Slides)||Module 26 in Courseworks.||Questions, Solutions|
|Week 13 (April 24th-30th)||Brown Clustering, and Word2Vec (Brown Clustering Slides), (Word2Vec Slides).||Module 18 in Courseworks.||Word2Vec paper (we will cover in the flipped classroom section)||Questions, Solutions|
|Week 14 (May 1st-6th)||TBD|