Translation Models (2/3/2022)
Lecture: (by Graham Neubig)
- Sequence-to-sequence models w/ attention
 - Decoding strategies
 - Transformers
 
Language in 10: Luo
Slides: Translation Models Slides 
Discussion: No discussion, but code walk of the annotated transformer and introduction of assignment 2. 
References:
- Highly Recommended: The Annotated Transformer
 - Reference: Attentional NMT (Bahdanau et al. 2015)
 - Reference: Effective Approaches to Attention (Luong et al. 2015)
 - Reference: Self Attention (Cheng et al. 2016)
 - Reference: Attention is All You Need (Vaswani et al. 2017)
 - Reference: Coverage Embedding Models (Mi et al. 2016)