In the first part, the general problem of machine translation (automatic translation of text from one language to another) will be discussed, as well as the history of research into machine translation. We will then briefly consider older approaches to machine translation (before the current focus on machine learning). Then, some particular challenges for natural language processing that must be solved on the way to general approaches for machine translation will be presented. Finally, we will discuss the important topic of evaluation of machine translation systems.
In the second part, we will look at statistical machine translation (SMT), which became the dominant paradigm in translation from about 2000 to 2015, and is still the core of many industrial systems. The related concepts of translational equivalence (established through word alignment), simple statistical models and search algorithms will be introduced.
In the third and last part of the lecture, we will consider the deep learning approaches used in so-called neural machine translation (NMT). We will briefly introduce the concepts of word embeddings and deep learning before moving on to provide a high-level overview of recurrent neural networks (RNNs) and Long Short-Term Memory (LSTM) approaches to translation, and then follow up with the state-of-the-art Transformer approach, and talk about transfer learning (with applications beyond NMT).
Goals
Theoretical understanding of the challenges of machine translation and the models used to solve them.
Goals
Practical experience in solving sub-problems of machine translation, as well as familiarity with the data used for training statistical models.
Email Address: SubstituteMyLastName@cis.uni-muenchen.de
Tuesdays, 16 to 18 (c.t.). ONLINE WITH ZOOM. Link will be sent to students listed in LSF.
Wednesdays, 14 to 16 (c.t.). ONLINE WITH ZOOM. Link will be sent to students listed in LSF.
Date | Topic | Reading (DO AFTER THE MEETING!) | Slides | Video |
April 14th | Orientation and Introduction to Machine Translation | mp4 | ||
April 20th | Introduction to Statistical Machine Translation | ppt pdf | mp4 | |
April 27th | Bitext alignment (extracting lexical knowledge from parallel corpora) | ppt pdf | mp4 | |
Optional: read about Model 1 in Koehn and/or Knight (see below) | ||||
April 28th | Many-to-many alignments and Phrase-based model | ppt pdf | mp4 | |
May 4th | Log-linear model and Minimum Error Rate Training | ppt pdf | mp4 | |
May 5th | Decoding | mp4 | ||
May 11th | Exercise 1 Released. Due Monday May 17th at 15:00. | exercise1.txt | mp4 | |
May 12th | Linear Models | pptx pdf | part1 mp4 part2 mp4 | |
May 18th | Exercise 2 Released. Due Monday May 31st at 15:00. | exercise2.html | mp4 tamchyna_acl_2016_slides.pdf tamchyna_acl_2016_slides.pptx |
|
May 19th | Neural Networks (and Word Embeddings) | mp4 (skip first 60 seconds) | ||
May 25th | Pfingstdienstag (holiday) | |||
May 26th | Bilingual Word Embeddings and Unsupervised SMT (Viktor Hangya) | mp4 | ||
June 1st | Training and RNN/LSTMs | mp4 | ||
June 2nd | Exercise 3 Released. Due Monday June 7th at 15:00. | exercise3.pdf | mp4 | |
June 8th | Exercise 4 Released. Due Monday June 14th at 15:00. | exercise4.pdf | mp4 | |
June 9th | Encoder-Decoder and Attention (Jindřich Libovický) | mp4 | ||
June 15th | Exercise 4 review | mp4 | ||
June 16th | Transformer (and Document NMT) | mp4 | ||
June 22nd | Unsupervised NMT | (see previous) | mp4 | |
June 23rd | Transfer Learning for Unsupervised NMT (Alexandra Chronopoulou) | mp4 | ||
June 29th | Exercise 5 Released. Due Monday July 5th at 15:00. | exercise5_updated.pdf | mp4 | |
June 30th | Overcoming Sparsity in NMT (research talk) | mp4 | ||
July 6th | Exercise 6 (pytorch NLP tutorial) released, not collected, recommended to be done on your own during the summer vacation. | exercise6.pdf | mp4 | |
July 7th | Operation Sequence Model and OOV Translation | 14_part1_OSM.pdf 14_part2_OOV.pdf | mp4 | |
July 14th | Review and Dry run for Zoom Exam (you need a working webcam) | trial_exam.doc | mp4 | |
July 20th | Exam, live on zoom (you need a working webcam) | exam_2021.doc |
Literature:
Philipp Koehn's book Statistical Machine Translation.
Kevin Knight's tutorial on SMT (particularly look at IBM Model 1)
Philipp Koehn's other book Neural Machine Translation.