Deep Learning For NLP
Organization
- Lecture: Dienstag 13:30-15:10, Room S101/A02
- Practice Class: Dienstag 15:20-17:00, Room S207/167
- Integrierte Lehrveranstaltung (TUCaN-ID 20-00-0947-iv)
The learning material is available from the Moodle eLeaning platform.
The required passcode will be distributed during the lecture.
The first lecture and the first practice class will be on April 16.
Course Content
The lecture provides an introduction to the foundational concepts of deep learning and their application to problems in the area of natural language processing (NLP)
Main aspects:
- Foundations of deep learning (e.g. feed-forward networks, backpropagation, activation functions, loss functions)
- Word embeddings: theory, different approaches and models, application as features for machine learning
- Different architectures of neuronal networks (e.g. recurrent NN, convolutional NN, encoder-decoder models) and their application for groups of NLP problems such as document classification (e.g. spam detection), sequence labeling (e.g. POS-tagging, Named Entity Recognition) and more high-level NLP tasks (e.g. machine translation, argumentation mining, text generation)
Requirements:
foundations in math (linear algebra) and programming (python), ability to quickly acquire machine learning concepts
Expectations
What you can expect from us:
- interactive lecture with integrated tutorials
- problem-based and explorative learning
- stimulating environment
What we expect from you:
- commitment
- feedback
- active participation
Literature
- Lecture by Richard Socher: Deep Learning for Natural Language Processing
- Ian Goodfellow, Yoshua Bengio, and Aaron Courville: Deep Learning, Book in preparation for MIT Press
- Yoav Goldberg (2015): A Primer on Neural Network Models for Natural Language Processing, published on arxiv
Teaching Staff
- Steffen Eger (no fixed office hour, please just ask for an appointment per mail)