Bruce12138
Modern NLP
Data Visualization
Large-Scale Data Science
Theory of Computation
  • Modern NLP
  • Data Visualization
  • Large-Scale Data Science
  • Theory of Computation
  • Archived

    • Machine Learning
    • Algorithm II
    • Graph Theory
Tech Notes
Modern NLP
Data Visualization
Large-Scale Data Science
Theory of Computation
  • Modern NLP
  • Data Visualization
  • Large-Scale Data Science
  • Theory of Computation
  • Archived

    • Machine Learning
    • Algorithm II
    • Graph Theory
Tech Notes
  • Modern NLP

    • Course Homepage
    • Course Slides

Course Homepage

Course Description

Natural language processing is ubiquitous in modern intelligent technologies, serving as a foundation for language translators, virtual assistants, search engines, and many more. In this course, we cover the foundations of modern methods for natural language processing, such as word embeddings, recurrent neural networks, transformers, and pretraining, and how they can be applied to important tasks in the field, such as machine translation and text classification. We also cover issues with these state-of-the-art approaches (such as robustness, interpretability, sensitivity), identify their failure modes in different NLP applications, and discuss analysis and mitigation techniques for these issues.

Quick access links:

  • Platforms
  • Lecture Schedule
  • Exercise Schedule
  • Contact

Class

PlatformWhere & when
LecturesWednesdays: 11:15-13:00 [STCC - Cloud C] & Thursdays: 13:15-14:00 [CE16]
Exercises SessionThursdays: 14:15-16:00 [CE11]
Project Assistance
(not every week)
Wednesdays: 13:15-14:00 [STCC - Cloud C]
QA Forum & AnnoucementsEd Forum [link]
GradesMoodle [link]
Lecture RecordingsMediaspace [link]

All lectures will be given in person and live streamed on Zoom. The link to the Zoom is available on the Ed Forum (pinned post). Beware that, in the event of a technical failure during the lecture, continuing to accompany the lecture live via zoom might not be possible.

Recording of the lectures will be made available on Mediaspace. We will reuse some of last year's recordings and we may record a few new lectures in case of different lecture contents.

Lecture Schedule

WeekDateTopicSuggested ReadingInstructor
Week 118 Feb
19 Feb
Introduction | Building a simple neural classifier [slides] [video]
Word embeddings [slides] [video]
  • Introduction to natural language processing, chapter 3.1 - 3.3 & chapter 14.5 - 14.6
  • Efficient Estimation of Word Representations in Vector Space
  • GloVe: Global Vectors for Word Representation
  • Enriching word vectors with subword information
  • Advances in pre-training distributed word representations
Antoine Bosselut
Week 225 Feb
26 Feb
Classical LMs | Neural LMs: Fixed Context Models [slides] [video]
Neural LMs: RNNs [slides] [video]
Suggested reading:
  • Introduction to natural language processing, chapter 6.1-6.4
  • A Neural Probabilistic Language Model
  • On the difficulty of training recurrent neural networks
  • Introduction to natural language processing, chapter 3.1 - 3.3 & chapter 18.3, 18.4
Antoine Bosselut
Week 34 Mar
5 Mar
Sequence-to-sequence Models | Transformers [slides]
Tokenization [slides]
Suggested reading:
  • Attention Is All You Need
  • The Annotated Transformer
  • The illustrated transformer
  • Between words and characters: A brief history of open-vocabulary modeling and tokenization in NLP
  • Byt5: Towards a token-free future with pre-trained byte-to-byte models
  • Neural machine translation of rare words with subword units
Antoine Bosselut

Exercise Schedule

WeekRelease DateExercise Session DateTopicInstructor
Week 119 Feb26 FebIntro + SetupMadhur Panwar
Week 226 Feb5 MarLMs + Neural LMs: fixed-context models
Language and Sequence-to-sequence models
Badr AlKhamissi
Week 35 Mar12 MarAttention + Transformers + TokenizationBadr AlKhamissi
Week 412 Mar19 MarPretrained LLMsBadr AlKhamissi
Week 519 Mar26 MarTransfer LearningMadhur Panwar
Week 626 Mar2 AprText GenerationMadhur Panwar
Week 71 Apr2 AprIn-context Learning + Post-trainingTBD

Contacts

Please email us at nlp-cs552-spring2026-ta-team [at] groupes [dot] epfl [dot] ch for any administrative questions, rather than emailing TAs individually. All course content questions need to be asked via Ed.

Lecturer: Antoine Bosselut

Teaching assistants: Madhur Panwar, Badr AlKhamissi, Zeming (Eric) Chen, Sepideh Mamooler, Ayush Tarun, Lazar Milikic

最近更新: 2026/3/5 11:49
Contributors: Zhixiang Dai, github-actions[bot]
Next
Course Slides