ELLIS header
University of Stuttgart Logo
Max Planck Institute for Intelligent Systems Logo
03.06.2026 - Distinguished Lecture Series: Rico Sennrich (University of Zurich)

03.06.2026 - Distinguished Lecture Series: Rico Sennrich (University of Zurich)

We are pleased to announce our upcoming Distinguished Lecture Series talk by Rico Sennrich (University of Zurich)! The talk will take place on Wednesday, June 3, 9:45 - 11:15 CET. in room UN32.101. Prof. Sennrich will be available for Meetings. If you are interested in a Meeting please email ellis-office@uni-stuttgart.de

Rico Sennrich is associate professor in Computational Linguistics at the University of Zurich, where he has worked since 2019, and Honorary Fellow at the University of Edinburgh. His research covers various areas of NLP, with a special focus on high-quality machine translation, multilingual and low-resource NLP, and multimodal models. Some of his work, such as BPE tokenization and RMS layer normalisation, has found wide use in modern LLMs.

Title: Beyond More Data: Advancing Low-Resource Natural Language Processing from Tokenization to Inference

Beyond More Data: Advancing Low-Resource Natural Language Processing from Tokenization to Inference

In a field where the state of the art is often advanced by scale - building larger models on more data - I will make the argument that a surprising amount of progress can be achieved by modifying other parts of the NLP pipeline, especially for low-resource languages. I will introduce a parity-aware modification of byte-pair encoding that is optimized towards cross-lingual fairness in tokenization length, and can yield more equitable models where length translates directly into costs, while also performing better on cross-lingual benchmarks. I will also discuss the task of machine translation, where massively multilingual models and large language models have been shown to handle many translation directions, but which still suffer from problems such as hallucinations or translations in the wrong language. I will show how these issues can be reduced massively with contrastive decoding methods that pair each input with appropriate contrastive inputs, and sketch wider applications of this strategy to interacting with LLMs.

Date: Wednesday, June 3, 2026
Time: 9:45 - 11:15 CET
Place: Universitätstraße 32.101, Campus Vaihingen of the University of Stuttgart.

Looking forward to seeing you all there! No registration necessary.