Skip to content

Unsupervised Cross-Lingual Representation Learning

Photo of Rahel Jhirad
Hosted By
Rahel J.
Unsupervised Cross-Lingual Representation Learning

Details

I am really pleased to welcome Sebastian Ruder. Sebastian is a research scientist at Deepmind.

Abstract: Research in natural language processing (NLP) has seen many advances over the recent years, from word embeddings to pretrained language models. However, most of these approaches rely on large labelled datasets, which has constrained their success to languages where such data is plentiful (mostly English). In this talk, I will give an overview of approaches that learn cross-lingual representations and enable us to scale NLP models to more of the world's 7,000 languages. I will cover the spectrum of such cross-lingual representations, from word embeddings to deep pretrained models, with a focus on unsupervised approaches. The talk will conclude with a discussion of the cutting-edge of learning such representations, their limitations, and future directions.

Bio: Sebastian is a research scientist at DeepMind, London. He completed his PhD in Natural Language Processing at the National University of Ireland while working as a research scientist at a Dublin-based NLP startup. Previously, he studied Computational Linguistics at the University of Heidelberg, Germany and at Trinity College, Dublin. His main research interests are transfer and cross-lingual learning. He is also interested in helping make ML and NLP more accessible. You can find him at his blog http://ruder.io/.

Follow him on Twitter here: twitter.com/seb_ruder

Photo of Economics and Big Data group
Economics and Big Data
See more events
NYU Courant Institute
251 Mercer St., Room 109 · New York, NY