26 September 2025
DESY
Europe/Berlin timezone

 

This workshop will focus on the transformer architecture and its underlying (self-)attention mechanisms that gained substantial interest in recent years. Despite their empirical success and groundbreaking advances in natural language processing, computer vision, and scientific computing, the mathematical understanding of transformers is still in its infancy, with many fundamental questions only starting to be posed and addressed. 

We aim to bring together researchers with backgrounds in multi-agent dynamics, optimal transport, and PDEs, to initiate discussions on a variety of aspects connected to the theoretical principles governing transformers. By fostering discussions, we seek to advance this young and rapidly evolving research field, uncovering new mathematical perspectives on transformer models.

 

Confirmed speakers

 

This is a satellite event to the Conference on Mathematics of Machine Learning 2025 that takes place at TUHH from September 22nd-25th 2025.

We gratefully acknowledge support by the DFG funded priority programme Theoretical Foundations of Deep Learning and Helmholtz Imaging.

Starts
Ends
Europe/Berlin
DESY
Building 1b, Seminar Room 4ab
Notkestraße 85 22607 Hamburg Germany
Go to map

Organized by

Martin Burger (Helmholtz Imaging, DESY and University of Hamburg)
Samira Kabri (Helmholtz Imaging, DESY)
Konstantin Riedl (University of Oxford)
Tim Roith (Helmholtz Imaging, DESY)
Lukas Weigand (Helmholtz Imaging, DESY)

Registration
Registration for this event is currently open.