Understanding Transformers: Intuition and mathematics of attention

We give a thorough introduction to the intuition and equations of self-attention and transformers. No prior experience needed. This discussion serves to get us on the same page for further discussions on large language models (hardware implementations, new algorithm design, few-shot learning, is it AGI?).

Go to group wiki Go to wiki users Info

Timetable

Day Time Location
Fri, 05.05.2023 15:00 - 17:00 Lecture room
Thu, 04.05.2023 17:00 - 18:00 Lecture room

Moderator

Jens Egholm Pedersen
Mihai Petrovici
Mark Schoene

Members

Christian Pehle
Yannik Stradmann