Transformers and Attention
Details
To get the most out of the session, it is highly recommended that you read the following:
* https://jalammar.github.io/visualizing-neural-machine-translation-mechanics-of-seq2seq-models-with-attention/
* https://jalammar.github.io/illustrated-transformer/
In this study group session, we will cover transformers and attention, the workhorse behind ChatGPT, Gemini, Claude etc... Please be prepared by reading the resources and coming in with questions!