Professional Documents
Culture Documents
Transformers Explained "Attention Is All You Need."
Transformers Explained "Attention Is All You Need."
Just:
1. Get a lot of GPUs
2. Get a LOT of data
3. Results will blow your mind.
How big?
GPT-3
But how does it work?
GPT-3
The three pillars
GPT-3
✨Transformers✨
Positional
Encoding
I love science
Positional Encoding
(tokenization)
1 2 3
I love science
GPT-3
✨Transformers✨
Today, transformers are
used for:
Today, anyone can train
models on unlabelled data
References
Morgan, Abby. "Explainable AI: Visualizing Attention in
Transformers." Comet, 16 July 2023,
www.comet.com/site/blog/explainable-ai-for-transformers/.