Learning Journal
2025-08-02
#self-attention #multi-head-attention
My notes after reading on the attention mechanism behind the transformer introduced in 'Attention is All You Need'.