How to explain Q, K and V of Self Attention in Transformers (BERT)?

Discover AI
Discover AI
11.6 هزار بار بازدید - 2 سال پیش - How to explain Q, K
How to explain Q, K and V of Self Attention in Transformers (BERT)?
Thought about it and present here my most general approach to explain the history of the notation of Query, Key and Values and how they combine for classical self attention mechanism in transformers.

#ai
#self_attention
#bert
2 سال پیش در تاریخ 1401/07/02 منتشر شده است.
11,664 بـار بازدید شده
... بیشتر