Quantcast
Channel: Data Science, Analytics and Big Data discussions - Latest topics
Viewing all articles
Browse latest Browse all 4448

Understanding Transformers Self attention calculation

$
0
0

Regarding this link: https://www.analyticsvidhya.com/blog/2019/06/understanding-transformers-nlp-state-of-the-art-models/?utm_source=blog&utm_medium=demystifying-bert-groundbreaking-nlp-framework#comment-160771

What is the value of Key, Value in the self attention calculation of Transformer model. ?
Query vector is embedding vector for the word that is queried, is that right?
Is attention calculated in RNN is different from self attention in Transformer?

1 post - 1 participant

Read full topic


Viewing all articles
Browse latest Browse all 4448

Trending Articles