Skip to main content

6 docs tagged with "transformers"

View all tags

What Is Self-Attention?

Self-attention allows a model to look at all the other words in a sentence (or a document, or code...) and decide how important each of them is for understanding a particular word.