massOfai

Self-Attention

NLP & Text

Relating positions in sequence to each other

What is Self-Attention?

Each position attends to all positions to compute representation. Core of transformers.

Real-World Examples

  • Transformers
  • BERT
  • GPT
  • Understanding long-range dependencies

When to Use This

Fundamental mechanism in modern NLP architectures