AI Blog

AI Blog

by Michele Laurelli

Self-Attention

/sɛlf əˈtɛnʃən/
Technique
Definition

An attention mechanism used in deep learning models that allows a neural network to weigh the importance of different parts of an input relative to each other.

Self-attention is a fundamental technique in language and vision models. It allows models to consider relationships between elements within a single sequence, greatly improving comprehension and generation capabilities.

Examples

1

Machine translation models using self-attention

2

Recommendation systems

3

Text generation models

Michele Laurelli - AI Research & Engineering