AI Blog

AI Blog

by Michele Laurelli

Attention Mechanism

/əˈtɛnʃən ˈmɛkənɪzəm/
Technique
Definition

A technique allowing models to focus on specific parts of the input when producing output.

Attention assigns importance weights to different input elements. Used in machine translation, image captioning, and transformers. Enables models to handle long sequences effectively.

Examples

1

Focusing on relevant words in translation

2

Highlighting important image regions

3

Multi-head attention in transformers

Michele Laurelli - AI Research & Engineering