Zubnet AIAprenderWiki › Attention Visualization
Fundamentos

Attention Visualization

Attention Maps, Attention Heatmap
Visualizar a qué un modelo Transformer «atiende» mostrando los pesos de atención como heatmaps. Para cada token de query, el mapa de atención muestra cuánto peso asigna a cada otro token. Pesos altos (puntos brillantes) indican atención fuerte — el modelo considera esos tokens altamente relevantes a la computación actual.

Por qué importa

La visualización de atención es la forma más intuitiva de asomarse dentro de un Transformer y entender su razonamiento. Cuando un modelo traduce «le chat noir» a «the black cat», los mapas de atención muestran que «black» atiende fuertemente a «noir» y «cat» a «chat». Esto ayuda a debuguear comportamiento del modelo, entender fallas, y construir intuición sobre cómo funciona la atención.

Deep Dive

The attention weight matrix is (seq_len × seq_len) for each head and layer. To visualize: pick a layer and head, display the matrix as a heatmap where row i shows which tokens token i attends to. Bright cells mean high attention. For multi-head attention, you can visualize individual heads (each specializes in different patterns) or average across heads (overall attention distribution).

What Attention Maps Show (and Don't)

Attention maps show which tokens a head considers when computing its output, but they don't directly show what the model "understands" or why it made a decision. High attention doesn't mean "important" — some heads attend to punctuation or positional patterns without semantic meaning. Attention maps are descriptive (what the model looked at) not explanatory (why it made its decision). They're a useful debugging tool, not a complete explanation.

Herramientas

BertViz provides interactive attention visualizations for Transformer models. Ecco and Captum offer attention-based interpretability for PyTorch models. For LLMs accessed via API, some providers return attention weights or log-probabilities that enable partial visualization. In image generation, cross-attention maps show which image regions correspond to which prompt words — useful for understanding why the model placed objects where it did.

Conceptos relacionados

← Todos los términos
ESC