Zubnet AIAprenderWiki › Attention Visualization
Fundamentos

Attention Visualization

Attention Maps, Attention Heatmap
Visualizar a que um modelo Transformer “atende” exibindo os pesos de atenção como heatmaps. Para cada token de query, o mapa de atenção mostra quanto peso ele atribui a cada outro token. Pesos altos (pontos brilhantes) indicam atenção forte — o modelo considera esses tokens altamente relevantes à computação atual.

Por que importa

Visualização de atenção é a forma mais intuitiva de espiar dentro de um Transformer e entender seu raciocínio. Quando um modelo traduz “le chat noir” para “the black cat”, os mapas de atenção mostram que “black” atende fortemente a “noir” e “cat” a “chat”. Isso ajuda a debugar comportamento do modelo, entender falhas, e construir intuição sobre como atenção funciona.

Deep Dive

The attention weight matrix is (seq_len × seq_len) for each head and layer. To visualize: pick a layer and head, display the matrix as a heatmap where row i shows which tokens token i attends to. Bright cells mean high attention. For multi-head attention, you can visualize individual heads (each specializes in different patterns) or average across heads (overall attention distribution).

What Attention Maps Show (and Don't)

Attention maps show which tokens a head considers when computing its output, but they don't directly show what the model "understands" or why it made a decision. High attention doesn't mean "important" — some heads attend to punctuation or positional patterns without semantic meaning. Attention maps are descriptive (what the model looked at) not explanatory (why it made its decision). They're a useful debugging tool, not a complete explanation.

Ferramentas

BertViz provides interactive attention visualizations for Transformer models. Ecco and Captum offer attention-based interpretability for PyTorch models. For LLMs accessed via API, some providers return attention weights or log-probabilities that enable partial visualization. In image generation, cross-attention maps show which image regions correspond to which prompt words — useful for understanding why the model placed objects where it did.

Conceitos relacionados

← Todos os termos
ESC