Influencing Token Attention
The lever you have — and its limits. Interactive companion to "Influencing Token Attention Through Prompting"
Note: Attention weights and vector distances shown are conceptual illustrations, not empirical measurements. Actual attention weights are computed inside models and are not directly inspectable during normal use. These visualisations demonstrate the direction of effects, not measured magnitudes.
"Summarise this medical report"
"Summarise this medical report. Sample size and methodology limitations are as important as efficacy findings."
What changed: The steering tokens ("sample size," "methodology," "limitations") now exist in the context. During attention computation, "small" and "trial" have higher similarity to these new tokens. The weights shift — limitation-related content gets more attention, headline numbers get less. Same input, different emphasis.
"This document contains important qualifications in sections 3 and 7 that are often overlooked. Read carefully for nuance before summarising."
[The actual document to be summarised...]
"Now summarise, preserving those qualifications."
Why position matters: Earlier tokens undergo more attention processes (causal attention accumulates). The priming tokens — "qualifications," "overlooked," "nuance" — get positional advantage. When the model later encounters qualifying content in the document, those attractor tokens are already established as reference points. Closing instructions reinforce at the other high-attention position.
The key insight: In trained vector space, "cat" and "democracy" are far apart — they rarely co-occur meaningfully in training data. But if you explicitly state their relationship matters for your analysis, you create a contextual bridge. The model now has tokens establishing cat-democracy relevance that participate in attention calculations. You're not retraining — the underlying distances remain — but you're creating local influence that the trained weights alone wouldn't produce.
The honest position: You have a lever. It's real and it works. But it's influence, not control. The model descends gradients shaped by billions of parameters learned from massive training. Your prompt adds local constraints that reshape which minimum it finds. That's genuinely useful — but it's not magic, and it doesn't replace verification.