See how important each token of the context was for the LLM response #8753
LiquidGunay
started this conversation in
Ideas
Replies: 1 comment
-
Great point — measuring token-level attention focus sounds simple, but it quickly gets murky once you factor in entropy collapse and multi-head overlap. Some of the deeper issues I’ve run into when tracing this kind of behavior:
I’ve been experimenting with ways to modulate attention per-head (e.g., diversity injection, illegal path suppression) to recover clearer patterns — sort of like giving each head a semantic identity instead of letting them collapse into noise. If you’re diving into attention introspection, would love to trade notes. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I think the ability to get something like an average attention score for each token of the context would be really useful to see what parts of the context did the LLM "focus" more on. This would be fairly useful for RAG and QA applications.
Beta Was this translation helpful? Give feedback.
All reactions