Interpreting Context Look-ups in Transformers: Investigating Attention-MLP Interactions

We probe how attention heads activate specialized "next-token" neurons in LLMs. Prompting GPT-4 reveals heads recognizing contexts tied to predicting tokens, activating neurons via residual connections. This elucidates context-dependent specialization in LLMs.

An interview with

"
Interpreting Context Look-ups in Transformers: Investigating Attention-MLP Interactions
" was written by

Author contribution

No items found.

Citation

Send feedback

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Media kit

Quotes

No items found.

All figures

No items found.