Hessian-Enhanced Token Attribution (HETA): Interpreting Autoregressive LLMs

📰 ArXiv cs.AI

arXiv:2604.13258v1 Announce Type: cross Abstract: Attribution methods seek to explain language model predictions by quantifying the contribution of input tokens to generated outputs. However, most existing techniques are designed for encoder-based architectures and rely on linear approximations that fail to capture the causal and semantic complexities of autoregressive generation in decoder-only models. To address these limitations, we propose Hessian-Enhanced Token Attribution (HETA), a novel a

Published 16 Apr 2026
Read full paper → ← Back to Reads