Hessian-Enhanced Token Attribution (HETA): Interpreting Autoregressive LLMs
📰 ArXiv cs.AI
arXiv:2604.13258v1 Announce Type: cross Abstract: Attribution methods seek to explain language model predictions by quantifying the contribution of input tokens to generated outputs. However, most existing techniques are designed for encoder-based architectures and rely on linear approximations that fail to capture the causal and semantic complexities of autoregressive generation in decoder-only models. To address these limitations, we propose Hessian-Enhanced Token Attribution (HETA), a novel a
DeepCamp AI