Can Cross-Layer Transcoders Replace Vision Transformer Activations? An Interpretable Perspective on Vision

📰 ArXiv cs.AI

arXiv:2604.13304v1 Announce Type: cross Abstract: Understanding the internal activations of Vision Transformers (ViTs) is critical for building interpretable and trustworthy models. While Sparse Autoencoders (SAEs) have been used to extract human-interpretable features, they operate on individual layers and fail to capture the cross-layer computational structure of Transformers, as well as the relative significance of each layer in forming the last-layer representation. Alternatively, we introdu

Published 16 Apr 2026
Read full paper → ← Back to Reads