Zero-Shot Quantization via Weight-Space Arithmetic
📰 ArXiv cs.AI
arXiv:2604.03420v1 Announce Type: cross Abstract: We show that robustness to post-training quantization (PTQ) is a transferable direction in weight space. We call this direction the quantization vector: extracted from a donor task by simple weight-space arithmetic, it can be used to patch a receiver model and improve robustness to PTQ-induced noise by as much as 60%, without receiver-side quantization-aware training (QAT). Because the method requires no receiver training data, it provides a zero
DeepCamp AI