Zero-Shot Quantization via Weight-Space Arithmetic

📰 ArXiv cs.AI

Zero-Shot Quantization via Weight-Space Arithmetic improves robustness to post-training quantization by up to 60% without receiver-side quantization-aware training

advanced Published 7 Apr 2026
Action Steps
  1. Extract the quantization vector from a donor task using weight-space arithmetic
  2. Apply the quantization vector to a receiver model to improve robustness to PTQ-induced noise
  3. Evaluate the performance of the patched receiver model on a target task
  4. Fine-tune the receiver model if necessary to further improve performance
Who Needs to Know This

AI engineers and researchers can benefit from this method as it provides a way to improve model robustness to quantization-induced noise without requiring additional training data or computational resources. This can be particularly useful in deployment scenarios where data is limited or expensive to obtain

Key Insight

💡 The quantization vector is a transferable direction in weight space that can be used to improve robustness to post-training quantization without requiring receiver-side quantization-aware training

Share This
🚀 Improve model robustness to quantization-induced noise by up to 60% with Zero-Shot Quantization via Weight-Space Arithmetic! 🤖
Read full paper → ← Back to Reads