ZeroS: Zero‑Sum Linear Attention for Efficient Transformers

Published in NeurIPS 2025 (Spotlight), 2025

A zero‑sum reparameterization of linear attention that avoids uniform accumulation bias and allows contrastive operations within a single layer, narrowing the gap with softmax attention under linear complexity.