PLASH: Provably Linear-Time Attention with Selective Higher-Order Feature Sketching

Published in International Conference on Machine Learning (ICML), 2026

Authors: Yuwen Huang and Xiang Pan

Proposes PLASH, an attention block with provably linear-time complexity that preserves the standard attention interface using learned representatives and selective higher-order feature sketching.