A. Alcalde Zafra, G. Fantuzzi, E. Zuazua (2025)Exact Sequence Classification with Hardmax Transformers
Abstract. We prove that hardmax attention transformers perfectly classify datasets of $N$ labeled sequences in , . Specifically, given sequences with an arbitrary but finite length in , we construct a transformer with blocks and parameters perfectly classifying this dataset. Our construction achieves the best complexity estimate to date, independent of the length of the sequences, by innovatively alternating feed-forward and self-attention layers and by capitalizing on the clustering effect inherent to the latter. Our novel constructive method also uses low-rank parameter matrices within the attention mechanism, a common practice in real-life transformer implementations. Consequently, our analysis holds twofold significance: it substantially advances the mathematical theory of transformers and it rigorously justifies their exceptional real-world performance in sequence classification tasks.
arxiv: 2502.02270