Algebraic Positional Encodings
Paper in proceeding, 2024

We introduce a novel positional encoding strategy for Transformer-style models, addressing the shortcomings of existing, often ad hoc, approaches. Our framework implements a flexible mapping from the algebraic specification of a domain to a positional encoding scheme where positions are interpreted as orthogonal operators. This design preserves the structural properties of the source domain, thereby ensuring that the end-model upholds them. The framework can accommodate various structures, including sequences, grids and trees, but also their compositions. We conduct a series of experiments demonstrating the practical applicability of our method. Our results suggest performance on par with or surpassing the current state of the art, without hyper-parameter optimizations or task search'' of any kind

Author

Konstantinos Kogkalidis

Aalto University

University of Bologna

Jean-Philippe Bernardy

Chalmers, Computer Science and Engineering (Chalmers), Computing Science

University of Gothenburg

V. Garg

Aalto University

YaiYai

Advances in Neural Information Processing Systems

10495258 (ISSN)

Vol. 37

Advances in Neural Information Processing Systems 38
Vancouver, Canada,

Subject Categories (SSIF 2025)

Computer and Information Sciences

Related datasets

Code Algebraic Positional Encodings [dataset]

URI: https://aalto-quml.github.io/ape/

More information

Latest update

4/3/2025 8