Home  | Publications | SMK+26

Language Models Learn Universal Representations of Numbers and Here's Why You Should Care

MCML Authors

Abstract

Prior work has shown that large language models (LLMs) often converge to accurate input embedding for numbers, based on sinusoidal representations. In this work, we quantify that these representations are in fact strikingly systematic, to the point of being almost perfectly universal: different LLM families develop equivalent sinusoidal structures, and number representations are broadly interchangeable in a large swathe of experimental setups. We show that properly factoring in this characteristic is crucial when it comes to assessing how accurately LLMs encode numeric and other ordinal information, and that mechanistically enhancing this sinusoidality can also lead to reductions of LLMs' arithmetic errors.

inproceedings SMK+26


ACL 2026

64th Annual Meeting of the Association for Computational Linguistics. San Diego, CA, USA, Jul 02-07, 2026. To be published. Preprint available.
Conference logo
A* Conference

Authors

M. Štefánik • T. Mickus • M. Kadlčík • B. Højer • M. Spiegel • R. Vázquez • A. Sinha • J. Kuchař • P. Mondorf • P. Stenetorp

Links

URL GitHub

Research Area

 B2 | Natural Language Processing

BibTeXKey: SMK+26

Back to Top