TY - GEN
T1 - Laying Anchors
T2 - 2024 Findings of the Association for Computational Linguistics: NAACL 2024
AU - Sharma, Mandar
AU - Taware, Rutuja Murlidhar
AU - Koirala, Pravesh
AU - Muralidhar, Nikhil
AU - Ramakrishnan, Naren
N1 - Publisher Copyright:
© 2024 Association for Computational Linguistics.
PY - 2024
Y1 - 2024
N2 - Off-the-shelf pre-trained language models have become the de facto standard in NLP pipelines for a multitude of downstream tasks. However, the inability of these models to properly encode numerals limits their performance on tasks requiring numeric comprehension. We introduce strategies to semantically prime numerals in any corpus by generating anchors governed by the distribution of numerals in said corpus, thereby enabling mathematically grounded representations of these numeral tokens. We establish the superiority of our proposed techniques through evaluation on a range of numeracy tasks for both in-domain (seen) and out-domain (unseen) numerals. Further, we expand our empirical evaluations to numerals ranging from 1 to 10 billion, a significantly broader range compared to previous studies of the same nature, and we demonstrate significant improvements in the mathematical grounding of our learned embeddings.
AB - Off-the-shelf pre-trained language models have become the de facto standard in NLP pipelines for a multitude of downstream tasks. However, the inability of these models to properly encode numerals limits their performance on tasks requiring numeric comprehension. We introduce strategies to semantically prime numerals in any corpus by generating anchors governed by the distribution of numerals in said corpus, thereby enabling mathematically grounded representations of these numeral tokens. We establish the superiority of our proposed techniques through evaluation on a range of numeracy tasks for both in-domain (seen) and out-domain (unseen) numerals. Further, we expand our empirical evaluations to numerals ranging from 1 to 10 billion, a significantly broader range compared to previous studies of the same nature, and we demonstrate significant improvements in the mathematical grounding of our learned embeddings.
UR - http://www.scopus.com/inward/record.url?scp=85197934735&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85197934735&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85197934735
T3 - Findings of the Association for Computational Linguistics: NAACL 2024 - Findings
SP - 2653
EP - 2660
BT - Findings of the Association for Computational Linguistics
A2 - Duh, Kevin
A2 - Gomez, Helena
A2 - Bethard, Steven
Y2 - 16 June 2024 through 21 June 2024
ER -