Accessibility navigation


Transformers, contextualism, and polysemy

Grindrod, J. ORCID: https://orcid.org/0000-0001-8684-974X (2025) Transformers, contextualism, and polysemy. Ergo. ISSN 2330-4014 (In Press)

[img] Text - Accepted Version
· Restricted to Repository staff only
· The Copyright of this document has not been checked yet. This may affect its availability.

75kB

It is advisable to refer to the publisher's version if you intend to cite from this work. See Guidance on citing.

Abstract/Summary

The transformer architecture, introduced by Vaswani et al. (2017), is at the heart of the remarkable recent progress in the development of language models, including widely-used chatbots such as Chat-GPT and Claude. In this paper, I argue that we can extract from the way the transformer architecture works a theory of the relationship between context and meaning. I call this the transformer theory, and I argue that it is novel with regard to two related philosophical debates: the contextualism debate regarding the extent of context-sensitivity across natural language, and the polysemy debate regarding how polysemy should be captured within an account of word meaning.

Item Type:Article
Refereed:Yes
Divisions:Arts, Humanities and Social Science > School of Humanities > Philosophy
ID Code:120383
Publisher:Michigan Publishing

University Staff: Request a correction | Centaur Editors: Update this record

Page navigation