abstract: Lexical iconicity, a direct relation between a word's meaning and its form, is an important aspect of every natural language, most commonly manifesting through sound-meaning associations. Since Large language models' (LLMs') access to both meaning and sound of text is only mediated (meaning through textual context, sound through written representation, further complicated by tokenization), we might expect that the encoding of iconicity in LLMs would be either insufficient or significantly different from human processing. This study addresses this hypothesis by having GPT-4 generate highly iconic pseudowords in artificial languages. To verify that these words actually carry iconicity, we had their meanings guessed by Czech and German participants (n=672) and subsequently by LLM-based participants (generated by GPT-4 and Claude 3.5 Sonnet). The results revealed that humans can guess the meanings of pseudowords in the generated iconic language more accurately than words in distant natural languages and that LLM-based participants are even more successful than humans in this task. This core finding is accompanied by several additional analyses concerning the universality of the generated language and the cues that both human and LLM-based participants utilize.


I helped in the creation of a jspsych plugin and in the data collection for the article. It can be found here.


Download the bibliographic data or copy it from here:
@misc{Marklova-Iconicity-in-Large-2025,
 abstract = {Lexical iconicity, a direct relation between a word's meaning and its form,
is an important aspect of every natural language, most commonly manifesting
through sound-meaning associations. Since Large language models' (LLMs') access
to both meaning and sound of text is only mediated (meaning through textual
context, sound through written representation, further complicated by
tokenization), we might expect that the encoding of iconicity in LLMs would be
either insufficient or significantly different from human processing. This
study addresses this hypothesis by having GPT-4 generate highly iconic
pseudowords in artificial languages. To verify that these words actually carry
iconicity, we had their meanings guessed by Czech and German participants
(n=672) and subsequently by LLM-based participants (generated by GPT-4 and
Claude 3.5 Sonnet). The results revealed that humans can guess the meanings of
pseudowords in the generated iconic language more accurately than words in
distant natural languages and that LLM-based participants are even more
successful than humans in this task. This core finding is accompanied by
several additional analyses concerning the universality of the generated
language and the cues that both human and LLM-based participants utilize.},
 archiveprefix = {arXiv},
 author = {Marklov{\'a}, Anna and Mili{\v c}ka, Ji{\v r}{\'i} and Ryvkin, Leonid and Bennet, {\v L}udmila Lackov{\'a} and Korman{\'i}kov{\'a}, Libu{\v s}e},
 eprint = {2501.05643},
 primaryclass = {cs.CL},
 title = {Iconicity in Large Language Models},
 year = {2025}
}