TY - JOUR
T1 - Communication Drives the Emergence of Language Universals in Neural Agents
T2 - Evidence from the Word-order/Case-marking Trade-off
AU - Lian, Yuchen
AU - Bisazza, Arianna
AU - Verhoef, Tessa
N1 - Publisher Copyright:
© 2023 Association for Computational Linguistics. Distributed under a CC-BY 4.0 license.
PY - 2023
Y1 - 2023
N2 - Artificial learners often behave differently from human learners in the context of neural agent-based simulations of language emergence and change. A common explanation is the lack of appropriate cognitive biases in these learners. However, it has also been proposed that more naturalistic settings of language learning and use could lead to more humanlike results. We investigate this latter account, focusing on the word-order/case-marking trade-off, a widely attested language universal that has proven particularly hard to simulate. We propose a new Neural-agent Language Learning and Communication framework (NeLLCom) where pairs of speaking and listening agents first learn a miniature language via supervised learning, and then optimize it for communication via reinforcement learning. Following closely the setup of earlier human experiments, we succeed in replicating the trade-off with the new framework without hard-coding specific biases in the agents. We see this as an essential step towards the investigation of language universals with neural learners.
AB - Artificial learners often behave differently from human learners in the context of neural agent-based simulations of language emergence and change. A common explanation is the lack of appropriate cognitive biases in these learners. However, it has also been proposed that more naturalistic settings of language learning and use could lead to more humanlike results. We investigate this latter account, focusing on the word-order/case-marking trade-off, a widely attested language universal that has proven particularly hard to simulate. We propose a new Neural-agent Language Learning and Communication framework (NeLLCom) where pairs of speaking and listening agents first learn a miniature language via supervised learning, and then optimize it for communication via reinforcement learning. Following closely the setup of earlier human experiments, we succeed in replicating the trade-off with the new framework without hard-coding specific biases in the agents. We see this as an essential step towards the investigation of language universals with neural learners.
UR - https://www.scopus.com/pages/publications/85177476588
U2 - 10.1162/tacl_a_00587
DO - 10.1162/tacl_a_00587
M3 - Article
AN - SCOPUS:85177476588
SN - 2307-387X
VL - 11
SP - 1033
EP - 1047
JO - Transactions of the Association for Computational Linguistics
JF - Transactions of the Association for Computational Linguistics
ER -