Volltext-Downloads (blau) und Frontdoor-Views (grau)

Human languages trade off complexity against efficiency

  • From a cross-linguistic perspective, language models are interesting because they can be used as idealised language learners that learn to produce and process language by being trained on a corpus of linguistic input. In this paper, we train different language models, from simple statistical models to advanced neural networks, on a database of 41 multilingual text collections comprising a wide variety of text types, which together include nearly 3 billion words across more than 6,500 documents in over 2,000 languages. We use the trained models to estimate entropy rates, a complexity measure derived from information theory. To compare entropy rates across both models and languages, we develop a quantitative approach that combines machine learning with semiparametric spatial filtering methods to account for both language- and document-specific characteristics, as well as phylogenetic and geographical language relationships. We first establish that entropy rate distributions are highly consistent across different language models, suggesting that the choice of model may have minimal impact on cross-linguistic investigations. On the basis of a much broader range of language models than in previous studies, we confirm results showing systematic differences in entropy rates, i.e. text complexity, across languages. These results challenge the long-held notion that all languages are equally complex. We then show that higher entropy rate tends to co-occur with shorter text length, and argue that this inverse relationship between complexity and length implies a compensatory mechanism whereby increased complexity is offset by increased efficiency. Finally, we introduce a multi-model multilevel inference approach to show that this complexity-efficiency trade-off is partly influenced by the social environment in which languages are used: languages spoken by larger communities tend to have higher entropy rates while using fewer symbols to encode messages.

Download full text files

Export metadata

Additional Services

Search Google Scholar

Statistics

frontdoor_oas
Metadaten
Author:Alexander KoplenigGND, Sascha WolferORCiDGND, Jan Oliver RüdigerORCiDGND, Peter MeyerORCiD
URN:urn:nbn:de:bsz:mh39-127860
DOI:https://doi.org/10.31219/osf.io/8xgqz
Publisher:OSF Preprints, Center for Open Science
Place of publication:Charlottesville, VA
Document Type:Preprint
Language:English
Year of first Publication:2024
Date of Publication (online):2024/08/27
Publishing Institution:Leibniz-Institut für Deutsche Sprache (IDS)
Publicationstate:Veröffentlichungsversion
Tag:computational linguistics; entropy; information theory; language complexity; language model; linguistic typology; machine learning; quantitative linguistics; spatial filtering
GND Keyword:Datenanalyse; Korpus <Linguistik>; Mehrsprachigkeit; Modell; Statistik
Page Number:57
DDC classes:400 Sprache / 400 Sprache, Linguistik
Open Access?:ja
Linguistics-Classification:Computerlinguistik
Program areas:Lexik
Licence (English):License LogoCreative Commons - Attribution 4.0 International