Neural language models are increasingly valued in computational psycholinguistics, due to their ability to provide conditional probability distributions over the lexicon that are predictive of human processing times. Given the vast array of available models, it is of both theoretical and methodological importance to assess what features of a model influence its psychometric quality. In this work we focus on parameter size, showing that larger Transformer-based language models generate probabilistic estimates that are less predictive of early eye-tracking measurements reflecting lexical access and early semantic integration. However, relatively bigger models show an advantage in capturing late eye-tracking measurements that reflect the full semantic and syntactic integration of a word into the current language context. Our results are supported by eye movement data in ten languages and consider four models, spanning from 564M to 4.5B parameters.

de Varda, A., Marelli, M. (2023). Scaling in Cognitive Modelling: a Multilingual Approach to Human Reading Times. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp.139-149). Association for Computational Linguistics (ACL).

Scaling in Cognitive Modelling: a Multilingual Approach to Human Reading Times

de Varda, AG;Marelli, M
2023

Abstract

Neural language models are increasingly valued in computational psycholinguistics, due to their ability to provide conditional probability distributions over the lexicon that are predictive of human processing times. Given the vast array of available models, it is of both theoretical and methodological importance to assess what features of a model influence its psychometric quality. In this work we focus on parameter size, showing that larger Transformer-based language models generate probabilistic estimates that are less predictive of early eye-tracking measurements reflecting lexical access and early semantic integration. However, relatively bigger models show an advantage in capturing late eye-tracking measurements that reflect the full semantic and syntactic integration of a word into the current language context. Our results are supported by eye movement data in ten languages and consider four models, spanning from 564M to 4.5B parameters.
paper
cognitive modelling, inverse scaling, sentence reading, eyetracking
English
61st Annual Meeting of the Association for Computational Linguistics, ACL 2023 - 9 July 2023 through 14 July 2023
2023
Proceedings of the Annual Meeting of the Association for Computational Linguistics
9781959429715
2023
2
139
149
none
de Varda, A., Marelli, M. (2023). Scaling in Cognitive Modelling: a Multilingual Approach to Human Reading Times. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp.139-149). Association for Computational Linguistics (ACL).
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/467169
Citazioni
  • Scopus 7
  • ???jsp.display-item.citation.isi??? ND
Social impact