Graded grammatical expectations in transformer models
In: Proceedings of the Annual Meeting of the Cognitive Science Society; vol 45, iss 45; (2023)
Online
Elektronische Ressource
Large language models (LLMs) can be reasonably thought of as models of idealized statistical learners. Thus, the extent to which they grasp the grammar of the language they are trained on suggests how much of it can be learned from memorization, abstraction, and generalization of linguistic input. However, the knowledge of LMMs’ grammar has largely been gleaned from examples of their outputs or datasets not designed to assess how native-like its knowledge is. In this study, we probed the knowledge of an LLM, GPT-3, with a graded grammatical acceptability task previously normed on humans. GPT-3’s ratings were correlated with human ratings, even with minimal examples. Moreover, GPT-3’s deviation from the human norms was predicted by the between-subject variation for each item, and these deviations were rarely outside of the range of human ratings. Follow-up analyses tested the extent to which local probabilistic structure drives these judgments using n-gram models.
Titel: |
Graded grammatical expectations in transformer models
|
---|---|
Autor/in / Beteiligte Person: | Contreras Kallens, Pablo |
Link: | |
Quelle: | Proceedings of the Annual Meeting of the Cognitive Science Society; vol 45, iss 45; (2023) |
Veröffentlichung: | 2023 |
Medientyp: | Elektronische Ressource |
Schlagwort: |
|
Sonstiges: |
|