padraiglindrome Profile Banner
Patrick Haller Profile
Patrick Haller

@padraiglindrome

Followers
290
Following
2K
Media
15
Statuses
76

PhD student in Computational Linguistics @cl_uzh. Interested in language modeling, human language processing, drag race, you name it. he/him ๐Ÿณ๏ธโ€๐ŸŒˆ

Zurich, Switzerland
Joined June 2015
Don't wanna be here? Send us removal request.
@padraiglindrome
Patrick Haller
10 months
RT @glnmario: How should computational (psycho)linguists properly apply token-level language models to the fieldโ€™s inherently character-levโ€ฆ.
0
7
0
@padraiglindrome
Patrick Haller
11 months
Excited for tomorrow to present our work on LLM response stability in the context of political bias assessment at @COLM_conf! Stop by between 9 and 11 at poster #39. Joint work w/@j_vamvas and @LenaAJaeger. ๐Ÿ”—
Tweet media one
2
8
47
@grok
Grok
4 days
Join millions who have switched to Grok.
173
334
3K
@padraiglindrome
Patrick Haller
1 year
But why do LMs emulate readers with low rather than high vocabulary size? We answer this question and discuss the implications of our results for the cognitive mechanisms of language processing in our paper: ๐Ÿ“„ Thanks for reading!.
0
0
1
@padraiglindrome
Patrick Haller
1 year
We find that LM surprisal is more predictive of reading time (RT) for readers scoring low in verbal intelligence and vocabulary size. LM entropy on the other hand predicts RT more accurately for readers with high memory updating and operation span scores.
Tweet media one
1
0
0
@padraiglindrome
Patrick Haller
1 year
(2) For each test, we assess the difference in predictive power of surprisal and entropy on reading times for the group or readers that performed above average vs the group of readers that performed below average.
Tweet media one
1
0
0
@padraiglindrome
Patrick Haller
1 year
We find that for most tests, readers that obtained higher scores show smaller surprisal effects (see Table) as well as smaller entropy effects.
Tweet media one
1
0
0
@padraiglindrome
Patrick Haller
1 year
(1) For each psychometric score, we fit a linear-mixed regression model with reading time as target variable and word length, lexical frequency, surprisal, entropy, psychometric score, and either surprisal x psychometric score or entropy x psychometric score as predictors.
Tweet media one
1
0
0
@padraiglindrome
Patrick Haller
1 year
๐Ÿ’กin this paper, we test (1) whether readers exhibit surprisal and entropy effects relative to their cognitive capacities assessed via psychometric scores and (2) whether surprisal/entropy has higher predictive power on reading times for certain groups of readers.
1
0
0
@padraiglindrome
Patrick Haller
1 year
๐Ÿ‘€๐Ÿ“– Excited to share our #ACL2024 findings paper on how language models emulate readers with specific cognitive capacities. See ๐Ÿงต . We'll present this work next week in Bangkok during the findings poster session 4 and at CMCL. w/@lsbolliger and @LenaAJaeger
Tweet media one
2
5
25
@padraiglindrome
Patrick Haller
1 year
RT @DiLi_lab: ๐Ÿ“–๐Ÿ‘€ We're excited to present PoTeC -- The Potsdam Textbook Corpus! PoTeC is a German naturalistic eye-tracking-while-reading cโ€ฆ.
0
5
0
@padraiglindrome
Patrick Haller
2 years
RT @j_vamvas: Really excited about our new preprint: Generating high-quality machine translations can be accelerateโ€ฆ.
0
7
0
@padraiglindrome
Patrick Haller
2 years
RT @tannonk: ๐Ÿ”Looking for some #multilingual #LLM reading for the holidays or just that last minute stocking filler? ๐ŸŽ… . ๐Ÿ‘€Look no further!โ€ฆ.
Tweet card summary image
arxiv.org
The vast majority of today's large language models (LLMs) are English-centric, having been pretrained predominantly on English text. Yet, in order to meet user expectations, models need to be able...
0
10
0
@padraiglindrome
Patrick Haller
2 years
RT @lsbolliger: ๐Ÿ“ข Interested in #eyemovements for #NLP? Or wanna chat about cognitive enhancement and interpretability of #LMs? .๐Ÿ’ƒ Come byโ€ฆ.
0
5
0
@padraiglindrome
Patrick Haller
2 years
RT @noeminaepli: how do you evaluate systems that generate non-standardized #dialects? .check out our WMT23 paper ๐Ÿ“œ.
0
12
0
@padraiglindrome
Patrick Haller
2 years
2โƒฃ We should investigate phenomena that have not been at the center of current theories of individual-level sentence processing so far. For more details, check out the preprint!.๐Ÿ“„
1
0
3
@padraiglindrome
Patrick Haller
2 years
This finding is particularly important for theories of sentence processing that generalize to the individual level!.1โƒฃ We might have to revisit theories of individual-level sentence processing incorporating effects that are stable on the group level.
1
0
3
@padraiglindrome
Patrick Haller
2 years
Our results show that individual sensitivities to word length are stable across experimental sessions and methodologies. However, most syntactic predictors and predictability effects have neither been found to be temporally nor cross-methodologically test-retest reliable.
Tweet media one
1
0
2
@padraiglindrome
Patrick Haller
2 years
Therefore, in our work, we assess the test-retest measurement reliability of individual differences for several psycholinguistic predictors: dependency distance, number of left dependents, surprisal, lexical frequency, and word length.
1
0
2
@padraiglindrome
Patrick Haller
2 years
Importantly, the test-retest measurement reliability of individual-level effects cannot be taken as a given due to the "reliability paradox"! It states that measurement reliability at the individual level is lower for manipulations with high between-subjects reliability.
Tweet media one
1
0
3