<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<!--<?xml-stylesheet type="text/xsl" href="article.xsl"?>-->
<article article-type="research-article" dtd-version="1.2" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="issn">2767-0279</journal-id>
<journal-title-group>
<journal-title>Glossa Psycholinguistics</journal-title>
</journal-title-group>
<issn pub-type="epub">2767-0279</issn>
<publisher>
<publisher-name>eScholarship Publishing</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.5070/G6011131</article-id>
<article-categories>
<subj-group>
<subject>Regular article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Comparing infrared and webcam eye tracking in the Visual World Paradigm</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Vos</surname>
<given-names>Myrte</given-names>
</name>
<email>myrte.vos@uit.no</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Minor</surname>
<given-names>Serge</given-names>
</name>
<email>sergey.minor@uit.no</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Ramchand</surname>
<given-names>Gillian</given-names>
</name>
<email>gillian.ramchand@uit.no</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
</contrib-group>
<aff id="aff-1"><label>1</label>UiT the Arctic University of Norway, NO</aff>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2022-08-05">
<day>05</day>
<month>08</month>
<year>2022</year>
</pub-date>
<pub-date pub-type="collection">
<year>2022</year>
</pub-date>
<volume>1</volume>
<issue>1</issue>
<elocation-id>9</elocation-id>
<permissions>
<copyright-statement>Copyright: &#x00A9; 2022 The Author(s)</copyright-statement>
<copyright-year>2022</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. See <uri xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</uri>.</license-p>
</license>
</permissions>
<self-uri xlink:href="https://glossapsycholinguistics.journalpub.escholarship.org/articles/10.5070/G6011131/"/>
<abstract>
<p>Visual World eye tracking is a temporally fine-grained method of monitoring attention, making it a popular tool in the study of online sentence processing. Recently, while infrared eye tracking was mostly unavailable, various web-based experiment platforms have rapidly developed webcam eye tracking functionalities, which are now in urgent need of testing and evaluation. We replicated a recent Visual World study on the incremental processing of verb aspect in English using &#8216;out of the box&#8217; webcam eye tracking software (jsPsych; <xref ref-type="bibr" rid="B19">de Leeuw, 2015</xref>) and crowdsourced participants, and fully replicated both the offline and online results of the original study. We furthermore discuss factors influencing the quality and interpretability of webcam eye tracking data, particularly with regards to temporal and spatial resolution; and conclude that remote webcam eye tracking can serve as an affordable and accessible alternative to lab-based infrared eye tracking, even for questions probing the time-course of language processing.</p>
</abstract>
</article-meta>
</front>
<body>
<sec>
<title>1 Introduction</title>
<p>This paper presents a method study of webcam eye tracking as an alternative mode of collecting data in the Visual World Paradigm, a popular experimental paradigm in psycholinguistics. While webcam eye tracking technology has been under development and in use in some form or other for nearly a decade, two confluent factors have prompted several web-based experiment platforms creating eye tracking tools for behavioral research. One, major improvements to its accuracy and accessibility through the open-source browser tool WebGazer; and the other, the urgent need for a remote alternative to lab-based infrared eye tracking following the outbreak of COVID-19. We evaluate one of these tools, developed within the jsPsych library (version 6.3 <xref ref-type="bibr" rid="B19">De Leeuw, 2015</xref>), by replicating one of our own recent Visual World studies with a fully web-based experiment.</p>
<sec>
<title>1.1 Webcam eye tracking</title>
<p>Most eye tracking systems used in behavioural research laboratories are infrared eye trackers: in brief, they project near-infrared light onto the pupils, which creates a corneal reflection (also known as a Purkinje image) that can be used to triangulate the visual angle of gaze. Eye trackers marketed towards scientists are often bundled together with proprietary stimuli presentation and data pre-processing software: though convenient and more user-friendly on the one hand, this can hinder the researcher in customizing the technology and experiment design beyond the options provided by the vendor, or accessing the raw data. Infrared eye trackers are also expensive, costing several thousand dollars at minimum; indeed, before the COVID-19 pandemic, innovation in eye tracking technology was mostly driven by a need to make it cheaper and more portable &#8211; not only for researchers, but for consumer-grade eye tracking applications and devices. As internet speed increased and crowdsourced workers became easily accessible through companies like Qualtrics and Amazon Mechanical Turk, the past decade saw the debut of many different &#8216;neuromarketing&#8217; applications: e.g. Turkergaze (<xref ref-type="bibr" rid="B75">Xu et al., 2015</xref>), GazeParser (<xref ref-type="bibr" rid="B68">Sogo, 2013</xref>), WebGazer (<xref ref-type="bibr" rid="B58">Papoutsaki et al., 2016</xref>), GazeHawk, GazeRecorder, EyesDecide, RealEye, EyeSee, etc. Of these, WebGazer has emerged as the clear favourite for use in browser-based research in cognitive science.</p>
<p>Unlike most other eye tracking tools, WebGazer<xref ref-type="fn" rid="n1">1</xref> maps eye features onto positions on the screen using dynamic, mouseclick-based calibration, taking advantage of the rule of thumb that users navigating a web page will look directly at where they click (<xref ref-type="bibr" rid="B15">Chen et al., 2001</xref>; <xref ref-type="bibr" rid="B33">Hauger et al., 2011</xref>; <xref ref-type="bibr" rid="B35">Huang et al., 2012</xref>). This reliance on natural browsing behaviour makes it better suited to User Interaction research than more &#8216;traditional&#8217; behavioural research paradigms (<xref ref-type="bibr" rid="B57">Papoutsaki et al., 2017</xref>). However, WebGazer has several advantages that make it an attractive tool for behavioural scientists: it is fully integrated in the browser, without requiring users to download software; it computes and outputs gaze data in the form of [<italic>x,y,t</italic>] coordinates on the client browser, without transmitting video data to the experiment server; its design is modular, making it easy to substitute alternatives for the default facial recognition algorithm and ridge regression model; and the fact that it is free, open source, and actively maintained.</p>
<p>Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) were the first to report a method study evaluating the usefulness of WebGazer in cognitive research. They conducted three common eye tracking tasks (fixation, pursuit, and free viewing) with custom-written experiment software integrating WebGazer, testing both in-lab and remote participants. They found an average spatial offset of, respectively, 15% (approx. 4&#176; visual angle) and 18% of screen size, and an average saccade duration of 450 ms and 750 ms,<xref ref-type="fn" rid="n2">2</xref> with significantly more variance in the remote sample. For comparison, a commercial infrared eye tracker sampling at &gt;120 Hz can be expected to record saccade durations of 200 ms or less, with spatial offsets between 0.1&#176; and 0.5&#176; (<xref ref-type="bibr" rid="B26">Ehinger et al., 2019</xref>; <xref ref-type="bibr" rid="B55">Ooms et al., 2015</xref>).</p>
<p>However, despite the noisier, lower-resolution data, they were able to replicate a well-known eye tracking result: namely, that Western participants learning and categorizing human faces pay particular attention to the eye region (in contrast to participants from other cultural backgrounds) (<xref ref-type="bibr" rid="B9">Blais et al., 2008</xref>; and others). In the wake of their cautiously optimistic assessment, a handful of WebGazer-based experiments followed: Federico &amp; Brandimonte (<xref ref-type="bibr" rid="B29">2019</xref>) used WebGazer in a lab setting through a commercial platform and a consumer-grade webcam; e.g. Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>), Degen et al. (<xref ref-type="bibr" rid="B23">2021</xref>), and Madsen et al. (<xref ref-type="bibr" rid="B47">2021</xref>) integrated WebGazer into their own experiment code to run remote eye tracking experiments in the browser. Though their data were encouraging, generally replicating effects found with infrared eye tracking, they were also much noisier due to the differences in computer hardware, operating system, processing capacity, and lighting quality between participants. In addition, programming and hosting these experiments required considerable time, effort, and specialized skills. The incentive for cognitive scientists to invest in webcam eye tracking therefore remained low.</p>
<p>Since the outbreak of the COVID-19 pandemic, which largely precluded in-lab research and infrared eye tracking, several popular behavioural experiment software programs and libraries (at last count: PCIbex, Gorilla, jsPsych, and PsychoPy) have developed webcam eye tracking functionalities, most<xref ref-type="fn" rid="n3">3</xref> of which rely on WebGazer. This, in tandem with a recent proliferation of researcher-friendly web hosting solutions (e.g. JATOS, Pushkin) and companies that combine experiment building graphical user interfaces and web hosting (e.g. Gorilla, Pavlovia, FindingFive), has made conducting webcam eye tracking experiments much more accessible. With the new wealth of possibilities comes the need to map out its caveats and limitations: it is already evident that dependent measures requiring very fine-grained temporal and spatial resolution, such as eye movements during reading, cannot be usefully investigated using webcam eye tracking. But the most fine-grained resolution at which it <italic>can</italic> be useful has not yet been pinned down, especially as the technology improves; and one experimental method where it almost certainly can at least supplement infrared eye tracking, is the Visual World Paradigm (<xref ref-type="bibr" rid="B23">Degen et al., 2021</xref>; <xref ref-type="bibr" rid="B66">Slim &amp; Hartsuiker, 2021b</xref>).</p>
</sec>
<sec>
<title>1.2 The Visual World Paradigm</title>
<p>The Visual World Paradigm is one of the most productive methods in online language processing research, owing to the fact that human visual attention is tightly coupled with linguistic processing (<xref ref-type="bibr" rid="B17">Cooper, 1974</xref>). Given a &#8216;visual world&#8217;, i.e. a display of scenes or objects, and an auditory linguistic stimulus, participants&#8217; eye movements will gravitate towards those parts of the display that are associated in some way with what they hear (<xref ref-type="bibr" rid="B70">Tanenhaus et al., 1995</xref>; <xref ref-type="bibr" rid="B1">Allopenna et al., 1998</xref>; see <xref ref-type="bibr" rid="B28">Falk Huettig &amp; Meyer, 2011</xref> for review). These fixations are closely time-locked to the linguistic stimulus, often occurring before or within 200 ms of the target word&#8217;s offset;<xref ref-type="fn" rid="n4">4</xref> they have also been found to reflect predictive processing, in cases where the selectional restrictions of an earlier word constrain the possible targets in the visual display. In Altmann &amp; Kamide (<xref ref-type="bibr" rid="B2">1999</xref>)&#8217;s eminent example, &#8220;The boy will eat the cake&#8221; triggered looks towards a cake (the only edible object in the display) <italic>before</italic> onset of the noun. There are several possible linking hypotheses for the relationship between eye movements and linguistic processing (see e.g. <xref ref-type="bibr" rid="B28">Falk Huettig &amp; Meyer, 2011</xref> and <xref ref-type="bibr" rid="B49">Magnuson, 2019</xref> for discussion), and the formulation of a model integrating visual processing, linguistic processing, eye movement mechanics and high-level discourse and nonlinguistic cognitive factors is a priority for this paradigm (see e.g. <xref ref-type="bibr" rid="B38">Huettig et al., 2020</xref>; <xref ref-type="bibr" rid="B13">Chabal et al., 2022</xref>; <xref ref-type="bibr" rid="B23">Degen et al., 2021</xref>). For our purposes, it will suffice to say that Visual World Paradigm studies have shown, to quote Magnuson (<xref ref-type="bibr" rid="B49">2019</xref>), that &#8220;listeners are sensitive to every potentially useful (i.e., predictive) constraint that has been tested as early as we can measure.&#8221; (p.134) The constraint that we investigated in the current studies is grammatical aspect: we give a theoretical motivation for this work in the section below, but readers who are interested primarily in the methodological results may take our word for it and proceed to section 2.</p>
</sec>
<sec>
<title>1.3 The original study</title>
<p>The experiment design is drawn from a Visual World eye tracking experiment we developed and conducted in Russian, English, and Spanish, with both adults and children of various ages, between 2018&#8211;2020 (<xref ref-type="bibr" rid="B54">Minor et al., 2022b</xref>). The aim of this study was a cross-linguistic comparison of three typologically different aspectual systems using the same picture stimuli and experiment design: in order to tease out subtle differences in the semantic representation of (in particular) the <italic>perfective</italic> verb forms in these systems, which are often grouped together under the same formal denotation, but are found to carve up narrative time in ways that are not easily captured by offline<xref ref-type="fn" rid="n5">5</xref> judgments and truth conditions alone. The version of this study that we chose to replicate, namely the English, exemplifies a case where online processing data can help illuminate a muddled semantic landscape.</p>
<p>We contrasted the &#8216;imperfective&#8217; English Past Progressive (e.g. <italic>was baking, was painting</italic>) with the &#8216;perfective&#8217; Simple Past (e.g. <italic>baked, painted</italic>). The imperfective/perfective contrast is not binary in English to the degree that it is in, for example, Slavic languages (<xref ref-type="bibr" rid="B32">Gvozdanovi&#263;, 2012</xref>); its grammatical rendering is somewhat lopsided, with the Past Progressive marking imperfective with an inflected <italic>be</italic>-<sc>aux</sc> and a participial verb, and the Simple Past bearing only a tense suffix and no overt aspectual marker. The grammatical, or &#8216;viewpoint&#8217;, aspect of the Past Progressive is non-habitual continuous: it highlights the ongoing part of the event, and does not entail that the result state, or <italic>telos</italic>, of the event is ever reached (<xref ref-type="bibr" rid="B22">de Swart, 2012</xref>).</p>
<p>The interpretation of the Simple Past is less clear-cut: it is generally considered perfective (<xref ref-type="bibr" rid="B72">Van Hout, 2011</xref>), though stative verbs form an exception, and various semanticists recognize that the Simple Past does not always entail the culmination of an event (see e.g. <xref ref-type="bibr" rid="B73">van Hout, 2018</xref>; <xref ref-type="bibr" rid="B52">Martin et al., 2020</xref>; <xref ref-type="bibr" rid="B51">Martin &amp; Demirdache, 2020</xref>). De Swart (<xref ref-type="bibr" rid="B21">1998</xref>) analyses the English Simple Past as aspectually neutral, with the (im)perfectivity of the verb being determined by its Aktionsart. When the Aktionsart is an accomplishment, however (as it is in this study), the Simple Past is interpreted as a perfective &#8211; after all, &#8220;culmination entailments are typically taken to be a diagnostic criterion for defining this aspectual class.&#8221; (<xref ref-type="bibr" rid="B51">Martin &amp; Demirdache, 2020</xref>).</p>
<p>This reading is supported by experimental work: Madden &amp; Zwaan (<xref ref-type="bibr" rid="B46">2003</xref>), whose paper laid the foundation for the stimulus design of our study, found that the Simple Past constrained the mental representation of events. Magliano &amp; Schleich (<xref ref-type="bibr" rid="B48">2000</xref>) found that the mental activation of events decayed faster if they were presented in the Simple Past form; and Bott &amp; Hamm (<xref ref-type="bibr" rid="B10">2014</xref>) found that coercion of a (Simple) Past accomplishment predicate into an activity reading caused processing difficulty in English, but not in German.</p>
<p>However, there is also some evidence hinting that Simple Past accomplishment predicates do not <italic>have</italic> to be perfective: in a pragmatics study contrasting the telicity of particle verbs (e.g. <italic>eat the apple up</italic>) with that of corresponding simplex verbs (<italic>eat the apple</italic>), Jeschull (<xref ref-type="bibr" rid="B39">2007</xref>) found that adults&#8217; preference was at ceiling for a completion interpretation of particle verbs, but at chance for simplex verbs. In a study exploring the perfective interpretations of simple and complex verb forms describing change-of-state events in Hindi and English, Arunachalam &amp; Kothari (<xref ref-type="bibr" rid="B6">2011</xref>) report that English speakers accepted partial-completion interpretations for Simple Past verbs approximately 50% of the time (patterning with the Hindi simple verb form, which does not entail event completion). In short: the jury is still out on the Simple Past, in both the theoretical and the experimental literature.</p>
<p>In order to better understand how the mental representation of accomplishment events is modulated by aspect in real-time, we chose to conduct a Visual World eye tracking study. We presented participants with two pictures of the same event: one in which the event is ongoing, and one where it has been completed. While viewing the pictures (the &#8216;Visual World&#8217;), participants heard a sentence describing the event, in which the grammatical aspect of the verb was manipulated. Participants chose which picture best matches the sentence they heard (the offline result), and their approximate gaze fixations were measured throughout the trial. Two previous studies of this kind, Zhou et al. (<xref ref-type="bibr" rid="B78">2014</xref>) (Mandarin) and Minor et al. (<xref ref-type="bibr" rid="B53">2022a</xref>) (Russian), found that participants reacted to aspectual morphemes by looking towards their corresponding pictures immediately after hearing the morpheme, without waiting to hear all of the verb&#8217;s arguments. In the case of Russian, this looking preference became statistically significant even before verb offset. This method therefore allowed us to see whether a more complex picture of online processing hides behind the varying offline judgments of the perfectivity of the Simple Past.</p>
<p>Our reasons for replicating this particular study were primarily practical: we were able to use the same stimuli and adhere to the original experimental design and data analysis as closely as possible, and conducting it in English allowed easy recruitment from a pool of over 40,000 eligible participants via Prolific.ac. However, the results of this study also made it an attractive candidate for replication: there was a stark and statistically significant difference between the two aspectual conditions, and there was a somewhat unexpected and intriguing <italic>lack</italic> of an effect of aspect in the Simple Past condition, with no detectable preference for either event type in both the online and offline results. Given that all the events were accomplishments, that the Simple Past is commonly analysed as a perfective, and that the design of the experiment, if anything, encouraged participants to interpret the Simple Past in complementary distribution to the Past Progressive, we were surprised by the aspectual &#8216;neutrality&#8217; (or ambivalence) of the Simple Past, and were interested in replicating this finding for its own sake.<xref ref-type="fn" rid="n6">6</xref></p>
<p>Our aim was to assess whether webcam eye tracking data performs well enough, in terms of temporal and spatial accuracy, to serve as a viable alternative to infrared eye tracking for Visual World experiments. In addition, we wanted to build our replication study using &#8216;out-of-box&#8217; and open-source software tools with minimal customization, as a proof of concept that this kind of &#8216;do it yourself&#8217; eye tracking experiment can realistically be built and run by anyone, without extensive programming experience or access to commercial platforms.</p>
<p>In the rest of the paper, we present a detailed Method of both studies, taking the original study as a default and specifying any adaptations made for the web-based version as and where appropriate. We then present the results, followed by a discussion of methodological factors affecting the temporal and spatial resolution of webcam eye tracking data, as well as participant retention and data quality.</p>
</sec>
</sec>
<sec sec-type="methods">
<title>2 Methods</title>
<p>All data, analysis scripts, materials, and experiment software are available at the Open Science Framework (<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/m395q/">https://osf.io/m395q/</ext-link>).</p>
<sec>
<title>2.1 Materials and design</title>
<p>The experiment included 24 test trials and 24 filler trials, all consisting of a visual display and an audio stimulus. In the test trials, two pictures were presented side-by-side on a screen, representing two stages of the same event: one where the event is ongoing (cf. <xref ref-type="fig" rid="F1">Figure 1(a)</xref>, and one in which it is completed (cf. <xref ref-type="fig" rid="F1">Figure 1(b)</xref>).</p>
<fig id="F1">
<label>Figure 1</label>
<caption>
<p>Visual display: &#8216;An old man demolishing a house&#8217;.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g1.png"/>
</fig>
<p>The audio stimuli included a preamble and a target sentence, and were recorded by a female native speaker of British English in a sound-proof booth. The preamble was a short sentence in the past tense meant to set up a narrative context for the subsequent target sentence (e.g. <italic>It was a crisp winter morning, There were many people shopping in town</italic>, etc.). The target sentence was a transitive clause containing a subject NP (<italic>grandma, grandpa, a girl</italic> or <italic>a boy</italic>), a past-tense verb and an object NP consisting of an adjective and a noun (e.g. <italic>a pretty flower, a new shirt</italic>, etc.). The experimental manipulation was grammatical aspect, and as such there were two versions of each target sentence: one in which the verb had the Past Progressive form, and one in which the verb had the Simple Past form, cf. example 1.</p>
<list list-type="gloss">
<list-item>
<list list-type="wordfirst">
<list-item><p>(1)</p></list-item>
</list>
<list list-type="wordfirst">
<list-item><p>a.</p></list-item>
</list>
</list-item>
<list-item>
<list list-type="sentence-gloss">
<list-item>
<list list-type="final-sentence">
<list-item><p>Grandpa was demolishing an old house.</p></list-item>
</list>
</list-item>
</list>
</list-item>
</list>
<list list-type="gloss">
<list-item>
<list list-type="wordfirst">
<list-item><p>&#160;</p></list-item>
</list>
<list list-type="wordfirst">
<list-item><p>b.</p></list-item>
</list>
</list-item>
<list-item>
<list list-type="sentence-gloss">
<list-item>
<list list-type="final-sentence">
<list-item><p>Grandma demolished an old house.</p></list-item>
</list>
</list-item>
</list>
</list-item>
</list>
<p>All items in the experiment involved telic accomplishments (cf. <xref ref-type="bibr" rid="B74">Vendler, 1967</xref>; <xref ref-type="bibr" rid="B25">Dowty, 1979</xref>), i.e. events with a process stage and a well-defined result stage. The design of the experiment relies on the assumption that these two stages can be mentally represented as &#8216;snapshots&#8217;, and are quickly and easily recognized as such by participants.</p>
<p>In the filler items, the visual display featured two pictures of <italic>different</italic> events (e.g. <italic>Grandpa chopped down a tree</italic> vs <italic>Grandpa blew out a big candle</italic>), but with the same subject, in various combinations of ongoing event versus completed event. The preamble and target sentences in the fillers were similar to those in the test items. To counteract the experimental bias that would arise if hearing the auxiliary <italic>be</italic> would always uniquely predict the ongoing event picture, half of the filler items included a construction with the auxiliary <italic>be</italic> that described a completed event (e.g. <italic>Grandma was successful in cracking open the nut, The boy was done with taking apart the wooden stool</italic>); the other half of the fillers included a past tense verb describing the initial or intermediate stages of an event (e.g. <italic>The girl began to drink a glass of milk, Grandpa occupied himself in the strawberry patch</italic>).</p>
<p>Two lists were created such that each test item appeared once in each list: in the Past Progressive form in one list, and the Simple Past form in the other. The position of the ongoing/completed event picture and the target picture was balanced across lists, and participants were randomly assigned to one of the two lists. Each list began with a filler item in order to acclimate the participants to the procedure; there were no practice items.</p>
<sec>
<title>2.1.1 Web-based replication</title>
<p>The materials and design of the web-based replication were kept identical where possible, and minimally adapted to the constraints of the browser environment where necessary. To ensure that the pictures would have the same relative size and placement on the screen regardless of the display size,<xref ref-type="fn" rid="n7">7</xref> the browser window was divided into a grid with 12 vertical columns (using the Bootstrap CSS framework), wherein each picture was centered in a container 5 columns wide, with a 2-column neutral space between them. The height and width in pixels of each container was recorded in the data output.</p>
<p>In the original infrared eye tracking study, the trial order was fixed to avoid any clustering of trials that could create a habituation effect; never more than two consecutive trials with the same target event type, or target picture presentation side, as well as alternating test and filler items. Because we expected the web-based replication to have a longer duration, and so possibly a stronger effect of boredom or habituation amid the overall higher level of noise, we pseudo-randomized the trials by ordering them in a list such that they formed blocks of four items that met all the balancing criteria we used for the original study. The order between these 8 blocks (but not within blocks) was randomized by participant.</p>
<p>The experiment was programmed using jsPsych (<xref ref-type="bibr" rid="B19">De Leeuw, 2015</xref>), which debuted its WebGazer-based eye tracking functionality with the release of version 6.3, in February 2021. The jsPsych framework organizes the various parts and functionalities of a behavioural experiment into modular scripts, or &#8216;plugins&#8217;; the eye tracking functionality is designed as an &#8216;extension&#8217; that can be added to and run in the background of any other plugin. Additionally, the package includes a plugin that initializes the webcam and locates the face and eyes in the center of the video feed; a calibration plugin, which trains a regression model to predict gaze location based on eye position; and a validation plugin, with which the accuracy of the prediction model is assessed. jsPsych version 6.3.1, released in April, also includes a forked version of WebGazer which was adapted to improve temporal resolution (see section 4.1.3 for a more in-depth discussion). Wherever possible, we used jsPsych&#8217;s plugins and API &#8216;out-of-box&#8217;; we lightly adapted the <monospace>audio-button-response</monospace> plugin to program the trials themselves, but were otherwise able to use the tools provided by jsPsych without customization.</p>
<p>The study was hosted on a JATOS<xref ref-type="fn" rid="n8">8</xref> server owned by UiT &#8211; the Arctic university of Norway.</p>
</sec>
</sec>
<sec>
<title>2.2 Procedure</title>
<sec>
<title>2.2.1 Original infrared eye tracking study</title>
<p>Instructions were delivered verbally by the experimenter. Participants were calibrated once at the start of the experiment, using a 9-point calibration grid, which was validated by fixating a randomly presented succession of points on that grid. Each trial had a preamble phase and a target phase. During the preamble phase participants were shown a picture of a smiley face at the center of the screen and heard the preamble sentence. After that the trial proceeded to the target phase where two pictures were presented side by side on the screen. After a 500 ms preview the participants heard the target sentence, and chose one of the pictures by raising the corresponding hand (left or right). Participants&#8217; eye movements were recorded using an SMI RED500 eye tracker with an integrated 22-inch monitor, at a sampling rate of 120 Hz; offline responses were recorded manually by the experimenter. The experiment lasted approximately 6 minutes.</p>
</sec>
<sec>
<title>2.2.2 Web-based replication</title>
<p>After completing the demographic survey, granting permission to access the webcam and passing a browser and equipment check, participants were encouraged to ensure that they would be undisturbed for at least 15 minutes; their face would be brightly and evenly lit; and that they were sitting comfortably (see <xref ref-type="fig" rid="F2">Figure 2</xref>).<xref ref-type="fn" rid="n9">9</xref> They were not instructed to sit at a particular distance from their screen (though see section 3.2.2 for a possible approach to managing this in future work.)</p>
<fig id="F2">
<label>Figure 2</label>
<caption>
<p>Instructions for posture and head positioning (the first page in the study).</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g2.png"/>
</fig>
<p>Participants were then directed through a calibration phase. 15 points, 30px wide, were presented consecutively and in random order across the entire screen (between 10&#8211;90% of the screen dimensions). Participants were instructed to look at and click on each point; when clicked, the point would vanish and the next would appear. The validation of this calibration phase consisted of two consecutively presented points, 30px wide, centered approximately where the trial pictures would later be located on the screen. Each point was visible for 3 seconds: participants were instructed to simply look at the points without moving their head. Gaze predictions were generated starting 500 ms after each point appeared, to allow the eyes to saccade. For the participant to &#8216;pass&#8217; the calibration, &gt;50% of gaze predictions for each point had to fall within a 200px tolerance radius of that point.</p>
<p>Following the validation, participants received visual feedback on their performance: the tolerance radius around each point was made visible, and the raw gaze data of the validation was plotted onto the screen as green (&#8216;hit&#8217;) and red (&#8216;miss&#8217;) dots.<xref ref-type="fn" rid="n10">10</xref> If the 50% threshold was reached for both validation points, the participant could proceed to the experiment; if not, they were looped back to the start of the calibration phase. In addition, the sampling rate (i.e. the rate at which WebGazer generates gaze predictions) had to be at least 5 samples per second. If calibration was not successful within 5 attempts, the study was aborted.</p>
<p>This calibration procedure was repeated another three times throughout the experiment (once every 12 trials) to compensate for small head movements and resultant decay of the accuracy of the gaze prediction model (see section 3.2 for further discussion).</p>
<p>Following calibration, the experiment proceeded exactly as it had in the original study, with the exception that participants had to interact with the web page to advance to the next trial. The preamble phase of each trial began with a green fixation point at center screen, which had to be clicked to play the preamble audio. When the audio finished playing, the trial advanced automatically to the two-picture display, the target sentence started playing, and the participants&#8217; cursor was hidden. When the target audio ended, the cursor reappeared, and the participant had to select one of the pictures by clicking on it, which triggered the start of the next trial. Participants&#8217; eye movements during the target phase of the trial, and their picture choice, were recorded. The study duration was 15.32 minutes on average.</p>
</sec>
</sec>
<sec>
<title>2.3 Participants</title>
<sec>
<title>2.3.1 Original infrared eye tracking study</title>
<p>35 adult monolingual English speakers were tested in Edinburgh (Scotland), in December 2019. A further 31 adult monolingual English speakers were recruited and tested in Norway (Trondheim and Troms&#248;) in September 2020, giving a total of 66 participants. All the participants tested in Norway had spent less than 5 years in Norway prior to the experiment, and attested to having only elementary conversational proficiency in Norwegian.</p>
<p>All participants had normal or corrected-to-normal vision. Written consent was obtained from all the participants prior to testing; as compensation, the participants tested in Edinburgh received &#163;5, and the participants tested in Norway received a cinema voucher or a gift card worth 120 Norwegian kroner (&#126;$13.50).</p>
</sec>
<sec>
<title>2.3.2 Web-based replication</title>
<p>124 adult monolingual English speakers were recruited via Prolific.ac. The sample size was determined on the basis of the results of our pilot studies.<xref ref-type="fn" rid="n11">11</xref> Several filters were applied on Prolific, to restrict who could access the study: participants had to be English speaking monolinguals who had spent most of their time before turning 18 in the United Kingdom; and they could only participate with a desktop computer (as opposed to a tablet or phone<xref ref-type="fn" rid="n12">12</xref>) and a webcam.</p>
<p>Consent was obtained electronically by clicking a button labeled &#8220;I agree and Start&#8221; at the bottom of a reloadable information and consent page. Participants were paid &#163;4 for an estimated study duration of 20 minutes (the actual study duration was &#126;15 minutes) if they successfully completed the entire study; they were paid &#163;2 if they completed part of the study after a successful initial calibration, but were barred from finishing it after a failed recalibration. If they were unable to calibrate and start the study, they received no compensation. Participants were informed of this conditional payment structure in the information and consent letter. In total, 197 people started the experiment; of these, 39 (19.8%) dropped out before calibrating, usually because their browser, webcam, or audio output did not work. 16 (8.1%) dropped out after failing their initial calibration, 8 (4%) dropped out after failing a recalibration, and 124 (62.9%) successfully completed the experiment. The remainder refreshed the web page during the experiment, which blocked them from further participation.</p>
</sec>
</sec>
<sec>
<title>2.4 Trial exclusions and data preparation</title>
<p>We inspected participants&#8217; accuracy in their picture choices in the filler trials to determine whether they merited exclusion. All 124 participants were &gt;85% accurate, and so none were excluded. As in the original study, we excluded trials with &gt;50% track loss (infrared version: 2 trials, 0.13% data loss): in the replication, this meant trials with &gt;50% gaze predictions located outside the participant&#8217;s screen dimensions (47 trials, 1.6% data loss).</p>
</sec>
<sec>
<title>2.5 Analysis</title>
<p>We coded the selection of the Ongoing Event picture in the Progressive condition and the Completed Event picture in the Simple Past condition as &#8216;target&#8217;, and the opposite choice as &#8216;competitor&#8217;. To test whether the proportion of &#8216;target&#8217; picture selections was significantly above chance in either condition, we fit two mixed effects logistic regressions (using the R package lme4(<xref ref-type="bibr" rid="B8">Bates et al., 2014</xref>; <xref ref-type="bibr" rid="B71">R Core Team, 2019</xref>) estimating the log-odds of a target response in the Past Progressive and Simple Past trials, with random intercepts for participants and items.<xref ref-type="fn" rid="n13">13</xref></p>
<p>To identify the time windows in which the probability of fixating on the target picture was significantly above chance, we performed a cluster-based permutation analysis for each condition (see e.g. <xref ref-type="bibr" rid="B36">Huang &amp; Snedeker, 2020</xref>; <xref ref-type="bibr" rid="B76">Yang et al., 2020</xref>). One advantage of this analysis over the more common growth curve analysis is that it gives an estimate of the time window (the titular &#8216;cluster&#8217; of time bins) in which an effect is significant, without the researcher pre-defining which time windows to analyze (which can seriously affect the statistical outcomes; see e.g. <xref ref-type="bibr" rid="B59">Peelle &amp; Van Engen, 2021</xref>; <xref ref-type="bibr" rid="B36">Huang &amp; Snedeker, 2020</xref>).</p>
<p>We selected the data starting from verb onset to 2000 ms after verb onset; binned the data into 50 ms time bins; calculated the proportion of fixations on the target picture in each time bin; and then binarized that data by rounding up to 1 or down to 0. Next, we fit a mixed effects logistic regression for each time bin, to estimate the log-odds of fixations on the target picture. Items and participants were included as random intercepts, and an intercept term was included to represent the difference between the log-odd of fixations on the target picture and 0, which corresponds to chance (0.5) probability. Next, we clustered together consecutive time bins where the probability of fixating on the target picture was significant at &#945; = 0.08, on the assumption that these all exhibit the same effect;<xref ref-type="fn" rid="n14">14</xref> and summed up their <italic>z</italic>-values to create a sum statistic for each cluster. Finally, we estimated how likely these clusters would be to occur by chance, under the null hypothesis that the probability of fixating on the target versus the competitor picture was at chance. We did this by creating a permutation distribution, whereby we randomly permuted the picture labels (target vs competitor) by participant and then repeated the regression and the clustering steps. This procedure was repeated 1000 times, yielding a distribution of sum statistics against which the statistics of the original clusters were compared. Clusters with <italic>p</italic> &lt; 0.05 were considered significant.</p>
</sec>
<sec>
<title>2.6 Results</title>
<p><xref ref-type="table" rid="T1">Table 1</xref> shows the offline responses in both the original infrared eye tracking study and the WebGazer replication. In both studies, the preference for the &#8216;target&#8217; Ongoing Event picture in the Past Progressive condition was almost at-ceiling, but the preference for either picture in the Simple Past condition hovers around chance level. The log-odds of selecting the target picture were significantly higher than 0 in the Past Progressive condition in both the original study (intercept <italic>B</italic> = 6.24, <italic>SE</italic> = 0.77, <italic>Z</italic> = 8.09, <italic>p</italic> &lt; 0.001) and the replication (intercept <italic>B</italic> = 5.25, <italic>SE</italic> = 0.59, <italic>Z</italic> = 8.86, <italic>p</italic> &lt; 0.001). In the Simple Past condition, the log-odds were not significant in the original study (intercept <italic>B</italic> = 0.26, <italic>SE</italic> = 0.3, <italic>Z</italic> = 0.86, <italic>p</italic> = 0.39); and although the proportion of selections of the competitor picture in this condition increased by 10 percentage points in the replication, the log-odds of selecting the target picture still did not significantly deviate from 0 (intercept <italic>B</italic> = &#8211;0.46, <italic>SE</italic> = 0.32, <italic>Z</italic> = &#8211;1.41, <italic>p</italic> = 0.16).</p>
<table-wrap id="T1">
<label>Table 1</label>
<caption>
<p>Offline responses in the original study and the replication.</p>
</caption>
<table>
<thead>
<tr>
<td colspan="5"><hr/></td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top" colspan="2"><bold>Infrared</bold></td>
<td align="left" valign="top" colspan="2"><bold>WebGazer</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td colspan="5"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">Event type</td>
<td align="left" valign="top">Prog</td>
<td align="left" valign="top">SPast</td>
<td align="left" valign="top">Prog</td>
<td align="left" valign="top">SPast</td>
</tr>
<tr>
<td colspan="5"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">Ongoing Event</td>
<td align="left" valign="top">95%</td>
<td align="left" valign="top">46%</td>
<td align="left" valign="top">98%</td>
<td align="left" valign="top">56%</td>
</tr>
<tr>
<td colspan="5"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">Completed Event</td>
<td align="left" valign="top">5%</td>
<td align="left" valign="top">54%</td>
<td align="left" valign="top">2%</td>
<td align="left" valign="top">44%</td>
</tr>
<tr>
<td colspan="5"><hr/></td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> presents the online results from both the original study (a, b) and the WebGazer replication (c, d), starting from, and ending 2000 ms after, lexical verb onset. The dashed vertical lines mark average lexical verb offset, and the shading in (a) and (c) represents the time windows in which the probability of looking towards the target picture was significantly above chance. In these graphs, looks that fell outside either picture were filtered out: in the original study, that constituted 6.36% of gaze data, but in the replication, it was 27.95% (see section 4.2 for discussion).</p>
<fig id="F3">
<label>Figure 3</label>
<caption>
<p>Proportion of looks to the target picture in the Progressive condition ((a) and (c)), and in the Simple Past condition ((b) and (d)). Data in <bold>(a)</bold> and <bold>(b)</bold> were collected with an infrared eye tracker, data in <bold>(c)</bold> and <bold>(d)</bold> using WebGazer. The colored ribbons around the graph lines represent the standard error of the mean. Grey shading represents the time bins where probability of looks to the target picture was significantly above chance. The dashed vertical lines mark average lexical verb offset (559 ms in the Simple Past condition, 674 ms in the Past Progressive condition).</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g3.png"/>
</fig>
<p>We used a cluster-based permutation analysis (&#945; = 0.08) to identify clusters of 50 ms time bins where the probability of fixating on the target picture was significantly above chance. In the original study, this analysis revealed one cluster from 500 to 2000 ms after lexical verb onset (sum <italic>Z</italic> = 103.57, <italic>p</italic> &lt; 0.001) in the Past Progressive condition; and no clusters in the Simple Past condition. In the replication study, the analysis identified one cluster from 550 to 2000 ms (sum <italic>Z</italic> = 133.8, <italic>p</italic> &lt; 0.001) in the Past Progressive condition, and no clusters in the Simple Past condition.</p>
</sec>
</sec>
<sec>
<title>3 Discussion</title>
<p>We replicated a Visual World eye tracking study using browser-based experiment software and webcam eye tracking tools, and remote participants. We were able to fully replicate the results of the original study, including the approximate onset of the time window in which the probability of fixating on the target picture was significant (which was one time bin, or 50 ms, later in the replication). This is a marked improvement on the outcomes of earlier WebGazer replications of eye tracking tasks. In the following sections, we will try to contextualize and account for this improvement, discussing technical factors (particularly spatial and temporal resolution) on the one hand, and methodological factors affecting participant retention and overall data quality on the other.</p>
<sec>
<title>3.1 Temporal resolution</title>
<p>Moreso than large spatial offsets (of which the Visual World Paradigm is much more forgiving than, say, eye tracking during reading), the primary concern with webcam eye tracking has been its low sampling rates and variable inter-sampling intervals &#8211; in other words, its poor temporal resolution. Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) noted a higher temporal error when data was collected remotely, on participants&#8217; own laptops and browsers where processing load and hardware performance could not be controlled for. In their WebGazer replications of Visual World studies, Slim &amp; Hartsuiker (<xref ref-type="bibr" rid="B65">2021a</xref>) and Degen et al. (<xref ref-type="bibr" rid="B23">2021</xref>) found 300&#8211;700 ms delays in the onset of the replicated effect, which they reasonably concluded would have to disqualify this eye tracking technique for use in any time-sensitive experiments.</p>
<p>There are methodological caveats to each of these studies that may, to some extent, account for their sluggish effect onsets &#8211; Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) and Calabrich et al. (<xref ref-type="bibr" rid="B12">2021</xref>) recalibrated multiple times but only analysed the data of 28 and 14 participants respectively, Slim &amp; Hartsuiker (2021) and Degen et al. (<xref ref-type="bibr" rid="B23">2021</xref>) had large datasets but no recalibrations, and so on. However, we think (and several of the aforementioned authors have indeed also speculated) that the biggest source of temporal noise in the data of these studies may have been courtesy of WebGazer itself. In their replication of a decision-making task with eye tracking (<xref ref-type="bibr" rid="B43">Krajbich et al., 2010</xref>), Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>) found that as processing demands on the participant&#8217;s browser and hardware increased over the course of the experiment, the time interval between gaze predictions increased dramatically, peaking at 972 ms (<italic>SD</italic> = 107 ms). They made an adjustment to the WebGazer software itself, whereby the process that generated gaze predictions was decoupled from the main process that updated with every new animation frame &#8211; a process that is highly vulnerable to timing delays when the browser has to juggle several intensive tasks. With this adjustment, they were able to achieve much higher and more stable sampling rates. Shortly after Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>)&#8217;s results became available, an update to jsPsych was released (version 6.3.1, April 10 2021) which included a forked and modified version of WebGazer: like Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>), the developers had found that the method WebGazer relies on to generate gaze predictions created a processing bottleneck that caused serious temporal errors, and adjusted the code to resolve the problem.<xref ref-type="fn" rid="n15">15</xref></p>
<p>As we had been running pilots of this study using custom eye tracking plugins written in the jsPsych framework, we were able to switch to jsPsych v6.3.1 immediately after it came out. Like the other replication studies, the experimental effects in those pilots resembled those of the original, but slower, weaker, and noisier. The temporal resolution of our data and the onset of our experimental effect improved dramatically as a result of implementing the experiment in jsPsych v6.3.1<xref ref-type="fn" rid="n16">16</xref>: due to several minor methodological improvements, but mostly, we expect, due to the modified WebGazer code.</p>
<p>By filtering out participants with a very low sampling rate (&lt;5 Hz), and relying on jsPsych&#8217;s version of WebGazer, our participants had an average sampling rate of 20.73 Hz (<italic>SD</italic> = 8.99): about one gaze prediction per 48 ms. Though, as can be seen in <xref ref-type="fig" rid="F4">Figure 4</xref>, the spread of our participants&#8217; sampling rates spans 5 to 45 Hz; and as a result, the number of data points per 50 ms time bin oscillates by as many as 2000 samples. (This oscillation could conceivably be why the significant effect window in <xref ref-type="fig" rid="F3">Figure 3(c)</xref>, as identified by the cluster-based permutation analysis, starts 50 ms later compared to the infrared study!)</p>
<fig id="F4">
<label>Figure 4</label>
<caption>
<p>Webcam study gaze sampling rates. a) Histogram and density plot of participants&#8217; mean sampling rates. The red vertical line represents the grand average sampling rate: 20.73 Hz (<italic>SD</italic> = 8.99). b) Total number of gaze samples per time bin. Due to participants&#8217; varying (but consistent) sampling rates, the number of gaze samples oscillates between time bins.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g4.png"/>
</fig>
<p>In webcam eye tracking, the sampling rate is effectively limited by the frames-per-second (fps) rate of the webcam &#8211; that is, WebGazer <italic>can</italic> generate predictions at a higher rate, but they may not reflect a &#8216;real&#8217; observation of the eyes. Most consumer-grade webcams sample at 15 to 30 fps (though more expensive ones can go up to 60 fps); the real-time sampling rate is, however, affected by the processing load of the participant&#8217;s device, so an actual fps and WebGazer sampling rate of 60 Hz is unlikely to occur. Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) reported mean sampling rates of 18.71 fps (<italic>SD</italic> = 1.44) in their in-lab dataset, and 14.04 fps (<italic>SD</italic> = 6.68) in their remotely collected dataset; Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>), using their modified version of WebGazer, report an average of 24.85 ms between gaze predictions (<italic>SD</italic> = 12.08), which converts to a 40.2 Hz sampling rate.</p>
<p>Given that infrared eye tracking systems are usually sampling at anywhere between 100&#8211;500 Hz, is this sampling rate sufficient? For a Visual World study, where the measure of interest is usually the proportion of <italic>fixations</italic> in a particular time window rather than saccadic eye movements, it appears that it is. The added value of very high sampling rates might even be doubtful in this paradigm: a fixation lasts 100&#8211;300 ms on average and even an express saccade will take at least 80 ms to launch, so how high a temporal resolution is really necessary? For instance, Dalmaijer (<xref ref-type="bibr" rid="B18">2014</xref>) conducted a method study with the EyeTribe eye tracker, which sampled at max. 60 Hz. He concluded that a 60 Hz sampling rate was good enough for research questions centered around fixation data. Ouzts &amp; Duchowski (<xref ref-type="bibr" rid="B56">2012</xref>) compared two eye tracking datasets with different sampling rates, and recommend downsampling to the lower rate rather than upsampling to the higher rate (as is common practice); higher does not always equal better, especially if it means padding the data by splicing data points. Andersson et al., (<xref ref-type="bibr" rid="B3">2010</xref>) simulated various sampling rates in experiments with varying demands for temporal precision, and show that error resulting from lower sampling rates can be mitigated with higher power.</p>
<sec>
<title>3.1.1 Temporal precision</title>
<p>Beyond WebGazer&#8217;s temporal resolution, there is the timing precision of the experiment itself to consider.</p>
<p>The <italic>accuracy</italic> of the software&#8217;s timing performance is not a grave concern for studies where the measure of interest is the difference between two or more conditions, such as the one presented here: any constant timing offset, or lag (usually arising from hardware characteristics) is canceled out. The <italic>precision</italic>, or variable error of the timing offset, does need to be accounted for. de Leeuw &amp; Motz (<xref ref-type="bibr" rid="B20">2016</xref>), comparing participant response times in a Javascript versus a Psychophysics Toolbox application, note that where there is a noticeable difference, it is mostly an increase in lag that&#8217;s relatively consistent across trials, and not an increase in variability between trials. In other words: poorer accuracy, but not poorer precision. Likewise, Slote &amp; Strand (<xref ref-type="bibr" rid="B67">2016</xref>) found that variation in the measurement error of audio stimulus onset in a Javascript experiment could be limited to less than 5 ms, even when processing load was high, by using WebAudio API to schedule audio presentation. Bridges et al. (<xref ref-type="bibr" rid="B11">2020</xref>) compared the timing performance of several popular behavioral science software packages, both in-lab and web-based, on a range of operating systems and browsers. They found that jsPsych showed an inter-trial variability of precision in the range of 3.2&#8211;8.4 ms in all browser/operating system configurations. More generally, they note that a problem which seems to affect all online software packages to various degrees is the exact synchronization of audio and visual stimuli, a task for which Javascript is not ideal (cf. <xref ref-type="bibr" rid="B4">Anwyl-Irvine et al., 2021a</xref> for additional data).</p>
<p>In future work, we may be able to use the participant sound card&#8217;s own estimation of the audio output latency to better understand the temporal accuracy and precision of this experimental set-up, but with a within-subjects experimental design and a statistical model taking random participant effects into account, timing offsets and variation of this size should not hinder a clear interpretation of the data.</p>
</sec>
</sec>
<sec>
<title>3.2 Spatial resolution</title>
<p>The other question to address in evaluating the performance of this web-based method is how accurately it captures gaze location. Though our results indicate that WebGazer&#8217;s spatial resolution is good enough to capture the expected effect in a two-picture paradigm, the fact that 28% of the replication data were looks outside either of the pictures (vs 6.3% in the original study) indicates that WebGazer remains a blunter instrument than infrared. This requires a balancing act: in order to minimize the risk of looks being misclassified, we had separated the two pictures by 20% of screen width, with the result that the majority of non-picture looks concentrated in this area (see <xref ref-type="fig" rid="F5">Figure 5(b)</xref>).</p>
<fig id="F5">
<label>Figure 5</label>
<caption>
<p>Density plots for <bold>(a)</bold> looks towards either Region Of Interest, and <bold>(b)</bold> looks outside either ROI (webcam study). Gaze and picture placement coordinates were computed as a percentage of the screen width and height. Two participants (#44 and #95) have been excluded from these graphs because their relative picture placements were not aligned with the others, possibly because they exited fullscreen mode and adjusted their browser window dimensions. Graphs <bold>(c)</bold> and <bold>(d)</bold> plot looks towards the target picture (red), the competitor picture (blue), the center column of the screen (yellow), and the remaining edges of the screen. The vertical dotted lines mark, from left to right: audio onset, verb onset, verb offset (in (c) only), and audio offset.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g5.png"/>
</fig>
<p>It could be (though this is pure speculation) that, because our participants&#8217; average screen size was much smaller than the display of our infrared eye tracker, they used more of their peripheral vision (which can and does contribute to object recognition and scene perception; see e.g. (<xref ref-type="bibr" rid="B60">Rosenholtz, 2016</xref>)) to perceive the pictures, resulting in more looks at center screen. With <xref ref-type="fig" rid="F5">Figures 5(c)</xref> and <xref ref-type="fig" rid="F5">5(d)</xref>, we can rule out that the high density of looks towards the center is driven by the reappearance of the cursor at audio offset: the proportion of looks to the center is high at the start of the trial (as expected), then drops sharply as soon as the audio stimulus begins, and remains at just over 20%.<xref ref-type="fn" rid="n17">17</xref></p>
<p>What <italic>is</italic> evident is that spatial resolution is not equal everywhere: Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>), Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>) and Slim &amp; Hartsuiker (<xref ref-type="bibr" rid="B66">2021b</xref>) found that fixation targets near the corners of the screen had significantly higher gaze offsets, and Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) report that gaze predictions for targets near the bottom of the screen have considerable offsets towards the top. This may be due to interference from eyelids or eyelashes, or simply a consequence of the positioning of the webcam (generally at the top of the screen).</p>
<p>Despite these caveats, replications of four-picture Visual World studies (<xref ref-type="bibr" rid="B23">Degen et al., 2021</xref>; <xref ref-type="bibr" rid="B66">Slim &amp; Hartsuiker, 2021b</xref>) so far indicate that WebGazer&#8217;s spatial resolution can accommodate a more crowded display &#8211; but this primarily depends on the quality of the calibration.</p>
<sec>
<title>3.2.1 Calibration</title>
<p>In conventional infrared eye tracking experiments, eye movement is tracked by reflecting a near-infrared light beam off the eye, and measuring the distance between the resulting glint (a.k.a. first Purkinje image) and the center of the pupil. This method is so accurate that for most experiments, the tracker is calibrated only once, at the start; provided the participant does not move and ambient light conditions remain relatively stable, recalibrations are usually not necessary. In eye tracking using visible light spectrum cameras, however, gaze prediction is inevitably less accurate: WebGazer does it by isolating the webcam image of the eyes as detected by a facial features recognition algorithm, reducing it to a 120-pixel grayscale eye feature vector, and supplying that to the gaze prediction model. This approach is more vulnerable to variable or uneven lighting, small head movements, etc., and so the accuracy of the gaze prediction model can be expected to decay significantly over the course of the experiment &#8211; see e.g. <xref ref-type="bibr" rid="B23">Degen et al., 2021</xref>&#8217;s webcam replication study, in which they did not recalibrate at any point during 54 trials. At the other extreme sit Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>), who recalibrated their participants before every block of trials, leading to about half of their study&#8217;s duration (<italic>M</italic> = 43.54 minutes) being spent on calibrating; they noted this was a somewhat arbitrary choice which seemed to wear out their participants, and marked the issue of how often to recalibrate as an important one to answer in future work.</p>
<p>Rather than set a fixed number of recalibrations every <italic>n</italic> trials, Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>) opted for conditional recalibrations: every 10 trials, their participants would see three validation dots (each visible for 2 seconds). If participants fell below the &#8216;hit&#8217; threshold (70% of gaze predictions within 130px of the validation dot) for four dots in two validations, they would recalibrate. Analysing the &#8216;hit&#8217; ratios of their validation trials, they found that the ratio dropped right after calibration, but declined very slowly at every successive validation trial.</p>
<p>In order to understand the rate of calibration accuracy decay in our own experiment design, and to determine the number of recalibrations needed for our replication experiment, we conducted a pilot wherein participants were calibrated only once at the start, and the calibration accuracy was measured with a validation after every second trial. <xref ref-type="fig" rid="F6">Figure 6</xref> presents the results of that pilot: the accuracy drops off immediately after calibration, and continues to decay quite rapidly thereafter.</p>
<fig id="F6">
<label>Figure 6</label>
<caption>
<p>Decay of the accuracy of the initial calibration (webcam study). The shaded ribbons represent the standard error; in this pilot (as in the replication study), the presentation order of the validation points was not randomized, and left was always presented first.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g6.png"/>
</fig>
<p>Wanting to mitigate this decay, but also to avoid exhausting our participants with frequent recalibrations, we chose to recalibrate 3 times, or once every 12 trials. Because it appears that the rate of decay varies by experiment design, we would recommend piloting any webcam eye tracking experiment with a similar procedure, to determine the optimal number of recalibrations.</p>
<p>It is worth noting that Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>)&#8217;s approach to tracking calibration accuracy decay was quite different from ours, and placed more performance demands on their participants: the consequence of too many validation &#8216;misses&#8217; was a recalibration, and validation success or failure was communicated through colour (the validation point turning green for a &#8216;hit&#8217;, and red for a &#8216;miss&#8217;). In our pilot, participants received no feedback on their performance during inter-trial validations, and experienced no consequences for slacking off. This could at least partially explain the steeper drop-off in calibration accuracy in our pilot. Having now established, through the collective effort of the various method studies cited here as well as our own, that <italic>technologically</italic> WebGazer can achieve the necessary spatial accuracy for Visual World studies, the development of <italic>behavioural</italic> best practices for calibration and validation during webcam eye tracking experiments would be a useful focus for future work.</p>
</sec>
<sec>
<title>3.2.2 Screen size and relative stimulus size</title>
<p>It is worth revisiting an experimental design flaw noted in footnote 6 (section 2.1.1): participants&#8217; screen size was not controlled for, and because we sized the screen contents relative to the browser window dimensions (in the case of the picture stimuli) or in pixels (in the case of the calibration points), the absolute size of the screen contents also varied between participants. We cannot guess at how great this variation is: because of a bug in our experiment software, we unfortunately did not collect accurate information about whether participants used a laptop or a PC with an external monitor. In a subsequent, near-identical study (wherein this bug was fixed), we found that 95% of participants used a laptop, giving reasonable hope that the size variation is modest.</p>
<p>How to address this problem in future work? In the case of the calibration points: thus far, the jsPsych calibration and validation plugins only accept number of pixels as a measure of point size. (Likewise the size of the tolerance radius around validation points.) While that means that the absolute size of the points will differ depending on screen resolution, the question is whether sizing them by some measure other than pixels will help. The eye feature vector constructed by WebGazer to track gaze is, after all, also measured in pixels. However, it should be possible to adjust the plugins to allow the size of the points (and of the tolerance radius around them) to be computed as a percentage of screen size, or some other bespoke measure.</p>
<p>Stimulus size can be more easily standardized. When we ran this study, this solution was not yet available, but jsPsych has since introduced a plugin for the Virtual Chinrest (based on Li et al. <xref ref-type="bibr" rid="B45">2020</xref>), which can be used to measure the distance between the participant and their screen, as well as standardize the the jsPsych page content to a known physical dimension. This could potentially be a good way to ensure all participants see pictures of the same absolute size, regardless of screen size. Though it does add another &#8216;hoop&#8217; for participants to jump through, an issue we&#8217;ll discuss in the next section.</p>
</sec>
</sec>
<sec>
<title>3.3 Participant retention and data quality</title>
<p>One of the major selling points of webcam eye tracking, as previously stated, is that data collection could potentially be much quicker and more efficient than its lab-based counterpart. Researchers no longer need to invite participants to the lab, to be tested one-by-one, or travel to reach their target demographic; this can also save a lot of money. Given a large remote participant pool (such as Prolific, particularly for English speakers), data collection may be completed within 1&#8211;2 hours as opposed to weeks or months. However, several webcam eye tracking studies (e.g. <xref ref-type="bibr" rid="B5">Anwyl-Irvine et al., 2021b</xref>; <xref ref-type="bibr" rid="B64">Semmelmann &amp; Weigelt, 2018</xref>; <xref ref-type="bibr" rid="B77">Yang &amp; Krajbich, 2021</xref>; and <xref ref-type="bibr" rid="B66">Slim &amp; Hartsuiker, 2021b</xref>) remark on their experiments&#8217; high participant attrition rates as a cause for concern: 62% in Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>), 61% in Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>), and 72% in Slim &amp; Hartsuiker (<xref ref-type="bibr" rid="B66">2021b</xref>), which Anwyl-Irvine et al., (<xref ref-type="bibr" rid="B5">2021b</xref>) cited as motivating their development of MouseView.js.<xref ref-type="fn" rid="n18">18</xref> Not only do high attrition rates undermine the time and cost efficiency of browser-based eye tracking for the researcher, they also suggest that the experiment is too difficult, uncomfortable, and/or long for the average participant &#8211; a problem worth resolving because the remaining sample may be skewed (&#8216;survival bias&#8217;), but also for its own sake. For our experiment, we therefore sought to improve the experience of taking part on the participants&#8217; side, while also filtering out participants with sub-optimal equipment set-ups as early as possible in the study flow. Of our 197 participants who began the study, 73 (37% attrition) did not complete it; if we remove the 39 participants who were prevented from advancing to the initial calibration because of equipment issues, the attrition rate drops to 21.5%. Here we consider a number of factors that we believe impact participant retention and overall data quality.<xref ref-type="fn" rid="n19">19</xref></p>
<p>Webcam eye tracking demands a lot more from a participant than the average survey or even reaction time experiment: for best results, they are asked to rest the computer on a flat surface, assume a posture they can comfortably maintain without moving for several minutes, adjust lighting if necessary, and close processing-heavy apps running in the background. For participants recruited via platforms such as Amazon Mechanical Turk or Prolific, wasted time means lower earnings; with an all-or-nothing renumeration policy (e.g. Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) only paid the participants that completed the entire study, namely $4 for an average study duration of 43.54 minutes), participants will quickly give up if they risk earning nothing after failing a recalibration. By paying our participants well above minimum wage, and by offering 50% payment if participants failed a recalibration, we hoped to convey appreciation for the concerted effort it takes to cooperate with the experiment design, and to incentivize that effort.</p>
<p>Beyond these pragmatic considerations, participants&#8217; tolerance for boredom may be lower. Many of the recent articles, webinars and blog posts reviewing methods and best practices of online research emphasize the limited &#8216;patience time window&#8217; of participants: the consensus, insofar as there is one, seems to be roughly 20 minutes (cf. e.g. <xref ref-type="bibr" rid="B42">Kochari, 2019</xref>; and a recent webinar on web-based eye and mouse tracking by Gorilla.sc). In a survey of 103 Germans, Sauter et al. (<xref ref-type="bibr" rid="B63">2020</xref>) found that 44% would abort an online study paying minimum wage if it took longer than 15 minutes; this figure rose to 79% for 30 minutes. On the other hand, Jun et al. (<xref ref-type="bibr" rid="B41">2017</xref>) and Chandler &amp; Kapelner (<xref ref-type="bibr" rid="B14">2013</xref>) find that if the study is considered interesting or meaningful that may mitigate effort, boredom, and fatigue. Without the performance pressure induced by a lab environment and direct supervision, the onus is on the researcher to design an experiment that is both short and pleasant to interact with.</p>
<p>In that regard, the amount of time and effort spent (re)calibrating is probably where there is most room for improvement. One perk of this process is that passing a calibration accuracy threshold amounts to a built-in gate-keeping mechanism: few bad faith participants will struggle through repeated calibrations only to deliberately ignore the instructions for the experimental trials. Nonetheless, Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) notes that &#8216;gamifying&#8217; the calibration procedures in webcam eye tracking experiments would do much to improve participants&#8217; enjoyment (and performance) &#8211; in fact, Xu et al. (<xref ref-type="bibr" rid="B75">2015</xref>) did just that: they developed two short video games based on well-known game formats. One game, based on Angry Birds, required a high degree of accuracy (the goal was to &#8220;train a powerful gaze-controlled gun&#8221; with which to take down the birds); the other, based on Whack-a-mole, had a more forgiving threshold for successful &#8216;hits&#8217;. Xu et al. (<xref ref-type="bibr" rid="B75">2015</xref>) used these games to advertise two versions of the same picture classification task: one longer, more demanding task yielding high-accuracy gaze data, and one shorter, easier task that yielded cruder data, but which was also much more popular on Amazon Mechanical Turk and attracted and retained more participants. By combining both data sets and post-hoc data processing, the authors were able to obtain satisfactory results. xLabs, a now-defunct company that offered webcam eye tracking for marketing research, calibrated users by letting them click on animated crawling ants or floating balloons, and validates by visualising their real-time gaze predictions as a &#8216;laser&#8217; with which to squash or pop them.<xref ref-type="fn" rid="n20">20</xref> This kind of gamified (re-)calibration process would also make the experiment design more suitable for children.</p>
<p>Finally, our aim was to mitigate some of the increased noise that is inevitably inherent to remote webcam eye tracking. Where possible, we opted for the thriftier approach of rejecting participants with sub-optimal equipment set-ups before starting the experiment, rather than removing them post-hoc. Potential participants can be filtered by browser features such as browser type and version, screen resolution, display refresh rate, and support for essential software libraries such as WebAudio API before starting the experiment.<xref ref-type="fn" rid="n21">21</xref> Furthermore, the <monospace>samples_per_sec</monospace> variable in the validation plugin&#8217;s data output can be used to filter out participants with a low sampling rate, which we take to be a symptom of an unsuitable setup &#8211; whether due to aged hardware, high CPU load, sub-optimal combination of operating system and browser, or some other factor. We set our threshold at 5 samples per second, but e.g. Madsen et al. (<xref ref-type="bibr" rid="B47">2021</xref>) set it at 15. In experiments with an expected smaller effect size or that require a more fine-grained spatial or temporal resolution, filtering by a relatively high sampling rate may be a prerequisite for obtaining interpretable data.</p>
<p>With regards to data post-processing, eye movement classification and raw gaze data smoothing algorithms constitute their own subfield within eye tracking research, and a review of that literature lies outside the scope of this paper (but see e.g. <xref ref-type="bibr" rid="B61">Salvucci &amp; Goldberg, 2000</xref>; <xref ref-type="bibr" rid="B69">Tafaj et al., 2012</xref>; and <xref ref-type="bibr" rid="B34">Hessels et al., 2017</xref>). However, we note that <xref ref-type="bibr" rid="B75">Xu et al., 2015</xref> extracted fixations from their raw gaze data through meanshift clustering, i.e. algorithmically identifying and assigning gaze data to spatio-temporal clusters and labeling the cluster center as one fixation. To evaluate this approach, they selected 1000 random pairs of images and participant gaze data from the infrared eye tracking dataset they were using for comparison subject/image pairs (<xref ref-type="bibr" rid="B40">Judd et al., 2009</xref>), and obtained &#8216;ground truth&#8217; fixation locations on the images from the gaze data. They then permuted that data to resemble webcam eye tracking data by subsampling it to 30 Hz and adding position noise; extracted fixations using their meanshift algorithm; and compared the results to their ground truth fixations. The algorithm was able to estimate these fixations reasonably well, which suggests that it is worth considering as a noise-reduction tool for webcam eye tracking data going forward.</p>
<p>Although we did not process our data beyond the procedure followed in the original infrared study, we suggest that there are several ways to filter out participants with low-quality data during data preparation: for example, by plotting and inspecting the distribution of a certain performance metric, and discarding participants below a certain cut-off point. Madsen et al. (<xref ref-type="bibr" rid="B47">2021</xref>)<xref ref-type="fn" rid="n22">22</xref> visualised the raw gaze data of their participants along the horizontal and the vertical axes, coding position as brightness and with time on the x-axis and subject on the y-axis; subjects were sorted top-to-bottom by their score on a comprehension test. In their plots, high-performing subjects clearly exhibited a stereotypical pattern of eye movements, which gradually fades out as performance drops. In a Visual World study such as the one we present here, with an equal or greater number of filler trials than experimental trials, performance on the fillers could serve as a similar heuristic. Though accuracy of participants&#8217; offline responses on the filler trials was at ceiling, and thus less effective as a metric, one could use a summary metric of the online filler data instead, e.g. the proportion of looks towards any Region of Interest.</p>
</sec>
</sec>
<sec>
<title>4 Conclusion</title>
<p>We have presented a web-based replication of a Visual World eye tracking study, demonstrating that it is possible to obtain results that approach laboratory-grade effect sizes and onsets, using free, open source and beginner-friendly software tools. We have also shown how with a few methodological and experiment design adjustments, the overall user experience and success of such a study can be improved. We thereby add to the rapidly growing body of work investigating the possibilities and limitations of WebGazer and remote webcam eye tracking studies, which in a short time has led and will undoubtedly continue to lead to better code, experimental protocols, participant experiences, and research outcomes.</p>
</sec>
<sec>
<title>Supplementary information and materials</title>
<sec>
<title>Demographic data</title>
<p>Before the start of the web-based experiment, we administered a short demographic survey and recorded participants&#8217; browser type and version,<xref ref-type="fn" rid="n23">23</xref> their operating system type and version, and their screen resolution. In order to limit the collection of personal data that has no well-motivated bearing on the research question, we chose not to record sex or gender. Nor did we record exact age, choosing instead to bin participants into 5 age groups: 18&#8211;30, 31&#8211;43, 44&#8211;56, 57&#8211;69, and 70+. The demographic data is given in <xref ref-type="table" rid="T2">Table 2</xref>, with the number of participants for each category given between parentheses.</p>
<table-wrap id="T2">
<label>Table 2</label>
<caption>
<p>Number of participants per category for Age group, Vision, Browser type, and Operating System.</p>
</caption>
<table>
<thead>
<tr>
<td colspan="4"><hr/></td>
</tr>
<tr>
<td align="left" valign="top"><bold>Age</bold></td>
<td align="left" valign="top"><bold>Vision</bold></td>
<td align="left" valign="top"><bold>Browser</bold></td>
<td align="left" valign="top"><bold>OS</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td colspan="4"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">18&#8211;30 (52)</td>
<td align="left" valign="top">Normal, uncorrected vision (67)</td>
<td align="left" valign="top">Chrome (113)</td>
<td align="left" valign="top">Windows (84)</td>
</tr>
<tr>
<td colspan="4"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">31&#8211;43 (35)</td>
<td align="left" valign="top">Glasses (49)</td>
<td align="left" valign="top">Firefox (11)</td>
<td align="left" valign="top">MacOS (31)</td>
</tr>
<tr>
<td colspan="4"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">44&#8211;56 (29)</td>
<td align="left" valign="top">Contact lenses (4)</td>
<td align="left" valign="top"></td>
<td align="left" valign="top">Chrome OS (7)</td>
</tr>
<tr>
<td colspan="4"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">57&#8211;69 (8)</td>
<td align="left" valign="top">Abnormal, uncorrected vision (4)</td>
<td align="left" valign="top"></td>
<td align="left" valign="top">Linux (2)</td>
</tr>
<tr>
<td colspan="4"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">70+ (0)</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td colspan="4"><hr/></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>Correlation of sampling rate and calibration accuracy</title>
<p>Spearman rank correlation of mean calibration accuracy (for successful calibration trials only) and mean sampling rate by participant (see <xref ref-type="fig" rid="F7">Figure 7</xref>). Because Slim &amp; Hartsuiker (<xref ref-type="bibr" rid="B66">2021b</xref>) found a strong correlation between their participants&#8217; calibration scores and frames-per-second rate (<italic>R</italic> = 0.852, <italic>p</italic> &lt; 0.001), we ran a similar correlation on our data. At <italic>R</italic> = 0.16, and <italic>p</italic> = 0.074, this correlation was not significant; but since Slim &amp; Hartsuiker (<xref ref-type="bibr" rid="B66">2021b</xref>)&#8217;s minimum threshold for calibration success was 5% for one validation point, and ours was 50% for two points, this is not surprising.</p>
<fig id="F7">
<label>Figure 7</label>
<caption>
<p>Spearman rank correlation of participants&#8217; mean sampling rate and mean calibration accuracy.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-7-131-g7.png"/>
</fig>
</sec>
<sec>
<title>Sentence stimuli</title>
<p>The sentence stimuli used in both the original and the replication study: see <xref ref-type="table" rid="T3">Table 3</xref> for the experimental items, and <xref ref-type="table" rid="T4">Table 4</xref> for the filler items.</p>
<table-wrap id="T3">
<label>Table 3</label>
<caption>
<p>Sentence stimuli (experimental trials).</p>
</caption>
<table>
<thead>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top"><bold>Preamble</bold></td>
<td align="left" valign="top"><bold>Sentence</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a crisp winter morning.</td>
<td align="left" valign="top">Grandpa was building a big snowman.<break/>Grandpa built a big snowman.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was playtime at the school.</td>
<td align="left" valign="top">The boy was coloring a pretty picture.<break/>The boy colored a pretty picture.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was time for lunch.</td>
<td align="left" valign="top">Grandma was slicing a juicy watermelon.<break/>Grandma sliced a juicy watermelon.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a bright sunny day.</td>
<td align="left" valign="top">Grandpa was digging a deep pit.<break/>Grandpa dug a deep pit.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was playtime at the school.</td>
<td align="left" valign="top">The girl was drawing a slender vase.<break/>The girl drew a slender vase.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There were jobs to do around the house.</td>
<td align="left" valign="top">Grandpa was drilling a big hole.<break/>Grandpa drilled a big hole.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a holiday weekend.</td>
<td align="left" valign="top">Grandpa was fixing the old fridge.<break/>Grandpa fixed the old fridge.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a holiday weekend.</td>
<td align="left" valign="top">Grandma was hanging a beautiful painting.<break/>Grandma hung a beautiful painting.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There was going to be a party.</td>
<td align="left" valign="top">Grandpa was ironing a clean shirt.<break/>Grandpa ironed a clean shirt.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a crisp winter morning.</td>
<td align="left" valign="top">Grandma was knitting a new jumper.<break/>Grandma knitted a new jumper.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a crisp winter morning.</td>
<td align="left" valign="top">Grandpa was lighting a cosy fire.<break/>Grandpa lit a cosy fire.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the afternoon.</td>
<td align="left" valign="top">Grandma was locking the side door.<break/>Grandma locked the side door.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">The weather was nice and warm.</td>
<td align="left" valign="top">The girl was opening a big window.<break/>The girl opened a big window.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There were jobs to do around the house.</td>
<td align="left" valign="top">The girl was painting a high wall.<break/>The girl painted a high wall.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a bright sunny day.</td>
<td align="left" valign="top">Grandma was planting a pretty flower.<break/>Grandma planted a pretty flower.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the first period (at school).</td>
<td align="left" valign="top">The boy was sharpening a thin pencil.<break/>The boy sharpened a thin pencil.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There was going to be a party.</td>
<td align="left" valign="top">The boy was sweeping the narrow corridor.<break/>The boy swept the narrow corridor.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was early in the morning.</td>
<td align="left" valign="top">The boy was cleaning the front room.<break/>The boy cleaned the front room.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">The weather was nice and warm.</td>
<td align="left" valign="top">Grandma was watering a green bush.<break/>Grandma watered a green bush.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a rainy day outside.</td>
<td align="left" valign="top">Grandma was baking a lovely cake.<break/>Grandma baked a lovely cake.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a dark night with no hint of a breeze.</td>
<td align="left" valign="top">The boy was burying a wooden chest.<break/>The boy buried a wooden chest.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There were many people shopping in town.</td>
<td align="left" valign="top">The girl was buying a new phone.<break/>The girl bought a new phone.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was time for lunch.</td>
<td align="left" valign="top">The girl was eating a tasty fish.<break/>The girl ate a tasty fish.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the day.</td>
<td align="left" valign="top">Grandpa was demolishing an old house.<break/>Grandpa demolished an old house.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
</tbody>
</table>
</table-wrap>
<table-wrap id="T4">
<label>Table 4</label>
<caption>
<p>Sentence stimuli (filler trials).</p>
</caption>
<table>
<thead>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top"><bold>Preamble</bold></td>
<td align="left" valign="top"><bold>Sentence</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was early in the morning.</td>
<td align="left" valign="top">Grandpa was satisfied that the candle was blown out.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the afternoon.</td>
<td align="left" valign="top">Grandma was successful in cracking open the nut.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There were jobs to do around the house.</td>
<td align="left" valign="top">The boy was done with taking apart the wooden stool.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a beautiful quiet evening.</td>
<td align="left" valign="top">Grandpa was unconcerned that the old bridge had been destroyed.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was early in the morning.</td>
<td align="left" valign="top">The girl was happy with her newly cut out flower.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the afternoon.</td>
<td align="left" valign="top">The boy was pleased with his super tall tower.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There was going to be a party later.</td>
<td align="left" valign="top">Grandma was impressed with the beautiful dress she had sewn.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a rainy day outside.</td>
<td align="left" valign="top">Grandma was halfway through cutting the sleeves off the shirt.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">The weather was nice and warm.</td>
<td align="left" valign="top">Grandpa was tired after chopping down the tree.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a bright and sunny day.</td>
<td align="left" valign="top">The girl was proud that she managed to swim across the river.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was break time at the school.</td>
<td align="left" valign="top">The girl was finished with her glass of milk.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the afternoon.</td>
<td align="left" valign="top">The girl was ready to eat an orange.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">The weather was nice and warm.</td>
<td align="left" valign="top">The boy enjoyed himself photographing nature.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a holiday weekend.</td>
<td align="left" valign="top">Grandpa concentrated on preparing dessert.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was early in the morning.</td>
<td align="left" valign="top">The girl worked on cutting out a flower.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">There were jobs to do around the house.</td>
<td align="left" valign="top">The boy wanted to take apart the old wooden stool.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the afternoon.</td>
<td align="left" valign="top">Grandpa relaxed and read a book.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">The weather was nice and warm.</td>
<td align="left" valign="top">Grandpa occupied himself in the strawberry patch.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a beautiful quiet evening.</td>
<td align="left" valign="top">The boy planned to saw up the log for the fire.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the first period at school.</td>
<td align="left" valign="top">The boy got started on constructing a tower out of blocks.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was playtime at the school.</td>
<td align="left" valign="top">The girl wanted to put together the pretty toy castle.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was late in the afternoon.</td>
<td align="left" valign="top">The girl decided to burn a blue notebook.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was the middle of the day.</td>
<td align="left" valign="top">The girl began to drink a glass of milk.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
<tr>
<td align="left" valign="top">It was a bright and sunny day.</td>
<td align="left" valign="top">The girl started to blow up the green balloon.</td>
</tr>
<tr>
<td colspan="2"><hr/></td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
</sec>
<sec>
<title>Data availability</title>
<p>All materials relating to this study, including the raw data files, the tidy, analysis-ready dataframe, the R script for the data analysis and visualisation, the stimuli, and the codebase for the experiment, are made available through the Open Science Framework (<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/m395q/">https://osf.io/m395q/</ext-link>).</p>
</sec>
</body>
<back>
<fn-group>
<fn id="n1"><p>The full name is WebGazer.js, also sometimes written as webgazer.js or WebGazer.js; in this paper we will refer to it simply as WebGazer.</p></fn>
<fn id="n2"><p>This was not (as an anonymous reviewer supposed) a typo: Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) use the word saccade to refer to the window of time during which their participants switch fixation targets, and measure its duration from the onset of the new fixation cross, to the moment the gaze &#8220;fully reached&#8221; the target.</p></fn>
<fn id="n3"><p>Not all; e.g. Labvanced (<xref ref-type="bibr" rid="B30">Finger et al., 2017</xref>) has its own proprietary eye tracking software; see e.g. B&#225;nki et al. (<xref ref-type="bibr" rid="B7">2022</xref>) and Chouinard et al. (<xref ref-type="bibr" rid="B16">2019</xref>) for studies using LabVanced for infant eye tracking research.</p></fn>
<fn id="n4"><p>200 ms being the average latency of a saccade from one visual target to another (<xref ref-type="bibr" rid="B62">Saslow, 1967</xref>); though see e.g. Magnuson et al. (<xref ref-type="bibr" rid="B50">2008</xref>) and Huettig &amp; Altmann (<xref ref-type="bibr" rid="B37">2011</xref>) for examples of how fixations can be delayed or suppressed depending on task conditions.</p></fn>
<fn id="n5"><p>To avoid confusion, this paper will use the terms &#8216;online&#8217; and &#8216;offline&#8217; only to refer to behavioral measures collected in real-time, and after processing has taken place, respectively. Despite the widespread use of &#8216;online&#8217; to mean &#8216;on the Internet&#8217;, we will refer to our webcam eye tracking study as being &#8216;web-based&#8217;, to mean conducted on the web/the Internet.</p></fn>
<fn id="n6"><p>Although it is arguably risky to try to replicate a null result using a novel, &#8216;noisier&#8217; method, the unambiguously positive result in the Past Progressive condition provided us with a clear benchmark. We also had clear positive results in the perfective condition of our experiments in Russian and Spanish; as well as in a follow-up study contrasting the Past Progressive with the Past Perfect. This reassured us that the null result in the Simple Past condition reflected an absence of (strong) perfectivity, rather than e.g. the Ongoing Event picture generally being a better representation of the event than the Completed Event picture.</p></fn>
<fn id="n7"><p>The disadvantage of this approach is that the absolute size of the pictures will vary between participants, a noise factor that is all the harder to control for because display size cannot be automatically recorded through the browser (and screen resolution, which can, is not correlated with display size). Short of asking participants to measure the diagonal of their screen with a tape measure, we cannot know. However, giving the pictures the same absolute size and placement would likely cause them to be displayed incompletely or incorrectly on some devices. See section 3.2.2 for further discussion.</p></fn>
<fn id="n8"><p>JATOS stands for &#8220;Just Another Tool for Online Studies&#8221; (<xref ref-type="bibr" rid="B44">Lange et al., 2015</xref>). It is a free, open-source backend tool for hosting and managing web-based studies.</p></fn>
<fn id="n9"><p>The instructional images for head positioning, posture, and lighting were taken from Semmelmann &amp; Weigelt (<xref ref-type="bibr" rid="B64">2018</xref>) (<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/jmz79/">https://osf.io/jmz79/</ext-link>).</p></fn>
<fn id="n10"><p>In the jsPsych documentation this is recommended as a testing and debugging tool, but we chose to keep it in hope that participants would be able to interpret the feedback and adjust accordingly (e.g., if a participant&#8217;s face is brightly lit on one side because they&#8217;re sat beside a window, the accuracy on that side of the screen is likely to be much lower). Feedback also bolsters intrinsic motivation and improves performance (see e.g. <xref ref-type="bibr" rid="B24">Dow et al., 2012</xref>): if participants can see that they have a healthy number of gaze samples, of which <italic>just</italic> under half appear to be within either tolerance radius, they will hopefully be motivated to keep going rather than discouraged because they failed to calibrate.</p></fn>
<fn id="n11"><p>Before running the replication study, we ran several smaller pilot studies in batches of 60 participants (the infrared study&#8217;s sample size) at a time. By merging the results of two pilots, we found that we had reached a similar standard error of the mean as in the infrared results (though still a weaker overall effect): once we switched to the new version of WebGazer and found a much clearer gaze pattern with as little as <italic>n</italic> = 24, we decided to set the sample size at <italic>n</italic> = 120.</p></fn>
<fn id="n12"><p>We leave the question of whether this type of study could also be taken via tablet or phone to future research.</p></fn>
<fn id="n13"><p>We did not fit maximally structured random effects in our models because some failed to converge with both random intercepts and slopes (singular fit), so we applied only random intercepts across all models.</p></fn>
<fn id="n14"><p>We chose to set the threshold at 0.08, rather than the customary 0.05, because relaxing the criteria for identifying clusters in the initial stage of the permutation analysis helps to find larger contiguous clusters. When we ran this analysis with a &#8216;traditional&#8217; alpha of 0.05, we found several clusters grouped closely together, separated only by 1 or 2 time bins where an effect was not found: we did not find it plausible that an effect of aspect in looking preference (well after verb offset) should blink in and out of existence. With an alpha of 0.08, these clusters merged into one. The risk of false positives is sidestepped by the re-sampling portion of the analysis: the initial, &#8220;real&#8221; sum statistic is tested against a distribution of 1000 permuted sum statistics that were <italic>also</italic> found with alpha = 0.08, so this final significance test is no less strict.</p></fn>
<fn id="n15"><p>Josh de Leeuw, the main developer of jsPsych, clarified on 15/6/2021 (discussion #1892 on the jspsych/jsPsych github forum): &#8220;As far as performance goes, I think we [Yang and de Leeuw] both applied similar modifications to webgazer. [&#8230;] The major change we both made to our respective forks is that we disabled webgazer&#8217;s automatic loop so that webgazer is no longer trying to provide an updated prediction with every animation frame, and instead we just invoke webgazer&#8217;s prediction algorithm at a regular interval. This seems to actually speed up the rate at which calculations can be done. And, perhaps even more importantly, using requestAnimationFrame was causing blocking in jsPsych&#8217;s timing, so if a participant had a particularly poor computer &#8211; or even a good one &#8211; the timing of experiments could become really bad really quickly (see issue #1700).&#8221;</p></fn>
<fn id="n16"><p>A note for users of Gorilla.sc: Will Webster, a software developer at Gorilla/Cauldron Science, confirmed that his team is aware of this issue and is working on forking, modifying, and integrating WebGazer into Gorilla&#8217;s own timing system (6/8/2021, personal correspondence).</p></fn>
<fn id="n17"><p>A minor but interesting note on <xref ref-type="fig" rid="F5">Figures 5(c)</xref> and <xref ref-type="fig" rid="F5">5(d)</xref> is that they show very different proportions of looks to the target vs. the competitor picture in the first &#126;200 ms of the trial. This could in part be due to spillover from the previous trial &#8211; experimental and filler items always alternate, and the preference for the target picture is much greater in the fillers. But as can be seen in <xref ref-type="fig" rid="F4">Figure 4(b)</xref>, the first 3 time bins of the trial have much fewer gaze samples than the rest &#8211; as if WebGazer has to &#8216;warm up&#8217; in the first 150 ms before settling into a regular sampling pattern. Something worth keeping in mind while designing experiments with WebGazer.</p></fn>
<fn id="n18"><p>MouseView.js is a Javascript library which blurs the display to mimic peripheral vision, but lets participants use their mouse pointer to move a sharp, fovea-like aperture.</p></fn>
<fn id="n19"><p>For a more general overview and cost-benefit analysis of conducting web-based behavioral studies, see e.g. Sauter et al. (<xref ref-type="bibr" rid="B63">2020</xref>), Eyal et al. (<xref ref-type="bibr" rid="B27">2021</xref>), and Gagn&#233; &amp; Franzen (<xref ref-type="bibr" rid="B31">2021</xref>).</p></fn>
<fn id="n20"><p>Much of the code behind these calibration games is still available via the company&#8217;s Github page.</p></fn>
<fn id="n21"><p>As of version 7.1, jsPsych has a dedicated plugin for this.</p></fn>
<fn id="n22"><p>See page 9 of their Supplementary Information.</p></fn>
<fn id="n23"><p>At the time this study was conducted, the only browser systems that it could reliably be conducted on were Chrome and Firefox. Likewise, Yang &amp; Krajbich (<xref ref-type="bibr" rid="B77">2021</xref>) report that of their 49 participants, 45 used Chrome and 4 used Firefox. The software libraries that webcam eye tracking experiments rely on &#8211; particularly, WebGazer and WebAudio API &#8211; are now becoming available to a wider range of browsers, such as Safari, Edge, and Opera.</p></fn>
</fn-group>
<sec>
<title>Ethics and consent</title>
<p>All procedures performed in studies involving human participants were in accordance with the ethical standards of the Norwegian Centre for Research Data and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards.</p>
<p>Informed consent was obtained from all individual participants involved in the study.</p>
</sec>
<ack>
<title>Acknowledgements</title>
<p>We gratefully acknowledge the work of Onur Ferhat, whose custom eye tracking plugins formed the basis of our pilot experiments. Thanks to Alain Schoorl for additional programming support; to the members of the &#216;yelab for their feedback and support throughout; to the editor and the reviewers for their helpful feedback; and to the hundreds of participants who contributed to these results.</p>
<p>Thanks also to the Linguistics departments at NTNU in Trondheim, and at the University of Edinburgh, for hosting us during our infrared eye tracking data collection trips. During our stay in Edinburgh, in December 2019, the academic staff were striking in protest against inequitable pay, extreme workloads, casualisation, and pension cuts: we stand in solidarity with their ongoing fight for a fairer and better workplace.</p>
</ack>
<sec>
<title>Funding information</title>
<p>The authors were funded by the Research Council of Norway FRIPRO project 275490 <italic>Modal Concepts and Compositionality: New Directions in Experimental Semantics</italic>.</p>
<p>This research was also supported through funding from the BLINK project (Marie Sk&#322;odowska-Curie Actions Individual Fellowship 2018-2021).</p>
</sec>
<sec>
<title>Competing interests</title>
<p>The authors have no competing interests to declare.</p>
</sec>
<sec>
<title>Author contributions</title>
<p>Conceptualization, M.V., S.M., and G.R.; methodology, M.V., S.M., and G.R.; software, M.V.; formal analysis, S.M.; visualization, S.M and M.V.; investigation, M.V.; data curation, M.V., S.M.; project administration, M.V.; validation, S.M.; writing &#8211; original draft, M.V.; writing &#8211; review and editing, M.V. and G.R.; supervision and funding acquisition, G.R.</p>
</sec>
<ref-list>
<ref id="B1"><label>1</label><mixed-citation publication-type="journal"><string-name><surname>Allopenna</surname>, <given-names>P. D.</given-names></string-name>, <string-name><surname>Magnuson</surname>, <given-names>J. S.</given-names></string-name>, &amp; <string-name><surname>Tanenhaus</surname>, <given-names>M. K.</given-names></string-name> (<year>1998</year>). <article-title>Tracking the time course of spoken word recognition using eye movements: Evidence for continuous mapping models</article-title>. <source>Journal of memory and language</source>, <volume>38</volume>(<issue>4</issue>), <fpage>419</fpage>&#8211;<lpage>439</lpage>. DOI: <pub-id pub-id-type="doi">10.1006/jmla.1997.2558</pub-id></mixed-citation></ref>
<ref id="B2"><label>2</label><mixed-citation publication-type="journal"><string-name><surname>Altmann</surname>, <given-names>G. T. M.</given-names></string-name>, &amp; <string-name><surname>Kamide</surname>, <given-names>Y.</given-names></string-name> (<year>1999</year>). <article-title>Incremental interpretation at verbs: Restricting the domain of subsequent reference</article-title>. <source>Cognition</source>, <volume>73</volume>(<issue>3</issue>), <fpage>247</fpage>&#8211;<lpage>264</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/S0010-0277(99)00059-1</pub-id></mixed-citation></ref>
<ref id="B3"><label>3</label><mixed-citation publication-type="journal"><string-name><surname>Andersson</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Nystr&#246;m</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Holmqvist</surname>, <given-names>K.</given-names></string-name> (<year>2010</year>). <article-title>Sampling frequency and eye-tracking measures: how speed affects durations, latencies, and more</article-title>. <source>Journal of Eye Movement Research</source>, <volume>3</volume>(<issue>3</issue>). DOI: <pub-id pub-id-type="doi">10.16910/jemr.3.3.6</pub-id></mixed-citation></ref>
<ref id="B4"><label>4</label><mixed-citation publication-type="journal"><string-name><surname>Anwyl-Irvine</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Dalmaijer</surname>, <given-names>E. S.</given-names></string-name>, <string-name><surname>Hodges</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Evershed</surname>, <given-names>Jo. K.</given-names></string-name> (<year>2021a</year>). <article-title>Realistic precision and accuracy of online experiment platforms, web browsers, and devices</article-title>. <source>Behavior research methods</source>, <volume>53</volume>(<issue>4</issue>), <fpage>1407</fpage>&#8211;<lpage>1425</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-020-01501-5</pub-id></mixed-citation></ref>
<ref id="B5"><label>5</label><mixed-citation publication-type="journal"><string-name><surname>Anwyl-Irvine</surname>, <given-names>A. L.</given-names></string-name>, <string-name><surname>Armstrong</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Dalmaijer</surname>, <given-names>E. S.</given-names></string-name> (<year>2021b</year>). <article-title>Mouseview. js: Reliable and valid attention tracking in web-based experiments using a cursor-directed aperture</article-title>. <source>Behavior research methods</source> (pp. <fpage>1</fpage>&#8211;<lpage>25</lpage>). DOI: <pub-id pub-id-type="doi">10.3758/s13428-021-01703-5</pub-id></mixed-citation></ref>
<ref id="B6"><label>6</label><mixed-citation publication-type="webpage"><string-name><surname>Arunachalam</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Kothari</surname>, <given-names>A.</given-names></string-name> (<year>2011</year>). <article-title>An experimental study of hindi and English perfective interpretation</article-title>. <source>Journal of South Asian Linguistics</source>, <volume>4</volume>(<issue>1</issue>), <fpage>27</fpage>&#8211;<lpage>42</lpage>. <uri>https://ojs.ub.unikonstanz.de/jsal/index.php/jsal/article/download/35/21/0</uri>.</mixed-citation></ref>
<ref id="B7"><label>7</label><mixed-citation publication-type="journal"><string-name><surname>B&#225;nki</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>de Eccher</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Falschlehner</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Hoehl</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Markova</surname>, <given-names>G.</given-names></string-name> (<year>2022</year>). <article-title>Comparing online webcam-and laboratory-based eye-tracking for the assessment of infants&#8217; audio-visual synchrony perception</article-title>. <source>Frontiers in Psychology</source>, <volume>6162</volume>. DOI: <pub-id pub-id-type="doi">10.3389/fpsyg.2021.733933</pub-id></mixed-citation></ref>
<ref id="B8"><label>8</label><mixed-citation publication-type="webpage"><string-name><surname>Bates</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>M&#228;chler</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Bolker</surname>, <given-names>B.</given-names></string-name>, &amp; <string-name><surname>Walker</surname>, <given-names>S.</given-names></string-name> (<year>2014</year>). <article-title>Fitting linear mixed-effects models using lme4</article-title>. <source>arXiv preprint arXiv:1406.5823</source>. <uri>https://arxiv.org/abs/1406.5823</uri>. DOI: <pub-id pub-id-type="doi">10.18637/jss.v067.i01</pub-id></mixed-citation></ref>
<ref id="B9"><label>9</label><mixed-citation publication-type="journal"><string-name><surname>Blais</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Jack</surname>, <given-names>R. E.</given-names></string-name>, <string-name><surname>Scheepers</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Fiset</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Caldara</surname>, <given-names>R.</given-names></string-name> (<year>2008</year>). <article-title>Culture shapes how we look at faces</article-title>. <source>PloS one</source>, <volume>3</volume>(<issue>8</issue>). <elocation-id>e3022</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1371/journal.pone.0003022</pub-id></mixed-citation></ref>
<ref id="B10"><label>10</label><mixed-citation publication-type="book"><string-name><surname>Bott</surname>, <given-names>O.</given-names></string-name>, &amp; <string-name><surname>Hamm</surname>, <given-names>F.</given-names></string-name> (<year>2014</year>). <chapter-title>Cross-linguistic variation in the processing of aspect</chapter-title>. In <source>Psycholinguistic approaches to meaning and understanding across languages</source> (pp. <fpage>83</fpage>&#8211;<lpage>109</lpage>). <publisher-name>Springer</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1007/978-3-319-05675-3_4</pub-id></mixed-citation></ref>
<ref id="B11"><label>11</label><mixed-citation publication-type="journal"><string-name><surname>Bridges</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Pitiot</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>MacAskill</surname>, <given-names>M. R.</given-names></string-name>, &amp; <string-name><surname>Peirce</surname>, <given-names>J. W.</given-names></string-name> (<year>2020</year>). <article-title>The timing mega-study: Comparing a range of experiment generators, both lab-based and online</article-title>. <source>PeerJ</source>, <volume>8</volume>, <elocation-id>e9414</elocation-id>. DOI: <pub-id pub-id-type="doi">10.7717/peerj.9414</pub-id></mixed-citation></ref>
<ref id="B12"><label>12</label><mixed-citation publication-type="webpage"><string-name><surname>Calabrich</surname>, <given-names>S. L.</given-names></string-name>, <string-name><surname>Oppenheim</surname>, <given-names>G. M.</given-names></string-name>, &amp; <string-name><surname>Jones</surname>, <given-names>M. W.</given-names></string-name> (<year>2021</year>). <chapter-title>Episodic memory cues in acquisition of novel visual-phonological associations: a webcam-based eye-tracking study</chapter-title>. In <source>Proceedings of the annual meeting of the cognitive science society</source>, <volume>43</volume>. <uri>https://escholarship.org/uc/item/76b3c54t</uri>.</mixed-citation></ref>
<ref id="B13"><label>13</label><mixed-citation publication-type="journal"><string-name><surname>Chabal</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Hayakawa</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Marian</surname>, <given-names>V.</given-names></string-name> (<year>2022</year>). <article-title>Language is activated by visual input regardless of memory demands or capacity</article-title>. <source>Cognition</source>, <volume>222</volume>, <fpage>104994</fpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2021.104994</pub-id></mixed-citation></ref>
<ref id="B14"><label>14</label><mixed-citation publication-type="journal"><string-name><surname>Chandler</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Kapelner</surname>, <given-names>A.</given-names></string-name> (<year>2013</year>). <article-title>Breaking monotony with meaning: Motivation in crowdsourcing markets</article-title>. <source>Journal of Economic Behavior &amp; Organization</source>, <volume>90</volume>, <fpage>123</fpage>&#8211;<lpage>133</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.jebo.2013.03.003</pub-id></mixed-citation></ref>
<ref id="B15"><label>15</label><mixed-citation publication-type="journal"><string-name><surname>Chen</surname>, <given-names>M. C.</given-names></string-name>, <string-name><surname>Anderson</surname>, <given-names>J. R.</given-names></string-name>, &amp; <string-name><surname>Sohn</surname>, <given-names>M. H.</given-names></string-name> (<year>2001</year>). <article-title>What can a mouse cursor tell us more? correlation of eye/mouse movements on web browsing</article-title>. In <source>Chi&#8217;01 extended abstracts on human factors in computing systems</source> (pp. <fpage>281</fpage>&#8211;<lpage>282</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/634067.634234</pub-id></mixed-citation></ref>
<ref id="B16"><label>16</label><mixed-citation publication-type="journal"><string-name><surname>Chouinard</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Scott</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Cusack</surname>, <given-names>R.</given-names></string-name> (<year>2019</year>). <article-title>Using automatic face analysis to score infant behaviour from video collected online</article-title>. <source>Infant Behavior and Development</source>, <volume>54</volume>, <fpage>1</fpage>&#8211;<lpage>12</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.infbeh.2018.11.004</pub-id></mixed-citation></ref>
<ref id="B17"><label>17</label><mixed-citation publication-type="webpage"><string-name><surname>Cooper</surname>, <given-names>R. M.</given-names></string-name> (<year>1974</year>). <article-title>The control of eye fixation by the meaning of spoken language: a new methodology for the real-time investigation of speech perception, memory, and language processing</article-title>. <source>Cognitive psychology</source>. <uri>https://psycnet.apa.org/doi/10.1016/0010-0285(74)90005-X</uri>. DOI: <pub-id pub-id-type="doi">10.1016/0010-0285(74)90005-X</pub-id></mixed-citation></ref>
<ref id="B18"><label>18</label><mixed-citation publication-type="webpage"><string-name><surname>Dalmaijer</surname>, <given-names>E.</given-names></string-name> (<year>2014</year>). <chapter-title>Is the low-cost eyetribe eye tracker any good for research?</chapter-title> <source>Tech. rep. PeerJ PrePrints</source>. <uri>https://peerj.com/preprints/585v1.pdf</uri>. DOI: <pub-id pub-id-type="doi">10.7287/peerj.preprints.585v1</pub-id></mixed-citation></ref>
<ref id="B19"><label>19</label><mixed-citation publication-type="journal"><string-name><surname>De Leeuw</surname>, <given-names>J. R.</given-names></string-name> (<year>2015</year>). <article-title>jspsych: A javascript library for creating behavioral experiments in a web browser</article-title>. <source>Behavior research methods</source>, <volume>47</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>12</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-014-0458-y</pub-id></mixed-citation></ref>
<ref id="B20"><label>20</label><mixed-citation publication-type="journal"><string-name><surname>de Leeuw</surname>, <given-names>J. R.</given-names></string-name>, &amp; <string-name><surname>Motz</surname>, <given-names>B. A.</given-names></string-name> (<year>2016</year>). <article-title>Psychophysics in a web browser? comparing response times collected with javascript and psychophysics toolbox in a visual search task</article-title>. <source>Behavior Research Methods</source>, <volume>48</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>12</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-015-0567-2</pub-id></mixed-citation></ref>
<ref id="B21"><label>21</label><mixed-citation publication-type="journal"><string-name><surname>De Swart</surname>, <given-names>H.</given-names></string-name> (<year>1998</year>). <article-title>Aspect shift and coercion</article-title>. <source>Natural language &amp; linguistic theory</source>, <volume>16</volume>(<issue>2</issue>), <fpage>347</fpage>&#8211;<lpage>385</lpage>. DOI: <pub-id pub-id-type="doi">10.1023/A:1005916004600</pub-id></mixed-citation></ref>
<ref id="B22"><label>22</label><mixed-citation publication-type="journal"><string-name><surname>de Swart</surname>, <given-names>H.</given-names></string-name> (<year>2012</year>). <article-title>Verbal aspect</article-title>. In <source>The oxford handbook of tense and aspect</source>. DOI: <pub-id pub-id-type="doi">10.1093/oxfordhb/9780195381979.013.0026</pub-id></mixed-citation></ref>
<ref id="B23"><label>23</label><mixed-citation publication-type="webpage"><string-name><surname>Degen</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Kursat</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name><surname>Leigh</surname>, <given-names>D.</given-names></string-name> (<year>2021</year>). <chapter-title>Seeing is believing: testing an explicit linking assumption for visual world eye-tracking in psycholinguistics</chapter-title>. Unpublished, retrieved 13/05/2021. <uri>https://github.com/thegricean/eyetracking_replications/blob/master/writing/2021_cogsci/sunbrehenyreplication.pdf</uri>.</mixed-citation></ref>
<ref id="B24"><label>24</label><mixed-citation publication-type="journal"><string-name><surname>Dow</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Kulkarni</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Klemmer</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Hartmann</surname>, <given-names>B.</given-names></string-name> (<year>2012</year>). <article-title>Shepherding the crowd yields better work</article-title>. In <source>Proceedings of the acm 2012 conference on computer supported cooperative work</source> (pp. <fpage>1013</fpage>&#8211;<lpage>1022</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/2145204.2145355</pub-id></mixed-citation></ref>
<ref id="B25"><label>25</label><mixed-citation publication-type="book"><string-name><surname>Dowty</surname>, <given-names>D. R.</given-names></string-name> (<year>1979</year>). <source>Word meaning and Montague grammar: The semantics of verbs and times in generative semantics and in Montague&#8217;s PTQ</source>. <publisher-loc>Dordrecht</publisher-loc>: <publisher-name>Reidel</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1007/978-94-009-9473-7</pub-id></mixed-citation></ref>
<ref id="B26"><label>26</label><mixed-citation publication-type="journal"><string-name><surname>Ehinger</surname>, <given-names>B. V.</given-names></string-name>, <string-name><surname>Gro&#223;</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Ibs</surname>, <given-names>I.</given-names></string-name>, &amp; <string-name><surname>K&#246;nig</surname>, <given-names>P.</given-names></string-name> (<year>2019</year>). <article-title>A new comprehensive eye-tracking test battery concurrently evaluating the pupil labs glasses and the eyelink 1000</article-title>. <source>PeerJ</source>, <volume>7</volume>, <elocation-id>e7086</elocation-id>. DOI: <pub-id pub-id-type="doi">10.7717/peerj.7086</pub-id></mixed-citation></ref>
<ref id="B27"><label>27</label><mixed-citation publication-type="journal"><string-name><surname>Eyal</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>David</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Andrew</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Zak</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Ekaterina</surname>, <given-names>D.</given-names></string-name> (<year>2021</year>). <article-title>Data quality of platforms and panels for online behavioral research</article-title>. <source>Behavior Research Methods</source> (pp. <fpage>1</fpage>&#8211;<lpage>20</lpage>). DOI: <pub-id pub-id-type="doi">10.3758/s13428-021-01694-3</pub-id></mixed-citation></ref>
<ref id="B29"><label>29</label><mixed-citation publication-type="journal"><string-name><surname>Federico</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Brandimonte</surname>, <given-names>M. A.</given-names></string-name> (<year>2019</year>). <article-title>Tool and object affordances: an ecological eye-tracking study</article-title>. <source>Brain and cognition</source>, <volume>135</volume>, <fpage>103582</fpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.bandc.2019.103582</pub-id></mixed-citation></ref>
<ref id="B30"><label>30</label><mixed-citation publication-type="journal"><string-name><surname>Finger</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>Goeke</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Diekamp</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Standvo&#223;</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>K&#246;nig</surname>, <given-names>P.</given-names></string-name> (<year>2017</year>). <article-title>Labvanced: a unified javascript framework for online studies</article-title>. In <source>International conference on computational social science (cologne)</source>.</mixed-citation></ref>
<ref id="B31"><label>31</label><mixed-citation publication-type="journal"><string-name><surname>Gagn&#233;</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Franzen</surname>, <given-names>L.</given-names></string-name> (<year>2021</year>). <article-title>How to run behavioural experiments online: best practice suggestions for cognitive psychology and neuroscience</article-title>. DOI: <pub-id pub-id-type="doi">10.31234/osf.io/nt67j</pub-id></mixed-citation></ref>
<ref id="B32"><label>32</label><mixed-citation publication-type="journal"><string-name><surname>Gvozdanovi&#263;</surname>, <given-names>J.</given-names></string-name> (<year>2012</year>). <article-title>Perfective and imperfective aspect</article-title>. In <source>The oxford handbook of tense and aspect</source>. DOI: <pub-id pub-id-type="doi">10.1093/oxfordhb/9780195381979.013.0027</pub-id></mixed-citation></ref>
<ref id="B33"><label>33</label><mixed-citation publication-type="book"><string-name><surname>Hauger</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Paramythis</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Weibelzahl</surname>, <given-names>S.</given-names></string-name> (<year>2011</year>). <chapter-title>Using browser interaction data to determine page reading behavior</chapter-title>. In <source>International conference on user modeling, adaptation, and personalization</source> (<fpage>147</fpage>&#8211;<lpage>158</lpage>). <publisher-name>Springer</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1007/978-3-642-22362-4_13</pub-id></mixed-citation></ref>
<ref id="B34"><label>34</label><mixed-citation publication-type="journal"><string-name><surname>Hessels</surname>, <given-names>R. S.</given-names></string-name>, <string-name><surname>Niehorster</surname>, <given-names>D. C.</given-names></string-name>, <string-name><surname>Kemner</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Hooge</surname>, <given-names>I. T. C.</given-names></string-name> (<year>2017</year>). <article-title>Noiserobust fixation detection in eye movement data: Identification by two-means clustering (i2mc)</article-title>. <source>Behavior research methods</source>, <volume>49</volume>(<issue>5</issue>), <fpage>1802</fpage>&#8211;<lpage>1823</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-016-0822-1</pub-id></mixed-citation></ref>
<ref id="B35"><label>35</label><mixed-citation publication-type="journal"><string-name><surname>Huang</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>White</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Buscher</surname>, <given-names>G.</given-names></string-name> (<year>2012</year>). <article-title>User see, user point: gaze and cursor alignment in web search</article-title>. In <source>Proceedings of the sigchi conference on human factors in computing systems</source> (pp. <fpage>1341</fpage>&#8211;<lpage>1350</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/2207676.2208591</pub-id></mixed-citation></ref>
<ref id="B36"><label>36</label><mixed-citation publication-type="journal"><string-name><surname>Huang</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name><surname>Snedeker</surname>, <given-names>J.</given-names></string-name> (<year>2020</year>). <article-title>Evidence from the visual world paradigm raises questions about unaccusativity and growth curve analyses</article-title>. <source>Cognition</source>, <volume>200</volume>, <fpage>104251</fpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2020.104251</pub-id></mixed-citation></ref>
<ref id="B37"><label>37</label><mixed-citation publication-type="journal"><string-name><surname>Huettig</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Altmann</surname>, <given-names>G. T. M.</given-names></string-name> (<year>2011</year>). <article-title>Looking at anything that is green when hearing &#8220;frog&#8221;: How object surface colour and stored object colour knowledge influence language-mediated overt attention</article-title>. <source>The Quarterly Journal of Experimental Psychology</source>, <volume>64</volume>(<issue>1</issue>), <fpage>122</fpage>&#8211;<lpage>145</lpage>. DOI: <pub-id pub-id-type="doi">10.1080/17470218.2010.481474</pub-id></mixed-citation></ref>
<ref id="B38"><label>38</label><mixed-citation publication-type="journal"><string-name><surname>Huettig</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Guerra</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Helo</surname>, <given-names>A.</given-names></string-name> (<year>2020</year>). <article-title>Towards understanding the task dependency of embodied language processing: the influence of colour during language-vision interactions</article-title>. <source>Journal of Cognition</source>, <volume>3</volume>(<issue>1</issue>). DOI: <pub-id pub-id-type="doi">10.5334/joc.135</pub-id></mixed-citation></ref>
<ref id="B28"><label>28</label><mixed-citation publication-type="journal"><string-name><surname>Huettig</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Rommers</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Meyer</surname>, <given-names>A. S.</given-names></string-name> (<year>2011</year>). <article-title>Using the visual world paradigm to study language processing: A review and critical evaluation</article-title>. <source>Acta psychologica</source>, <volume>137</volume>(<issue>2</issue>), <fpage>151</fpage>&#8211;<lpage>171</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.actpsy.2010.11.003</pub-id></mixed-citation></ref>
<ref id="B39"><label>39</label><mixed-citation publication-type="webpage"><string-name><surname>Jeschull</surname>, <given-names>L.</given-names></string-name> (<year>2007</year>). <chapter-title>The pragmatics of telicity and what children make of it</chapter-title>. In <source>Proceedings of the 2nd conference on generative approaches to language acquisition north america</source> (pp. <fpage>180</fpage>&#8211;<lpage>187</lpage>). <publisher-name>Citeseer</publisher-name>. <uri>http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.492.5503&amp;rep=rep1&amp;type=pdf</uri>.</mixed-citation></ref>
<ref id="B40"><label>40</label><mixed-citation publication-type="book"><string-name><surname>Judd</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Ehinger</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Durand</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Torralba</surname>, <given-names>A.</given-names></string-name> (<year>2009</year>). <chapter-title>Learning to predict where humans look</chapter-title>. In <source>2009 ieee 12th international conference on computer vision</source> (pp. <fpage>2106</fpage>&#8211;<lpage>2113</lpage>). <publisher-name>IEEE</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1109/ICCV.2009.5459462</pub-id></mixed-citation></ref>
<ref id="B41"><label>41</label><mixed-citation publication-type="journal"><string-name><surname>Jun</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Hsieh</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Reinecke</surname>, <given-names>K.</given-names></string-name> (<year>2017</year>). <article-title>Types of motivation affect study selection, attention, and dropouts in online experiments</article-title>. <source>Proceedings of the ACM on Human-Computer Interaction 1(CSCW)</source> (pp. <fpage>1</fpage>&#8211;<lpage>15</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/3134691</pub-id></mixed-citation></ref>
<ref id="B42"><label>42</label><mixed-citation publication-type="journal"><string-name><surname>Kochari</surname>, <given-names>A. R.</given-names></string-name> (<year>2019</year>). <article-title>Conducting web-based experiments for numerical cognition research</article-title>. <source>Journal of cognition</source>, <volume>2</volume>(<issue>1</issue>). DOI: <pub-id pub-id-type="doi">10.5334/joc.85</pub-id></mixed-citation></ref>
<ref id="B43"><label>43</label><mixed-citation publication-type="journal"><string-name><surname>Krajbich</surname>, <given-names>I.</given-names></string-name>, <string-name><surname>Armel</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Rangel</surname>, <given-names>A.</given-names></string-name> (<year>2010</year>). <article-title>Visual fixations and the computation and comparison of value in simple choice</article-title>. <source>Nature neuroscience</source>, <volume>13</volume>(<issue>10</issue>), <fpage>1292</fpage>&#8211;<lpage>1298</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/nn.2635</pub-id></mixed-citation></ref>
<ref id="B44"><label>44</label><mixed-citation publication-type="journal"><string-name><surname>Lange</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>K&#252;hn</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Filevich</surname>, <given-names>E.</given-names></string-name> (<year>2015</year>). <article-title>&#8220;just another tool for online studies&#8221;(jatos): An easy solution for setup and management of web servers supporting online studies</article-title>. <source>PloS one</source>, <volume>10</volume>(<issue>6</issue>), <elocation-id>e0130834</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1371/journal.pone.0130834</pub-id></mixed-citation></ref>
<ref id="B45"><label>45</label><mixed-citation publication-type="journal"><string-name><surname>Li</surname>, <given-names>Q.</given-names></string-name>, <string-name><surname>Joo</surname>, <given-names>S. J.</given-names></string-name>, <string-name><surname>Yeatman</surname>, <given-names>J. D.</given-names></string-name>, &amp; <string-name><surname>Reinecke</surname>, <given-names>K.</given-names></string-name> (<year>2020</year>). <article-title>Controlling for participants&#8217; viewing distance in large-scale, psychophysical online experiments using a virtual chinrest</article-title>. <source>Scientific reports</source>, <volume>10</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>11</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/s41598-019-57204-1</pub-id></mixed-citation></ref>
<ref id="B46"><label>46</label><mixed-citation publication-type="journal"><string-name><surname>Madden</surname>, <given-names>C. J.</given-names></string-name>, &amp; <string-name><surname>Zwaan</surname>, <given-names>R. A.</given-names></string-name> (<year>2003</year>). <article-title>How does verb aspect constrain event representations?</article-title> <source>Memory and Cognition</source>, <volume>31</volume>, <fpage>663</fpage>&#8211;<lpage>672</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/BF03196106</pub-id></mixed-citation></ref>
<ref id="B47"><label>47</label><mixed-citation publication-type="journal"><string-name><surname>Madsen</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Julio</surname>, <given-names>S. U.</given-names></string-name>, <string-name><surname>Gucik</surname>, <given-names>P. J.</given-names></string-name>, <string-name><surname>Steinberg</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Parra</surname>, <given-names>L. C.</given-names></string-name> (<year>2021</year>). <article-title>Synchronized eye movements predict test scores in online video education</article-title>. <source>Proceedings of the National Academy of Sciences</source>, <volume>118</volume>(<issue>5</issue>). DOI: <pub-id pub-id-type="doi">10.1073/pnas.2016980118</pub-id></mixed-citation></ref>
<ref id="B48"><label>48</label><mixed-citation publication-type="journal"><string-name><surname>Magliano</surname>, <given-names>J. P.</given-names></string-name>, &amp; <string-name><surname>Schleich</surname>, <given-names>M. C.</given-names></string-name> (<year>2000</year>). <article-title>Verb aspect and situation models</article-title>. <source>Discourse processes</source>, <volume>29</volume>(<issue>2</issue>), <fpage>83</fpage>&#8211;<lpage>112</lpage>. DOI: <pub-id pub-id-type="doi">10.1207/S15326950dp2902_1</pub-id></mixed-citation></ref>
<ref id="B49"><label>49</label><mixed-citation publication-type="journal"><string-name><surname>Magnuson</surname>, <given-names>J. S.</given-names></string-name> (<year>2019</year>). <article-title>Fixations in the visual world paradigm: where, when, why?</article-title> <source>Journal of Cultural Cognitive Science</source>, <volume>3</volume>(<issue>2</issue>), <fpage>113</fpage>&#8211;<lpage>139</lpage>. DOI: <pub-id pub-id-type="doi">10.1007/s41809-019-00035-3</pub-id></mixed-citation></ref>
<ref id="B50"><label>50</label><mixed-citation publication-type="journal"><string-name><surname>Magnuson</surname>, <given-names>J. S.</given-names></string-name>, <string-name><surname>Tanenhaus</surname>, <given-names>M. K.</given-names></string-name>, &amp; <string-name><surname>Aslin</surname>, <given-names>R. N.</given-names></string-name> (<year>2008</year>). <article-title>Immediate effects of form-class constraints on spoken word recognition</article-title>. <source>Cognition</source>, <volume>108</volume>(<issue>3</issue>), <fpage>866</fpage>&#8211;<lpage>873</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2008.06.005</pub-id></mixed-citation></ref>
<ref id="B51"><label>51</label><mixed-citation publication-type="journal"><string-name><surname>Martin</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Demirdache</surname>, <given-names>H.</given-names></string-name> (<year>2020</year>). <article-title>Partitive accomplishments across languages</article-title>. <source>Linguistics</source>, <volume>58</volume>(<issue>5</issue>), <fpage>1195</fpage>&#8211;<lpage>1232</lpage>. DOI: <pub-id pub-id-type="doi">10.1515/ling-2020-0201</pub-id></mixed-citation></ref>
<ref id="B52"><label>52</label><mixed-citation publication-type="journal"><string-name><surname>Martin</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Demirdache</surname>, <given-names>H.</given-names></string-name>, <string-name><surname>del Real</surname>, <given-names>I. G.</given-names></string-name>, <string-name><surname>Van Hout</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Kazanina</surname>, <given-names>N.</given-names></string-name> (<year>2020</year>). <article-title>Children&#8217;s non-adultlike interpretations of telic predicates across languages</article-title>. <source>Linguistics</source>, <volume>58</volume>(<issue>5</issue>), <fpage>1447</fpage>&#8211;<lpage>1500</lpage>. DOI: <pub-id pub-id-type="doi">10.1515/ling-2020-0182</pub-id></mixed-citation></ref>
<ref id="B53"><label>53</label><mixed-citation publication-type="webpage"><string-name><surname>Minor</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Mitrofanova</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Ramchand</surname>, <given-names>G.</given-names></string-name> (<year>2022a</year>). <article-title>Fine-grained time course of verb aspect processing</article-title>. <uri>https://drive.google.com/file/d/1BUExbQzd2fbllbr80O9reJRmjaFZeFjB/view?usp=sharing</uri>. DOI: <pub-id pub-id-type="doi">10.1371/journal.pone.0264132</pub-id></mixed-citation></ref>
<ref id="B54"><label>54</label><mixed-citation publication-type="webpage"><string-name><surname>Minor</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Mitrofanova</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Guajardo</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Vos</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Ramchand</surname>, <given-names>G.</given-names></string-name> (<year>2022b</year>). <article-title>Temporal information and event bounding across languages: Evidence from visual world eyetracking</article-title>. <source>Talk at Semantics and Linguistic Theory 32</source>. <uri>https://osf.io/tv3b8/</uri>.</mixed-citation></ref>
<ref id="B55"><label>55</label><mixed-citation publication-type="journal"><string-name><surname>Ooms</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Dupont</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Lapon</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name><surname>Popelka</surname>, <given-names>S.</given-names></string-name> (<year>2015</year>). <article-title>Accuracy and precision of fixation locations recorded with the low-cost Eye Tribe tracker in different experimental setups</article-title>. <source>JOURNAL OF EYE MOVEMENT RESEARCH</source>, <volume>8</volume>(<issue>1</issue>), <fpage>20</fpage>. DOI: <pub-id pub-id-type="doi">10.16910/jemr.8.1.5</pub-id></mixed-citation></ref>
<ref id="B56"><label>56</label><mixed-citation publication-type="journal"><string-name><surname>Ouzts</surname>, <given-names>A. D.</given-names></string-name>, &amp; <string-name><surname>Duchowski</surname>, <given-names>A. T.</given-names></string-name> (<year>2012</year>). <article-title>Comparison of eye movement metrics recorded at different sampling rates</article-title>. In <source>Proceedings of the symposium on eye tracking research and applications</source> (pp. <fpage>321</fpage>&#8211;<lpage>324</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/2168556.2168626</pub-id></mixed-citation></ref>
<ref id="B57"><label>57</label><mixed-citation publication-type="journal"><string-name><surname>Papoutsaki</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Laskey</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Huang</surname>, <given-names>J.</given-names></string-name> (<year>2017</year>). <article-title>Searchgazer: Webcam eye tracking for remote studies of web search</article-title>. In <source>Proceedings of the 2017 conference on conference human information interaction and retrieval</source> (pp. <fpage>17</fpage>&#8211;<lpage>26</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/3020165.3020170</pub-id></mixed-citation></ref>
<ref id="B58"><label>58</label><mixed-citation publication-type="webpage"><string-name><surname>Papoutsaki</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Sangkloy</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Laskey</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Daskalova</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Huang</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Hays</surname>, <given-names>J.</given-names></string-name> (<year>2016</year>). <chapter-title>Webgazer: Scalable webcam eye tracking using user interactions</chapter-title>. In <source>Proceedings of the 25th international joint conference on artificial intelligence (ijcai)</source> (pp. <fpage>3839</fpage>&#8211;<lpage>3845</lpage>). <publisher-name>AAAI</publisher-name>. <uri>https://par.nsf.gov/servlets/purl/10024076</uri>.</mixed-citation></ref>
<ref id="B59"><label>59</label><mixed-citation publication-type="journal"><string-name><surname>Peelle</surname>, <given-names>J. E.</given-names></string-name>, &amp; <string-name><surname>Van Engen</surname>, <given-names>K. J.</given-names></string-name> (<year>2021</year>). <article-title>Time stand still: Effects of temporal window selection on eye tracking analysis</article-title>. <source>Collabra: Psychology</source>, <volume>7</volume>(<issue>1</issue>), <fpage>25961</fpage>. DOI: <pub-id pub-id-type="doi">10.1525/collabra.25961</pub-id></mixed-citation></ref>
<ref id="B71"><label>71</label><mixed-citation publication-type="journal"><collab>R Core Team</collab>. (<year>2019</year>). <article-title>R: A language and environment for statistical computing (version 3.6.1)[computer software]. r foundation for statistical computing</article-title>. <source>Vienna, Austria</source>.</mixed-citation></ref>
<ref id="B60"><label>60</label><mixed-citation publication-type="journal"><string-name><surname>Rosenholtz</surname>, <given-names>R.</given-names></string-name> (<year>2016</year>). <article-title>Capabilities and limitations of peripheral vision</article-title>. <source>Annual review of vision science</source>, <volume>2</volume>, <fpage>437</fpage>&#8211;<lpage>457</lpage>. DOI: <pub-id pub-id-type="doi">10.1146/annurev-vision-082114-035733</pub-id></mixed-citation></ref>
<ref id="B61"><label>61</label><mixed-citation publication-type="journal"><string-name><surname>Salvucci</surname>, <given-names>D. D.</given-names></string-name>, &amp; <string-name><surname>Goldberg</surname>, <given-names>J. H.</given-names></string-name> (<year>2000</year>). <article-title>Identifying fixations and saccades in eye-tracking protocols</article-title>. In <source>Proceedings of the 2000 symposium on eye tracking research &amp; applications</source> (pp. <fpage>71</fpage>&#8211;<lpage>78</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/355017.355028</pub-id></mixed-citation></ref>
<ref id="B62"><label>62</label><mixed-citation publication-type="journal"><string-name><surname>Saslow</surname>, <given-names>M. G.</given-names></string-name> (<year>1967</year>). <article-title>Effects of components of displacement-step stimuli upon latency for saccadic eye movement</article-title>. <source>Josa</source>, <volume>57</volume>(<issue>8</issue>), <fpage>1024</fpage>&#8211;<lpage>1029</lpage>. DOI: <pub-id pub-id-type="doi">10.1364/JOSA.57.001024</pub-id></mixed-citation></ref>
<ref id="B63"><label>63</label><mixed-citation publication-type="journal"><string-name><surname>Sauter</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Draschkow</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Mack</surname>, <given-names>W.</given-names></string-name> (<year>2020</year>). <article-title>Building, hosting and recruiting: A brief introduction to running behavioral experiments online</article-title>. <source>Brain sciences</source>, <volume>10</volume>(<issue>4</issue>), <fpage>251</fpage>. DOI: <pub-id pub-id-type="doi">10.3390/brainsci10040251</pub-id></mixed-citation></ref>
<ref id="B64"><label>64</label><mixed-citation publication-type="journal"><string-name><surname>Semmelmann</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Weigelt</surname>, <given-names>S.</given-names></string-name> (<year>2018</year>). <article-title>Online webcam-based eye tracking in cognitive science: A first look</article-title>. <source>Behavior Research Methods</source>, <volume>50</volume>(<issue>2</issue>), <fpage>451</fpage>&#8211;<lpage>465</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-017-0913-7</pub-id></mixed-citation></ref>
<ref id="B65"><label>65</label><mixed-citation publication-type="webpage"><string-name><surname>Slim</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Hartsuiker</surname>, <given-names>R.</given-names></string-name> (<year>2021a</year>). <chapter-title>Online visual world eye-tracking using webcams</chapter-title>. In <source>Architectures and mechanisms for language processing</source>. <publisher-loc>Paris, France</publisher-loc>. <uri>https://amlap2021.github.io/program/148.pdf</uri>.</mixed-citation></ref>
<ref id="B66"><label>66</label><mixed-citation publication-type="journal"><string-name><surname>Slim</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Hartsuiker</surname>, <given-names>R.</given-names></string-name> (<year>2021b</year>). <article-title>Visual world eyetracking using webgazer.js</article-title>. DOI: <pub-id pub-id-type="doi">10.31234/osf.io/5adgf</pub-id></mixed-citation></ref>
<ref id="B67"><label>67</label><mixed-citation publication-type="journal"><string-name><surname>Slote</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Strand</surname>, <given-names>J. F.</given-names></string-name> (<year>2016</year>). <article-title>Conducting spoken word recognition research online: Validation and a new timing method</article-title>. <source>Behavior Research Methods</source>, <volume>48</volume>(<issue>2</issue>), <fpage>553</fpage>&#8211;<lpage>566</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-015-0599-7</pub-id></mixed-citation></ref>
<ref id="B68"><label>68</label><mixed-citation publication-type="journal"><string-name><surname>Sogo</surname>, <given-names>H.</given-names></string-name> (<year>2013</year>). <article-title>Gazeparser: an open-source and multiplatform library for low-cost eye tracking and analysis</article-title>. <source>Behavior research methods</source>, <volume>45</volume>(<issue>3</issue>), <fpage>684</fpage>&#8211;<lpage>695</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-012-0286-x</pub-id></mixed-citation></ref>
<ref id="B69"><label>69</label><mixed-citation publication-type="journal"><string-name><surname>Tafaj</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Kasneci</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Rosenstiel</surname>, <given-names>W.</given-names></string-name>, &amp; <string-name><surname>Bogdan</surname>, <given-names>M.</given-names></string-name> (<year>2012</year>). <article-title>Bayesian online clustering of eye movement data</article-title>. In <source>Proceedings of the symposium on eye tracking research and applications</source> (<fpage>285</fpage>&#8211;<lpage>288</lpage>). DOI: <pub-id pub-id-type="doi">10.1145/2168556.2168617</pub-id></mixed-citation></ref>
<ref id="B70"><label>70</label><mixed-citation publication-type="journal"><string-name><surname>Tanenhaus</surname>, <given-names>M. K.</given-names></string-name>, &amp; <string-name><surname>Soivey-Knowlton</surname>, <given-names>M. J.</given-names></string-name>, &amp; <string-name><surname>Eberhard</surname>, <given-names>K. M.</given-names></string-name>, &amp; <string-name><surname>Sedivy</surname>, <given-names>J. C.</given-names></string-name> (<year>1995</year>). <article-title>Integration of visual and linguistic information in spoken language comprehension</article-title>. <source>Science</source>, <volume>268</volume>(<issue>5217</issue>), <fpage>1632</fpage>&#8211;<lpage>1634</lpage>. DOI: <pub-id pub-id-type="doi">10.1126/science.7777863</pub-id></mixed-citation></ref>
<ref id="B72"><label>72</label><mixed-citation publication-type="book"><string-name><surname>Van Hout</surname>, <given-names>A.</given-names></string-name> (<year>2011</year>). <chapter-title>Past tense interpretations in dutch</chapter-title>. In <source>Organizing grammar</source> (pp. <fpage>241</fpage>&#8211;<lpage>251</lpage>). <publisher-name>De Gruyter Mouton</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1515/9783110892994</pub-id></mixed-citation></ref>
<ref id="B73"><label>73</label><mixed-citation publication-type="journal"><string-name><surname>van Hout</surname>, <given-names>A.</given-names></string-name> (<year>2018</year>). <article-title>On the acquisition of event culmination</article-title>. <source>Semantics in language acquisition</source> (pp. <fpage>96</fpage>&#8211;<lpage>121</lpage>). DOI: <pub-id pub-id-type="doi">10.1075/tilar.24.05hou</pub-id></mixed-citation></ref>
<ref id="B74"><label>74</label><mixed-citation publication-type="book"><string-name><surname>Vendler</surname>, <given-names>Z.</given-names></string-name> (<year>1967</year>). <source>Linguistics in philosophy</source>. <publisher-loc>Ithaca, N.Y.</publisher-loc>: <publisher-name>Cornell University Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.7591/9781501743726</pub-id></mixed-citation></ref>
<ref id="B75"><label>75</label><mixed-citation publication-type="webpage"><string-name><surname>Xu</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Ehinger</surname>, <given-names>K. A.</given-names></string-name>, <string-name><surname>Zhang</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Finkelstein</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Kulkarni</surname>, <given-names>S. R.</given-names></string-name>, &amp; <string-name><surname>Xiao</surname>, <given-names>J.</given-names></string-name> (<year>2015</year>). <article-title>Turkergaze: Crowdsourcing saliency with webcam based eye tracking</article-title>. <source>arXiv preprint arXiv:1504.06755</source> <uri>https://arxiv.org/abs/1504.06755</uri>.</mixed-citation></ref>
<ref id="B76"><label>76</label><mixed-citation publication-type="journal"><string-name><surname>Yang</surname>, <given-names>W.</given-names></string-name>, <string-name><surname>Chan</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Chang</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Kidd</surname>, <given-names>E.</given-names></string-name> (<year>2020</year>). <article-title>Four-year-old mandarin-speaking children&#8217;s online comprehension of relative clauses</article-title>. <source>Cognition</source>, <volume>196</volume>, <fpage>104103</fpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2019.104103</pub-id></mixed-citation></ref>
<ref id="B77"><label>77</label><mixed-citation publication-type="webpage"><string-name><surname>Yang</surname>, <given-names>X.</given-names></string-name>, &amp; <string-name><surname>Krajbich</surname>, <given-names>I.</given-names></string-name> (<year>2021</year>). <article-title>Webcam-based online eye-tracking for behavioral research</article-title>. <source>Judgment and Decision Making</source>, <volume>16</volume>(<issue>6</issue>), <fpage>1486</fpage>. <uri>http://journal.sjdm.org/21/210525/jdm210525.pdf</uri>.</mixed-citation></ref>
<ref id="B78"><label>78</label><mixed-citation publication-type="journal"><string-name><surname>Zhou</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Crain</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Zhan</surname>, <given-names>L.</given-names></string-name> (<year>2014</year>). <article-title>Grammatical aspect and event recognition in children&#8217;s online sentence comprehension</article-title>. <source>Cognition</source>, <volume>133</volume>(<issue>1</issue>), <fpage>262</fpage>&#8211;<lpage>276</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2014.06.018</pub-id></mixed-citation></ref>
</ref-list>
</back>
</article>