<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<!--<?xml-stylesheet type="text/xsl" href="article.xsl"?>-->
<article article-type="research-article" dtd-version="1.2" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="issn">2767-0279</journal-id>
<journal-title-group>
<journal-title>Glossa Psycholinguistics</journal-title>
</journal-title-group>
<issn pub-type="epub">2767-0279</issn>
<publisher>
<publisher-name>eScholarship Publishing</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.5070/G6011165</article-id>
<article-categories>
<subj-group>
<subject>Registered report</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Listeners&#8217; convergence towards an artificial agent in a joint phoneme categorization task</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<name>
<surname>Nguyen</surname>
<given-names>No&#235;l</given-names>
</name>
<email>noel.nguyen-trong@univ-amu.fr</email>
<xref ref-type="aff" rid="aff-1">1</xref>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Lancia</surname>
<given-names>Leonardo</given-names>
</name>
<email>leonardo.lancia@cnrs.fr</email>
<xref ref-type="aff" rid="aff-1">1</xref>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Huttner</surname>
<given-names>Lena</given-names>
</name>
<email>lena-marie.huttner@univ-amu.fr</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Schwartz</surname>
<given-names>Jean-Luc</given-names>
</name>
<email>jean-luc.schwartz@gipsa-lab.grenoble-inp.fr</email>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Diard</surname>
<given-names>Julien</given-names>
</name>
<email>julien.diard@univ-grenoble-alpes.fr</email>
<xref ref-type="aff" rid="aff-4">4</xref>
</contrib>
</contrib-group>
<aff id="aff-1"><label>1</label>Aix Marseille Univ, CNRS, LPL, Aix-en-Provence, France</aff>
<aff id="aff-2"><label>2</label>Institute for Language, Communication and the Brain, Aix Marseille University, France</aff>
<aff id="aff-3"><label>3</label>Universit&#233; Grenoble Alpes, CNRS, GIPSA-Lab, Grenoble, France</aff>
<aff id="aff-4"><label>4</label>Universit&#233; Grenoble Alpes, CNRS, Laboratoire de Psychologie et NeuroCognition, Grenoble, France</aff>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2024-03-04">
<day>04</day>
<month>03</month>
<year>2024</year>
</pub-date>
<pub-date pub-type="collection">
<year>2024</year>
</pub-date>
<volume>3</volume>
<issue>1</issue>
<elocation-id>20</elocation-id>
<permissions>
<copyright-statement>Copyright: &#x00A9; 2024 The Author(s)</copyright-statement>
<copyright-year>2024</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. See <uri xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</uri>.</license-p>
</license>
</permissions>
<self-uri xlink:href="https://glossapsycholinguistics.journalpub.escholarship.org/articles/10.5070/G6011165/"/>
<abstract>
<p>This study focuses on inter-individual convergence effects in the perception and categorization of speech sounds. We ask to what extent two listeners can come to establish a shared set of categorization criteria in a phoneme identification task that they accomplish together. Several hypotheses are laid out in the framework of a Bayesian model of speech perception that we have developed to account for how two listeners may each infer the parameters that govern their partner&#8217;s responses. In our experimental paradigm, participants were asked to perform a joint phoneme identification task with a partner that, unbeknownst to them, was an artificial agent, whose responses we manipulated along two dimensions, the location of the categorical boundary and the slope of the identification function. Convergence was found to arise for bias but not for slope. Numerical simulations suggested that lack of convergence in slope may stem from the listeners&#8217; prior level of confidence in the variance in VOT for the two phonemic categories. This study sheds new light on perceptual convergence between listeners in the categorization of speech sounds, a phenomenon that has received little attention so far in spite of its central importance for speech communication.</p>
</abstract>
</article-meta>
</front>
<body>
<sec>
<title>1. Introduction</title>
<p>In spoken language interactions, speech sounds must be perceptually categorized consistently across talkers for these to understand each other. One key objective in speech communication research is to explain how people can perceive speech sounds in a way that is similar enough to ensure mutual understanding. To achieve this remarkable feat, talkers must share a set of conventions on how to map speech sounds onto linguistically relevant categories.<xref ref-type="fn" rid="n1">1</xref> These conventions result from countless inter-individual interactions over entire generations of talkers, which resonate within each individual whenever she recognizes a vowel or consonant in the speech stream. The goal of this study was to contribute to the characterization of the cognitive mechanisms that preside over the formation of this shared perceptual space.</p>
<p>Computational models of the emergence of language (e.g., <xref ref-type="bibr" rid="B19">De Boer, 2000</xref>; <xref ref-type="bibr" rid="B62">Moulin-Frier et al., 2015</xref>) have shown that speech sound systems can arise at a collective scale as the byproduct of pairwise communication between agents. According to these models, local, unidirectional or bidirectional communicative exchanges engender the gradual formation of a globally shared speech code. In the COSMO model (<xref ref-type="bibr" rid="B62">Moulin-Frier et al., 2015</xref>) for example, a common repertoire of linguistic units for referring to objects progressively forms itself in a group of communicating agents through sequences of sensorimotor operations performed by pairs of agents. Likewise, in De Boer&#8217;s (<xref ref-type="bibr" rid="B19">2000</xref>) model, vowel systems emerge by virtue of a self-organization process from pairwise interactions between agents, each of which has to imitate the sounds produced by the other. In the experimental domain, researchers have used innovative designs to identify the conditions that may account for the emergence of language (<xref ref-type="bibr" rid="B84">Scott-Phillips &amp; Kirby, 2010</xref>; <xref ref-type="bibr" rid="B89">Verhoef et al., 2014</xref>) and these researchers too regard pairwise communicative exchanges as the building blocks upon which linguistic systems can deploy themselves. This study was carried out in the framework of a project that seeks to determine how shared conventions may arise in the perception of speech as a result of inter-individual communicative exchanges. We more specifically aimed to answer the following question: When communicating with one another, to what extent do people converge towards each other in the way they categorize speech sounds?</p>
<p>Much attention has been paid over the last two decades or so to inter-individual adaptation mechanisms in the production and perception of speech. One key mechanism is phonetic convergence, i.e., the tendency for a talker to partly imitate another person&#8217;s way of producing speech sounds when exposed to that person&#8217;s speech (<xref ref-type="bibr" rid="B6">Babel, 2011</xref>). Ever since Goldinger&#8217;s (<xref ref-type="bibr" rid="B33">1998</xref>) and Pardo&#8217;s (<xref ref-type="bibr" rid="B73">2006</xref>) seminal studies, phonetic convergence effects between speakers have been explored in both interactive (e.g., <xref ref-type="bibr" rid="B1">Abel &amp; Babel, 2016</xref>; <xref ref-type="bibr" rid="B40">Kim et al., 2011</xref>; <xref ref-type="bibr" rid="B73">Pardo, 2006</xref>) and non-interactive, laboratory (e.g., <xref ref-type="bibr" rid="B22">Delvaux &amp; Soquet, 2007</xref>; <xref ref-type="bibr" rid="B33">Goldinger, 1998</xref>; <xref ref-type="bibr" rid="B68">Nielsen, 2011</xref>) settings, by means of direct, acoustic measures (e.g., <xref ref-type="bibr" rid="B34">Harrington et al., 2019</xref>; <xref ref-type="bibr" rid="B63">Mukherjee et al., 2019</xref>; <xref ref-type="bibr" rid="B100">Zellou et al., 2016</xref>), indirect, perceptual evaluations performed by listeners (<xref ref-type="bibr" rid="B23">Dias &amp; Rosenblum, 2016</xref>; <xref ref-type="bibr" rid="B60">Miller et al., 2013</xref>; <xref ref-type="bibr" rid="B73">Pardo, 2006</xref>), or both (e.g., <xref ref-type="bibr" rid="B16">Clopper &amp; Dossey, 2020</xref>; <xref ref-type="bibr" rid="B74">Pardo et al., 2013a</xref>, <xref ref-type="bibr" rid="B75">2013b</xref>; see <xref ref-type="bibr" rid="B76">Pardo et al., 2017</xref>, for a review). Convergence effects in a talker as a consequence of her being exposed to other people&#8217;s speech may extend well beyond that exposure across the talker&#8217;s lifetime (<xref ref-type="bibr" rid="B35">Harrington et al., 2000</xref>), and it has also been assumed to play a central role at yet a larger time scale in the emergence and evolution of phonological systems (see <xref ref-type="bibr" rid="B66">Nguyen &amp; Delvaux, 2015</xref>, for a review). A central issue for the present study is whether convergence in speech production entails convergence in perception. In Pickering and Garrod&#8217;s integrated theory of language production and comprehension (<xref ref-type="bibr" rid="B78">Pickering &amp; Garrod, 2013</xref>, <xref ref-type="bibr" rid="B79">2021</xref>), imitating the interlocutor&#8217;s way of speaking contributes to making it easier to understand what that person is saying and to predict what she will say next. It may be assumed that convergence between talkers in production causes each talker to become more attuned to the phonetic characteristics of words produced by the other talker, via a perception-action resonance phenomenon. However, whether this may result in both talkers categorizing speech sounds in a more similar manner appears to remain an open question.</p>
<p>To study convergence in perception as a potential correlate of convergence in production, experiments must by definition combine convergence-in-production with convergence-in-perception tasks. To our knowledge, there have been very few studies in that category. Adank et al. (<xref ref-type="bibr" rid="B2">2010</xref>) exposed Dutch-speaking participants to a novel, artificially-created accent under different conditions during a training phase, and assessed comprehension of the accent before and after training (by measuring the signal-to-noise ratio at which listeners could repeat 50% of the key words in sentences heard with background noise). The results showed that accented speech comprehension was improved after training for participants whose task was to imitate the speaker&#8217;s accent in the training phase (but not for those who had to listen to the accented sentences, or to listen and transcribe them, or to listen and repeat them in the participant&#8217;s own accent, during training). In Nguyen et al.&#8217;s (<xref ref-type="bibr" rid="B67">2012</xref>) experiment, however, phonetic imitation did not have a significant impact on how listeners later recognized words in a non-native regional accent. The authors suggested that phonetic convergence may contribute to predicting upcoming words in sentences in adverse listening conditions, in accord with Adank et al.&#8217;s (<xref ref-type="bibr" rid="B2">2010</xref>) findings, but may play a more limited role in the recognition of single words. Importantly, both Adank et al.&#8217;s (<xref ref-type="bibr" rid="B2">2010</xref>) and Nguyen et al.&#8217;s (<xref ref-type="bibr" rid="B67">2012</xref>) studies examined whether phonetic convergence towards a model speaker can facilitate understanding that speaker, but did not ask whether convergence in production implies, or is conducive to, convergence in perception.</p>
<p>In recent work, Lancia &amp; Nguyen (<xref ref-type="bibr" rid="B50">2019</xref>) and Huttner &amp; Nguyen (<xref ref-type="bibr" rid="B36">2023</xref>) explored potential convergence effects in perception from a different angle. In both studies, the authors&#8217; goal was to find a way to provoke these effects regardless of whether they may or may not be connected with convergence in overt production. To do this, the authors used a joint-perception paradigm. Participants were asked to perform a phoneme identification task in a joint fashion and were explicitly instructed to respond in the same way as their partner(s). On each trial, each participant first responded individually to the stimulus, then was shown the response(s) of the other participant(s). The results showed that participants tended to increasingly agree with each other on how to categorize the stimuli as the experiment unfolded.</p>
<p>Joint perception appears to be a still developing field, but a very promising one. It has been mostly explored in the visual domain (<xref ref-type="bibr" rid="B7">Bahrami et al., 2010</xref>; <xref ref-type="bibr" rid="B44">Koriat, 2012</xref>; <xref ref-type="bibr" rid="B82">Richardson et al., 2012</xref>; <xref ref-type="bibr" rid="B85">Seow &amp; Fleming, 2019</xref>; <xref ref-type="bibr" rid="B87">Sorkin et al., 2001</xref>; <xref ref-type="bibr" rid="B91">Wahn et al., 2018</xref>). In these studies, one of the main objectives has been to determine whether &#8220;two heads are better than one&#8221;, i.e., to what extent two people that communicate with each other do better than individuals in perceptual decision-making tasks (<xref ref-type="bibr" rid="B7">Bahrami et al., 2010</xref>; <xref ref-type="bibr" rid="B44">Koriat, 2012</xref>; <xref ref-type="bibr" rid="B87">Sorkin et al., 2001</xref>) and, for more than two people, whether there is a group benefit in the accomplishment of these tasks (<xref ref-type="bibr" rid="B91">Wahn et al., 2018</xref>). Another central objective is to characterize the effect of social context on perceptual decision-making (<xref ref-type="bibr" rid="B82">Richardson et al., 2012</xref>; <xref ref-type="bibr" rid="B85">Seow &amp; Fleming, 2019</xref>), in a perspective that can be traced back to Asch&#8217;s (<xref ref-type="bibr" rid="B5">1951</xref>) landmark work. These studies all used tasks in which responses, whether produced by one or more people, can be classified as correct or incorrect. In Bahrami et al. (<xref ref-type="bibr" rid="B7">2010</xref>) for example, participants judged which of two briefly presented visual stimuli contained an oddball target. Lancia &amp; Nguyen (<xref ref-type="bibr" rid="B50">2019</xref>) and Huttner &amp; Nguyen (<xref ref-type="bibr" rid="B36">2023</xref>) applied the joint-perception paradigm to speech in a way that is novel in two respects: First, they extended this paradigm to perception of auditory speech. Second, and in both studies, pairs of participants were presented with speech sounds that ranged on an acoustic continuum between two endpoints associated with two different phonemic categories (/s/ and /&#643;/), as in a standard phoneme identification task. In such a task, and as is well known, stimuli between the two endpoints are perceived as ambiguous to various degrees, and there is no a priori correct response. These two studies therefore did not aim to determine whether performance increased when listeners did the task in pairs rather than individually. Rather, they asked to what extent two listeners can come to use the same criteria in mapping speech sounds onto phoneme categories.</p>
<p>Our research project stands across the modeling and experimental domains. We aim to develop a Bayesian model of convergence between listeners in speech perception, which we put to the test using novel experimental designs. The present work formed a first step towards this goal.</p>
<p>In the next three sections, we first lay out an initial version of the model (Section 2) and the predictions that can be made from it (Section 3). We then present our experiment (Section 4), which, building on Lancia &amp; Nguyen (<xref ref-type="bibr" rid="B50">2019</xref>), entailed participants performing a joint phoneme identification task with a partner. Unlike in this previous study, however, and unbeknownst to the participants, their partner was an artificial agent whose responses we manipulated in order to examine their effects on the participants&#8217; own responses. Our results are presented in Section 5. This is followed by the presentation of a set of simulations that we conducted with a view to accounting for these results (Section 6), and a general discussion (Section 7).</p>
</sec>
<sec>
<title>2 Towards modeling convergence between listeners in speech perception</title>
<sec>
<title>2.1 Theoretical background</title>
<p>As already indicated, we have undertaken to design our model in a Bayesian framework. Bayesian models, in the study of cognition and of the human brain in general, have had widespread application and success in the last decades. They are found at most description levels, from probabilistic computational neuroscience (e.g., <xref ref-type="bibr" rid="B27">Friston, 2010</xref>; <xref ref-type="bibr" rid="B81">Pouget et al., 2013</xref>), to probabilistic models implementing psychologically based or neuro-plausible theories of sensory processing (e.g., <xref ref-type="bibr" rid="B13">Chikkerur et al., 2010</xref>; <xref ref-type="bibr" rid="B32">Ginestet et al., 2022</xref>; <xref ref-type="bibr" rid="B52">Laurent et al., 2017</xref>; <xref ref-type="bibr" rid="B98">Yu et al., 2009</xref>), to computational-level accounts of cognitive functions (e.g., <xref ref-type="bibr" rid="B9">Brainard &amp; Freeman, 1997</xref>; <xref ref-type="bibr" rid="B39">Kersten et al., 2004</xref>; <xref ref-type="bibr" rid="B92">Weiss et al., 2002</xref>). They even connect seamlessly, and sometimes are close mathematical cousins to statistical methods and tools (e.g., <xref ref-type="bibr" rid="B15">Clayton, 2021</xref>; <xref ref-type="bibr" rid="B18">Dayan &amp; Abbott, 2001</xref>; <xref ref-type="bibr" rid="B53">Ma, 2012</xref>). They have also percolated through almost all subdomains of cognitive science, concerning most if not all sensory modalities and their combinations (e.g., <xref ref-type="bibr" rid="B3">Alais &amp; Burr, 2004</xref>; <xref ref-type="bibr" rid="B24">Ernst &amp; Banks, 2002</xref>; <xref ref-type="bibr" rid="B29">Geisler, 2008</xref>; <xref ref-type="bibr" rid="B56">Mamassian et al., 2003</xref>; <xref ref-type="bibr" rid="B92">Weiss et al., 2002</xref>; <xref ref-type="bibr" rid="B96">Wozny et al., 2008</xref>; <xref ref-type="bibr" rid="B99">Yuille &amp; Kersten, 2006</xref>; <xref ref-type="bibr" rid="B101">Zupan et al., 2002</xref>), abstract reasoning and learning (e.g., <xref ref-type="bibr" rid="B12">Chater et al., 2010</xref>; <xref ref-type="bibr" rid="B88">Tenenbaum et al., 2011</xref>), metacognition (<xref ref-type="bibr" rid="B26">Fleming &amp; Daw, 2017</xref>), motor control (e.g., <xref ref-type="bibr" rid="B77">Patri et al., 2018</xref>; <xref ref-type="bibr" rid="B94">Wolpert, 2007</xref>; <xref ref-type="bibr" rid="B95">Wolpert &amp; Ghahramani, 2000</xref>). Whatever their epistemological or ontological flavors, the common denominator of Bayesian models of cognition is their use of probabilities to model uncertain knowledge of the cognitive agent, and the use of Bayesian inference to model reasoning in the presence of incomplete and uncertain information (<xref ref-type="bibr" rid="B8">Bessi&#232;re et al., 2013</xref>; <xref ref-type="bibr" rid="B37">Jaynes, 2003</xref>). Such characteristics can be considered as key to model human or animal cognition, which have to reason with incomplete and uncertain knowledge. In the field of speech and language, work by Norris &amp; McQueen (<xref ref-type="bibr" rid="B69">2008</xref>) and Norris et al. (<xref ref-type="bibr" rid="B71">2016</xref>) (speech recognition), Sohoglu &amp; Davis (<xref ref-type="bibr" rid="B86">2020</xref>) (brain underpinnings of speech perception), Xu &amp; Tenenbaum (<xref ref-type="bibr" rid="B97">2007</xref>) (word learning), Carr et al. (<xref ref-type="bibr" rid="B11">2020</xref>) (language evolution), and Moulin-Frier et al. (<xref ref-type="bibr" rid="B62">2015</xref>) (emergence of phonological systems), among others, have shown the fertility of Bayesian approaches and the broad perspectives they offer.</p>
<p>In their application to speech, Bayesian approaches allow us to model phoneme identification as a probabilistic process that mathematically takes into account sensory and categorical uncertainty. They also make it possible to evaluate to what extent the listener&#8217;s <italic>prior beliefs</italic> come into play in her decision-making, along with the perceptual evidence that is available to her. In addition, and because Bayesian models, in essence, boil down to updating prior beliefs in the face of the available evidence, they are well fitted to modeling adaptation and learning processes. These are all characteristics that, in our view, are relevant to speech perception.</p>
<p>Our current model draws on previous work by Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>), Kronrod et al. (<xref ref-type="bibr" rid="B46">2016</xref>) and Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) (see also <xref ref-type="bibr" rid="B14">Clayards et al., 2008</xref>; <xref ref-type="bibr" rid="B41">Kleinschmidt, 2020</xref>). These models aim to account for how individual listeners identify speech sounds equally spaced along an acoustic dimension between two endpoints respectively and unambiguously associated with two phonemic categories in a two-alternative forced choice (2AFC) task. Taking these models as a starting point, we propose a simple extension to joint perception in dyads of listeners. In this first step, our goal is to model the listeners&#8217; asymptotic behavior. More specifically, we seek to determine the extent to which listeners may perceptually converge towards their partners, from the listeners&#8217; entire set of responses. We do not endeavor yet to model trial-by-trial changes in the listeners&#8217; response pattern.</p>
</sec>
<sec>
<title>2.2 The single-listener model</title>
<p>Like most Bayesian models of perception (<xref ref-type="bibr" rid="B31">Gifford et al., 2014</xref>; <xref ref-type="bibr" rid="B54">Ma et al., 2023</xref>; <xref ref-type="bibr" rid="B90">Vincent, 2015</xref>), ours has a generative (forward) component and an inferential (inverse) component. The generative component contains a characterization of how sounds distribute themselves in the acoustic domain for each category. This is specified by two conditional probability distributions, <italic>p</italic>(<italic>S&#124;c</italic><sub>1</sub>) and <italic>p</italic>(<italic>S&#124;c</italic><sub>2</sub>), where <italic>S</italic> refers to a representation of the acoustic space and <italic>c</italic><sub>1</sub> and <italic>c</italic><sub>2</sub> to the two categories, respectively. It is usually assumed that <italic>p</italic>(<italic>S&#124;c</italic><sub>1</sub>) and <italic>p</italic>(<italic>S&#124;c</italic><sub>2</sub>) are both normal distributions and, when <italic>S</italic> is assumed to be monodimensional, as is generally the case in 2AFC tasks, are each characterized by a mean and a variance:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq001-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant='script'>N</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant='script'>N</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M1">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
p(S|{c_1}) = {\mathcal N}({\mu _1},\sigma _{{c_1}}^2 + \sigma _S^2)\\
p(S|{c_2}) = {\mathcal N}({\mu _2},\sigma _{{c_2}}^2 + \sigma _S^2)
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e14.gif"/>
</alternatives>
</disp-formula>
<p>For each distribution, variance is a sum of two terms, <inline-formula>
<alternatives>
<mml:math id="Eq002-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M2">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _c^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e1.gif"/>
</alternatives>
</inline-formula>, a measure of dispersion of the intended target sound around the mean for the category, and <inline-formula>
<alternatives>
<mml:math id="Eq003-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M3">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _S^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e2.gif"/>
</alternatives>
</inline-formula>, which represents articulatory, acoustic and perceptual noise around the intended target sound independent of the category (<xref ref-type="bibr" rid="B25">Feldman et al., 2009</xref>; <xref ref-type="bibr" rid="B46">Kronrod et al., 2016</xref>).</p>
<p>As in both Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>) and Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>), the variances associated with the two categories are considered as equal:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq004-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M4">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _{{c_1}}^2 = \sigma _{{c_2}}^2 = \sigma _c^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e15.gif"/>
</alternatives>
</disp-formula>
<p>We further assume that the two distributions are in symmetric positions with respect to the midpoint of the continuum,<xref ref-type="fn" rid="n2">2</xref> i.e., at the same distance <italic>&#948;<sub>&#181;</sub></italic> from that midpoint, on either side of it. Both distributions are schematized in the middle panel of <xref ref-type="fig" rid="F1">Figure 1</xref>.</p>
<fig id="F1">
<caption>
<p><bold>Figure 1:</bold> Three main components of the single-listener model. Left panel: Bernoulli distribution associated with the prior probabilities <italic>p</italic>(<italic>c</italic><sub>1</sub>) and <italic>p</italic>(<italic>c</italic><sub>2</sub>) for the two phonemic categories. The two prior probabilities are here assumed to be equal, i.e., <italic>p</italic>(<italic>c</italic><sub>1</sub>) <italic>= p</italic>(<italic>c</italic><sub>2</sub>) = 0.5. Middle panel: Distributions of sounds in a one-dimensional acoustic space <italic>S</italic> for the two categories. Voice Onset Time (VOT), as one of the main acoustic cues to the /b/-/p/ phonemic contrast, is used here as the example for <italic>S</italic>. The /b/ category corresponds to shorter VOT values (on the left of the continuum) and the /p/ category to longer VOT values. Right panel: Posterior probability value for <italic>c</italic><sub>1</sub>, given the sound. Figure partly made using the R script associated with Kurumada &amp; Roettger (<xref ref-type="bibr" rid="B48">2022</xref>) and available at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/b75q9/">https://osf.io/b75q9/</ext-link>.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g1.png"/>
</fig>
<p>The generative component also includes the prior probabilities <italic>p</italic>(<italic>c</italic><sub>1</sub>) and <italic>p</italic>(<italic>c</italic><sub>2</sub>) for the two categories. This represents the listener&#8217;s prior beliefs, i.e., to what extent she expects the input sound to correspond to one category rather than the other, before hearing that sound. The prior probability can be related to the phoneme&#8217;s frequency of occurrence, among many other factors. Since there are two categories only, the prior follows a Bernoulli distribution Ber(<italic>p</italic>), i.e., <italic>p</italic>(<italic>c</italic><sub>1</sub>) <italic>= p</italic> and <italic>p</italic>(<italic>c</italic><sub>2</sub>) = 1 <italic>&#8211; p</italic> with 0 &#8804; <italic>p</italic> &#8804; 1 (see <xref ref-type="fig" rid="F1">Figure 1</xref>, left panel).</p>
<p>The model&#8217;s inferential component allows the probability value for each phoneme category given the input sound to be computed. This is done thanks to Bayes&#8217; theorem:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq005-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2009;</mml:mo><mml:mo>&#x007C;</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M5">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p({c_1}\,|S) = \frac{{p(S|{c_1})\:p({c_1})}}{{p(S|{c_1})\:p({c_1}) + p(S|{c_2})\:p({c_2})}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e16.gif"/>
</alternatives>
</disp-formula>
<p>which simplifies to (<xref ref-type="bibr" rid="B25">Feldman et al., 2009</xref>; <xref ref-type="bibr" rid="B42">Kleinschmidt &amp; Jaeger, 2015</xref>):</p>
<disp-formula>
<alternatives>
<mml:math id="Eq006-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2009;</mml:mo><mml:mo>&#x007C;</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M6">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p({c_1}\,|S) = \frac{1}{{1 + {e^{- gS + b}}}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e17.gif"/>
</alternatives>
</disp-formula>
<p>where, under both the same-variance and same-distance-from-mean assumptions (see Appendix 1 for further detail):</p>
<disp-formula>
<alternatives>
<mml:math id="Eq007-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M7">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
g = \frac{{{\mu _1} - {\mu _2}}}{{\sigma _c^2 + \sigma _S^2}}\\
b = \log \frac{{p({c_2})}}{{p({c_1})}} = \log \frac{{p({c_2})}}{{1 - p({c_2})}}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e18.gif"/>
</alternatives>
</disp-formula>
<p>The posterior <italic>p</italic>(<italic>c</italic><sub>1</sub><italic>&#124;S</italic>) thus takes the form of a logistic function governed by two parameters, <italic>g</italic> and <italic>b</italic>, as illustrated in the right panel of <xref ref-type="fig" rid="F1">Figure 1</xref>. The location of the categorical boundary along the continuum is given by <italic>b/g</italic>, and the slope<xref ref-type="fn" rid="n3">3</xref> of the logistic curve at this location is given by <italic>g/</italic>4. As we posit that <italic>&#956;</italic><sub>1</sub> is lower than <italic>&#956;</italic><sub>2</sub> and therefore that <italic>&#956;</italic><sub>1</sub> <italic>&#8211; &#956;</italic><sub>2</sub> is negative, so is <italic>g</italic>, and <italic>p</italic>(<italic>c</italic><sub>1</sub><italic>&#124;S</italic>) decreases as <italic>S</italic> gets closer to the endpoint associated with <italic>c</italic><sub>2</sub>.</p>
<p>The logistic curve gets steeper when the normal distributions for the two phoneme categories are further apart from each other (larger difference between <italic>&#956;</italic><sub>1</sub> and <italic>&#956;</italic><sub>2</sub>) and/or when these distributions are both narrower (lower category variance <inline-formula>
<alternatives>
<mml:math id="Eq008-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M8">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _c^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e1.gif"/>
</alternatives>
</inline-formula> and/or lower noise variance <inline-formula>
<alternatives>
<mml:math id="Eq009-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M9">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _S^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e2.gif"/>
</alternatives>
</inline-formula>). The parameter <italic>b</italic> assimilates to a log odds ratio, and is a measure of the relative prior probabilities of the two phoneme categories. When <italic>p</italic>(<italic>c</italic><sub>2</sub>) increases relative to <italic>p</italic>(<italic>c</italic><sub>1</sub>), this causes the categorical boundary to be shifted towards the <italic>c</italic><sub>1</sub> endpoint, i.e., corresponds to a greater bias for associating the input sound with <italic>c</italic><sub>2</sub>. In the following, we will refer to <italic>g</italic> as the Slope<xref ref-type="fn" rid="n4">4</xref> parameter and to <italic>b</italic> as the Bias parameter.</p>
</sec>
<sec>
<title>2.3 Extension to modeling perceptual convergence between listeners</title>
<p>We now turn to how this single-listener model can be extended to account for potential perceptual convergence effects across listeners. This is done in the context of a 2AFC phoneme identification task that is jointly performed by two listeners. On hearing each stimulus, listeners must try to predict their partner&#8217;s response and respond in the same way. Once both have responded, each listener&#8217;s response is communicated to the other listener. In such a task, we simply assume that each listener expects the other listener to behave like a Bayesian agent, and will undertake to infer the parameter distributions of her partner&#8217;s internal model, so as to get her own model to fit these distributions as well as possible. Inference is performed by each listener from the partner&#8217;s set of responses, and entails computing estimated distributions for both the Slope and Bias parameters.</p>
<p>For the listener, estimating the distribution of the Bias parameter amounts to asking herself whether her partner has a bias towards choosing one response over the other and, if so, which response and to what extent. To our knowledge, this issue has not been explored in previous work. There is, however, an extensive literature on post-perceptual biases in phoneme categorization tasks, on which we may rely to further characterize the potential impact of bias in our proposed experimental setting. In particular, both Connine &amp; Clifton (<xref ref-type="bibr" rid="B17">1987</xref>) and Pitt (<xref ref-type="bibr" rid="B80">1995</xref>) examined to what extent listeners were influenced by monetary payoff, by attributing them a reward or penalty depending on which phoneme category they chose in response to each stimulus. Connine &amp; Clifton (<xref ref-type="bibr" rid="B17">1987</xref>) and Pitt (<xref ref-type="bibr" rid="B80">1995</xref>) both found that the location of the categorical boundary shifted in accordance with monetary payoff. Because monetary payoff is an unequivocally post-perceptual bias and, in a joint phoneme categorization task, the partner&#8217;s bias can also be viewed as a post-perceptual one, links can be drawn between our experimental set-up and that of Connine &amp; Clifton (<xref ref-type="bibr" rid="B17">1987</xref>) and Pitt (<xref ref-type="bibr" rid="B80">1995</xref>), which we further develop below.</p>
<p>As indicated above, the Slope parameter depends on both the means of the normal distributions for the two phoneme categories and their variance, itself an addition of the category variance and noise variance. In the process of inferring phoneme categories from sounds, noise variance relates to trial-to-trial differences in the location of the stimulus in the acoustic space as perceived by the listener. It has been pointed out (e.g., <xref ref-type="bibr" rid="B38">Kapnoula et al., 2017</xref>; <xref ref-type="bibr" rid="B59">McMurray, 2022</xref>) that the 2AFC task does not allow noise variance to be disentangled from category variance, because listeners are requested to respond in a binary fashion. In our model, as in those proposed by Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>), Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) and Kronrod et al. (<xref ref-type="bibr" rid="B46">2016</xref>), category and noise variance are not estimated independently of each other. In a joint 2AFC task, therefore, we may seek to determine to what extent each listener is sensitive to noise+category variance as a whole, as reflected in the other listener&#8217;s response pattern.</p>
<p>McMurray (<xref ref-type="bibr" rid="B59">2022</xref>) underlines that categorical perception has long been seen as being characterized by a steep identification curve and that shallower curves were seen, by contrast, as indicative of decreased precision in listeners, due to an increased amount of sensory noise. Contrary to this traditional view, however, Kong &amp; Edwards (<xref ref-type="bibr" rid="B43">2016</xref>), Kapnoula et al. (<xref ref-type="bibr" rid="B38">2017</xref>), and Ou et al. (<xref ref-type="bibr" rid="B72">2021</xref>), among others, have argued that phoneme identification is intrinsically gradient and that gradiency contributes to making the speech perception system more efficient. It allows listeners to commit themselves to one phoneme category to a lesser degree in the face of more ambiguous stimuli, and make listeners more receptive to secondary cues to a phonemic contrast.</p>
<p>In that respect, Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) study is particularly relevant to our own piece of work. These authors asked to what extent listeners are sensitive to the variance of the probability distributions for voice onset time (VOT) as an acoustic cue to the voicing contrast, in word-initial bilabial voiced (e.g., <italic>beach</italic>) and voiceless (e.g., <italic>peach</italic>) stops. They exposed listeners to stimuli ranging on a VOT continuum between a voiced and a voiceless bilabial stop, and whose relative frequency of occurrence mirrored a mixture of two Gaussian distributions associated with the voiced and voiceless categories, respectively. Clayards et al. (<xref ref-type="bibr" rid="B14">2008</xref>) manipulated the variance of these distributions, which was either wide or narrow. In a Bayesian framework, as adopted by these authors, this amounted to manipulating the variance of the distributions <italic>p</italic>(<italic>S&#124;c</italic><sub>1</sub>) and <italic>p</italic>(<italic>S&#124;c</italic><sub>2</sub>) as defined above, where <italic>c</italic><sub>1</sub> and <italic>c</italic><sub>2</sub> refer to the voiced and voiceless stops, respectively. In accordance with the application of Bayes&#8217; theorem in the model, the posterior categorization curve <italic>p</italic>(<italic>c</italic><sub>1</sub><italic>&#124;S</italic>) was expected to be shallower in the wide-variance compared with the narrow-variance condition. Two groups of listeners performed a word-to-picture matching task in the wide-variance condition for one group and the narrow-variance condition for the other group, and their identification curves were consistent with this prediction.</p>
<p>In Clayards et al. &#8216;s (<xref ref-type="bibr" rid="B14">2008</xref>) study, the identification task was accomplished individually by each listener, and the variable of interest was the probability distribution of the stimulus in the acoustic space for each phoneme category. Our own focus is different and concerns the listener&#8217;s potential sensitivity to the probability distributions that may underlie another listener&#8217;s response pattern. However, Clayards et al.&#8217;s findings are of particular interest to us as they show that the listeners&#8217; degree of gradiency, or steepness as referred to here, is flexible and may change in an adaptive way and over a short time frame. Detail about Clayards et al.&#8217;s model parameters and obtained effect size is given in Appendix 2. In the following section, we present our experimental design and predictions.</p>
</sec>
</sec>
<sec>
<title>3 Experimental design and predictions</title>
<p>Our main goal was to determine to what extent listeners are sensitive to their partner&#8217;s bias and degree of gradiency in a joint phoneme identification task. More specifically, we sought to establish whether listeners would converge towards their partner in either or both of these two dimensions. To shed light on this issue, and unbeknownst to them, each participant performed the task not with another human participant, but with a virtual agent (a <italic>bot</italic>, hereafter). This allowed us to manipulate the bot&#8217;s response pattern in a systematic way and to examine to what extent these manipulations were mirrored in the participants&#8217; own response patterns.</p>
<p>Each of the two parameters took either of two values: Steep or Shallow for the Slope parameter, biased towards one or the other of two categories for the Bias parameter. This yielded four experimental conditions, which are illustrated in <xref ref-type="fig" rid="F2">Figure 2</xref>. There were four groups of participants, one for each condition. As also indicated in this figure, our experiment focused on the perception of the voicing contrast in stimuli ranging on a VOT continuum between a voiced bilabial stop and a voiceless one in syllable-initial position.</p>
<fig id="F2">
<caption>
<p><bold>Figure 2:</bold> Bot&#8217;s schematized response patterns in the four experimental conditions. Each plot shows the bot response probability for category <italic>c</italic><sub>1</sub>, <italic>p</italic>(<italic>c</italic>1<italic>&#124;S</italic>) (vertical axis) as a function of VOT on the /ba/-/pa/ continuum (horizontal axis). Categories <italic>c</italic><sub>1</sub> (/b/) and <italic>c</italic><sub>2</sub> (/p/) are associated with short and longer VOT values, respectively.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g2.png"/>
</fig>
<p>We put three main predictions to the test. Our first prediction was that participants would show a stronger bias towards the voiced category in the Voiced-Biased conditions compared with the Voiceless-Biased conditions. Our second prediction was that the slope of the participants&#8217; identification curves would be shallower in the Shallow conditions compared with the Steep conditions.</p>
<p>Our third prediction related to the link between Bias, on the one hand, and gradiency, on the other hand, in the setting of the location of the categorical boundary. As specified above (in 2.2), the location of the categorical boundary is given by the Bias-to-Slope ratio <italic>b/g</italic>. This means that, for the boundary to be shifted over a given interval across the continuum, Bias <italic>b</italic> must be modified to a lesser extent when the Slope <italic>g</italic> is shallower. This property was mentioned by Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>) in their single-listener model but, to our knowledge, it has not been empirically assessed yet. In our joint perception setting, we predicted that Bias would be larger in the establishment of the categorical boundary in the Shallow compared with the Steep condition.</p>
</sec>
<sec sec-type="methods">
<title>4. Method</title>
<sec>
<title>4.1 Materials</title>
<p>We built up a set of nine stimuli that ranged at equal intervals on a VOT continuum between /ba/ and /pa/. These stimuli originated from two natural tokens of /ba/ and /pa/, as spoken by a male native speaker of Southern British English. We used recordings that were made for a previous study in the sound-proof room of the Phonetics Laboratory of the University of Cambridge, UK, using high-quality equipment. The acoustic signal was low-pass filtered and digitized at a sampling rate of 16,000 Hz. We generated the stimuli from these two recordings by means of the progressive cutback and replacement method as implemented in Winn&#8217;s (<xref ref-type="bibr" rid="B93">2020</xref>) Praat script.<xref ref-type="fn" rid="n5">5</xref> VOT increased from 6 to 38 ms in 4-ms steps from Stimulus 1 to Stimulus 9. These values cover the range of VOT durations that have been used in previous experiments on the role of VOT in the perception of the voicing contrast in bilabial stops in English (e.g., <xref ref-type="bibr" rid="B14">Clayards et al., 2008</xref>; <xref ref-type="bibr" rid="B38">Kapnoula et al., 2017</xref>; <xref ref-type="bibr" rid="B72">Ou et al., 2021</xref>; <xref ref-type="bibr" rid="B93">Winn, 2020</xref>). We set the onset F0 frequency to a fixed value of 114 Hz for all stimuli, halfway between the onset F0 value for the original /ba/ (104 Hz) and that for the original /pa/ (126 Hz).</p>
<p>We conducted a preliminary test to assess the stimuli&#8217;s quality and ensure that the listeners&#8217; responses would show the expected pattern (continuum endpoints categorized as voiced and voiceless, respectively; categorical boundary in the vicinity of the continuum&#8217;s midpoint). The results, which overall confirmed that the stimuli were adequate for our proposed experiment, are presented in Appendix 3. However, because the average proportion of /ba/ responses across participants was close to 0 for both Stimulus 8 (VOT: 34 ms) and Stimulus 9 (VOT: 38 ms), we discarded Stimulus 9 and used Stimuli 1&#8211;8 only. Within that series of stimuli, the midpoint, arithmetically halfway between the two endpoints, was therefore located at 20 ms on the VOT scale.</p>
</sec>
<sec>
<title>4.2 Participants</title>
<p>We recruited 320 participants (balanced in gender, age range: 20&#8211;40 years old) online through the Prolific crowdsourcing website. An announcement was sent to Prolific-registered participants that responded to the following criteria: be born and live in England; have English as first language; have no or little proficiency in other languages; have no hearing difficulties and normal or corrected-to-normal vision; have access to a computer and headphones or earphones.</p>
<p>80 participants (40 female) were assigned to each of the four experimental groups. The number of participants was established on the basis of a data simulation, see Appendix 4. Each participant received a fee of 3 euros upon completion of the experiment.</p>
</sec>
<sec>
<title>4.3 Experimental design and set-up</title>
<p>The experiment was implemented by means of jsPsych (<xref ref-type="bibr" rid="B20">de Leeuw, 2015</xref>) and deployed on the MindProbe (mindprobe.eu) JATOS server. Participants were directed to MindProbe from Prolific and invited to take the experiment online through a web browser. They were asked to use a computer (as opposed to a tablet or smartphone), alone in a quiet room, and to wear headphones or earphones connected to their computer.</p>
<p>We first presented participants with a consent form that they were asked to digitally agree to, and in which we informed them that their responses would be recorded in a form that would not allow participants to be identified, and would only be used if they completed the experiment. Participants were also told that they could stop the experiment at any moment before the end.</p>
<p>Participants were then requested to take Milne et al.&#8217;s (<xref ref-type="bibr" rid="B61">2021</xref>) headphone screening test. Those who provided less than five correct responses out of the six trials were not allowed to continue and were replaced by other participants.</p>
<p>Next, participants were told that they would be presented with a sequence of speech sounds that may be identified as &#8220;ba&#8221; or &#8220;pa&#8221;. After hearing each sound, the participants&#8217; task was to say whether that sound corresponds to &#8220;ba&#8221; or &#8220;pa&#8221; by clicking on one of two buttons displayed on their computer screen. They were instructed to respond as both accurately and fast as possible, and to try to always provide a response even if in doubt. The respective positions of the &#8220;ba&#8221; and &#8220;pa&#8221; buttons on the screen was counterbalanced across participants.</p>
<p>In a first, training phase, participants individually performed the task on six sounds, which corresponded to either of the two VOT continuum endpoints, and were therefore expected to be unambiguously associated with /ba/ or /pa/ (three repetitions per endpoint, randomized order). Once having responded to each stimulus, participants were told whether or not their response was the correct one. Those who provided less than five correct responses out of the six trials were not allowed to continue and were replaced by other participants.</p>
<p>We then informed the participants that, in the following phase (referred to as the test phase hereafter), they would have to identify a sequence of English speech sounds as &#8220;ba&#8221; or &#8220;pa&#8221; again. Rather than performing the task individually, however, they had to do it together with another participant. This participant was presented to them as being, like them, a native speaker of English as spoken in England. Once having responded to each stimulus, they would be told whether their partner had provided the same response, or the opposite one. Participants were asked to aim to respond in the same way as their partner. Both the participant and her partner would earn one point if their responses were identical.</p>
<p>Participants heard ten repetitions of each of the eight auditory stimuli on the VOT continuum, in a fully randomized order. In both the training and test phase, and at the onset of each trial, a cross was displayed at the center of the screen for 750 ms. This was followed by the auditory stimulus and, simultaneously, the display of the two response buttons, labelled <italic>ba</italic> and <italic>pa</italic>, respectively, on either side of the screen center. Participants had 3,000 ms to respond. The partner&#8217;s response, as well as the cumulated number of points earned by both the participant and the partner from the onset of the test phase, were then shown on the screen for 2,500 ms. A 15-s pause was made at the end of the first half of the test. The test phase lasted about 10 min.</p>
<p>At the end of the experiment, participants were asked to fill out a questionnaire that comprised the three following questions: 1) How would you rate the level of difficulty of the test, on a scale from 1 (very easy) to 5 (very hard)? 2) How would you rate the level of agreement with your partner, on a scale from 1 (minimal) to 5 (maximal)? 3) During the experiment, did it occur to you that your partner might not be a human, but an artificial system? (two-alternative forced choice: a) I believed my partner was a human, or b) I believed my partner was an artificial system).</p>
<p><xref ref-type="table" rid="T1">Table 1</xref> contains the values that were used for Bias <italic>b</italic>, Slope <italic>g</italic>, and categorical boundary position <italic>b/g</italic> to control the bot&#8217;s response function in each of the four Slope &#215; Bias conditions. The values for <italic>g</italic> correspond to those assigned to <italic>g</italic> in the narrow-variance and wide-variance conditions in Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) study, which we use as a reference (see Appendix 2). The values assigned to the bot&#8217;s categorical boundary location <italic>b/g</italic> correspond to a 10-ms interval centered at the midpoint of the VOT scale, namely, 20 ms. This amounted to shifting the bot&#8217;s categorical boundary relative to the midpoint by &#8211;5 ms in the /p/-Biased condition (boundary at 15 ms), and by +5 ms in the /b/-Biased condition (boundary at 25 ms). Target values for <italic>b</italic> were derived from those for <italic>g</italic> and <italic>b/g</italic>.</p>
<table-wrap id="T1">
<caption>
<p><bold>Table 1:</bold> Values assigned to parameters <italic>g</italic> (ms<sup>&#8211;1</sup>), <italic>b</italic>, and <italic>b/g</italic> (ms) in the bot&#8217;s identification function in the four Slope &#215; Bias conditions. The categorical boundary location <italic>b/g</italic> is given with respect to the 20-ms midpoint on the VOT scale.</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top">Parameter</td>
<td align="left" valign="top" colspan="4">Slope condition</td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top" colspan="2"><italic>Shallow</italic></td>
<td align="left" valign="top" colspan="2"><italic>Steep</italic></td>
</tr>
<tr>
<td align="left" valign="top"><italic>g</italic></td>
<td align="left" valign="top" colspan="2">&#8211;0.26</td>
<td align="left" valign="top" colspan="2">&#8211;0.78</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top" colspan="2">Bias condition</td>
<td align="left" valign="top" colspan="2">Bias condition</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><italic>/b/-Biased</italic></td>
<td align="left" valign="top"><italic>/p/-Biased</italic></td>
<td align="left" valign="top"><italic>/b/-Biased</italic></td>
<td align="left" valign="top"><italic>/p/-Biased</italic></td>
</tr>
<tr>
<td align="left" valign="top"><italic>b</italic></td>
<td align="left" valign="top">&#8211;1.30</td>
<td align="left" valign="top">1.30</td>
<td align="left" valign="top">&#8211;3.90</td>
<td align="left" valign="top">3.90</td>
</tr>
<tr>
<td align="left" valign="top"><italic>b/g</italic></td>
<td align="left" valign="top">5.00</td>
<td align="left" valign="top">&#8211;5.00</td>
<td align="left" valign="top">5.00</td>
<td align="left" valign="top">&#8211;5.00</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The bot&#8217;s response to each stimulus <italic>S</italic> was either voiced (coded as 1) or voiceless (coded as 0). In each of the four Slope &#215; Bias conditions, the distribution of the bot&#8217;s responses was established so that the proportion of voiced responses over the entire set of trials corresponded to that defined in the model, given <italic>b</italic> and <italic>g</italic>:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq010-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2009;</mml:mo><mml:mo>&#x007C;</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M10">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p({c_1}\,|S) = \frac{1}{{1 + {e^{- gS + b}}}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e19.gif"/>
</alternatives>
</disp-formula>
<p>rounded to the nearest integer.</p>
<p>Note that the parameters of the bot&#8217;s response function were fixed and did not evolve in the course of the experiment depending on the participant&#8217;s own responses. In other words, the bot did not adapt itself to the participant&#8217;s response pattern. We aim to explore perceptual convergence in a stepwise fashion, and the goal of this study was to provide a first characterization of how a human participant may converge towards her partner. The use of adaptive bots will be considered in subsequent studies.</p>
<p>Our expectations as regards the listeners&#8217; adaptation to the bot&#8217;s response pattern can be characterized as follows. As an estimate of the expected decrease in <italic>g</italic> in the Steep-Slope relative to the Shallow-Slope conditions, we used Clayards et al.&#8217;s obtained difference in <italic>g</italic> between their Narrow vs. Wide conditions, namely, &#8211;0.12. As an estimate of the expected decrease in <italic>b/g</italic> in the /p/-Biased compared with the /b/-Biased conditions, we used Connine &amp; Clifton&#8217;s (<xref ref-type="bibr" rid="B17">1987</xref>) obtained difference between their voiceless vs. voiced bias conditions, namely, &#8211;3 ms. Our estimates of the expected changes in <italic>b</italic> in the /p/-Biased relative to the /b/-Biased conditions were computed from <italic>g</italic> and <italic>b/g</italic>.</p>
</sec>
<sec>
<title>4.4 Statistical analysis</title>
<p>One participant took the experiment twice in two different conditions, and we set her responses to the second testing aside. Data for four other participants (1.2% of the 320 initial participants) were also left aside, because the proportion of /ba/ responses in each of these participants was equal to or lower than 50% to Stimulus 1, and/or was equal to or higher than 50% to Stimulus 8. As a result, our analyses were conducted on the data for 315 participants (female: 154, male: 159, gender unspecified: 2; mean age: 29 years, 10 months, minimum: 20 years, maximum: 40 years), with 79 participants in each group, except the /b/-Biased/Steep-Slope group (78).</p>
<p>We submitted the data to a Bayesian logistic regression analysis by means of the <monospace>brms</monospace> R package (<xref ref-type="bibr" rid="B10">B&#252;rkner, 2017</xref>; see <xref ref-type="bibr" rid="B41">Kleinschmidt, 2020</xref>, for the same approach). The <monospace>brms</monospace> formula was the following:</p>
<p><monospace>resp &#126; 1 + bias_cond * slope_cond * vot_s + (1 + vot_s &#124; subj_id)</monospace></p>
<p>where <monospace>resp</monospace> is the participant&#8217;s response to the stimulus (0: /p/, 1: /b/), <monospace>bias_cond</monospace> refers to the Bias condition (0: bias for /b/, 1: bias for /p/), <monospace>slope_cond</monospace> refers to the Slope condition (0: Shallow, 1: Steep), <monospace>vot_s</monospace> refers to the VOT value for the stimulus, standardized by subtracting the mean VOT (i.e., the midpoint value on the VOT scale, namely, 20 ms), and <monospace>subj_id</monospace> refers to the participant&#8217;s identification number. As can be seen, Bias condition, Slope condition, and standardized VOT were used as population-level predictors, in combination with two group-level terms, namely, an intercept and slope for each participant. The participant&#8217;s response was treated as a Bernoulli random variable, and the link function was the logit.</p>
<p>This involved estimating the values of eight population-level coefficients <italic>&#946;</italic><sub>0</sub><italic>, &#8230;, &#946;</italic><sub>7</sub>, and two group-level coefficients <italic>u</italic><sub>0</sub><italic><sub>i</sub>, u</italic><sub>1</sub><italic><sub>i</sub></italic>, which allowed us to predict the response <italic>resp<sub>ij</sub></italic> from Participant <italic>i</italic> to stimulus <italic>vot_s<sub>j</sub></italic> in each of the four Bias &#215; Slope conditions as follows:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq011-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mtext mathvariant="italic">res</mml:mtext><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mtext mathvariant="italic">ij</mml:mtext></mml:mrow></mml:msub><mml:mo>&#126;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>2</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>3</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>0</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;</mml:mo><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>4</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>5</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>6</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>7</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:mtext mathvariant="italic">vot</mml:mtext><mml:mo>&#x005F;</mml:mo><mml:msub><mml:mi>s</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M11">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
res{p_{ij}}\sim({\beta _0} + {1_{{\rm{/p/}}}}(bias)\:{\beta _1} + {1_{{\rm{steep}}}}(slope)\:{\beta _2} + {1_{{\rm{/p/}}}}(bias)\:{1_{{\rm{steep}}}}(slope)\:{\beta _3} + {u_{0i}})\\
\,\,\,\,\,\,\,\,\,\,\,\, + ({\beta _4} + {1_{{\rm{/p/}}}}(bias)\:{\beta _5} + {1_{{\rm{steep}}}}(slope)\:{\beta _6} + {1_{{\rm{/p/}}}}(bias)\:{1_{{\rm{steep}}}}(slope)\:{\beta _7} + {u_{1i}})\:vot\_{s_j}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e20.gif"/>
</alternatives>
</disp-formula>
<p>where</p>
<list list-type="bullet">
<list-item><p>1<sub>/p/</sub>(<italic>bias</italic>) is an indicator function set to 0 in the /b/-Biased conditions and 1 in the /p/-Biased conditions</p></list-item>
<list-item><p>1<sub>steep</sub>(<italic>slope</italic>) is an indicator function set to 0 in the Shallow-Slope conditions and 1 in the Steep-Slope conditions</p></list-item>
<list-item><p><italic>&#946;</italic><sub>0</sub> is the intercept at 0 on the standardized VOT continuum in the /b/-Biased/Shallow-Slope condition</p></list-item>
<list-item><p><italic>&#946;</italic><sub>1</sub>, <italic>&#946;</italic><sub>2</sub>, <italic>&#946;</italic><sub>3</sub> are the offsets added to <italic>&#946;</italic><sub>0</sub> in the other three conditions, as follows</p>
<p><inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g10.png"/></p></list-item>
<list-item><p><italic>&#946;</italic><sub>4</sub> is the Slope parameter of the logistic function in the /b/-Biased/Shallow-Slope condition</p></list-item>
<list-item><p><italic>&#946;</italic><sub>5</sub>, <italic>&#946;</italic><sub>6</sub>, <italic>&#946;</italic><sub>7</sub> are the offsets added to <italic>&#946;</italic><sub>4</sub> in the other three conditions, as follows</p>
<p><inline-graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g11.png"/></p></list-item>
<list-item><p><italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> is the random intercept for Participant <italic>i</italic></p></list-item>
<list-item><p><italic>u</italic><sub>1</sub><italic><sub>i</sub></italic> is the random slope for Participant <italic>i</italic></p></list-item>
</list>
<p>Importantly, a direct correspondence can be established between the population-level coefficients and the <italic>b</italic> and <italic>g</italic> parameters in our model:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq012-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x00D7;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>2</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>3</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>4</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>5</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>6</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>/p/</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">bias</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mn>1</mml:mn><mml:mrow><mml:mtext>steep</mml:mtext></mml:mrow></mml:msub><mml:mo stretchy='false'>(</mml:mo><mml:mtext mathvariant="italic">slope</mml:mtext><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>7</mml:mn></mml:msub></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M12">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
b = - 1 \times ({\beta _0} + {1_{{\rm{/p/}}}}(bias)\:{\beta _1} + {1_{{\rm{steep}}}}(slope)\:{\beta _2} + {1_{{\rm{/p/}}}}(bias)\:{1_{{\rm{steep}}}}(slope)\:{\beta _3})\\
g = {\beta _4} + {1_{{\rm{/p/}}}}(bias)\:{\beta _5} + {1_{{\rm{steep}}}}(slope)\:{\beta _6} + {1_{{\rm{/p/}}}}(bias)\:{1_{{\rm{steep}}}}(slope)\:{\beta _7}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e21.gif"/>
</alternatives>
</disp-formula>
<p>This permitted us to estimate the distributions for <italic>b</italic> and <italic>g</italic> from the logistic regression.</p>
<p>The population-level parameters were given weakly-informative prior distributions. For the intercept <italic>&#946;</italic><sub>0</sub>, the prior distribution was <inline-formula>
<alternatives>
<mml:math id="Eq013-mml">
<mml:mi mathvariant='script'>N</mml:mi>
</mml:math>
<tex-math id="M13">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mathcal N}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e3.gif"/>
</alternatives>
</inline-formula>(0,1), i.e., a normal distribution with a mean of 0 and a standard deviation of 1. This amounted to having the probability for the stimulus to be perceived as /b/ centered at 0.5 at the continuum midpoint,<xref ref-type="fn" rid="n6">6</xref> but with large variations both above and below 0.5 at the midpoint. Likewise, we assigned <italic>&#946;</italic><sub>1</sub>, namely, the extent to which the intercept <italic>&#946;</italic><sub>0</sub> changes in the /p/-Biased compared with the /b/-Biased condition, a prior distribution defined as <inline-formula>
<alternatives>
<mml:math id="Eq014-mml">
<mml:mi mathvariant='script'>N</mml:mi>
</mml:math>
<tex-math id="M14">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mathcal N}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e3.gif"/>
</alternatives>
</inline-formula>(0,1). For the Slope parameter <italic>&#946;</italic><sub>4</sub>, the prior distribution was a normal distribution with a mean of &#8211;0.5 and a standard deviation of 1, i.e., <inline-formula>
<alternatives>
<mml:math id="Eq015-mml">
<mml:mi mathvariant='script'>N</mml:mi>
</mml:math>
<tex-math id="M15">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mathcal N}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e3.gif"/>
</alternatives>
</inline-formula>(&#8211;0.5,1). The &#8211;0.5 value corresponded to a decrease of (&#8211;0.5/4) &#215; 100 = 12.5% in the proportion of /b/ responses over a 1-ms VOT interval across the categorical boundary.<xref ref-type="fn" rid="n7">7</xref> The standard deviation of 1 unit caused the prior distribution to encompass a large range of values both above and below the &#8211;0.5 mean. Finally, the prior distribution for <italic>&#946;</italic><sub>6</sub>, i.e., the amount of change in Slope <italic>&#946;</italic><sub>4</sub> in the Steep-Slope compared with the Shallow-Slope condition, was a normal distribution <inline-formula>
<alternatives>
<mml:math id="Eq016-mml">
<mml:mi mathvariant='script'>N</mml:mi>
</mml:math>
<tex-math id="M16">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mathcal N}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e3.gif"/>
</alternatives>
</inline-formula>(0,1). To sum up, prior distributions for the population-level terms were centered on mean values that reflected the expected location of the categorical boundary and Slope parameter of the identification function across that boundary in a standard 2AFC phoneme identification task, but which were compatible with large variations around these mean values. Prior distributions for the other population-level terms and the group-level terms were the <monospace>brms</monospace> default ones.</p>
</sec>
</sec>
<sec>
<title>5. Results</title>
<p>A summary of the <monospace>brms</monospace> model&#8217;s output is displayed in <xref ref-type="table" rid="T2">Table 2</xref>. The <monospace>brms</monospace> model&#8217;s reference condition is the /b/-Biased, Shallow-Slope condition (with <italic>&#946;</italic><sub>0</sub>: intercept at 0 on the standardized VOT continuum and <italic>&#946;</italic><sub>4</sub>: Slope parameter of the categorization function, in that condition). <xref ref-type="table" rid="T2">Table 2</xref> shows that the estimate for the <italic>&#946;</italic><sub>1</sub> coefficient is negative and that the upper bound of the 95% credible interval for that coefficient is well below 0. This is consistent with a lower proportion of /ba/ responses when the bot showed a bias towards /pa/ as opposed to /ba/ in the Shallow-Slope condition. The estimate for <italic>&#946;</italic><sub>5</sub> is positive and its 95% CI appears to be above 0, which is indicative of the participants&#8217; categorization function tending to be shallower in the /p/-Biased compared with the reference condition.</p>
<table-wrap id="T2">
<caption>
<p><bold>Table 2:</bold> Summary of the <monospace>brms</monospace> logistic regression model. Group-level effects: <italic>&#964;</italic><sub>0</sub> and <italic>&#964;</italic><sub>1</sub> refer to the estimates of the standard deviations associated with the by-participant random intercepts <italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> and random slopes <italic>u</italic><sub>1</sub><italic><sub>i</sub></italic>, respectively; <italic>&#961;</italic> is the estimate of the correlation coefficient between <italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> and <italic>u</italic><sub>1</sub><italic><sub>i</sub></italic>. Est. Error: estimated error; l-95% and u-95% CI: lower and upper bound of credible interval, respectively; Rhat: information on the convergence of the algorithm (see <xref ref-type="bibr" rid="B10">B&#252;rkner, 2017</xref>).</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top" colspan="6"><bold>Population-Level Effects</bold></td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold>Estimate</bold></td>
<td align="left" valign="top"><bold>Est. Error</bold></td>
<td align="left" valign="top"><bold>l-95% CI</bold></td>
<td align="left" valign="top"><bold>u-95% CI</bold></td>
<td align="left" valign="top"><bold>Rhat</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>0</sub></td>
<td align="left" valign="top">1.64</td>
<td align="left" valign="top">0.17</td>
<td align="left" valign="top">1.33</td>
<td align="left" valign="top">1.99</td>
<td align="left" valign="top">1.01</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>1</sub></td>
<td align="left" valign="top">&#8211;1.51</td>
<td align="left" valign="top">0.24</td>
<td align="left" valign="top">&#8211;1.98</td>
<td align="left" valign="top">&#8211;1.04</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>2</sub></td>
<td align="left" valign="top">0.12</td>
<td align="left" valign="top">0.24</td>
<td align="left" valign="top">&#8211;0.32</td>
<td align="left" valign="top">0.59</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>3</sub></td>
<td align="left" valign="top">&#8211;0.27</td>
<td align="left" valign="top">0.33</td>
<td align="left" valign="top">&#8211;0.88</td>
<td align="left" valign="top">0.37</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>4</sub></td>
<td align="left" valign="top">&#8211;0.67</td>
<td align="left" valign="top">0.03</td>
<td align="left" valign="top">&#8211;0.73</td>
<td align="left" valign="top">&#8211;0.61</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>5</sub></td>
<td align="left" valign="top">0.11</td>
<td align="left" valign="top">0.04</td>
<td align="left" valign="top">0.03</td>
<td align="left" valign="top">0.19</td>
<td align="left" valign="top">1.01</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>6</sub></td>
<td align="left" valign="top">0.01</td>
<td align="left" valign="top">0.04</td>
<td align="left" valign="top">&#8211;0.07</td>
<td align="left" valign="top">0.08</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#946;</italic><sub>7</sub></td>
<td align="left" valign="top">&#8211;0.06</td>
<td align="left" valign="top">0.06</td>
<td align="left" valign="top">&#8211;0.18</td>
<td align="left" valign="top">0.05</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top" colspan="6"><bold>Group-Level Effects</bold></td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold>Estimate</bold></td>
<td align="left" valign="top"><bold>Est. Error</bold></td>
<td align="left" valign="top"><bold>l-95% CI</bold></td>
<td align="left" valign="top"><bold>u-95% CI</bold></td>
<td align="left" valign="top"><bold>Rhat</bold></td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#964;</italic><sub>0</sub></td>
<td align="left" valign="top">1.37</td>
<td align="left" valign="top">0.07</td>
<td align="left" valign="top">1.23</td>
<td align="left" valign="top">1.51</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#964;</italic><sub>1</sub></td>
<td align="left" valign="top">0.19</td>
<td align="left" valign="top">0.01</td>
<td align="left" valign="top">0.17</td>
<td align="left" valign="top">0.22</td>
<td align="left" valign="top">1.00</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#961;</italic></td>
<td align="left" valign="top">&#8211;0.11</td>
<td align="left" valign="top">0.08</td>
<td align="left" valign="top">&#8211;0.28</td>
<td align="left" valign="top">0.05</td>
<td align="left" valign="top">1.00</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The estimate for the <italic>&#946;</italic><sub>6</sub> coefficient is very close to 0, and this indicates that the Slope parameter of the participants&#8217; categorization function showed little or no variation in the Steep-Slope condition relative to the reference one.</p>
<p>Although the estimate for the <italic>&#946;</italic><sub>3</sub> coefficient is negative, the 95% CI encompasses both negative and positive values, and this suggests that the proportion of /ba/ responses changed according to Bias to about the same extent in the Steep-Slope compared with the Shallow-Slope condition. The estimate for the <italic>&#946;</italic><sub>7</sub> coefficient also straddles the 0 value, and is consistent with the Slope parameter of the participants&#8217; categorization function showing no observable variation depending on Bias in the Steep-Slope relative to the Shallow-Slope condition.</p>
<p><xref ref-type="fig" rid="F3">Figure 3</xref> contains a graphical representation of the <monospace>brms</monospace> model&#8217;s output. Each orange curve represents the estimated mean participants&#8217; categorization function in a given experimental condition as computed from the mean values of the posterior distributions of the model&#8217;s population-level parameters. The highest-density interval around the categorization function, as estimated from the posterior distributions of the model&#8217;s population-level and group-level parameters, is also shown, as well as the bot&#8217;s categorization function (in blue).</p>
<fig id="F3">
<caption>
<p><bold>Figure 3:</bold> Estimated mean participants&#8217; categorization function (orange curve) and corresponding highest-density interval (orange stripe) in each of the four experimental conditions. The bot&#8217;s categorization functions are also displayed in blue. Phonemic categories /b/ and /p/ are associated with short and longer VOT values, respectively.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g3.png"/>
</fig>
<p>The leftward shift in the location of the categorical boundary in the /p/-Biased relative to the /b/-Biased conditions can be clearly seen. By contrast, the participants&#8217; categorization function displays little or no visible change in Slope in the Steep- compared with the Shallow-Slope conditions.</p>
<p>Let us now turn to the link between the <monospace>brms</monospace> population-level parameters and both Bias <italic>b</italic> and Slope <italic>g</italic> in our model. <xref ref-type="fig" rid="F4">Figure 4</xref> shows the posterior distributions of <italic>&#946;</italic><sub>0</sub>, <italic>&#946;</italic><sub>1</sub>, <italic>&#946;</italic><sub>2</sub>, <italic>&#946;</italic><sub>3</sub> as associated with Bias <italic>b</italic>, and of <italic>&#946;</italic><sub>4</sub>, <italic>&#946;</italic><sub>5</sub>, <italic>&#946;</italic><sub>6</sub>, <italic>&#946;</italic><sub>7</sub> as associated with Slope <italic>g</italic>. The distributions for <italic>b</italic> and <italic>g</italic> in the four experimental conditions, as computed from these parameters (see 4.4) are also shown.</p>
<fig id="F4">
<caption>
<p><bold>Figure 4:</bold> Left panel: Posterior distributions of the <monospace>brms</monospace> population-level parameters <italic>&#946;</italic><sub>0</sub>, &#8230;, <italic>&#946;</italic><sub>3</sub> and associated distributions of Bias <italic>b</italic> in the four experimental conditions. B/SHAL: /b/-Biased/Shallow Slope; P/SHAL: /p/-Biased/Shallow Slope; B/STEE: /b/-Biased/Steep Slope; P/STEE: /p/-Biased/Steep Slope. Right panel: Posterior distributions of the <monospace>brms</monospace> population-level parameters <italic>&#946;</italic><sub>4</sub>, &#8230;, <italic>&#946;</italic><sub>7</sub> and associated distributions of Slope <italic>g</italic> in the four experimental conditions. Thin horizontal bars: intervals from quantile at <italic>p =</italic> 0.001 to quantile at <italic>p =</italic> 0.999. Thick horizontal bars: 95% highest density intervals. Filled circles: modes of distributions.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g4.png"/>
</fig>
<p>The distributions of the population-level parameters are linked to the summary statistics provided in <xref ref-type="table" rid="T2">Table 2</xref> and discussed above. The distributions for <italic>b</italic> display a clear difference between the /p/-Biased vs. /b/-Biased conditions. Conversely, there is a large overlap in the distributions for <italic>g</italic> in the Steep-Slope conditions relative to the Shallow-Slope ones.</p>
<p>The estimated location of the /ba/-/pa/ categorical boundary on the unstandardized VOT continuum in each of the four experimental conditions, computed as the ratio <italic>b/g</italic>, is presented in <xref ref-type="table" rid="T3">Table 3</xref>.</p>
<table-wrap id="T3">
<caption>
<p><bold>Table 3:</bold> Estimated location of the /ba/-/pa/ categorical boundary on the unstandardized VOT continuum (in ms), in each of the four experimental conditions.</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"><bold>Slope condition</bold></td>
<td align="left" valign="top" colspan="3"><bold>Bias condition</bold></td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold>/b/-Biased</bold></td>
<td align="left" valign="top"><bold>/p/-Biased</bold></td>
<td align="left" valign="top"><bold>Diff.</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Shallow</td>
<td align="left" valign="top">22.47</td>
<td align="left" valign="top">20.24</td>
<td align="left" valign="top">&#8211;2.23</td>
</tr>
<tr>
<td align="left" valign="top">Steep</td>
<td align="left" valign="top">22.68</td>
<td align="left" valign="top">19.98</td>
<td align="left" valign="top">&#8211;2.70</td>
</tr>
<tr>
<td align="left" valign="top">Diff.</td>
<td align="left" valign="top">0.21</td>
<td align="left" valign="top">&#8211;0.26</td>
<td align="left" valign="top"></td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The shift towards the /ba/ endpoint in the /p/-Biased relative to the /b/-Biased condition was between 2.2 and 2.7 ms. There were very limited changes in the location of the categorical boundary depending on the Slope condition. Importantly, and because <italic>g</italic> displayed little variation across conditions, movements of the categorical boundary in the /p/-Biased vs. /b/-Biased conditions can be mostly attributed to Bias <italic>b</italic> in our model.</p>
<p>Finally, the participants&#8217; responses to our post-test questionnaire can be summarized as follows. The test&#8217;s perceived level of difficulty had a mean value of 2.3 out of 5 (minimum: 1, maximum: 4); the mean perceived level of agreement with the participant&#8217;s partner was at 3.7 out of 5 (minimum: 2, maximum: 5); 105 (33 %) participants responded that they believed their partner was a human, whereas 210 (67 %) said they believed their partner was an artificial system.</p>
</sec>
<sec>
<title>6. Simulations</title>
<p>The lack of convergence in the Slope parameter observed in our experiment could at least in part be ascribed to the characteristics of the 2AFC task. It has been pointed out that, in a phoneme categorization task, the precise shape of the categorization function may depend on how listeners are asked to respond to stimuli (e.g., <xref ref-type="bibr" rid="B58">Massaro &amp; Cohen, 1983</xref>; <xref ref-type="bibr" rid="B59">McMurray, 2022</xref>). Specifically, the 2AFC task may yield categorization functions that have a steeper slope in the vicinity of the categorical boundary, compared with continuous categorization tasks (<xref ref-type="bibr" rid="B4">Apfelbaum et al., 2022</xref>). This should be particularly true if the listener&#8217;s choice between the two proposed categories is based on a winner-take-all mechanism that consists in always opting for the category with the highest probability value (<xref ref-type="bibr" rid="B65">Nearey &amp; Hogan, 1986</xref>). In such a scenario, adaptive changes in slope that listeners may have shown could have been filtered out at the forced-choice decision stage. In other words, responses produced by listeners in the 2AFC task may be too coarse-grained to reflect such adaptive effects. To circumvent this problem, it would be possible to have both the listener and her partner perform a continuous categorization task, such as the visual analog scale task (<xref ref-type="bibr" rid="B4">Apfelbaum et al., 2022</xref>; <xref ref-type="bibr" rid="B38">Kapnoula et al., 2017</xref>), to determine whether this allows convergence in slope to be brought to light. This is an avenue to pursue in future work.</p>
<p>In the present section, we examine the potential role of two other factors in the listeners&#8217; lack of adaptation to the bot&#8217;s response patterns with respect to Slope. To do so, we ran a series of numerical simulations whose results are presented below.</p>
<p>The first of these factors relates to the listeners&#8217; amount of exposure to both the stimuli and the bot&#8217;s responses. In the experiment, each of the eight stimuli and the following bot&#8217;s response were presented ten times to the listeners in each experimental condition. Although this proved sufficient for the listeners to display convergence towards the bot with respect to Bias, it may be the case that a larger number of trials per stimulus would have been needed for convergence in Slope to occur. Intuitively, listeners may be able to accurately estimate the size and direction of a bias in the bot&#8217;s responses by keeping track of the overall number of responses in each category, whereas estimating the Slope parameter may entail listeners monitoring variations in the bot&#8217;s response across stimuli on the acoustic continuum. A more limited amount of evidence may be needed for the former than for the latter.</p>
<p>To check this, we simply asked to what extent both Bias and Slope could be accurately estimated on the basis of 10 trials for each of the eight stimuli, in each of the four experimental conditions. We generated 80 simulated responses from the bot to a random sequence of 10 presentations of each of the eight stimuli by randomly drawing samples from the Bernoulli distribution Ber(<italic>p</italic>), where <italic>p</italic> is the probability for the bot to opt for the /ba/ response given the stimulus as characterized earlier (<xref ref-type="fig" rid="F2">Figure 2</xref> and <xref ref-type="fig" rid="F4">4.3</xref>). This process was repeated 100 times and thus yielded 100 80-response sequences. We then submitted each 80-response sequence to a Bayesian logistic regression in order to estimate both Bias <italic>b</italic> and Slope <italic>g</italic> from that sequence. The results are displayed in <xref ref-type="fig" rid="F5">Figure 5</xref>.</p>
<fig id="F5">
<caption>
<p><bold>Figure 5:</bold> Fitting the bot&#8217;s response pattern on the basis of the bot&#8217;s simulated responses to 10 repetitions of each of the eight stimuli. Blue curves: bot&#8217;s response probability for /ba/ given the stimulus&#8217; VOT value in each experimental condition. Orange curves: logistic functions constructed from the posterior values of parameters <italic>b</italic> and <italic>g</italic> as extracted from each 80-response sequence by means of a Bayesian logistic regression, and averaged over 100 sequences. Orange stripes: 95% highest-density intervals.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g5.png"/>
</fig>
<p>As can be seen, both the Bias and Slope parameters of the bot&#8217;s categorization function are well captured by the logistic regression on the basis of the bot&#8217;s responses to 10 repetitions of each stimulus. Thus, it does not seem that lack of convergence in Slope was due to the listeners&#8217; being provided with too limited evidence for them to be able to accurately infer the bot&#8217;s underlying distribution for Slope.</p>
<p>We now turn to a second factor that may account for the lack of convergence in Slope, namely, the listeners&#8217; degree of confidence in their prior beliefs. Support for this potential account can be found in Kleinschmidt &amp; Jaeger&#8217;s (<xref ref-type="bibr" rid="B42">2015</xref>) modeling and experimental work on adaptation in speech perception. A central question in Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) is how listeners, when exposed to a particular phonetic realization of a phonemic contrast, may recalibrate their internal representations for the two phoneme categories so as to infer in the best possible way the phoneme associated with the sound that is presented to them. To answer that question, Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) (K&amp;J, hereafter) have developed a Bayesian model of speech perception with which our own proposed model has close links. The likelihood function &#8211; the probability distribution of the stimuli in a one-dimensional acoustic space for each of the two phoneme categories &#8211; has the same basic form in both models, namely, <inline-formula>
<alternatives>
<mml:math id="Eq017-mml">
<mml:mi mathvariant='script'>N</mml:mi>
</mml:math>
<tex-math id="M17">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mathcal N}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e3.gif"/>
</alternatives>
</inline-formula>(<italic>&#956;<sub>c<sub>1</sub></sub></italic>, &#963;<sup>2</sup>) and <inline-formula>
<alternatives>
<mml:math id="Eq018-mml">
<mml:mi mathvariant='script'>N</mml:mi>
</mml:math>
<tex-math id="M18">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mathcal N}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e3.gif"/>
</alternatives>
</inline-formula>(<italic>&#956;<sub>c<sub>2</sub></sub></italic>, &#963;<sup>2</sup>), where <italic>c</italic><sub>1</sub> and <italic>c</italic><sub>2</sub> refer to the two categories, respectively. In the K&amp;J model, perceptual recalibration can occur by means of two main mechanisms: category shift and category expansion. Category shift involves shifting the means <italic>&#956;<sub>c<sub>1</sub></sub>, &#956;<sub>c<sub>2</sub></sub></italic> of both distributions<xref ref-type="fn" rid="n8">8</xref> along the acoustic continuum. Category expansion involves increasing (or, conversely, decreasing) the variance <italic>&#963;<sup>2</sup></italic> for both categories. While a category shift causes the location of the categorical boundary to move along the acoustic continuum, category expansion affects the Slope parameter of the categorization function in the vicinity of the categorical boundary: that Slope becomes shallower when the variance increases, and steeper when the variance decreases.</p>
<p>To achieve perceptual recalibration, the listener must update her prior beliefs in either the means or variance, or both, in the face of the evidence she is exposed to.<xref ref-type="fn" rid="n9">9</xref> In the K&amp;J model, both the means and variance have their own prior distributions, which are governed by a set of hyperparameters. These hyperparameters determine the prior values for the means and variance, but also and quite importantly the listener&#8217;s level of confidence in these prior values, i.e., how strongly she believes that such prior values should be assigned to the means and variance. Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) and Kleinschmidt (<xref ref-type="bibr" rid="B41">2020</xref>) used Bayesian inference to estimate the listeners&#8217; prior beliefs and updated (posterior) values for the means and variance from the listeners&#8217; responses in a number of 2AFC phoneme categorization tasks. These estimates were consistent with listeners using either the category shift or category expansion mechanism to perform perceptual recalibration. However, both Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) and Kleinschmidt (<xref ref-type="bibr" rid="B41">2020</xref>) also provided evidence suggesting that listeners tended to believe in their prior value for the variance to a greater extent than in those for the means. This, in turn, suggests that the listener&#8217;s preferred mechanism to perform perceptual recalibration was shifting the category means, rather than enlarging/shrinking the category variances.</p>
<p>As indicated above, there is a direct link between category variance and the Slope of the categorization function. In the K&amp;J model as well as our model, Slope is defined as <italic>g</italic> = (<italic>&#956;<sub>c<sub>2</sub></sub></italic> &#8211; <italic>&#956;<sub>c<sub>1</sub></sub></italic>) /&#963;<sup>2</sup>. Changes in category variance therefore engender variations in Slope. In addition, and because the distance between the two means <italic>&#181;<sub>c<sub>2</sub></sub></italic> &#8211; <italic>&#956;<sub>c<sub>1</sub></sub></italic> is fixed in the K&amp;J model, Slope only depends on category variance. If we extend the K&amp;J model to our joint phoneme categorization task, our data should be consistent with a listener that is moderately confident in her prior values for the category means, but highly confident in her prior values for the category variances. We therefore implemented a simplified version of the K&amp;J model in R to test that hypothesis.</p>
<p>K&amp;J use a joint conjugate prior distribution for the mean and variance of each category, namely, the normal-inverse-chi-squared distribution, <inline-formula>
<alternatives>
<mml:math id="Eq019-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mo>&#x03BC;</mml:mo><mml:mo>,</mml:mo><mml:msup><mml:mo>&#x03C3;</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant='script'>N</mml:mi><mml:mtext>-</mml:mtext><mml:mtext mathvariant="italic">Inv-</mml:mtext><mml:msup><mml:mo>&#x03C7;</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mo>&#x03BC;</mml:mo><mml:mo>,</mml:mo><mml:msup><mml:mo>&#x03C3;</mml:mo><mml:mn>2</mml:mn></mml:msup><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2009;</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mn>0</mml:mn><mml:mn>2</mml:mn></mml:msubsup><mml:mo>/</mml:mo><mml:msub><mml:mo>&#x03BA;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>;</mml:mo><mml:mo>&#x2009;</mml:mo><mml:msub><mml:mo>&#x03BD;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2009;</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mn>0</mml:mn><mml:mn>2</mml:mn></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mrow>
</mml:math>
<tex-math id="M19">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p(\mu,{\sigma ^2}) = {\mathcal N} - Inv - {\chi ^2}(\mu,{\sigma ^2}|{\mu _0},\,\sigma _0^2/{\kappa _0};\,{\nu _0},\,\sigma _0^2)
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e4.gif"/>
</alternatives>
</inline-formula>, which allows the parameters of the posterior distribution to be computed easily and in an analytical fashion (<xref ref-type="bibr" rid="B30">Gelman et al., 2021</xref>; <xref ref-type="bibr" rid="B49">Lambert, 2018</xref>; <xref ref-type="bibr" rid="B64">Murphy, 2007</xref>). The hyperparameters <italic>&#954;</italic><sub>0</sub> and <italic>&#957;</italic><sub>0</sub> can be interpreted as representing the strength of the listener&#8217;s belief in the mean and variance, respectively. They are seen as pseudo-counts, i.e., the number of observations needed for the listener to start overcoming her prior beliefs. In our simulation, we set the value for <italic>&#954;</italic><sub>0</sub> to either 50 (a moderately low value, compared with the total number of trials in the experiment, namely, 80) or 1000 (an arbitrarily high value). Likewise, the value for <italic>&#957;</italic><sub>0</sub> was set to either 50 or 1000. The model&#8217;s prior values for the category means and variance were derived from our data and, more specifically, from the listeners&#8217; average classification function across all experimental conditions, as computed from the posterior distributions of the fixed and random effects in the logistic regression. We then made the model converge towards the bot&#8217;s classification function as established in each of the four experimental conditions, using the K&amp;J belief updating procedure, with a simulated number of trials set to 80, as in our experiment. The results are shown in <xref ref-type="fig" rid="F6">Figure 6</xref>.</p>
<p>In the simulations represented in the upper left panel, confidence in prior beliefs was low for both the category means and variance, and the model was expected to show convergence towards the bot with respect to both the location of the categorical boundary and the Slope parameter of the categorization function at this location. This is indeed what occurred: the model&#8217;s categorization function shifted towards the /b/ endpoint in the /p/-Biased relative to the /b/-Biased condition, and was steeper in the Steep-Slope compared with the Shallow-Slope condition. By contrast, the lower right panel illustrates the results obtained when prior confidence is high for both the category means and variance. As expected, the model behaved in a conservative manner, and showed very limited adaptation to the bot&#8217;s response patterns. The upper right panel corresponds to a setting that makes the model conservative for the category means but flexible for the variance, as reflected in the fact that changes in the model&#8217;s categorization function mainly occur with respect to Slope. Finally, the lower left panel shows the results of the model&#8217;s belief-updating process when prior confidence is low for the category means but high for the category variance.</p>
<fig id="F6">
<caption>
<p><bold>Figure 6:</bold> Results of the simulation carried out using a simplified version of Kleinschmidt &amp; Jaeger&#8217;s (<xref ref-type="bibr" rid="B42">2015</xref>) belief-updating Bayesian model. Confidence in the prior values for the category means and variance was set to either a low or high level. Orange curves: categorization functions of the model at the end of the simulated 80-trial experiment. Blue curves: bot&#8217;s categorization functions. Categories /b/ and /p/ are associated with short and longer VOT values, respectively.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g6.png"/>
</fig>
<p>As can be seen, it is this last case that displays the closest fit with the results of the experiment (<xref ref-type="fig" rid="F3">Figure 3</xref>). This lends support to the hypothesis that listeners may require more exposure to the stimuli and partner&#8217;s responses than was the case in the experiment, for them to overcome their prior beliefs and show convergence in Slope. Further work will be needed to fully assess that hypothesis.</p>
</sec>
<sec>
<title>7. General discussion</title>
<p>Adaptive mechanisms in phoneme categorization have been a central topic in research on speech perception for decades. Seminal work on lexical influence in the categorization of phonemes (<xref ref-type="bibr" rid="B28">Ganong, 1980</xref>), perceptual compensation for coarticulation (<xref ref-type="bibr" rid="B57">Mann &amp; Repp, 1981</xref>), perceptual learning (<xref ref-type="bibr" rid="B45">Kraljic &amp; Samuel, 2005</xref>; <xref ref-type="bibr" rid="B70">Norris et al., 2003</xref>), adaptation to distributional statistics of phonetic cues (<xref ref-type="bibr" rid="B14">Clayards et al., 2008</xref>), to cite but a few, and the many studies that followed, have provided us with major insights about the processing mechanisms that listeners deploy to identify phonemes, given the idiosyncratic characteristics of the speakers and the context in which speech sounds are produced. At the heart of this vast body of research lies a question which, laid out in a Bayesian framework, can be stated as follows: how do listeners infer the speaker&#8217;s intended phoneme category, given both the sound that speaker has produced, and the listener&#8217;s prior beliefs about the mapping of phonemes onto sounds? The focus of our own work, however, is different. While previous research has centered on perceptual adaptation to the speaker, as performed by listeners in an individual fashion, we seek to determine to what extent one listener can converge towards another listener in the categorization of speech sounds. In Bayesian terms, this amounts to asking whether one listener can infer the way in which another listener herself infers which phoneme was produced by the speaker, given the sound that both listeners have heard, and both listeners&#8217; prior beliefs. To the best of our knowledge, if a great deal of attention has been devoted to listener-to-speaker adaptation in speech perception, adaptation between listeners has not been studied so far.</p>
<p>In this experiment, participants were presented with stimuli ranging from /ba/ to /pa/ on a VOT continuum in a 2AFC task jointly performed with an artificial agent presented to the participants as a human partner. We manipulated the artificial agent&#8217;s response pattern with respect to both Bias and Slope, in a four-condition between-participant design. In agreement with our first prediction, participants were found to converge towards the artificial agent with respect to Bias. Contrary to our second prediction, however, participants did not show a convergence effect with respect to Slope. Because Prediction 3 focused on a link between change in Bias and change in Slope, and in the absence of evidence for the latter, that prediction did not apply to our data. Thus, convergence was found to arise for Bias but not for Slope. Numerical simulations showed that the number of trials used in the experiment was sufficient for Slope to be accurately estimated using a standard Bayesian logistic-regression classifier. These simulations suggest that lack of convergence in Slope may stem from the listeners&#8217; prior level of confidence in the variance in VOT for the two phonemic categories, which may require more exposure to the stimuli and partner&#8217;s responses to be overcome.</p>
<p>The present experimental confirmation of the first prediction is clearly a new result. It shows that individuals can shift their perceptual judgement to make it more consistent with the judgement expressed by an interacting partner. Quantitatively, the shift in category boundary between the /p/ and /b/ biased conditions amounts to about 2 ms (see <xref ref-type="fig" rid="F3">Figure 3</xref>), to be compared with the 10-ms shift displayed by the partner. According to the responses to the post-test questionnaire, the partner was believed to be an artificial system by two-thirds of the participants. This may have caused convergence to be reduced, compared with a situation in which participants believe their partner to be human. In addition, the bot did not adapt itself to the participant&#8217;s own responses, and this may have led participants to converge towards the bot to a lesser extent than they would have done had convergence been reciprocal. Recent research (e.g., <xref ref-type="bibr" rid="B55">Mahmoodi et al., 2018</xref>) has shown that inter-individual reciprocity in social influence plays an important role in perceptual judgment, and is obliterated when people believe they interact with a computer. However, it is difficult to determine whether participants formed that belief in the course of the experiment itself, or only after, on seeing the possibility that the partner was an artificial system explicitly raised in the questionnaire. We plan to further explore the potential effect on perceptual convergence of the partner&#8217;s perceived nature as a human being vs. artificial system in future studies.</p>
<p>Remarkably, the effect of Bias was restricted to the more ambiguous stimuli and did not extend to the endpoint stimuli, which were consistently identified as /ba/ and /pa/, respectively (see <xref ref-type="fig" rid="F3">Figure 3</xref>). This means that participants did not simply favor one or the other of the two proposed responses regardless of the stimulus. Had this been the case, the participants&#8217; categorization functions would have differed across Bias conditions over the entirety of the VOT continuum. In that respect, the participants proved able to closely imitate the bot&#8217;s response pattern, whose variations across Bias conditions were also confined to the more ambiguous stimuli. To what extent the effect of Bias was post-perceptual, akin to the monetary payoff in Connine &amp; Clifton (<xref ref-type="bibr" rid="B17">1987</xref>) and Pitt (<xref ref-type="bibr" rid="B80">1995</xref>), remains to be established. In any case, and quite importantly, the location of the voiced-voiceless categorical boundary differed in the expected direction depending on Bias: that boundary was closer to the /pa/ endpoint in the /b/-Biased conditions relative to the /p/-Biased conditions. This indicates that adjustments in Bias may form a quick and efficient mechanism employed by listeners to align themselves with their partner in the laying out of categorical boundaries in the acoustic space. The present study appears to be the first one to provide evidence for listeners&#8217; convergence in Bias towards their partner in a joint phoneme identification task.</p>
<p>Lack of convergence with respect to Slope could be interpreted, in line with the simulations in Section 6, as pointing to listeners&#8217; having greater confidence in their prior beliefs for category variance compared with category mean. Longer exposure to the evidence would hence be required for listeners to overcome their priors and adapt themselves to their partner&#8217;s response pattern in variance and, consequently, slope. It could also be assumed that adaptation in variance and slope is actually less useful than adaptation in category means for efficient communication between interacting partners. Clayards et al. (<xref ref-type="bibr" rid="B14">2008</xref>) examined to what extent listeners are sensitive to the shape of the distribution of acoustic stimuli across a VOT continuum in the categorization of voiced vs. voiceless bilabial stops in word-initial position in English. In that study, listeners were presented with stimuli whose distribution with respect to VOT originated from a mixture of two Gaussians with either narrow or wide variance, in a sound-to-picture mapping task. The results showed that the listeners&#8217; categorization function was shallower in the wide-variance compared with the narrow-variance condition. Variations were therefore observed in the Slope parameter of the listeners&#8217; categorization function between these two conditions. There is, however, a major difference between Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) and our study. Clayards et al.&#8217;s results revealed perceptual adaptation in a single-listener categorization task to the <italic>stimuli distribution</italic>, whose form could be established by listeners in a direct manner, on the basis of the number of repetitions for each stimulus on the VOT continuum. Our own findings showed lack of listener&#8217;s adaptation for Slope to the <italic>partner&#8217;s response patterns</italic> in a joint categorization task. If we assume that listeners regard these patterns as relying on the partner&#8217;s own internal distributions for the two phonemic categories, access to these distributions can only be gained indirectly by listeners, and by means of an inference process. This suggests that recovering the speech sound distributions associated with two phoneme categories in another listener is substantially more difficult than direct recovery of the distributions for the two categories from the relative frequencies of the speech sounds.</p>
<p>The model we used in this study was based on the single-listener Bayesian models of phoneme identification previously proposed by Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>), Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) and Kronrod et al. (<xref ref-type="bibr" rid="B46">2016</xref>). We extended this modeling framework to a two-listener categorization task in a simple fashion, by assuming that each listener would expect her partner to behave like a Bayesian agent, and would undertake to infer the parameter distributions of her partner&#8217;s internal model, so as to get her own model to fit these distributions as well as possible. Inference was expected to be performed by the listener from her partner&#8217;s set of responses, and to entail computing estimated distributions for both the Slope and Bias parameters. An important difference between our model and both Kleinschmidt &amp; Jaeger&#8217;s (<xref ref-type="bibr" rid="B42">2015</xref>) and Kronrod et al.&#8217;s (<xref ref-type="bibr" rid="B46">2016</xref>)&#8217;s models lies in the fact that we allowed the prior probabilities for the voiced and voiceless categories to differ from each other. This led us to predict that convergence towards the listener&#8217;s partner would extend to Bias, a prediction for which our data provided support, as already mentioned. However, it is clear that our model still requires major developments if it is to become a full-fledged model of joint perception. In particular, these developments should make it possible for us to account for how a listener <italic>combines</italic> her own prior beliefs and internally-represented probability distribution for each phoneme category with those of her partner, and which respective weights she attributes to her and her partner&#8217;s categorization device. Another central issue relates to the dynamics of between-listener adaptation in a joint categorization task. Work is in progress to expand our model along these lines.</p>
<p>That perceptual convergence across listeners appears to have been overlooked raises two questions: does such a phenomenon occur outside the laboratory? And if so, what can it be useful for? To the first question, we suggest that the answer is yes. There are many real-life situations that spring to mind and in which perceptual convergence may be sought and achieved. For example, when several people are listening to someone giving a talk, it may happen that the speaker produces a word that one listener is not sure having correctly identified. That listener may turn to her neighbor and ask: &#8220;Did [the speaker] say <italic>pin</italic> or <italic>bin</italic>?&#8221;. On being told by the neighbor that it was most likely the word <italic>bin</italic>, the listener may then adjust her perceptual boundary between voiced and voiceless stops accordingly. Classrooms of students learning a foreign language may also give rise to perceptual convergence effects. If the students are being trained to identify melodic contours in that language, for example, interactions can take place between them (&#8220;I clearly heard a falling contour, didn&#8217;t you?&#8221;) that may contribute to shaping their perception of the contours. In a military context, several people may have to ensure that they have understood in the same way verbal instructions transmitted to them through some communication channel, prior to executing these instructions. In a forensic context, several people may be asked to listen to an audio recording and come up with a common transcription, which entails mutual adaptation in the mapping of sounds onto phonemes. In all these situations, it seems difficult for a standard one-to-one speaker-listener model to fully account for how perceptual boundaries between sounds may be pushed around in each individual, and listener-to-listener connections should in our view be recognized as having a significant influence.</p>
<p>As to our second question, we believe that being able to infer how other people categorize speech sounds, may have an important role in speech communication for each member of a language community, as both speaker and listener. For speakers, being endowed with the capacity to perform such inferences is clearly central, if we take the view that perceptual targets are brought into play in speech production (<xref ref-type="bibr" rid="B83">Schwartz et al., 2012</xref>), and if speakers are to predict the way in which the sounds they produced will be processed by their interlocutors. For listeners, one important aim may be to ensure that other listeners perceive speech sounds in the same way, if speech is to fulfill its function as a communication device. In short, we contend that perceptual convergence between listeners in speech perception has important implications for theories of speech production and perception and should be better understood. The present piece of work is a first step in that direction.</p>
</sec>
<sec>
<title>Appendix 1: Computation of the model&#8217;s parameters</title>
<p>The way in which sounds distribute themselves in the acoustic domain for each category is specified by two conditional probability distributions, <italic>p</italic>(<italic>S&#124;c</italic><sub>1</sub>) and <italic>p</italic>(<italic>S&#124;c</italic><sub>2</sub>), where <italic>S</italic> refers to the sound and <italic>c</italic><sub>1</sub> and <italic>c</italic><sub>2</sub> to the two categories, respectively. It is assumed that <italic>p</italic>(<italic>S&#124;c</italic><sub>1</sub>) and <italic>p</italic>(<italic>S&#124;c</italic><sub>1</sub>) are both normal distributions and, as such, are each characterized by a mean and a variance:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq020-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant='script'>N</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi mathvariant='script'>N</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub><mml:mo>,</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M20">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
p(S|{c_1}) = {\mathcal N}({\mu _1},\sigma _{{c_1}}^2 + \sigma _S^2)\\
p(S|{c_2}) = {\mathcal N}({\mu _2},\sigma _{{c_2}}^2 + \sigma _S^2)
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e22.gif"/>
</alternatives>
</disp-formula>
<p>For each distribution, variance is a sum of two terms, <inline-formula>
<alternatives>
<mml:math id="Eq021-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M21">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _c^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e6.gif"/>
</alternatives>
</inline-formula>, a measure of dispersion of the intended target sound around the mean for the category, and <inline-formula>
<alternatives>
<mml:math id="Eq022-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M22">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _S^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e7.gif"/>
</alternatives>
</inline-formula>, which refers to sensory-motor variance around the intended target sound independent of the category (<xref ref-type="bibr" rid="B25">Feldman et al., 2009</xref>; <xref ref-type="bibr" rid="B46">Kronrod et al., 2016</xref>).</p>
<p>As in both Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>) and Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>), the variances of the two distributions are considered as equal:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq023-mml">
<mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mrow><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow>
</mml:math>
<tex-math id="M23">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\sigma _{{c_1}}^2 = \sigma _{{c_2}}^2 = \sigma _c^2
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e23.gif"/>
</alternatives>
</disp-formula>
<p>We further assume that the two distributions are in symmetric positions with respect to the midpoint of the continuum <italic>&#956;</italic><sub>0</sub>, i.e., at the same distance <italic>&#948;<sub>&#181;</sub></italic> from that midpoint, on either side of it:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq024-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03B4;</mml:mo><mml:mo>&#x03BC;</mml:mo></mml:msub></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mo>&#x03B4;</mml:mo><mml:mo>&#x03BC;</mml:mo></mml:msub></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M24">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
{\mu _1} = {\mu _0} - {\delta _\mu }\\
{\mu _2} = {\mu _0} + {\delta _\mu }
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e24.gif"/>
</alternatives>
</disp-formula>
<p>The same-variance and same-distance-from-midpoint assumptions are both limitations that may be overcome in a more elaborated version of the model. However, they are acceptable in an experimental setting. Their advantage is that they allow the listener&#8217;s predicted responses to be computed easily and in an analytical way, as shown below. Note that constraints on <italic>&#956;</italic><sub>1</sub>, <italic>&#956;</italic><sub>2</sub>, or both, were also introduced in previous models (preestablished values used for <italic>&#956;</italic><sub>1</sub> in both Feldman et al. (<xref ref-type="bibr" rid="B25">2009</xref>) and Kronrod et al. (<xref ref-type="bibr" rid="B46">2016</xref>), and for the <italic>&#956;</italic><sub>1</sub> <italic>&#8211; &#956;</italic><sub>2</sub> distance in Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>)).</p>
<p>The probability that the phonemic category is <italic>c</italic><sub>1</sub> given S is given by the posterior probability value <italic>p</italic>(<italic>c</italic><sub>1</sub><italic>&#124;S</italic>), in accordance with Bayes&#8217; theorem:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq025-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x007C;</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo>&#x007C;</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M25">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p({c_1}|S) = \frac{{p(S|{c_1})p({c_1})}}{{p(S|{c_1})p({c_1}) + p(S|{c_2})p({c_2})}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e25.gif"/>
</alternatives>
</disp-formula>
<p>which simplifies to:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq026-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x007C;</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M26">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p({c_1}|S) = \frac{1}{{1 + {e^{- gS + b}}}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e26.gif"/>
</alternatives>
</disp-formula>
<p>where</p>
<disp-formula>
<alternatives>
<mml:math id="Eq027-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msubsup><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn><mml:mn>2</mml:mn></mml:msubsup></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:mo stretchy='false'>(</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo>+</mml:mo><mml:msubsup><mml:mo>&#x03C3;</mml:mo><mml:mi>S</mml:mi><mml:mn>2</mml:mn></mml:msubsup><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac><mml:mo>+</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M27">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
g = \frac{{{\mu _1} - {\mu _2}}}{{\sigma _c^2 + \sigma _S^2}}\\
b = \frac{{\mu _1^2 - \mu _2^2}}{{2(\sigma _c^2 + \sigma _S^2)}} + \log \frac{{p({c_2})}}{{p({c_1})}}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e27.gif"/>
</alternatives>
</disp-formula>
<p>The priors <italic>p</italic>(<italic>c</italic><sub>1</sub>) and <italic>p</italic>(<italic>c</italic><sub>2</sub>) contribute to controlling the location of the category boundary <italic>b/g</italic> along the continuum: when <italic>p</italic>(<italic>c</italic><sub>2</sub>) increases relative to <italic>p</italic>(<italic>c</italic><sub>1</sub>) &#8211; all other things being equal &#8211; this causes the boundary to move towards the <italic>c</italic><sub>1</sub> endpoint. In Kleinschmidt &amp; Jaeger (<xref ref-type="bibr" rid="B42">2015</xref>) and Kronrod et al. (<xref ref-type="bibr" rid="B46">2016</xref>), the priors are both set to <italic>p</italic>(<italic>c</italic><sub>1</sub>) <italic>= p</italic>(<italic>c</italic><sub>2</sub>) = 0.5, and this causes them to cancel each other out in the computation of the posterior <italic>p</italic>(<italic>c</italic><sub>1</sub><italic>&#124;S</italic>). Because we are interested in exploring the effect of unequal priors on the listener&#8217;s responses, we allow <italic>p</italic>(<italic>c</italic><sub>1</sub>) and <italic>p</italic>(<italic>c</italic><sub>2</sub>) to differ from each other.</p>
<p>Taking the origin along the stimulus&#8217; acoustic dimension as the midpoint <italic>&#956;</italic><sub>0</sub> between <italic>&#956;</italic><sub>1</sub> and <italic>&#956;</italic><sub>2</sub>, we obtain:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq028-mml">
<mml:mrow><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow>
</mml:math>
<tex-math id="M28">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mu _1} = - {\mu _2}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e28.gif"/>
</alternatives>
</disp-formula>
<p>It follows that:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq029-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x2009;&#x2009;&#x2009;&#x2009;</mml:mo><mml:mo>=</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M29">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
b = \log \frac{{p({c_2})}}{{p({c_1})}}\\
\,\,\,\, = \log \frac{{p({c_2})}}{{1 - p({c_2})}}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e29.gif"/>
</alternatives>
</disp-formula>
<p>From <italic>b</italic>, as empirically measured by means of a logistic regression from a set of data, the values of the priors <italic>p</italic>(<italic>c</italic><sub>1</sub>) and <italic>p</italic>(<italic>c</italic><sub>2</sub>) can be computed by application of the inverse logit function:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq030-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msup><mml:mi>e</mml:mi><mml:mi>b</mml:mi></mml:msup></mml:mrow><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mi>b</mml:mi></mml:msup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>&#x2212;</mml:mo><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M30">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
p({c_2}) = \frac{{{e^b}}}{{1 + {e^b}}}\\
p({c_1}) = 1 - p({c_2})
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e30.gif"/>
</alternatives>
</disp-formula>
</sec>
<sec>
<title>Appendix 2: Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) model parameters and obtained effect size</title>
<p>Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) model aimed to account for how listeners identify acoustic stimuli as voiced vs. voiceless bilabial stops in a 2AFC task. These stimuli ranged on a VOT continuum from &#8211;30 to 80 ms in twelve 10-ms steps. The voiced and voiceless categories were associated with normal probability distributions centered on <italic>&#956;</italic><sub>1</sub> = 0 ms and <italic>&#956;</italic><sub>2</sub> = 50 ms, respectively, and whose standard deviation <italic>&#963;</italic> was set to 8 ms in the narrow-variance condition, and to 14 ms in the wide-variance condition. We here indicate how Slope <italic>g</italic>, Bias <italic>b</italic>, and associated parameters, can be derived from these values.</p>
<p>As seen in Appendix 1, values for <italic>g</italic> and <italic>b</italic> can be computed as follows:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq031-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:mi>g</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub></mml:mrow><mml:mrow><mml:msup><mml:mo>&#x03C3;</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msubsup><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn><mml:mn>2</mml:mn></mml:msubsup></mml:mrow><mml:mrow><mml:mn>2</mml:mn><mml:msup><mml:mo>&#x03C3;</mml:mo><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>+</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M31">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
g = \frac{{{\mu _1} - {\mu _2}}}{{{\sigma ^2}}}\\
b = \frac{{\mu _1^2 - \mu _2^2}}{{2{\sigma ^2}}} + \log \frac{{p({c_2})}}{{p({c_1})}}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e31.gif"/>
</alternatives>
</disp-formula>
<p>Since Clayards et al. assume that <italic>&#956;</italic><sub>1</sub> and <italic>&#956;</italic><sub>2</sub> are located at the same distance <italic>&#948;<sub>&#181;</sub></italic> (25 ms) on either side of the midpoint <italic>&#956;</italic><sub>0</sub> (+25 ms) of the VOT scale, and if we standardize the VOT scale by subtraction of <italic>&#956;</italic><sub>0</sub>, we have</p>
<disp-formula>
<alternatives>
<mml:math id="Eq032-mml">
<mml:mrow><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03B4;</mml:mo><mml:mo>&#x03BC;</mml:mo></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn></mml:msub><mml:mo>=</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mo>&#x03B4;</mml:mo><mml:mo>&#x03BC;</mml:mo></mml:msub><mml:mo>,</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mtext>thus</mml:mtext><mml:mo>&#x2009;&#x2009;</mml:mo><mml:msubsup><mml:mo>&#x03BC;</mml:mo><mml:mn>1</mml:mn><mml:mn>2</mml:mn></mml:msubsup><mml:mo>&#x2212;</mml:mo><mml:msubsup><mml:mo>&#x03BC;</mml:mo><mml:mn>2</mml:mn><mml:mn>2</mml:mn></mml:msubsup><mml:mo>=</mml:mo><mml:mn>0</mml:mn><mml:mo>,</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mtext>hence</mml:mtext><mml:mo>&#x2009;&#x2009;</mml:mo><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mtext>log</mml:mtext><mml:mfrac><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>2</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mrow><mml:mi>p</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mi>c</mml:mi><mml:mn>1</mml:mn></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M32">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{\mu _1} = - {\delta _\mu },{\mu _2} = + {\delta _\mu },\;{\rm{thus}}\;\;\mu _1^2 - \mu _2^2 = 0,\;{\rm{hence}}\;\;b = \log \frac{{p({c_2})}}{{p({c_1})}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e32.gif"/>
</alternatives>
</disp-formula>
<p>Given that Clayards et al. implicitly assume that the two categories are assigned identical prior probabilities, i.e., that <italic>p</italic>(<italic>c</italic><sub>1</sub>) <italic>= p</italic>(<italic>c</italic><sub>2</sub>) = 0.5, we have</p>
<disp-formula>
<alternatives>
<mml:math id="Eq033-mml">
<mml:mrow><mml:mi>b</mml:mi><mml:mo>=</mml:mo><mml:mtext>log&#x00A0;</mml:mtext><mml:mn>1</mml:mn><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow>
</mml:math>
<tex-math id="M33">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
b = \log 1 = 0
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e33.gif"/>
</alternatives>
</disp-formula>
<p>And</p>
<disp-formula>
<alternatives>
<mml:math id="Eq034-mml">
<mml:mrow><mml:mfrac><mml:mi>b</mml:mi><mml:mi>g</mml:mi></mml:mfrac><mml:mo>=</mml:mo><mml:mn>0</mml:mn></mml:mrow>
</mml:math>
<tex-math id="M34">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\frac{b}{g} = 0
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e34.gif"/>
</alternatives>
</disp-formula>
<p>The derivative of the logistic function <inline-formula>
<alternatives>
<mml:math id="Eq035-mml">
<mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle scriptlevel='+1'><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow>
</mml:math>
<tex-math id="M35">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
f(S) = {\textstyle{1 \over {1 + {e^{- gS + b}}}}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e8.gif"/>
</alternatives>
</inline-formula> at the 0.5 cross-over point, <inline-formula>
<alternatives>
<mml:math id="Eq036-mml">
<mml:mrow><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mi>f</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mn>0.5</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi>b</mml:mi><mml:mo>/</mml:mo><mml:mi>g</mml:mi></mml:mrow>
</mml:math>
<tex-math id="M36">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
S = {f^{- 1}}(0.5) = b/g
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e9.gif"/>
</alternatives>
</inline-formula>, is given by</p>
<disp-formula>
<alternatives>
<mml:math id="Eq037-mml">
<mml:mtable columnalign='left'><mml:mtr><mml:mtd><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mfrac><mml:mi>g</mml:mi><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mfrac><mml:mi>b</mml:mi><mml:mi>g</mml:mi></mml:mfrac><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mfrac><mml:mi>g</mml:mi><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mfrac><mml:mi>b</mml:mi><mml:mi>g</mml:mi></mml:mfrac><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;</mml:mo><mml:mo>=</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mn>0</mml:mn></mml:msup><mml:mfrac><mml:mi>g</mml:mi><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mn>0</mml:mn></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac></mml:mtd></mml:mtr><mml:mtr><mml:mtd><mml:mo>&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;&#x2009;</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mi>g</mml:mi><mml:mn>4</mml:mn></mml:mfrac></mml:mtd></mml:mtr></mml:mtable>
</mml:math>
<tex-math id="M37">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
\begin{array}{l}
{e^{- gS + b}}\frac{g}{{{{(1 + {e^{- gS + b}})}^2}}} &amp;= {e^{- g\frac{b}{g} + b}}\frac{g}{{{{(1 + {e^{- g\frac{b}{g} + b}})}^2}}}\\
&amp;= {e^0}\frac{g}{{{{(1 + {e^0})}^2}}}\\
&amp;= \frac{g}{4}
\end{array}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e35.gif"/>
</alternatives>
</disp-formula>
<p><xref ref-type="table" rid="T4">Table 4</xref> contains the values for <italic>g, b</italic>, and associated parameters, as derived from <italic>&#956;</italic><sub>1</sub>, <italic>&#956;</italic><sub>2</sub>, and <italic>&#963;</italic> according to the above formula, for the Narrow and Wide conditions.</p>
<p>Note that, in the Narrow condition, the probability distributions for the voiced and voiceless categories were well separated, and this resulted in a sharp optimal response curve (see <xref ref-type="fig" rid="F1">Figure 1</xref> in <xref ref-type="bibr" rid="B14">Clayards et al., 2008</xref>). The value of the derivative at the 0.5 cross-over point corresponds to a decrease of 20% in the percentage of voiced responses over a 1-ms increase in VOT.<xref ref-type="fn" rid="n10">10</xref> Note also that the listeners&#8217; responses were not expected to vary with respect to <italic>b/g</italic> across the two conditions.</p>
<table-wrap id="T4">
<caption>
<p><bold>Table 4:</bold> Values assigned to <italic>g</italic> (in ms<sup>&#8211;1</sup>), <italic>b</italic> and associated parameters in Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) narrow-variance and wide-variance conditions, and observed effect size. VOT scale standardized by subtraction of the VOT midpoint value (+25 ms).</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"><bold>Parameter</bold></td>
<td align="left" valign="top" colspan="2"><bold>Variance condition</bold></td>
<td align="left" valign="top"><bold>Observed</bold></td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold><italic>Narrow</italic></bold></td>
<td align="left" valign="top"><bold><italic>Wide</italic></bold></td>
<td align="left" valign="top"><bold>effect size</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top"><italic>g</italic> (ms<sup>&#8211;1</sup>)</td>
<td align="left" valign="top">&#8211;0.78</td>
<td align="left" valign="top">&#8211;0.26</td>
<td align="left" valign="top">&#8211;0.12</td>
</tr>
<tr>
<td align="left" valign="top"><italic>derivative</italic></td>
<td align="left" valign="top">&#8211;0.20</td>
<td align="left" valign="top">&#8211;0.06</td>
<td align="left" valign="top">&#8211;0.03</td>
</tr>
<tr>
<td align="left" valign="top"><italic>b</italic></td>
<td align="left" valign="top">0.00</td>
<td align="left" valign="top">0.00</td>
<td align="left" valign="top">&#8212;</td>
</tr>
<tr>
<td align="left" valign="top"><italic>b/g</italic> (ms)</td>
<td align="left" valign="top">0.00</td>
<td align="left" valign="top">0.00</td>
<td align="left" valign="top">&#8212;</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Clayards et al. (<xref ref-type="bibr" rid="B14">2008</xref>) present their results in the form of a measure referred to as <italic>&#946;</italic>, and which corresponds to the reciprocal of <italic>-g</italic> as defined here, i.e., <italic>g = &#8211;</italic>1/<italic>&#946;</italic>. <italic>&#946;</italic> was found to have an average value of 3.5 (<italic>g = &#8211;</italic>0.29) in the Narrow condition and 6.2 (<italic>g =</italic> &#8211;0.16) in the Wide condition. <xref ref-type="table" rid="T4">Table 4</xref> displays the observed effect size (&#8211;0.12) expressed as a difference in <italic>g</italic>&#8217;s average value between the two conditions. This corresponds to a difference of &#8211;0.03 in the derivative. In other terms, at the categorical boundary, the proportion of voiced responses decreased by an additional 3% over a 1-ms VOT time unit, in the Narrow compared with the Wide condition.</p>
<p>Also note that the observed difference in <italic>g</italic> between conditions (namely, &#8211;0.12) is about 4 times lower than the difference between the values assigned to <italic>g</italic> in these two conditions (&#8211;0.78 &#8211; (&#8211;26) = &#8211;0.52). Clayards et al. observe that &#8220;as predicted, listeners are less certain than the optimal observer given either of the distributions&#8221; (<xref ref-type="bibr" rid="B14">2008, p 806</xref>). One potential way of accounting for this greater uncertainty is by assuming that the listeners&#8217; responses are affected by sensorymotor variance, in addition to the variance associated with the voiced and voiceless categories. This, in turn, opens up the interesting possibility that sensorymotor variance be estimated as the difference between the expected and observed effect sizes.</p>
</sec>
<sec>
<title>Appendix 3: Preliminary assessment of the stimuli</title>
<p>We conducted a preliminary study in which the stimuli were presented to participants in a standard, two-alternative forced-choice test, performed individually by each participant. This was done to ensure that the stimuli would yield response patterns with the desired primary characteristics (two endpoint stimuli unambiguously or close to unambiguously identified as /ba/ and /pa/, respectively; categorical boundary located in the vicinity of the midpoint on the VOT continuum). We also aimed to collect data that would allow us to estimate the amount of inter-individual variability in both the location of the categorical boundary and the Slope parameter of the psychometric curve at this point, as these estimates were required for the data simulation (Appendix 4).</p>
<p>The study was implemented in jsPsych 6.3.1 (<xref ref-type="bibr" rid="B20">de Leeuw, 2015</xref>), jatosified using JATOS version 3.7.2 (<xref ref-type="bibr" rid="B51">Lange et al., 2015</xref>), and deployed on the MindProbe<xref ref-type="fn" rid="n11">11</xref> JATOS server.</p>
<p>We recruited participants online and through the Prolific<xref ref-type="fn" rid="n12">12</xref> crowdsourcing website. To be preselected, participants had to fulfill the following criteria: be born and live in England; have English as first language; have no hearing difficulties and normal or corrected-to-normal vision.</p>
<p>We also asked participants to take the experiment on a computer (as opposed to a tablet or smartphone), in a quiet room, away from any distractions, and to wear headphones or earphones connected to their computer.</p>
<p>Preselected participants were directed from Prolific to the MindProbe server. They were informed that their data would be collected, stored and processed in a fully anonymous manner and were asked to digitally agree to a consent form. Next, and to ensure that they were equipped with headphones/earphones, participants were required to pass the online headphone screening test designed by Milne et al. (<xref ref-type="bibr" rid="B61">2021</xref>). Participants who did not respond correctly in at least five of the six trials were not allowed to continue.</p>
<p>Participants were then told that they would be presented with a sequence of speech sounds that may be identified as &#8220;ba&#8221; or &#8220;pa&#8221;. After hearing each sound, the participants&#8217; task was to say whether that sound corresponded to &#8220;ba&#8221; or &#8220;pa&#8221; by clicking on one of two buttons displayed on their computer screen. Participants were asked to respond both as accurately and fast as possible, and to try to always provide a response even if in doubt.</p>
<p>In a first, training phase, participants performed the task on six sounds, which corresponded to either of the two VOT continuum endpoints, and were therefore expected to be unambiguously associated with /ba/ or /pa/ (three repetitions per endpoint, randomized order). Once they had responded to each stimulus, participants were told whether or not their response was the expected one. In the subsequent, test phase, participants heard 10 repetitions of each of the 9 auditory stimuli on the VOT continuum, in a fully randomized order. No feedback was given after each response.</p>
<p>In both the training and test phase, and at the onset of each trial, a cross was displayed at the center of the screen for 750 ms. This was immediately followed by the auditory stimulus, and the display of the two response buttons, labelled <italic>ba</italic> and <italic>pa</italic>, respectively, on either side of the screen center. Participants had 3,000 ms to respond. A 15-s pause was made at the end of the first half of the test. The average duration of the test phase was 5&#8217;30&#8221;.</p>
<p>30 participants (16 female, mean age: 42.5 years old, min: 21 years old, max: 66 years old) performed the experiment to the end. They were paid &#163;2.5 for their participation.</p>
<p>We made the stimuli and dataset available on OSF under a Creative Commons licence at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/gj9c4/">https://osf.io/gj9c4/</ext-link>.</p>
<sec>
<title>Bayesian logistic regression</title>
<p>We submitted the data to a Bayesian logistic regression using the <monospace>brms</monospace> R package (<xref ref-type="bibr" rid="B10">B&#252;rkner, 2017</xref>). The <monospace>brms</monospace> model was designed as follows:</p>
<p><monospace>resp &#126; 1 + vot_s + (1 + vot_s &#124; subj_id)</monospace></p>
<p>where <monospace>resp</monospace> is the participant&#8217;s response to the stimulus (0: /p/, 1: /b/), <monospace>vot_s</monospace> refers to the VOT value for the stimulus, standardized by subtracting the mean VOT, namely, 22 ms, and <monospace>subj_id</monospace> refers to the participant&#8217;s identification number. This model therefore comprises one population-level intercept and one population-level slope, and two group-level terms, the intercept and slope for each participant. Response was defined as a Bernoulli random variable and the link function was the logit.</p>
<p>This amounts to predicting the response <italic>resp<sub>ij</sub></italic> from Participant <italic>i</italic> to Stimulus <italic>j</italic> as follows:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq038-mml">
<mml:mrow><mml:mtext mathvariant="italic">res</mml:mtext><mml:msub><mml:mi>p</mml:mi><mml:mrow><mml:mtext mathvariant="italic">ij</mml:mtext></mml:mrow></mml:msub><mml:mo>&#126;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>0</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>+</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mtext mathvariant="italic">vot</mml:mtext><mml:mo>&#x005F;</mml:mo><mml:msub><mml:mi>s</mml:mi><mml:mi>j</mml:mi></mml:msub></mml:mrow>
</mml:math>
<tex-math id="M38">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
res{p_{ij}}\sim({\beta _0} + {u_{0i}}) + ({\beta _1} + {u_{1i}})\;vot\_{s_j}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e36.gif"/>
</alternatives>
</disp-formula>
<p>where <italic>&#946;</italic><sub>0</sub> is the intercept at 0 on the standardized VOT continuum, <italic>&#946;</italic><sub>1</sub> is the Slope parameter of the logistic function, and <italic>u</italic><sub>0</sub><italic><sub>i</sub>, u</italic><sub>1</sub><italic><sub>i</sub></italic> the random intercept and slope for Participant <italic>i</italic>, respectively.</p>
<p>We assigned <italic>&#946;</italic><sub>0</sub> a prior normal distribution with a mean of 0 and a standard deviation of 1, and <italic>&#946;</italic><sub>1</sub> a prior normal distribution with a mean of &#8211;0.5 and a standard deviation of 1. The other <monospace>brms</monospace> priors were set to their default values.</p>
<p>The <italic>b</italic> and <italic>g</italic> parameters as defined in our model can be directly derived from the <italic>&#946;</italic><sub>0</sub> and <italic>&#946;</italic><sub>1</sub> coefficients in the following way: <italic>b = &#8211;&#946;</italic><sub>0</sub>, <italic>g = &#946;</italic><sub>1</sub>.<xref ref-type="fn" rid="n13">13</xref></p>
<p><xref ref-type="table" rid="T5">Table 5</xref> contains the summary statistics for both the population-level and group-level effects, as computed from 2,000 draws extracted from the parameters&#8217; posterior distributions. Slope <italic>g</italic> has a mean value of &#8211;0.73. Bias <italic>b</italic> contributes to setting the location of the categorical boundary <italic>b/g</italic>, whose mean value is &#8211;1.37 ms on the standardized VOT scale (i.e., 20.63 ms on the original scale), with a standard deviation of 0.48 ms.</p>
<table-wrap id="T5">
<caption>
<p><bold>Table 5:</bold> Summary statistics associated with the logistic regression.</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"><italic>Parameter</italic></td>
<td align="left" valign="top"><bold>Mean</bold></td>
<td align="left" valign="top"><bold>SD</bold></td>
<td align="left" valign="top"><bold>HDI lower limit</bold></td>
<td align="left" valign="top"><bold>HDI upper limit</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top"><italic>g</italic></td>
<td align="left" valign="top">&#8211;0.73</td>
<td align="left" valign="top">0.08</td>
<td align="left" valign="top">&#8211;0.86</td>
<td align="left" valign="top">&#8211;0.61</td>
</tr>
<tr>
<td align="left" valign="top"><italic>derivative</italic></td>
<td align="left" valign="top">&#8211;0.18</td>
<td align="left" valign="top">0.02</td>
<td align="left" valign="top">&#8211;0.22</td>
<td align="left" valign="top">&#8211;0.15</td>
</tr>
<tr>
<td align="left" valign="top"><italic>b</italic></td>
<td align="left" valign="top">1.00</td>
<td align="left" valign="top">0.37</td>
<td align="left" valign="top">0.40</td>
<td align="left" valign="top">1.61</td>
</tr>
<tr>
<td align="left" valign="top"><italic>b/g</italic></td>
<td align="left" valign="top">&#8211;1.37</td>
<td align="left" valign="top">0.48</td>
<td align="left" valign="top">&#8211;2.13</td>
<td align="left" valign="top">&#8211;0.57</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#964;</italic><sub>0</sub></td>
<td align="left" valign="top">2.04</td>
<td align="left" valign="top">0.34</td>
<td align="left" valign="top">1.54</td>
<td align="left" valign="top">2.66</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#964;</italic><sub>1</sub></td>
<td align="left" valign="top">0.32</td>
<td align="left" valign="top">0.07</td>
<td align="left" valign="top">0.21</td>
<td align="left" valign="top">0.45</td>
</tr>
<tr>
<td align="left" valign="top"><italic>&#961;</italic></td>
<td align="left" valign="top">0.19</td>
<td align="left" valign="top">0.23</td>
<td align="left" valign="top">&#8211;0.22</td>
<td align="left" valign="top">0.55</td>
</tr>
</tbody>
</table>
</table-wrap>
<p><xref ref-type="table" rid="T5">Table 5</xref> also contains the summary statistics for the derivative of the logistic function <inline-formula>
<alternatives>
<mml:math id="Eq039-mml">
<mml:mrow><mml:mi>f</mml:mi><mml:mo stretchy='false'>(</mml:mo><mml:mi>S</mml:mi><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mstyle scriptlevel='+1'><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mstyle></mml:mrow>
</mml:math>
<tex-math id="M39">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
f(S) = {\textstyle{1 \over {1 + {e^{- gS + b}}}}}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e10.gif"/>
</alternatives>
</inline-formula> at the categorical boundary, <inline-formula>
<alternatives>
<mml:math id="Eq040-mml">
<mml:mrow><mml:mi>S</mml:mi><mml:mo>=</mml:mo><mml:msup><mml:mi>f</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup><mml:mo stretchy='false'>(</mml:mo><mml:mn>0.5</mml:mn><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mi>b</mml:mi><mml:mo>/</mml:mo><mml:mi>g</mml:mi></mml:mrow>
</mml:math>
<tex-math id="M40">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
S = {f^{- 1}}(0.5) = b/g
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e11.gif"/>
</alternatives>
</inline-formula>:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq041-mml">
<mml:mrow><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mfrac><mml:mi>g</mml:mi><mml:mrow><mml:msup><mml:mrow><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mi>g</mml:mi><mml:mi>S</mml:mi><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow><mml:mn>2</mml:mn></mml:msup></mml:mrow></mml:mfrac><mml:mo>=</mml:mo><mml:mfrac><mml:mi>g</mml:mi><mml:mn>4</mml:mn></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M41">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
{e^{- gS + b}}\frac{g}{{{{(1 + {e^{- gS + b}})}^2}}} = \frac{g}{4}
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e37.gif"/>
</alternatives>
</disp-formula>
<p>This represents the rate of change of the logistic function at the 0.5 crossover point and is expressed as a decrease in the proportion of /ba/ responses over an increase of 1 ms in VOT. The derivative allows us to express rate of change with respect to a 1-ms time unit, i.e., independently of the width of the interval between two adjacent stimuli on the VOT continuum, and this makes it easier to make comparisons with predicted/observed identification functions in other studies. In particular, we provide derivative values computed from Clayards et al.&#8217;s (<xref ref-type="bibr" rid="B14">2008</xref>) model and data in Appendix 2.</p>
<p>As measures of inter-individual variability, standard deviations <italic>&#964;</italic><sub>0</sub> and <italic>&#964;</italic><sub>1</sub> associated with the by-participant random intercepts <italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> and random slopes <italic>u</italic><sub>1</sub><italic><sub>i</sub></italic>, respectively, and the correlation coefficient <italic>&#961;</italic> between <italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> and <italic>u</italic><sub>1</sub><italic><sub>i</sub></italic>, were used in the data simulation (Appendix 4).</p>
<p>The proportions of /ba/ responses for each participant and each of the stimuli on the /ba/- /pa/ VOT continuum are displayed in <xref ref-type="fig" rid="F7">Figure 7</xref> as a scatter plot, with a small jitter on both the horizontal and vertical axes to improve legibility. The orange line represents mean values of the posterior predictive distribution for each stimulus, computed over 2000 draws for each participant and pooled across participants. The orange stripe represents the 95% highest-density interval.</p>
<fig id="F7">
<caption>
<p><bold>Figure 7:</bold> Listeners&#8217; individual proportions of /ba/ responses to each of the stimuli on the /ba/-/pa/ VOT continuum. Orange line: mean values of the posterior predictive distribution, orange stripe: 95% highest-density interval. Phonemic categories /b/ and /p/ are associated with short and longer VOT values, respectively.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g7.png"/>
</fig>
<p>Kernel density plots for <italic>g</italic>, derivative, <italic>b</italic> and <italic>b/g</italic> are shown in <xref ref-type="fig" rid="F8">Figure 8</xref>, together with the mean value and highest density interval for each distribution. Note that for <italic>b/g</italic>, the horizontal scale represents the standardized VOT in ms, with 0 corresponding to the continuum midpoint (22 ms on the original scale).</p>
<fig id="F8">
<caption>
<p><bold>Figure 8:</bold> Kernel density plots for <italic>g</italic> (upper left panel), derivative (upper right panel), <italic>b</italic> (lower left panel) and <italic>b/g</italic> (lower right panel). For each distribution, the mean value is displayed as a black circle and the 95% highest density interval is displayed in darker blue.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g8.png"/>
</fig>
</sec>
</sec>
<sec>
<title>Appendix 4: Data simulation</title>
<p>In this section, we present the main results of a series of numerical computations whose aim was to simulate the expected changes in the participants&#8217; responses in the /p/- vs. /b/-Biased condition, and the Steep- vs. Shallow-Slope condition, on the basis of both our pilot study and our model. In designing the data simulation procedure, we drew on the blueprint proposed by DeBruine &amp; Barr (<xref ref-type="bibr" rid="B21">2021</xref>), which we extended to Bayesian logistic regression. The simulations were carried out using R on a dedicated RStudio server.</p>
<p>We performed a series of virtual experiments each of which entailed simulating the responses provided by <italic>N</italic> participants. Each response consisted in mapping one stimulus, as characterized by its VOT duration, onto either /b/ or /p/. Following the experimental design laid out in 4.3, we used 8 VOT durations equally spaced from 6 to 34 ms, which were each presented 10 times to each participant.</p>
<p>Each virtual experiment first involved setting values for 1) Bias <italic>b</italic> and Slope <italic>g</italic> across participants, and 2) standard deviation in participants&#8217; random intercept <italic>&#964;</italic><sub>0</sub>, standard deviation in participants&#8217; random slope <italic>&#964;</italic><sub>1</sub>, and correlation <italic>&#961;</italic> between <italic>&#964;</italic><sub>0</sub> and <italic>&#964;</italic><sub>1</sub>, as measures of between-participant variation in both Bias and Slope. These values were randomly extracted from the parameters&#8217; posterior distributions, as established in our pilot study (see <xref ref-type="table" rid="T5">Table 5</xref>). We then generated a random intercept <italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> and random slope <italic>u</italic><sub>1</sub><italic><sub>i</sub></italic> for each participant <italic>i</italic> from <italic>&#964;</italic><sub>0</sub>, <italic>&#964;</italic><sub>1</sub> and <italic>&#961;</italic>, using the <monospace>rnorm_multi</monospace> function in DeBruine&#8217;s <monospace>faux</monospace> R package. As a result, parameters <italic>b, g, &#964;</italic><sub>0</sub>, <italic>&#964;</italic><sub>1</sub> and <italic>&#961;</italic> had constant values in each experiment (as in <xref ref-type="bibr" rid="B21">DeBruine &amp; Barr, 2021</xref>) but varied across experiments, while parameters <italic>u</italic><sub>0</sub><italic><sub>i</sub></italic> and <italic>u</italic><sub>1</sub><italic><sub>i</sub></italic> varied across participants within each experiment.</p>
<p>We subdivided each set of <italic>N</italic> participants into four subsets that corresponded to our four experimental conditions. For each of the subsets, Bias <italic>b</italic> and Slope <italic>g</italic> were modified by an amount equal to half of the effect size that was expected in the given experimental condition, as indicated in 4.3.</p>
<p>We then computed the probability <italic>p</italic> for each stimulus to be identified as /b/ by Participant <italic>i</italic> as follows:</p>
<disp-formula>
<alternatives>
<mml:math id="Eq042-mml">
<mml:mrow><mml:mi>p</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mn>1</mml:mn><mml:mrow><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>g</mml:mi><mml:mo>&#x205F;</mml:mo><mml:mo>+</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>1</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo><mml:mo>&#x205F;</mml:mo><mml:mtext mathvariant="italic">vot</mml:mtext><mml:mo>&#x005F;</mml:mo><mml:mi>s</mml:mi><mml:mo>&#x2009;</mml:mo><mml:mo>+</mml:mo><mml:mo>&#x2009;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mi>b</mml:mi><mml:mo>&#x205F;</mml:mo><mml:mo>+</mml:mo><mml:mo>&#x205F;</mml:mo><mml:msub><mml:mi>u</mml:mi><mml:mrow><mml:mn>0</mml:mn><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup></mml:mrow></mml:mfrac></mml:mrow>
</mml:math>
<tex-math id="M42">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
p = \frac{1}{{1 + {e^{- (g\: + \:{u_{1i}})\:vot\_s\; + \;(b\: + \:{u_{0i}})}}}}\]
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e38.gif"/>
</alternatives>
</disp-formula>
<p>The response to the stimulus was generated as a random sample of the Bernoulli distribution Ber(<italic>p</italic>).</p>
<p>Two series of 100 experiments were simulated, with a number of virtual participants per experiment that was set to 200 in the first series, and 320 in the second series. We submitted each data set (<italic>N</italic> participants &#215; 8 stimuli &#215; 10 repetitions = <italic>N &#215; 80</italic> trials) to a Bayesian logistic regression by means of the <monospace>brms</monospace> R package, using the same <monospace>brms</monospace> formula as in our proposed statistical analysis (4.4). Summary statistics for each of the model&#8217;s parameters were then computed from 2,000 draws extracted from the parameters&#8217; posterior distributions.</p>
<p>In the following, we focus on the differences in Bias <italic>b</italic>, Slope <italic>g</italic> and categorical boundary <italic>b/g</italic> as a function of the experimental conditions. These differences, as estimated by the <monospace>brms</monospace> model in each experiment, then averaged over the 100 experiments in each series, are shown in <xref ref-type="table" rid="T6">Table 6</xref>. As can be seen, they are close to or equal to those that we entered in the data generation procedure in both series. For example, the mean estimate of the shift in the location <italic>b/g</italic> of the categorical boundary in the /p/-Biased compared with the /b/-Biased conditions is close to the input value, namely, &#8211;3 ms. Likewise, the additional decrease in Slope <italic>g</italic> in the Steep-Slope compared with the Shallow-Slope conditions is equal to the input value, &#8211;0.12. The difference in Bias <italic>b</italic> in the /p/-Biased relative to the /b/-Biased condition within both the Shallow- and Steep-Slope conditions is itself very close to the input value. Thus, as would be expected, the Bayesian logistic regression allows us to retrieve the values that we assigned to the changes in the Bias and Slope parameters as a function of experimental conditions in the generation of these synthetic data sets.</p>
<table-wrap id="T6">
<caption>
<p><bold>Table 6:</bold> Differences in Bias <italic>b</italic>, Slope <italic>g</italic> and categorical boundary <italic>b/g</italic> as a function of the four experimental conditions, as estimated by <monospace>brms</monospace> and averaged over 100 simulations in Series 1 (200 virtual participants per simulation) and Series 2 (320 virtual participants per simulation).</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"><bold>Series</bold></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold><monospace>brms</monospace> models&#8217; parameters</bold></td>
<td align="left" valign="top"><bold>Explanation</bold></td>
<td align="left" valign="top"><bold>Simulated mean value</bold></td>
<td align="left" valign="top"><bold>Mean estimate from models</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">1</td>
<td align="left" valign="top">Shallow-Slope cond.</td>
<td align="left" valign="top"><italic>&#8211;&#946;</italic><sub>1</sub></td>
<td align="left" valign="top">Shift in <italic>b</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">+2.01</td>
<td align="left" valign="top">+1.94</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"><italic>&#8211;&#946;</italic><sub>1</sub> <italic>/ &#946;</italic><sub>4</sub></td>
<td align="left" valign="top">Shift in <italic>b/g</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">&#8211;3.00</td>
<td align="left" valign="top">&#8211;2.94</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top">Steep-Slope cond.</td>
<td align="left" valign="top">&#8211;<italic>(&#946;</italic><sub>1</sub> <italic>+ &#946;</italic><sub>3</sub><italic>)</italic></td>
<td align="left" valign="top">Shift in <italic>b</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">+2.37</td>
<td align="left" valign="top">+2.35</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top">&#8211;<italic>(&#946;</italic><sub>1</sub> <italic>+ &#946;</italic><sub>3</sub><italic>) / (&#946;</italic><sub>4</sub> <italic>+ &#946;</italic><sub>5</sub><italic>)</italic></td>
<td align="left" valign="top">Shift in <italic>b/g</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">&#8211;3.00</td>
<td align="left" valign="top">&#8211;3.00</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top"><italic>+&#946;</italic><sub>4</sub></td>
<td align="left" valign="top">Shift in <italic>g</italic>, Steep- vs. Shallow-Slope cond.</td>
<td align="left" valign="top">&#8211;0.12</td>
<td align="left" valign="top">&#8211;0.12</td>
</tr>
<tr>
<td align="left" valign="top">2</td>
<td align="left" valign="top">Shallow-Slope cond.</td>
<td align="left" valign="top"><italic>&#8211;&#946;</italic><sub>1</sub></td>
<td align="left" valign="top">Shift in <italic>b</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">+2.01</td>
<td align="left" valign="top">+1.99</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"><italic>&#8211;&#946;</italic><sub>1</sub> <italic>/ &#946;</italic><sub>4</sub></td>
<td align="left" valign="top">Shift in <italic>b/g</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">&#8211;3.00</td>
<td align="left" valign="top">&#8211;3.09</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top">Steep-Slope cond.</td>
<td align="left" valign="top">&#8211;(<italic>&#946;</italic><sub>1</sub> <italic>+ &#946;</italic><sub>3</sub>)</td>
<td align="left" valign="top">Shift in <italic>b</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">+2.37</td>
<td align="left" valign="top">+2.25</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top">&#8211;(<italic>&#946;</italic><sub>1</sub> <italic>+ &#946;</italic><sub>3</sub>) <italic>/</italic> (<italic>&#946;</italic><sub>4</sub> <italic>+ &#946;</italic><sub>5</sub>)</td>
<td align="left" valign="top">Shift in <italic>b/g</italic>, /p/- vs. /b/-Biased cond.</td>
<td align="left" valign="top">&#8211;3.00</td>
<td align="left" valign="top">&#8211;2.93</td>
</tr>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top"><italic>+&#946;</italic><sub>4</sub></td>
<td align="left" valign="top">Shift in <italic>g</italic>, Steep- vs. Shallow-Slope cond.</td>
<td align="left" valign="top">&#8211;0.12</td>
<td align="left" valign="top">&#8211;0.12</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>Schematized distributions of the differences in categorical boundary <italic>b/g</italic> and Slope <italic>g</italic> as a function of experimental condition in each simulation, as estimated using <monospace>brms</monospace>, are shown in <xref ref-type="fig" rid="F9">Figure 9</xref>. As already pointed out, the mean estimates are close to the values assigned to these differences in the data generation procedure. However, one can also see that Slope <italic>g</italic> got close to 0 in a number of simulations in the first series (upper right panel). More precisely, 0 was within the 5%&#8211;95% percentile range of the posterior distribution for <italic>g</italic> in 31% of cases (i.e., 31 simulations out of 100). That proportion fell to 10% in the second series, i.e., when the number of virtual participants was raised to 320 (lower right panel). Thus, given our model, priors, and results from our pilot experiment, the difference in <italic>g</italic> between the Steep-Slope and Shallow-Slope conditions was very likely to be negative when the number of participants was set to 320 or more. As a consequence, and in our experiment, we had 320 participants.</p>
<fig id="F9">
<caption>
<p><bold>Figure 9:</bold> Changes in categorical boundary <italic>b/g</italic> and Slope <italic>g</italic> as a function of experimental condition, as estimated by means of a Bayesian logistic regression applied to the responses generated in each simulated experiment. (a) Estimated change in <italic>b/g</italic> in the /p/-Biased condition relative to the /b/-Biased condition, within the Shallow-Slope condition. (b) Estimated change in <italic>b/g</italic> in the /p/-Biased condition relative to the /b/-Biased condition, within the Steep-Slope condition. (c) Estimated change in <italic>g</italic> in the Steep-Slope condition relative to the Shallow-Slope condition. Simulations were performed 100 times with either 200 virtual participants (upper panel) or 320 participants (lower panel) per simulation. In each panel, the mean value (black circle) and extent of the 5%&#8211;95% percentile range (horizontal grey line) are displayed as estimated in each simulation from the posterior distributions of the parameters.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-3-1-165-g9.png"/>
</fig>
</sec>
<sec>
<title>Data accessibility statement</title>
<p>The data and R scripts associated with this paper are available at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/6xv3c/">https://osf.io/6xv3c/</ext-link>.</p>
</sec>
</body>
<back>
<fn-group>
<fn id="n1"><p>In this work, and in a manner that is common in the field, we will consider that these categories are phonemic.</p></fn>
<fn id="n2"><p>In practice, and in a standard phoneme identification task, listeners are presented with a finite number of stimuli equally spaced between two endpoints along a given acoustic dimension. The two endpoints are expected to be unambiguously identified as Category <italic>c<sub>1</sub></italic> and <italic>c<sub>2</sub></italic>, respectively. The midpoint refers to the value arithmetically halfway between the two endpoints on the chosen acoustic dimension.</p></fn>
<fn id="n3"><p>More precisely, the first derivative.</p></fn>
<fn id="n4"><p>As already indicated, the slope of the logistic curve at the location of the categorical boundary is mathematically equal to <italic>g/</italic>4. However, in keeping with the literature in both Bayesian statistics (e.g., <xref ref-type="bibr" rid="B47">Kruschke, 2014, Chap 21</xref>) and Bayesian models of speech perception (e.g., <xref ref-type="bibr" rid="B14">Clayards et al., 2008</xref>), we will still refer to <italic>g</italic> as the Slope parameter, but with a capital S to distinguish this term from slope (with a lower-case s) as designating the first derivative <italic>g/</italic>4.</p></fn>
<fn id="n5"><p>Version 32, as available at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://github.com/ListenLab/VOT">https://github.com/ListenLab/VOT</ext-link>.</p></fn>
<fn id="n6"><p>Given that this probability has an estimated mean value of <inline-formula>
<alternatives>
<mml:math id="Eq043-mml">
<mml:mrow><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>0</mml:mn></mml:msub></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mn>0</mml:mn></mml:msup><mml:mo stretchy='false'>)</mml:mo><mml:mo>=</mml:mo><mml:mn>0.5</mml:mn></mml:mrow>
</mml:math>
<tex-math id="M43">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
1/(1 + {e^{- {\beta _0}}}) = 1/(1 + {e^0}) = 0.5
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e5.gif"/>
</alternatives>
</inline-formula> at the midpoint.</p></fn>
<fn id="n7"><p>Note that the derivative of the logistic function at the location of the categorical boundary is <italic>g/</italic>4; see Appendix 3 for further detail.</p></fn>
<fn id="n8"><p>In K&amp;J&#8217;s model, the distance between the two means <italic>&#956;<sub>c<sub>2</sub></sub></italic> &#8211; <italic>&#956;<sub>c<sub>1</sub></sub></italic> is fixed and, as a consequence, both means are bound to move around in the same direction and by the same extent.</p></fn>
<fn id="n9"><p>Note that, in the K&amp;J model, the prior probabilities <italic>p</italic>(<italic>c<sub>1</sub></italic>) and <italic>p</italic>(<italic>c<sub>2</sub></italic>) assigned by the listener to the two categories <italic>c<sub>1</sub></italic> and <italic>c<sub>2</sub></italic> are fixed and both set to 0.5. As a result, they are not expected to have an influence on the shape of the listener&#8217;s categorization function.</p></fn>
<fn id="n10"><p>In Clayards et al. (<xref ref-type="bibr" rid="B14">2008</xref>), the chosen dependent variable is the proportion of voiceless responses, and the listener&#8217;s response curve is modeled as a logistic function with a positive slope. We have opted to use the proportion of voiced responses as the dependent variable, which we model as a logistic function with a negative slope. This is why both <italic>g</italic> and the derivative have negative values in <xref ref-type="table" rid="T4">Table 4</xref>.</p></fn>
<fn id="n11"><p><ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://mindprobe.eu">https://mindprobe.eu</ext-link>.</p></fn>
<fn id="n12"><p><ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://www.prolific.co/">https://www.prolific.co/</ext-link>.</p></fn>
<fn id="n13"><p>Note that the generic logistic function that links a predictor <italic>x</italic> to a response <italic>y</italic> is defined as <inline-formula>
<alternatives>
<mml:math id="Eq044-mml">
<mml:mrow><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>0</mml:mn></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mo>&#x03B2;</mml:mo><mml:mn>1</mml:mn></mml:msub><mml:mi>x</mml:mi><mml:mo stretchy='false'>)</mml:mo></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow>
</mml:math>
<tex-math id="M44">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
y = 1/(1 + {e^{- ({\beta _0} + {\beta _1}x)}})
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e12.gif"/>
</alternatives>
</inline-formula> in <monospace>brms</monospace>, whereas we used the variant <inline-formula>
<alternatives>
<mml:math id="Eq045-mml">
<mml:mrow><mml:mi>y</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn><mml:mo>/</mml:mo><mml:mo stretchy='false'>(</mml:mo><mml:mn>1</mml:mn><mml:mo>+</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mtext mathvariant="italic">gx</mml:mtext><mml:mo>+</mml:mo><mml:mi>b</mml:mi></mml:mrow></mml:msup><mml:mo stretchy='false'>)</mml:mo></mml:mrow>
</mml:math>
<tex-math id="M45">
\documentclass[10pt]{article}
\usepackage{wasysym}
\usepackage[substack]{amsmath}
\usepackage{amsfonts}
\usepackage{amssymb}
\usepackage{amsbsy}
\usepackage[mathscr]{eucal}
\usepackage{mathrsfs}
\usepackage{pmc}
\usepackage[Euler]{upgreek}
\pagestyle{empty}
\oddsidemargin -1.0in
\begin{document}
\[
y = 1/(1 + {e^{- gx + b}})
\]
\end{document}
</tex-math>
<graphic xlink:href="glossapx-3-1-165-e13.gif"/>
</alternatives>
</inline-formula>. This is why <italic>b</italic> in the model is the opposite of <italic>&#946;<sub>0</sub></italic> as computed in <monospace>brms</monospace>.</p></fn>
</fn-group>
<sec>
<title>Ethics and consent</title>
<p>The experiment received the approval of the Ethics Committee of Aix-Marseille University (approval number: 2022-02-24-009).</p>
</sec>
<ack>
<title>Acknowledgements</title>
<p>Thanks are due to Ladislas Nalborczyk, Elin Runnqvist and Kristof Strijkers for fruitful discussions. We are also grateful to Sharon Peperkamp, Timo Roettger, and two anonymous reviewers, for helpful comments and suggestions.</p>
<p>This work was carried out with the support of the Laboratoire Parole et Langage, Aix-Marseille University, France and the French National Center for Scientific Research. Support from the Institute for Language, Communication and the Brain (ILCB, Grant ANR-16-CONV-0002) at Aix-Marseille University, the Excellence Initiative of Aix-Marseille University (A*MIDEX), and the Institut Carnot Cognition, is also gratefully acknowledged.</p>
</ack>
<sec>
<title>Competing Interests</title>
<p>The authors have no competing interests to declare.</p>
</sec>
<sec>
<title>Authors&#8217; contributions</title>
<p>NN: conceptualization, methodology, software, data collection, statistical analyses, writing, review &amp; editing, funding acquisition</p>
<p>LL: conceptualization, conduction of a pilot experiment, methodology, review</p>
<p>LH: conduction of a pilot experiment</p>
<p>JLS: conceptualization, methodology, writing, review &amp; editing</p>
<p>JD: conceptualization, methodology, writing, review &amp; editing</p>
</sec>
<ref-list>
<ref id="B1"><label>1</label><mixed-citation publication-type="journal"><string-name><surname>Abel</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Babel</surname>, <given-names>M.</given-names></string-name> (<year>2016</year>). <article-title>Cognitive load reduces perceived linguistic convergence between dyads</article-title>. <source>Language and Speech</source>, <volume>60</volume>(<issue>3</issue>), <fpage>1</fpage>&#8211;<lpage>24</lpage>. DOI: <pub-id pub-id-type="doi">10.1177/0023830916665652</pub-id></mixed-citation></ref>
<ref id="B2"><label>2</label><mixed-citation publication-type="journal"><string-name><surname>Adank</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Hagoort</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Bekkering</surname>, <given-names>H.</given-names></string-name> (<year>2010</year>). <article-title>Imitation improves language comprehension</article-title>. <source>Psychological Science</source>, <volume>21</volume>(<issue>12</issue>), <fpage>1903</fpage>&#8211;<lpage>1909</lpage>. DOI: <pub-id pub-id-type="doi">10.1177/0956797610389192</pub-id></mixed-citation></ref>
<ref id="B3"><label>3</label><mixed-citation publication-type="journal"><string-name><surname>Alais</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Burr</surname>, <given-names>D.</given-names></string-name> (<year>2004</year>). <article-title>The ventriloquist effect results from near-optimal bimodal integration</article-title>. <source>Current Biology</source>, <volume>14</volume>, <fpage>257</fpage>&#8211;<lpage>262</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cub.2004.01.029</pub-id></mixed-citation></ref>
<ref id="B4"><label>4</label><mixed-citation publication-type="journal"><string-name><surname>Apfelbaum</surname>, <given-names>K. S.</given-names></string-name>, <string-name><surname>Kutlu</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>McMurray</surname>, <given-names>B.</given-names></string-name>, &amp; <string-name><surname>Kapnoula</surname>, <given-names>E. C.</given-names></string-name> (<year>2022</year>). <article-title>Don&#8217;t force it! Gradient speech categorization calls for continuous categorization tasks</article-title>. <source>The Journal of the Acoustical Society of America</source>, <volume>152</volume>(<issue>6</issue>), <fpage>3728</fpage>&#8211;<lpage>3745</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/10.0015201</pub-id></mixed-citation></ref>
<ref id="B5"><label>5</label><mixed-citation publication-type="journal"><string-name><surname>Asch</surname>, <given-names>S.</given-names></string-name> (<year>1951</year>). <article-title>Effects of group pressure upon the modification and distortion of judgments</article-title>. <source>Organizational Influence Processes</source>, <volume>58</volume>, <fpage>295</fpage>&#8211;<lpage>303</lpage>.</mixed-citation></ref>
<ref id="B6"><label>6</label><mixed-citation publication-type="webpage"><string-name><surname>Babel</surname>, <given-names>M.</given-names></string-name> (<year>2011</year>). <article-title>Imitation in speech</article-title>. <source>Acoustics Today</source>, <volume>7</volume>(<issue>4</issue>), <fpage>16</fpage>&#8211;<lpage>23</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/1.3684224</pub-id></mixed-citation></ref>
<ref id="B7"><label>7</label><mixed-citation publication-type="journal"><string-name><surname>Bahrami</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Olsen</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Latham</surname>, <given-names>P. E.</given-names></string-name>, <string-name><surname>Roepstorff</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Rees</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Frith</surname>, <given-names>C. D.</given-names></string-name> (<year>2010</year>). <article-title>Optimally interacting minds</article-title>. <source>Science</source>, <volume>329</volume>(<issue>5995</issue>), <fpage>1081</fpage>&#8211;<lpage>1085</lpage>. DOI: <pub-id pub-id-type="doi">10.1126/science.1185718</pub-id></mixed-citation></ref>
<ref id="B8"><label>8</label><mixed-citation publication-type="book"><string-name><surname>Bessi&#232;re</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Mazer</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Ahuactzin</surname>, <given-names>J. M.</given-names></string-name>, &amp; <string-name><surname>Mekhnacha</surname>, <given-names>K.</given-names></string-name> (<year>2013</year>). <source>Bayesian programming</source>. <publisher-name>CRC Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1201/b16111</pub-id></mixed-citation></ref>
<ref id="B9"><label>9</label><mixed-citation publication-type="journal"><string-name><surname>Brainard</surname>, <given-names>D. H.</given-names></string-name>, &amp; <string-name><surname>Freeman</surname>, <given-names>W. T.</given-names></string-name> (<year>1997</year>). <article-title>Bayesian color constancy</article-title>. <source>Journal of the Optical Society of America A</source>, <volume>14</volume>(<issue>7</issue>), <fpage>1393</fpage>&#8211;<lpage>1411</lpage>. DOI: <pub-id pub-id-type="doi">10.1364/JOSAA.14.001393</pub-id></mixed-citation></ref>
<ref id="B10"><label>10</label><mixed-citation publication-type="journal"><string-name><surname>B&#252;rkner</surname>, <given-names>P.-C.</given-names></string-name> (<year>2017</year>). <article-title>Brms: An R package for Bayesian multilevel models using Stan</article-title>. <source>Journal of Statistical Software</source>, <volume>80</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>28</lpage>. DOI: <pub-id pub-id-type="doi">10.18637/jss.v080.i01</pub-id></mixed-citation></ref>
<ref id="B11"><label>11</label><mixed-citation publication-type="journal"><string-name><surname>Carr</surname>, <given-names>J. W.</given-names></string-name>, <string-name><surname>Smith</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Culbertson</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Kirby</surname>, <given-names>S.</given-names></string-name> (<year>2020</year>). <article-title>Simplicity and informativeness in semantic category systems</article-title>. <source>Cognition</source>, <volume>202</volume>, <elocation-id>104289</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2020.104289</pub-id></mixed-citation></ref>
<ref id="B12"><label>12</label><mixed-citation publication-type="journal"><string-name><surname>Chater</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Oaksford</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Hahn</surname>, <given-names>U.</given-names></string-name>, &amp; <string-name><surname>Heit</surname>, <given-names>E.</given-names></string-name> (<year>2010</year>). <article-title>Bayesian models of cognition</article-title>. <source>WIREs Cognitive Science</source>, <volume>1</volume>(<issue>6</issue>), <fpage>811</fpage>&#8211;<lpage>823</lpage>. DOI: <pub-id pub-id-type="doi">10.1002/wcs.79</pub-id></mixed-citation></ref>
<ref id="B13"><label>13</label><mixed-citation publication-type="journal"><string-name><surname>Chikkerur</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Serre</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Tan</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Poggio</surname>, <given-names>T.</given-names></string-name> (<year>2010</year>). <article-title>What and where: A Bayesian inference theory of attention</article-title>. <source>Vision Research</source>, <volume>50</volume>(<issue>22</issue>), <fpage>2233</fpage>&#8211;<lpage>2247</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.visres.2010.05.013</pub-id></mixed-citation></ref>
<ref id="B14"><label>14</label><mixed-citation publication-type="journal"><string-name><surname>Clayards</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Tanenhaus</surname>, <given-names>M. K.</given-names></string-name>, <string-name><surname>Aslin</surname>, <given-names>R. N.</given-names></string-name>, &amp; <string-name><surname>Jacobs</surname>, <given-names>R. A.</given-names></string-name> (<year>2008</year>). <article-title>Perception of speech reflects optimal use of probabilistic speech cues</article-title>. <source>Cognition</source>, <volume>108</volume>(<issue>3</issue>), <fpage>804</fpage>&#8211;<lpage>809</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cognition.2008.04.004</pub-id></mixed-citation></ref>
<ref id="B15"><label>15</label><mixed-citation publication-type="book"><string-name><surname>Clayton</surname>, <given-names>A.</given-names></string-name> (<year>2021</year>). <source>Bernoulli&#8217;s fallacy: Statistical illogic and the crisis of modern science</source>. <publisher-name>Columbia University Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.7312/clay19994</pub-id></mixed-citation></ref>
<ref id="B16"><label>16</label><mixed-citation publication-type="journal"><string-name><surname>Clopper</surname>, <given-names>C. G.</given-names></string-name>, &amp; <string-name><surname>Dossey</surname>, <given-names>E.</given-names></string-name> (<year>2020</year>). <article-title>Phonetic convergence to Southern American English: Acoustics and perception</article-title>. <source>The Journal of the Acoustical Society of America</source>, <volume>147</volume>(<issue>1</issue>), <fpage>671</fpage>&#8211;<lpage>683</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/10.0000555</pub-id></mixed-citation></ref>
<ref id="B17"><label>17</label><mixed-citation publication-type="journal"><string-name><surname>Connine</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Clifton</surname>, <given-names>C.</given-names></string-name> (<year>1987</year>). <article-title>Interactive use of lexical information in speech perception</article-title>. <source>Journal of Experimental Psychology: Human Perception and Performance</source>, <volume>13</volume>(<issue>2</issue>), <fpage>291</fpage>&#8211;<lpage>299</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0096-1523.13.2.291</pub-id></mixed-citation></ref>
<ref id="B18"><label>18</label><mixed-citation publication-type="book"><string-name><surname>Dayan</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Abbott</surname>, <given-names>L.</given-names></string-name> (<year>2001</year>). <source>Theoretical neuroscience</source>. <publisher-name>The MIT Press</publisher-name>.</mixed-citation></ref>
<ref id="B19"><label>19</label><mixed-citation publication-type="webpage"><string-name><surname>De Boer</surname>, <given-names>B.</given-names></string-name> (<year>2000</year>). <article-title>Self-organization in vowel systems</article-title>. <source>Journal of Phonetics</source>, <volume>28</volume>(<issue>4</issue>), <fpage>441</fpage>&#8211;<lpage>465</lpage>. <uri>http://www.sciencedirect.com/science/article/pii/S0095447000901256</uri>. DOI: <pub-id pub-id-type="doi">10.1006/jpho.2000.0125</pub-id></mixed-citation></ref>
<ref id="B20"><label>20</label><mixed-citation publication-type="journal"><string-name><surname>de Leeuw</surname>, <given-names>J. R.</given-names></string-name> (<year>2015</year>). <article-title>jsPsych: A JavaScript library for creating behavioral experiments in a Web browser</article-title>. <source>Behavior Research Methods</source>, <volume>47</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>12</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-014-0458-y</pub-id></mixed-citation></ref>
<ref id="B21"><label>21</label><mixed-citation publication-type="journal"><string-name><surname>DeBruine</surname>, <given-names>L. M.</given-names></string-name>, &amp; <string-name><surname>Barr</surname>, <given-names>D. J.</given-names></string-name> (<year>2021</year>). <article-title>Understanding mixed-effects models through data simulation</article-title>. <source>Advances in Methods and Practices in Psychological Science</source>, <volume>4</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>15</lpage>. DOI: <pub-id pub-id-type="doi">10.1177/2515245920965119</pub-id></mixed-citation></ref>
<ref id="B22"><label>22</label><mixed-citation publication-type="journal"><string-name><surname>Delvaux</surname>, <given-names>V.</given-names></string-name>, &amp; <string-name><surname>Soquet</surname>, <given-names>A.</given-names></string-name> (<year>2007</year>). <article-title>The influence of ambient speech on adult speech productions through unintentional imitation</article-title>. <source>Phonetica</source>, <volume>64</volume>(<issue>2&#8211;3</issue>), <fpage>145</fpage>&#8211;<lpage>173</lpage>. DOI: <pub-id pub-id-type="doi">10.1159/000107914</pub-id></mixed-citation></ref>
<ref id="B23"><label>23</label><mixed-citation publication-type="journal"><string-name><surname>Dias</surname>, <given-names>J. W.</given-names></string-name>, &amp; <string-name><surname>Rosenblum</surname>, <given-names>L. D.</given-names></string-name> (<year>2016</year>). <article-title>Visibility of speech articulation enhances auditory phonetic convergence</article-title>. <source>Attention, Perception, &amp; Psychophysics</source>, <volume>78</volume>(<issue>1</issue>), <fpage>317</fpage>&#8211;<lpage>333</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13414-015-0982-6</pub-id></mixed-citation></ref>
<ref id="B24"><label>24</label><mixed-citation publication-type="journal"><string-name><surname>Ernst</surname>, <given-names>M. O.</given-names></string-name>, &amp; <string-name><surname>Banks</surname>, <given-names>M. S.</given-names></string-name> (<year>2002</year>). <article-title>Humans integrate visual and haptic information in a statistically optimal fashion</article-title>. <source>Nature</source>, <volume>415</volume>(<issue>6870</issue>), <fpage>429</fpage>&#8211;<lpage>33</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/415429a</pub-id></mixed-citation></ref>
<ref id="B25"><label>25</label><mixed-citation publication-type="journal"><string-name><surname>Feldman</surname>, <given-names>N. H.</given-names></string-name>, <string-name><surname>Griffiths</surname>, <given-names>T. L.</given-names></string-name>, &amp; <string-name><surname>Morgan</surname>, <given-names>J. L.</given-names></string-name> (<year>2009</year>). <article-title>The influence of categories on perception: Explaining the perceptual magnet effect as optimal statistical inference</article-title>. <source>Psychological Review</source>, <volume>116</volume>(<issue>4</issue>), <fpage>752</fpage>&#8211;<lpage>782</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/a0017196</pub-id></mixed-citation></ref>
<ref id="B26"><label>26</label><mixed-citation publication-type="journal"><string-name><surname>Fleming</surname>, <given-names>S. M.</given-names></string-name>, &amp; <string-name><surname>Daw</surname>, <given-names>N. D.</given-names></string-name> (<year>2017</year>). <article-title>Self-evaluation of decision-making: A general Bayesian framework for metacognitive computation</article-title>. <source>Psychological Review</source>, <volume>124</volume>(<issue>1</issue>), <fpage>91</fpage>&#8211;<lpage>114</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/rev0000045</pub-id></mixed-citation></ref>
<ref id="B27"><label>27</label><mixed-citation publication-type="journal"><string-name><surname>Friston</surname>, <given-names>K.</given-names></string-name> (<year>2010</year>). <article-title>The free-energy principle: A unified brain theory?</article-title> <source>Nature Reviews Neuroscience</source>, <volume>11</volume>, <fpage>127</fpage>&#8211;<lpage>138</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/nrn2787</pub-id></mixed-citation></ref>
<ref id="B28"><label>28</label><mixed-citation publication-type="journal"><string-name><surname>Ganong</surname>, <given-names>W.</given-names></string-name> (<year>1980</year>). <article-title>Phonetic categorization in auditory word perception</article-title>. <source>Journal of Experimental Psychology: Human Perception and Performance</source>, <volume>6</volume>, <fpage>110</fpage>&#8211;<lpage>125</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0096-1523.6.1.110</pub-id></mixed-citation></ref>
<ref id="B29"><label>29</label><mixed-citation publication-type="journal"><string-name><surname>Geisler</surname>, <given-names>W. S.</given-names></string-name> (<year>2008</year>). <article-title>Visual perception and the statistical properties of natural scenes</article-title>. <source>Annual Review of Psychology</source>, <volume>59</volume>, <fpage>167</fpage>&#8211;<lpage>92</lpage>. DOI: <pub-id pub-id-type="doi">10.1146/annurev.psych.58.110405.085632</pub-id></mixed-citation></ref>
<ref id="B30"><label>30</label><mixed-citation publication-type="book"><string-name><surname>Gelman</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Carlin</surname>, <given-names>J. B.</given-names></string-name>, <string-name><surname>Stern</surname>, <given-names>H. S.</given-names></string-name>, <string-name><surname>Dunson</surname>, <given-names>D. B.</given-names></string-name>, <string-name><surname>Vehtari</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Rubin</surname>, <given-names>D. B.</given-names></string-name> (<year>2021</year>). <source>Bayesian Data Analysis</source>. <publisher-name>Chapman &amp; Hall</publisher-name>.</mixed-citation></ref>
<ref id="B31"><label>31</label><mixed-citation publication-type="journal"><string-name><surname>Gifford</surname>, <given-names>A. M.</given-names></string-name>, <string-name><surname>Cohen</surname>, <given-names>Y. E.</given-names></string-name>, &amp; <string-name><surname>Stocker</surname>, <given-names>A. A.</given-names></string-name> (<year>2014</year>). <article-title>Characterizing the impact of category uncertainty on human auditory categorization behavior</article-title> (<string-name><given-names>T. D.</given-names> <surname>Griffiths</surname></string-name>, Ed.). <source>PLoS Computational Biology</source>, <volume>10</volume>(<issue>7</issue>), <elocation-id>e1003715</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1371/journal.pcbi.1003715</pub-id></mixed-citation></ref>
<ref id="B32"><label>32</label><mixed-citation publication-type="journal"><string-name><surname>Ginestet</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Valdois</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Diard</surname>, <given-names>J.</given-names></string-name> (<year>2022</year>). <article-title>Probabilistic modeling of orthographic learning based on visuo-attentional dynamics</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>29</volume>(<issue>5</issue>), <fpage>1649</fpage>&#8211;<lpage>1672</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13423-021-02042-4</pub-id></mixed-citation></ref>
<ref id="B33"><label>33</label><mixed-citation publication-type="journal"><string-name><surname>Goldinger</surname>, <given-names>S.</given-names></string-name> (<year>1998</year>). <article-title>Echoes of echoes? An episodic theory of lexical access</article-title>. <source>Psychological Review</source>, <volume>105</volume>(<issue>2</issue>), <fpage>251</fpage>&#8211;<lpage>279</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0033-295X.105.2.251</pub-id></mixed-citation></ref>
<ref id="B34"><label>34</label><mixed-citation publication-type="journal"><string-name><surname>Harrington</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Gubian</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Stevens</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Schiel</surname>, <given-names>F.</given-names></string-name> (<year>2019</year>). <article-title>Phonetic change in an Antarctic winter</article-title>. <source>Journal of the Acoustical Society of America</source>, <volume>146</volume>(<issue>5</issue>), <fpage>3327</fpage>&#8211;<lpage>3332</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/1.5130709</pub-id></mixed-citation></ref>
<ref id="B35"><label>35</label><mixed-citation publication-type="webpage"><string-name><surname>Harrington</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Palethorpe</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Watson</surname>, <given-names>C. I.</given-names></string-name> (<year>2000</year>). <article-title>Does the Queen speak the Queen&#8217;s English?</article-title> <source>Nature</source>, <volume>408</volume>(<issue>21/28</issue>), <fpage>927</fpage>&#8211;<lpage>928</lpage>. <uri>http://www.nature.com/nature/journal/v408/n6815/abs/408927a0.html</uri>. DOI: <pub-id pub-id-type="doi">10.1038/35050160</pub-id></mixed-citation></ref>
<ref id="B36"><label>36</label><mixed-citation publication-type="journal"><string-name><surname>Huttner</surname>, <given-names>L.-M.</given-names></string-name>, &amp; <string-name><surname>Nguyen</surname>, <given-names>N.</given-names></string-name> (<year>2023</year>). <article-title>Between-listener convergence in speech sound categorization</article-title>. <source>Ms. submitted for publication</source>.</mixed-citation></ref>
<ref id="B37"><label>37</label><mixed-citation publication-type="book"><string-name><surname>Jaynes</surname>, <given-names>E. T.</given-names></string-name> (<year>2003</year>). <source>Probability theory: The logic of science</source>. <publisher-name>Cambridge University Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1017/CBO9780511790423</pub-id></mixed-citation></ref>
<ref id="B38"><label>38</label><mixed-citation publication-type="journal"><string-name><surname>Kapnoula</surname>, <given-names>E. C.</given-names></string-name>, <string-name><surname>Winn</surname>, <given-names>M. B.</given-names></string-name>, <string-name><surname>Kong</surname>, <given-names>E. J.</given-names></string-name>, <string-name><surname>Edwards</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>McMurray</surname>, <given-names>B.</given-names></string-name> (<year>2017</year>). <article-title>Evaluating the sources and functions of gradiency in phoneme categorization: An individual differences approach</article-title>. <source>Journal of Experimental Psychology: Human Perception and Performance</source>, <volume>43</volume>(<issue>9</issue>), <fpage>1594</fpage>&#8211;<lpage>1611</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/xhp0000410</pub-id></mixed-citation></ref>
<ref id="B39"><label>39</label><mixed-citation publication-type="journal"><string-name><surname>Kersten</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Mamassian</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Yuille</surname>, <given-names>A.</given-names></string-name> (<year>2004</year>). <article-title>Object perception as Bayesian inference</article-title>. <source>Annual Review of Psychology</source>, <volume>55</volume>, <fpage>271</fpage>&#8211;<lpage>304</lpage>. DOI: <pub-id pub-id-type="doi">10.1146/annurev.psych.55.090902.142005</pub-id></mixed-citation></ref>
<ref id="B40"><label>40</label><mixed-citation publication-type="journal"><string-name><surname>Kim</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Horton</surname>, <given-names>W. S.</given-names></string-name>, &amp; <string-name><surname>Bradlow</surname>, <given-names>A. R.</given-names></string-name> (<year>2011</year>). <article-title>Phonetic convergence in spontaneous conversations as a function of interlocutor language distance</article-title>. <source>Laboratory Phonology</source>, <volume>2</volume>(<issue>1</issue>), <fpage>125</fpage>&#8211;<lpage>156</lpage>. DOI: <pub-id pub-id-type="doi">10.1515/labphon.2011.004</pub-id></mixed-citation></ref>
<ref id="B41"><label>41</label><mixed-citation publication-type="webpage"><string-name><surname>Kleinschmidt</surname>, <given-names>D. F.</given-names></string-name> (<year>2020</year>). <source>What constrains distributional learning in adults?</source> (Tech. rep.). PsyArXiv, <uri>https://psyarxiv.com/6yhbe/</uri>. DOI: <pub-id pub-id-type="doi">10.31234/osf.io/6yhbe</pub-id></mixed-citation></ref>
<ref id="B42"><label>42</label><mixed-citation publication-type="journal"><string-name><surname>Kleinschmidt</surname>, <given-names>D. F.</given-names></string-name>, &amp; <string-name><surname>Jaeger</surname>, <given-names>T. F.</given-names></string-name> (<year>2015</year>). <article-title>Robust speech perception: Recognize the familiar, generalize to the similar, and adapt to the novel</article-title>. <source>Psychological Review</source>, <volume>122</volume>(<issue>2</issue>), <fpage>148</fpage>&#8211;<lpage>203</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/a0038695</pub-id></mixed-citation></ref>
<ref id="B43"><label>43</label><mixed-citation publication-type="journal"><string-name><surname>Kong</surname>, <given-names>E. J.</given-names></string-name>, &amp; <string-name><surname>Edwards</surname>, <given-names>J.</given-names></string-name> (<year>2016</year>). <article-title>Individual differences in categorical perception of speech: Cue weighting and executive function</article-title>. <source>Journal of Phonetics</source>, <volume>59</volume>, <fpage>40</fpage>&#8211;<lpage>57</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.wocn.2016.08.006</pub-id></mixed-citation></ref>
<ref id="B44"><label>44</label><mixed-citation publication-type="journal"><string-name><surname>Koriat</surname>, <given-names>A.</given-names></string-name> (<year>2012</year>). <article-title>When are two heads better than one and why?</article-title> <source>Science</source>, <volume>336</volume>, <fpage>360</fpage>&#8211;<lpage>362</lpage>. DOI: <pub-id pub-id-type="doi">10.1126/science.1216549</pub-id></mixed-citation></ref>
<ref id="B45"><label>45</label><mixed-citation publication-type="journal"><string-name><surname>Kraljic</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Samuel</surname>, <given-names>A.</given-names></string-name> (<year>2005</year>). <article-title>Perceptual learning for speech: Is there a return to normal?</article-title> <source>Cognitive Psychology</source>, <volume>51</volume>, <fpage>141</fpage>&#8211;<lpage>178</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.cogpsych.2005.05.001</pub-id></mixed-citation></ref>
<ref id="B46"><label>46</label><mixed-citation publication-type="journal"><string-name><surname>Kronrod</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Coppess</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Feldman</surname>, <given-names>N. H.</given-names></string-name> (<year>2016</year>). <article-title>A unified account of categorical effects in phonetic perception</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>23</volume>(<issue>6</issue>), <fpage>1681</fpage>&#8211;<lpage>1712</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13423-016-1049-y</pub-id></mixed-citation></ref>
<ref id="B47"><label>47</label><mixed-citation publication-type="book"><string-name><surname>Kruschke</surname>, <given-names>J. K.</given-names></string-name> (<year>2014</year>). <source>Doing Bayesian data analysis: A tutorial with R, JAGS, and Stan</source>. <publisher-name>Academic Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1016/B978-0-12-405888-0.00008-8</pub-id></mixed-citation></ref>
<ref id="B48"><label>48</label><mixed-citation publication-type="journal"><string-name><surname>Kurumada</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Roettger</surname>, <given-names>T. B.</given-names></string-name> (<year>2022</year>). <article-title>Thinking probabilistically in the study of intonational speech prosody</article-title>. <source>WIREs Cognitive Science</source>, <volume>13</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>27</lpage>. DOI: <pub-id pub-id-type="doi">10.1002/wcs.1579</pub-id></mixed-citation></ref>
<ref id="B49"><label>49</label><mixed-citation publication-type="book"><string-name><surname>Lambert</surname>, <given-names>B.</given-names></string-name> (<year>2018</year>). <source>A student&#8217;s guide to Bayesian statistics</source>. <publisher-name>Sage</publisher-name>.</mixed-citation></ref>
<ref id="B50"><label>50</label><mixed-citation publication-type="journal"><string-name><surname>Lancia</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name><surname>Nguyen</surname>, <given-names>N.</given-names></string-name> (<year>2019</year>). <article-title>The joint perception and categorization of speech sounds: A pilot study</article-title>. <source>Proceedings of the 7th Joint Action Meeting</source>.</mixed-citation></ref>
<ref id="B51"><label>51</label><mixed-citation publication-type="journal"><string-name><surname>Lange</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>K&#252;hn</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Filevich</surname>, <given-names>E.</given-names></string-name> (<year>2015</year>). <article-title>&#8220;Just Another Tool for Online Studies&#8221; (JATOS): An easy solution for setup and management of Web servers supporting online studies</article-title> (<string-name><given-names>D.</given-names> <surname>Margulies</surname></string-name>, Ed.). <source>PLoS ONE</source>, <volume>10</volume>(<issue>6</issue>), <elocation-id>e0130834</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1371/journal.pone.0130834</pub-id></mixed-citation></ref>
<ref id="B52"><label>52</label><mixed-citation publication-type="journal"><string-name><surname>Laurent</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Barnaud</surname>, <given-names>M.-L.</given-names></string-name>, <string-name><surname>Schwartz</surname>, <given-names>J.-L.</given-names></string-name>, <string-name><surname>Bessiere</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Diard</surname>, <given-names>J.</given-names></string-name> (<year>2017</year>). <article-title>The complementary roles of auditory and motor information evaluated in a Bayesian perceptuo-motor model of speech perception</article-title>. <source>Psychological Review</source>, <volume>124</volume>(<issue>5</issue>), <fpage>572</fpage>&#8211;<lpage>602</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/rev0000069</pub-id></mixed-citation></ref>
<ref id="B53"><label>53</label><mixed-citation publication-type="journal"><string-name><surname>Ma</surname>, <given-names>W. J.</given-names></string-name> (<year>2012</year>). <article-title>Organizing probabilistic models of perception</article-title>. <source>Trends in Cognitive Sciences</source>, <volume>16</volume>(<issue>10</issue>), <fpage>511</fpage>&#8211;<lpage>518</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.tics.2012.08.010</pub-id></mixed-citation></ref>
<ref id="B54"><label>54</label><mixed-citation publication-type="book"><string-name><surname>Ma</surname>, <given-names>W. J.</given-names></string-name>, <string-name><surname>Kording</surname>, <given-names>K. P.</given-names></string-name>, &amp; <string-name><surname>Goldreich</surname>, <given-names>D.</given-names></string-name> (<year>2023</year>). <source>Bayesian models of perception and action: An introduction</source>. <publisher-name>MIT Press</publisher-name>.</mixed-citation></ref>
<ref id="B55"><label>55</label><mixed-citation publication-type="journal"><string-name><surname>Mahmoodi</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Bahrami</surname>, <given-names>B.</given-names></string-name>, &amp; <string-name><surname>Mehring</surname>, <given-names>C.</given-names></string-name> (<year>2018</year>). <article-title>Reciprocity of social influence</article-title>. <source>Nature Communications</source>, <volume>9</volume>(<issue>1</issue>), <fpage>2474</fpage>. DOI: <pub-id pub-id-type="doi">10.1038/s41467-018-04925-y</pub-id></mixed-citation></ref>
<ref id="B56"><label>56</label><mixed-citation publication-type="book"><string-name><surname>Mamassian</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Landy</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Maloney</surname>, <given-names>L. T.</given-names></string-name> (<year>2003</year>). <chapter-title>Bayesian modelling of visual perception</chapter-title>. In <string-name><given-names>R. P. N.</given-names> <surname>Rao</surname></string-name>, <string-name><given-names>B. A.</given-names> <surname>Olshausen</surname></string-name>, &amp; <string-name><given-names>M. S.</given-names> <surname>Lewicki</surname></string-name> (Eds.), <source>Probabilistic models of the brain: Perception and neural function</source> (pp. <fpage>13</fpage>&#8211;<lpage>36</lpage>). <publisher-name>MIT Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.7551/mitpress/5583.003.0005</pub-id></mixed-citation></ref>
<ref id="B57"><label>57</label><mixed-citation publication-type="journal"><string-name><surname>Mann</surname>, <given-names>V.</given-names></string-name>, &amp; <string-name><surname>Repp</surname>, <given-names>B.</given-names></string-name> (<year>1981</year>). <article-title>Influence of preceding fricative on stop consonant perception</article-title>. <source>Journal of the Acoustical Society of America</source>, <volume>69</volume>, <fpage>548</fpage>&#8211;<lpage>558</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/1.385483</pub-id></mixed-citation></ref>
<ref id="B58"><label>58</label><mixed-citation publication-type="journal"><string-name><surname>Massaro</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Cohen</surname>, <given-names>M.</given-names></string-name> (<year>1983</year>). <article-title>Categorical or continuous speech perception: A new test</article-title>. <source>Speech Communication</source>, <volume>2</volume>, <fpage>15</fpage>&#8211;<lpage>35</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/0167-6393(83)90061-4</pub-id></mixed-citation></ref>
<ref id="B59"><label>59</label><mixed-citation publication-type="journal"><string-name><surname>McMurray</surname>, <given-names>B.</given-names></string-name> (<year>2022</year>). <article-title>The myth of categorical perception</article-title>. <source>Journal of the Acoustical Society of America</source>, <volume>152</volume>(<issue>6</issue>), <fpage>3819</fpage>&#8211;<lpage>3842</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/10.0016614</pub-id></mixed-citation></ref>
<ref id="B60"><label>60</label><mixed-citation publication-type="journal"><string-name><surname>Miller</surname>, <given-names>R. M.</given-names></string-name>, <string-name><surname>Sanchez</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Rosenblum</surname>, <given-names>L. D.</given-names></string-name> (<year>2013</year>). <article-title>Is speech alignment to talkers or tasks?</article-title> <source>Attention, Perception, &amp; Psychophysics</source>, <volume>75</volume>(<issue>8</issue>), <fpage>1817</fpage>&#8211;<lpage>1826</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13414-013-0517-y</pub-id></mixed-citation></ref>
<ref id="B61"><label>61</label><mixed-citation publication-type="journal"><string-name><surname>Milne</surname>, <given-names>A. E.</given-names></string-name>, <string-name><surname>Bianco</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Poole</surname>, <given-names>K. C.</given-names></string-name>, <string-name><surname>Zhao</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Oxenham</surname>, <given-names>A. J.</given-names></string-name>, <string-name><surname>Billig</surname>, <given-names>A. J.</given-names></string-name>, &amp; <string-name><surname>Chait</surname>, <given-names>M.</given-names></string-name> (<year>2021</year>). <article-title>An online headphone screening test based on dichotic pitch</article-title>. <source>Behavior Research Methods</source>, <volume>53</volume>, <fpage>1551</fpage>&#8211;<lpage>1562</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13428-020-01514-0</pub-id></mixed-citation></ref>
<ref id="B62"><label>62</label><mixed-citation publication-type="journal"><string-name><surname>Moulin-Frier</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Diard</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Schwartz</surname>, <given-names>J.-L.</given-names></string-name>, &amp; <string-name><surname>Bessiere</surname>, <given-names>P.</given-names></string-name> (<year>2015</year>). <article-title>COSMO (&#8220;Communicating about Objects using Sensory&#8211;Motor Operations&#8221;): A Bayesian modeling framework for studying speech communication and the emergence of phonological systems</article-title>. <source>Journal of Phonetics</source>, <volume>53</volume>, <fpage>5</fpage>&#8211;<lpage>41</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.wocn.2015.06.001</pub-id></mixed-citation></ref>
<ref id="B63"><label>63</label><mixed-citation publication-type="journal"><string-name><surname>Mukherjee</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Badino</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Hilt</surname>, <given-names>P. M.</given-names></string-name>, <string-name><surname>Tomassini</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Inuggi</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Fadiga</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Nguyen</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>D&#8217;Ausilio</surname>, <given-names>A.</given-names></string-name> (<year>2019</year>). <article-title>The neural oscillatory markers of phonetic convergence during verbal interaction</article-title>. <source>Human Brain Mapping</source>, <volume>40</volume>(<issue>1</issue>), <fpage>187</fpage>&#8211;<lpage>201</lpage>. DOI: <pub-id pub-id-type="doi">10.1002/hbm.24364</pub-id></mixed-citation></ref>
<ref id="B64"><label>64</label><mixed-citation publication-type="webpage"><string-name><surname>Murphy</surname>, <given-names>K. P.</given-names></string-name> (<year>2007</year>). <source>Conjugate Bayesian analysis of the Gaussian distribution</source>. <uri>https://www.cs.ubc.ca/~murphyk/Papers/bayesGauss.pdf</uri></mixed-citation></ref>
<ref id="B65"><label>65</label><mixed-citation publication-type="book"><string-name><surname>Nearey</surname>, <given-names>T. M.</given-names></string-name>, &amp; <string-name><surname>Hogan</surname>, <given-names>J.</given-names></string-name> (<year>1986</year>). <chapter-title>Phonological contrast in experimental phonetics: Relating distributions of production data to perceptual categorization curves</chapter-title>. In <string-name><given-names>J. J.</given-names> <surname>Ohala</surname></string-name> &amp; <string-name><given-names>J. J.</given-names> <surname>Jaeger</surname></string-name> (Eds.), <source>Experimental Phonology</source> (pp. <fpage>141</fpage>&#8211;<lpage>146</lpage>). <publisher-name>Academic Press</publisher-name>.</mixed-citation></ref>
<ref id="B66"><label>66</label><mixed-citation publication-type="journal"><string-name><surname>Nguyen</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Delvaux</surname>, <given-names>V.</given-names></string-name> (<year>2015</year>). <article-title>Role of imitation in the emergence of phonological systems</article-title>. <source>Journal of Phonetics</source>, <volume>53</volume>, <fpage>46</fpage>&#8211;<lpage>54</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.wocn.2015.08.004</pub-id></mixed-citation></ref>
<ref id="B67"><label>67</label><mixed-citation publication-type="journal"><string-name><surname>Nguyen</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Dufour</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Brunelliere</surname>, <given-names>A.</given-names></string-name> (<year>2012</year>). <article-title>Does imitation facilitate word recognition in a non-native regional accent?</article-title> <source>Frontiers in Psychology</source>, <volume>3</volume>, Article <elocation-id>480</elocation-id>. DOI: <pub-id pub-id-type="doi">10.3389/fpsyg.2012.00480</pub-id></mixed-citation></ref>
<ref id="B68"><label>68</label><mixed-citation publication-type="journal"><string-name><surname>Nielsen</surname>, <given-names>K.</given-names></string-name> (<year>2011</year>). <article-title>Specificity and abstractness of VOT imitation</article-title>. <source>Journal of Phonetics</source>, <volume>39</volume>(<issue>2</issue>), <fpage>132</fpage>&#8211;<lpage>142</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.wocn.2010.12.007</pub-id></mixed-citation></ref>
<ref id="B69"><label>69</label><mixed-citation publication-type="journal"><string-name><surname>Norris</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>McQueen</surname>, <given-names>J.</given-names></string-name> (<year>2008</year>). <article-title>Shortlist B: A Bayesian model of continuous speech recognition</article-title>. <source>Psychological Review</source>, <volume>115</volume>(<issue>2</issue>), <fpage>357</fpage>&#8211;<lpage>395</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0033-295X.115.2.357</pub-id></mixed-citation></ref>
<ref id="B70"><label>70</label><mixed-citation publication-type="journal"><string-name><surname>Norris</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>McQueen</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Cutler</surname>, <given-names>A.</given-names></string-name> (<year>2003</year>). <article-title>Perceptual learning in speech</article-title>. <source>Cognitive Psychology</source>, <volume>47</volume>(<issue>2</issue>), <fpage>204</fpage>&#8211;<lpage>238</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/S0010-0285(03)00006-9</pub-id></mixed-citation></ref>
<ref id="B71"><label>71</label><mixed-citation publication-type="journal"><string-name><surname>Norris</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>McQueen</surname>, <given-names>J. M.</given-names></string-name>, &amp; <string-name><surname>Cutler</surname>, <given-names>A.</given-names></string-name> (<year>2016</year>). <article-title>Prediction, Bayesian inference and feedback in speech recognition</article-title>. <source>Language, Cognition and Neuroscience</source>, <volume>31</volume>(<issue>1</issue>), <fpage>4</fpage>&#8211;<lpage>18</lpage>. DOI: <pub-id pub-id-type="doi">10.1080/23273798.2015.1081703</pub-id></mixed-citation></ref>
<ref id="B72"><label>72</label><mixed-citation publication-type="journal"><string-name><surname>Ou</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Yu</surname>, <given-names>A. C. L.</given-names></string-name>, &amp; <string-name><surname>Xiang</surname>, <given-names>M.</given-names></string-name> (<year>2021</year>). <article-title>Individual differences in categorization gradience as predicted by online processing of phonetic cues during spoken word recognition: Evidence from eye movements</article-title>. <source>Cognitive Science</source>, <volume>45</volume>, <elocation-id>e12948</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1111/cogs.12948</pub-id></mixed-citation></ref>
<ref id="B73"><label>73</label><mixed-citation publication-type="journal"><string-name><surname>Pardo</surname>, <given-names>J. S.</given-names></string-name> (<year>2006</year>). <article-title>On phonetic convergence during conversational interaction</article-title>. <source>Journal of the Acoustical Society of America</source>, <volume>119</volume>(<issue>4</issue>), <fpage>2382</fpage>&#8211;<lpage>2393</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/1.2178720</pub-id></mixed-citation></ref>
<ref id="B74"><label>74</label><mixed-citation publication-type="journal"><string-name><surname>Pardo</surname>, <given-names>J. S.</given-names></string-name>, <string-name><surname>Jay</surname>, <given-names>I. C.</given-names></string-name>, <string-name><surname>Hoshino</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Hasbun</surname>, <given-names>S. M.</given-names></string-name>, <string-name><surname>Sowemimo-Coker</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Krauss</surname>, <given-names>R. M.</given-names></string-name> (<year>2013a</year>). <article-title>Influence of role-switching on phonetic convergence in conversation</article-title>. <source>Discourse Processes</source>, <volume>50</volume>(<issue>4</issue>), <fpage>276</fpage>&#8211;<lpage>300</lpage>. DOI: <pub-id pub-id-type="doi">10.1080/0163853X.2013.778168</pub-id></mixed-citation></ref>
<ref id="B75"><label>75</label><mixed-citation publication-type="journal"><string-name><surname>Pardo</surname>, <given-names>J. S.</given-names></string-name>, <string-name><surname>Jordan</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Mallari</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Scanlon</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Lewandowski</surname>, <given-names>E.</given-names></string-name> (<year>2013b</year>). <article-title>Phonetic convergence in shadowed speech: The relation between acoustic and perceptual measures</article-title>. <source>Journal of Memory and Language</source>, <volume>69</volume>(<issue>3</issue>), <fpage>183</fpage>&#8211;<lpage>195</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.jml.2013.06.002</pub-id></mixed-citation></ref>
<ref id="B76"><label>76</label><mixed-citation publication-type="journal"><string-name><surname>Pardo</surname>, <given-names>J. S.</given-names></string-name>, <string-name><surname>Urmanche</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Wilman</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Wiener</surname>, <given-names>J.</given-names></string-name> (<year>2017</year>). <article-title>Phonetic convergence across multiple measures and model talkers</article-title>. <source>Attention, Perception, &amp; Psychophysics</source>, <volume>79</volume>, <fpage>1</fpage>&#8211;<lpage>23</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13414-016-1226-0</pub-id></mixed-citation></ref>
<ref id="B77"><label>77</label><mixed-citation publication-type="journal"><string-name><surname>Patri</surname>, <given-names>J.-F.</given-names></string-name>, <string-name><surname>Perrier</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Schwartz</surname>, <given-names>J.-L.</given-names></string-name>, &amp; <string-name><surname>Diard</surname>, <given-names>J.</given-names></string-name> (<year>2018</year>). <article-title>What drives the perceptual change resulting from speech motor adaptation? Evaluation of hypotheses in a Bayesian modeling framework</article-title>. <source>PLoS Computational Biology</source>, <volume>14</volume>(<issue>1</issue>), <elocation-id>e1005942</elocation-id>. DOI: <pub-id pub-id-type="doi">10.1371/journal.pcbi.1005942</pub-id></mixed-citation></ref>
<ref id="B78"><label>78</label><mixed-citation publication-type="journal"><string-name><surname>Pickering</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Garrod</surname>, <given-names>S.</given-names></string-name> (<year>2013</year>). <article-title>An integrated theory of language production and comprehension</article-title>. <source>Behavioral and Brain Sciences</source>, <volume>36</volume>(<issue>04</issue>), <fpage>329</fpage>&#8211;<lpage>347</lpage>. DOI: <pub-id pub-id-type="doi">10.1017/S0140525X12001495</pub-id></mixed-citation></ref>
<ref id="B79"><label>79</label><mixed-citation publication-type="book"><string-name><surname>Pickering</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Garrod</surname>, <given-names>S.</given-names></string-name> (<year>2021</year>). <source>Understanding dialogue: Language use and social interaction</source>. <publisher-name>Cambridge University Press</publisher-name>. DOI: <pub-id pub-id-type="doi">10.1017/9781108610728</pub-id></mixed-citation></ref>
<ref id="B80"><label>80</label><mixed-citation publication-type="journal"><string-name><surname>Pitt</surname>, <given-names>M. A.</given-names></string-name> (<year>1995</year>). <article-title>The locus of the lexical shift in phoneme identification</article-title>. <source>Journal of Experimental Psychology: Learning, Memory and Cognition</source>, <volume>21</volume>(<issue>4</issue>), <fpage>1037</fpage>&#8211;<lpage>1052</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0278-7393.21.4.1037</pub-id></mixed-citation></ref>
<ref id="B81"><label>81</label><mixed-citation publication-type="journal"><string-name><surname>Pouget</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Beck</surname>, <given-names>J. M.</given-names></string-name>, <string-name><surname>Ma</surname>, <given-names>W. J.</given-names></string-name>, &amp; <string-name><surname>Latham</surname>, <given-names>P. E.</given-names></string-name> (<year>2013</year>). <article-title>Probabilistic brains: Knowns and unknowns</article-title>. <source>Nature Neuroscience</source>, <volume>16</volume>(<issue>9</issue>), <fpage>1170</fpage>&#8211;<lpage>1178</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/nn.3495</pub-id></mixed-citation></ref>
<ref id="B82"><label>82</label><mixed-citation publication-type="journal"><string-name><surname>Richardson</surname>, <given-names>D. C.</given-names></string-name>, <string-name><surname>Street</surname>, <given-names>C. N. H.</given-names></string-name>, <string-name><surname>Tan</surname>, <given-names>J. Y. M.</given-names></string-name>, <string-name><surname>Kirkham</surname>, <given-names>N. Z.</given-names></string-name>, <string-name><surname>Hoover</surname>, <given-names>M. A.</given-names></string-name>, &amp; <string-name><surname>Ghane Cavanaugh</surname>, <given-names>A.</given-names></string-name> (<year>2012</year>). <article-title>Joint perception: Gaze and social context</article-title>. <source>Frontiers in Human Neuroscience</source>, <volume>6</volume>, Article <elocation-id>194</elocation-id>. DOI: <pub-id pub-id-type="doi">10.3389/fnhum.2012.00194</pub-id></mixed-citation></ref>
<ref id="B83"><label>83</label><mixed-citation publication-type="journal"><string-name><surname>Schwartz</surname>, <given-names>J.-L.</given-names></string-name>, <string-name><surname>Basirat</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Menard</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name><surname>Sato</surname>, <given-names>M.</given-names></string-name> (<year>2012</year>). <article-title>The Perception-for-Action-Control Theory (PACT): A perceptuo-motor theory of speech perception</article-title>. <source>Journal of Neurolinguistics</source>, <volume>25</volume>(<issue>5</issue>), <fpage>336</fpage>&#8211;<lpage>354</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.jneuroling.2009.12.004</pub-id></mixed-citation></ref>
<ref id="B84"><label>84</label><mixed-citation publication-type="journal"><string-name><surname>Scott-Phillips</surname>, <given-names>T. C.</given-names></string-name>, &amp; <string-name><surname>Kirby</surname>, <given-names>S.</given-names></string-name> (<year>2010</year>). <article-title>Language evolution in the laboratory</article-title>. <source>Trends in Cognitive Sciences</source>, <volume>14</volume>(<issue>9</issue>), <fpage>411</fpage>&#8211;<lpage>417</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.tics.2010.06.006</pub-id></mixed-citation></ref>
<ref id="B85"><label>85</label><mixed-citation publication-type="journal"><string-name><surname>Seow</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Fleming</surname>, <given-names>S. M.</given-names></string-name> (<year>2019</year>). <article-title>Perceptual sensitivity is modulated by what others can see</article-title>. <source>Attention, Perception, &amp; Psychophysics</source>, <volume>81</volume>(<issue>6</issue>), <fpage>1979</fpage>&#8211;<lpage>1990</lpage>. DOI: <pub-id pub-id-type="doi">10.3758/s13414-019-01724-5</pub-id></mixed-citation></ref>
<ref id="B86"><label>86</label><mixed-citation publication-type="journal"><string-name><surname>Sohoglu</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Davis</surname>, <given-names>M. H.</given-names></string-name> (<year>2020</year>). <article-title>Rapid computations of spectrotemporal prediction error support perception of degraded speech</article-title>. <source>eLife</source>, <volume>9</volume>, <fpage>1</fpage>&#8211;<lpage>25</lpage>. DOI: <pub-id pub-id-type="doi">10.7554/eLife.58077</pub-id></mixed-citation></ref>
<ref id="B87"><label>87</label><mixed-citation publication-type="journal"><string-name><surname>Sorkin</surname>, <given-names>R. D.</given-names></string-name>, <string-name><surname>Hays</surname>, <given-names>C. J.</given-names></string-name>, &amp; <string-name><surname>West</surname>, <given-names>R.</given-names></string-name> (<year>2001</year>). <article-title>Signal-detection analysis of group decision making</article-title>. <source>Psychological Review</source>, <volume>108</volume>(<issue>1</issue>), <fpage>183</fpage>&#8211;<lpage>203</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0033-295X.108.1.183</pub-id></mixed-citation></ref>
<ref id="B88"><label>88</label><mixed-citation publication-type="journal"><string-name><surname>Tenenbaum</surname>, <given-names>J. B.</given-names></string-name>, <string-name><surname>Kemp</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Griffiths</surname>, <given-names>T. L.</given-names></string-name>, &amp; <string-name><surname>Goodman</surname>, <given-names>N. D.</given-names></string-name> (<year>2011</year>). <article-title>How to grow a mind: Statistics, structure, and abstraction</article-title>. <source>Science</source>, <volume>331</volume>, <fpage>1279</fpage>&#8211;<lpage>1285</lpage>. DOI: <pub-id pub-id-type="doi">10.1126/science.1192788</pub-id></mixed-citation></ref>
<ref id="B89"><label>89</label><mixed-citation publication-type="journal"><string-name><surname>Verhoef</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Kirby</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>De Boer</surname>, <given-names>B.</given-names></string-name> (<year>2014</year>). <article-title>Emergence of combinatorial structure and economy through iterated learning with continuous acoustic signals</article-title>. <source>Journal of Phonetics</source>, <volume>43</volume>, <fpage>57</fpage>&#8211;<lpage>68</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.wocn.2014.02.005</pub-id></mixed-citation></ref>
<ref id="B90"><label>90</label><mixed-citation publication-type="journal"><string-name><surname>Vincent</surname>, <given-names>B. T.</given-names></string-name> (<year>2015</year>). <article-title>A tutorial on Bayesian models of perception</article-title>. <source>Journal of Mathematical Psychology</source>, <volume>66</volume>, <fpage>103</fpage>&#8211;<lpage>114</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.jmp.2015.02.001</pub-id></mixed-citation></ref>
<ref id="B91"><label>91</label><mixed-citation publication-type="journal"><string-name><surname>Wahn</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Kingstone</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Konig</surname>, <given-names>P.</given-names></string-name> (<year>2018</year>). <article-title>Group benefits in joint perceptual tasks &#8211; a review</article-title>. <source>Annals of the New York Academy of Sciences</source>, <volume>1426</volume>(<issue>1</issue>), <fpage>166</fpage>&#8211;<lpage>178</lpage>. DOI: <pub-id pub-id-type="doi">10.1111/nyas.13843</pub-id></mixed-citation></ref>
<ref id="B92"><label>92</label><mixed-citation publication-type="journal"><string-name><surname>Weiss</surname>, <given-names>Y.</given-names></string-name>, <string-name><surname>Simoncelli</surname>, <given-names>E. P.</given-names></string-name>, &amp; <string-name><surname>Adelson</surname>, <given-names>E. H.</given-names></string-name> (<year>2002</year>). <article-title>Motion illusions as optimal percepts</article-title>. <source>Nature Neuroscience</source>, <volume>5</volume>, <fpage>598</fpage>&#8211;<lpage>604</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/nn0602-858</pub-id></mixed-citation></ref>
<ref id="B93"><label>93</label><mixed-citation publication-type="journal"><string-name><surname>Winn</surname>, <given-names>M. B.</given-names></string-name> (<year>2020</year>). <article-title>Manipulation of voice onset time in speech stimuli: A tutorial and flexible Praat script</article-title>. <source>Journal of the Acoustical Society of America</source>, <volume>147</volume>(<issue>2</issue>), <fpage>852</fpage>&#8211;<lpage>866</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/10.0000692</pub-id></mixed-citation></ref>
<ref id="B94"><label>94</label><mixed-citation publication-type="journal"><string-name><surname>Wolpert</surname>, <given-names>D. M.</given-names></string-name> (<year>2007</year>). <article-title>Probabilistic models in human sensorimotor control</article-title>. <source>Human Movement Science</source>, <volume>26</volume>(<issue>4</issue>), <fpage>511</fpage>&#8211;<lpage>524</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.humov.2007.05.005</pub-id></mixed-citation></ref>
<ref id="B95"><label>95</label><mixed-citation publication-type="journal"><string-name><surname>Wolpert</surname>, <given-names>D. M.</given-names></string-name>, &amp; <string-name><surname>Ghahramani</surname>, <given-names>Z.</given-names></string-name> (<year>2000</year>). <article-title>Computational principles of movement neuroscience</article-title>. <source>Nature Neuroscience</source>, <volume>3</volume>, <fpage>1212</fpage>&#8211;<lpage>1217</lpage>. DOI: <pub-id pub-id-type="doi">10.1038/81497</pub-id></mixed-citation></ref>
<ref id="B96"><label>96</label><mixed-citation publication-type="journal"><string-name><surname>Wozny</surname>, <given-names>D. R.</given-names></string-name>, <string-name><surname>Beierholm</surname>, <given-names>U. R.</given-names></string-name>, &amp; <string-name><surname>Shams</surname>, <given-names>L.</given-names></string-name> (<year>2008</year>). <article-title>Human trimodal perception follows optimal statistical inference</article-title>. <source>Journal of Vision</source>, <volume>8</volume>(<issue>3</issue>), <elocation-id>24</elocation-id>, <fpage>1</fpage>&#8211;<lpage>11</lpage>. DOI: <pub-id pub-id-type="doi">10.1167/8.3.24</pub-id></mixed-citation></ref>
<ref id="B97"><label>97</label><mixed-citation publication-type="journal"><string-name><surname>Xu</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Tenenbaum</surname>, <given-names>J. B.</given-names></string-name> (<year>2007</year>). <article-title>Word learning as Bayesian inference</article-title>. <source>Psychological Review</source>, <volume>114</volume>(<issue>2</issue>), <fpage>245</fpage>&#8211;<lpage>272</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/0033-295X.114.2.245</pub-id></mixed-citation></ref>
<ref id="B98"><label>98</label><mixed-citation publication-type="journal"><string-name><surname>Yu</surname>, <given-names>A. J.</given-names></string-name>, <string-name><surname>Dayan</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Cohen</surname>, <given-names>J. D.</given-names></string-name> (<year>2009</year>). <article-title>Dynamics of attentional selection under conflict: Toward a rational Bayesian account</article-title>. <source>Journal of Experimental Psychology: Human Perception and Performance</source>, <volume>35</volume>(<issue>3</issue>), <fpage>700</fpage>&#8211;<lpage>717</lpage>. DOI: <pub-id pub-id-type="doi">10.1037/a0013553</pub-id></mixed-citation></ref>
<ref id="B99"><label>99</label><mixed-citation publication-type="journal"><string-name><surname>Yuille</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Kersten</surname>, <given-names>D.</given-names></string-name> (<year>2006</year>). <article-title>Vision as Bayesian inference: Analysis by synthesis?</article-title> <source>Trends in Cognitive Sciences, special issue: Probabilistic models of cognition</source>, <volume>10</volume>(<issue>7</issue>), <fpage>301</fpage>&#8211;<lpage>308</lpage>. DOI: <pub-id pub-id-type="doi">10.1016/j.tics.2006.05.002</pub-id></mixed-citation></ref>
<ref id="B100"><label>100</label><mixed-citation publication-type="journal"><string-name><surname>Zellou</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Scarborough</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Nielsen</surname>, <given-names>K.</given-names></string-name> (<year>2016</year>). <article-title>Phonetic imitation of coarticulatory vowel nasalization</article-title>. <source>Journal of the Acoustical Society of America</source>, <volume>140</volume>(<issue>5</issue>), <fpage>3560</fpage>&#8211;<lpage>3575</lpage>. DOI: <pub-id pub-id-type="doi">10.1121/1.4966232</pub-id></mixed-citation></ref>
<ref id="B101"><label>101</label><mixed-citation publication-type="journal"><string-name><surname>Zupan</surname>, <given-names>L. H.</given-names></string-name>, <string-name><surname>Merfeld</surname>, <given-names>D. M.</given-names></string-name>, &amp; <string-name><surname>Darlot</surname>, <given-names>C.</given-names></string-name> (<year>2002</year>). <article-title>Using sensory weighting to model the influence of canal, otolith and visual cues on spatial orientation and eye movements</article-title>. <source>Biological Cybernetics</source>, <volume>86</volume>(<issue>3</issue>), <fpage>209</fpage>&#8211;<lpage>230</lpage>. DOI: <pub-id pub-id-type="doi">10.1007/s00422-001-0290-1</pub-id></mixed-citation></ref>
</ref-list>
</back>
</article>