<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<!--<?xml-stylesheet type="text/xsl" href="article.xsl"?>-->
<article article-type="research-article" dtd-version="1.2" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="issn">2767-0279</journal-id>
<journal-title-group>
<journal-title>Glossa Psycholinguistics</journal-title>
</journal-title-group>
<issn pub-type="epub">2767-0279</issn>
<publisher>
<publisher-name>eScholarship Publishing</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.5070/G601147021</article-id>
<article-categories>
<subj-group>
<subject>Regular article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Social associations between voices and semantics affect word learning in voice-AI contexts</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-9167-0744</contrib-id>
<name>
<surname>Zellou</surname>
<given-names>Georgia</given-names>
</name>
<email>gzellou@ucdavis.edu</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-7896-4801</contrib-id>
<name>
<surname>R&#225;cz</surname>
<given-names>P&#233;ter</given-names>
</name>
<email>racz.peter.marton@ttk.bme.hu</email>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-1552-083X</contrib-id>
<name>
<surname>Barreda</surname>
<given-names>Santiago</given-names>
</name>
<email>sbarreda@ucdavis.edu</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
</contrib-group>
<aff id="aff-1"><label>1</label>Linguistics Department, UC Davis</aff>
<aff id="aff-2"><label>2</label>Department of Cognitive Science, Faculty of Natural Sciences, Budapest University of Technology and Economics</aff>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2026-02-13">
<day>13</day>
<month>02</month>
<year>2026</year>
</pub-date>
<pub-date pub-type="collection">
<year>2026</year>
</pub-date>
<volume>5</volume>
<issue>1</issue>
<elocation-id>X</elocation-id>
<permissions>
<copyright-statement>Copyright: &#x00A9; 2026 The Author(s)</copyright-statement>
<copyright-year>2026</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. See <uri xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</uri>.</license-p>
</license>
</permissions>
<self-uri xlink:href="https://glossapsycholinguistics.journalpub.escholarship.org/articles/10.5070/G601147021/"/>
<abstract>
<p>The present study investigates whether the relationship between voice gender (female and male voices) and gender-associated semantics (e.g., <italic>flower</italic> [female] vs. <italic>football</italic> [male]) affects the learning of an artificial lexicon after brief auditory exposure. Participants were trained on 16 novel words in either a gender-aligning condition (i.e., female voices produced novel words with female-associated meanings and male voices produced novel words with male-associated meanings) or a gender-mismatching condition (e.g., female voices produced words with male-associated meanings, etc.). Listeners either heard voices that contain stereotypical or non-stereotypical gross acoustic patterns. Overall, listeners showed worse word learning in gender-mismatching conditions. Voice stereotypicality modulated these effects: flipping gross gender cues reduced learning differences across conditions. These results demonstrate how gender associations modulate social attention when learning a new language. The findings also have implications for gender biases in human-computer interaction and speech technology, since the experiment was designed under the guise of a language learning application.</p>
</abstract>
</article-meta>
</front>
<body>
<sec>
<title>1. Introduction</title>
<p>Second language acquisition is difficult for adult learners. While much work has focused on explanations rooted in cognitive plasticity and age-related constraints, another potential contributing factor is the influence of statistical biases from one&#8217;s first language and personal cultural experiences. This is an area that has received comparatively less attention and warrants further investigation. Statistical biases can arise from socio-indexical information. Linguistic patterns are socially structured, and this structure can shape how learners map sensory input onto linguistic forms. Prior work demonstrates how social associations affect first-language spoken word comprehension (e.g., <xref ref-type="bibr" rid="B30">Kim, 2016</xref>; <xref ref-type="bibr" rid="B36">McGowan, 2015</xref>) and can have societal consequences (e.g., accent discrimination, <xref ref-type="bibr" rid="B6">Baugh, 2016</xref>). The role of socio-indexical information in shaping adult language learning is also a growing area of research (cf. <xref ref-type="bibr" rid="B21">Feher et al., 2014</xref>; <xref ref-type="bibr" rid="B41">Needle &amp; Pierrehumbert, 2018</xref>; <xref ref-type="bibr" rid="B47">R&#225;cz et al., 2017</xref>; <xref ref-type="bibr" rid="B48">Roberts, 2017</xref>). Adults have expectations regarding what, for example, men and women of the same age sound like (e.g., <xref ref-type="bibr" rid="B55">Strand, 1999</xref>) and talk about (<xref ref-type="bibr" rid="B17">Eckert, 1989</xref>), based on statistical patterns in their experiences, and these expectations might influence how novel word forms are learned.</p>
<p>Among these socio-indexical cues, gendered expectations &#8211; rooted in socially constructed stereotypes &#8211; may play a particularly influential role in shaping how learners process and acquire new words. Our goal is to examine how these expectations function, not to endorse them. This is because gender is a highly salient social category that listeners consistently classify from voice and use to guide attention and interpretation during language processing. In the current study, we ask if social associations between semantic meaning and a speaker&#8217;s voice acoustics affect the learning of novel words. Specifically, we investigate whether the relationship between perceived voice gender and stereotypically gender-associated meanings (e.g., <italic>flower</italic> = strongly female-associated vs. <italic>football</italic> = strongly male-associated) influences the learning of an artificial lexicon after brief auditory exposure. We also manipulate the voices producing the words such that they contain either stereotypical or non-stereotypical gross acoustic patterns across different versions of the experiment, thereby additionally exploring the effect of voice gender typicality on word learning. Furthermore, since device-based language learning applications are widely used (e.g., Duolingo), the experiment was designed to resemble a language learning app. Since learning is a general cognitive process, the current study can address longstanding issues at the intersection of language, society, and learning.</p>
<sec>
<title>1.1 Socio-indexical information and language processing</title>
<p>Language contains socially structured variation: linguistic variants are not evenly distributed across speaker groups (e.g., <xref ref-type="bibr" rid="B32">Labov, 2006</xref>). Gender is a key variable in linguistic variation. Individuals across gender identities tend to produce systematically distinct acoustic, phonological, lexical, and discourse patterns within a given speech community (for reviews, see <xref ref-type="bibr" rid="B13">Cheshire, 2004</xref>; <xref ref-type="bibr" rid="B14">Coates, 2015</xref>; <xref ref-type="bibr" rid="B18">Eckert &amp; McConnell-Ginet, 2013</xref>). Voices provide a listener with multiple layers of indexical meaning about talkers and are used by listeners to perform social categorizations. Based on voice alone, listeners show consistent classification of a speaker&#8217;s gender (e.g., <xref ref-type="bibr" rid="B44">Poon &amp; Ng, 2015</xref>), among other attributes such as age (e.g., <xref ref-type="bibr" rid="B4">Barreda &amp; Assmann, 2021</xref>) and nativeness (e.g., <xref ref-type="bibr" rid="B23">Girard et al., 2008</xref>). Such social categorizations subsequently affect how language and speech are processed by the listener (e.g., <xref ref-type="bibr" rid="B16">Drager &amp; Kirtley, 2016</xref>; <xref ref-type="bibr" rid="B27">Hay et al., 2006</xref>; <xref ref-type="bibr" rid="B42">Niedzielski, 1999</xref>; <xref ref-type="bibr" rid="B55">Strand, 1999</xref>). For instance, listeners&#8217; belief that a gender-ambiguous voice is either a woman or a man shifts their categorizations of the fricatives /s/ and /&#643;/ (<xref ref-type="bibr" rid="B55">Strand, 1999</xref>). The proposed explanation for this is that listeners make a gender classification of the speaker, and then identify the speech sound based on gender-dependent acoustic priors for linguistic categories.</p>
<p>Socially-based expectations about how language variation patterns across different types of speakers impacts language comprehension: McGowan (<xref ref-type="bibr" rid="B36">2015</xref>) found that comprehension of foreign-accented speech was enhanced when viewing an Asian vs. a white face (an extension of earlier work by Rubin (<xref ref-type="bibr" rid="B50">1992</xref>) that showed lower comprehension of non-accented speech when viewing an image of an Asian face, compared to an image of a white face). For speaker age, Kim (<xref ref-type="bibr" rid="B30">2016</xref>) found that listeners respond faster and more accurately to Korean words rated as being stereotypically &#8220;older&#8221; and &#8220;younger&#8221; when produced by an adult voice matching in the age category; meanwhile, processing was slower and less accurate when word and age category mismatched. Most, Sorber, and Cunningham (<xref ref-type="bibr" rid="B38">2007</xref>) had participants perform an &#8220;auditory Stroop task&#8221; where they categorized the gender of voices producing names and words stereotypically associated with female and male gender roles (e.g., <italic>Rachel, football</italic>) and found slower reaction times in trials where voice and word gender-associations were incongruent. Such effects can be explained using existing frameworks of language processing which generally assume that contextual cues are stored alongside linguistic representations, which is why context (like semantic content or social attributes, like speaker gender) can cue representations (such as how acoustic patterns map onto phonetic categories based on experienced priors) and vice versa (e.g., <xref ref-type="bibr" rid="B28">Johnson &amp; Sjerps, 2021</xref>, discussed further below). There is evidence that these statistical patterns arise very early during language learning: For instance, caretakers produce higher rates of standard phonological forms when talking to girls than boys (Foulkes et al., 2005; <xref ref-type="bibr" rid="B52">Smith et al., 2009</xref>). Consequently, even very young children display gendered use of linguistic variants (<xref ref-type="bibr" rid="B39">Munson et al., 2022</xref>; <xref ref-type="bibr" rid="B45">Prystawski et al., 2020</xref>; <xref ref-type="bibr" rid="B49">Robertson &amp; Murachver, 2003</xref>; <xref ref-type="bibr" rid="B54">Stennes et al., 2005</xref>).</p>
<p>There is some prior work looking at how socio-indexical information affects novel word learning. R&#225;cz, Hay, and Pierrehumbert (<xref ref-type="bibr" rid="B47">2017</xref>) found that participants can learn an artificial language containing speaker-gender conditioned morpho-phonological variation; thus, listeners are sensitive to socially structured variation in the input during language learning. They also compared learning in the gender-correlated variation condition to a condition where variant use correlated with different spatial orientations of the talker (e.g., a forward-facing speaker used one alternant and a side-facing speaker used another). Participants learned gender-correlated variation (a socially-structured way that language variation occurs in the real world) better than spatially-oriented variation (not a socially-relevant cue in the real world), suggesting that perceivers are attending more strongly to socially interpretable cues when learning language. More evidence comes from Needle and Pierrehumbert (<xref ref-type="bibr" rid="B41">2018</xref>) who found that participants&#8217; gender identifications of real words matched those words&#8217; statistical distributions across a corpus of women&#8217;s and men&#8217;s writing (e.g., words more likely to appear in women&#8217;s writing were more likely identified as produced by a female author). Moreover, Needle and Pierrehumbert also found an effect for pseudowords that contained English-like derivational affixes that had gender-biased statistical distributions: for instance, the <italic>-case</italic> suffix (e.g., <italic>pillowcase, suitcase</italic>) is more likely to occur in women&#8217;s writing, and participants were more likely to identify the author as female when evaluating pseudowords like <italic>clumcase</italic>.</p>
<p>Theoretical models of processing which posit that attention guides the encoding and storage of information can account for, and make predictions about, how social and linguistic details in the input might influence the strength of learning. For example, predictive-processing theories propose that comprehenders continuously generate expectations about upcoming information and allocate greater attention to unexpected inputs; the resulting attention boost when stimuli contain expectations violations are proposed to facilitate encoding and generalization. This has been observed in artificial language studies, where items containing expectation violations are demonstrated to strengthen abstraction and generalization of newly learned structures by learners (e.g., <xref ref-type="bibr" rid="B8">Bovolenta &amp; Marsden, 2021</xref>). One might predict, then, that learners exposed to items containing social and semantic cues that violate expectations will display stronger learning than those exposed to matching cues, since they attract greater attentional engagement and, thus, lead to stronger encoding and generalization.</p>
<p>However, not all violations are beneficial to learning. Experiments that have manipulated social cues in discourse observe that incongruent turn-taking or prosody hinder the success of word learning (<xref ref-type="bibr" rid="B33">Lee &amp; Lew-Williams, 2022</xref>). More support comes from a series of large-scale experiments on cross-situational word learning which found that adults relied heavily on social cues (e.g., gaze) to reduce referential uncertainty (<xref ref-type="bibr" rid="B35">MacDonald et al., 2017</xref>). When these cues were less reliable, learners reverted to tracking multiple hypotheses instead of committing to a single mapping, which slowed learning. This suggests that violations of cue reliability expectations hinder efficient encoding. Thus, finding that mismatched social and semantic information leads to lower word learning in the present study would be consistent with models of processing where learners are assessing the reliability of social cues and use that to guide attention during learning.</p>
</sec>
<sec>
<title>1.2 Social factors in human-computer interaction</title>
<p>Humans are now in a new digital era where many people of all ages regularly talk to voice-activated artificially intelligent (AI) personal assistants, such as Siri, Alexa, and Google Assistant, that spontaneously produce interactive speech (<xref ref-type="bibr" rid="B1">Ammari et al., 2019</xref>). While language-based interaction with computers has long existed &#8211; primarily through typing &#8211; the advent of spoken language technology now enables users to operate devices using speech alone. The ease and efficiency in using speech to interact with machines to complete everyday tasks explains why people find speech-enabled functions natural and effortless, which means that voice-AI is being integrated into a larger proportion of our spoken interactions (<xref ref-type="bibr" rid="B15">De Renesse, 2017</xref>). It also suggests that many more humans will be spending a lot of time talking to machines in the future.</p>
<p>Social expectations are so robust that they even apply in contexts where listeners know they are not interacting with a real human. Indeed, there is much research showing that people anthropomorphize machines by attributing human-like traits and characteristics to them (<xref ref-type="bibr" rid="B58">Waytz et al., 2010</xref>). Moreover, prior research shows that people apply human-based behavioral norms to interactions with machines and computers, including norms shaped by gender stereotypes. For instance, Kuchenbrandt and colleagues (<xref ref-type="bibr" rid="B31">2014</xref>) tested whether the gender-typicality of a task with a robot would affect users&#8217; performance. They had participants complete either a stereotypically-female task (sorting a sewing kit) or a stereotypically-male task (sorting a toolbox) while being instructed by a robot with either a female or a male guise. They found participants performed worse with the robot in the context of a stereotypically-female task, for both the female and male guises. Kuchenbrandt and colleagues remarked that many of the activities we perform with robots are inherently related to societal gender roles, and this will influence how users interact with robots.</p>
<p>For voice-enabled devices in particular, just by the nature of their use of speech, it has been shown that people apply voice bias to machines. For instance, Nass, Moon, and Green (<xref ref-type="bibr" rid="B40">1997</xref>) had participants complete a tutoring session with a computer that produced either a female voice or a male voice and were presented with either stereotypically masculine topics (i.e., &#8220;technology&#8221;) or stereotypically feminine topics (i.e., &#8220;relationships and love&#8221;). They found that congruence between the apparent gender of the voice and the stereotypicality of the topic led to higher ratings of the computer&#8217;s competence than if there was a voice-topic mismatch. Ernst and Herm-Stapelberg (<xref ref-type="bibr" rid="B20">2020</xref>) investigated whether gender stereotyping influences the perceived likability of virtual voice assistants. They had participants interact with a virtual assistant, assigned either to a female-voiced condition or a male-voiced condition. Post-interaction surveys revealed that participants who interacted with the female-voiced assistant rated the system as more likable than the male-voiced assistant. These studies show that people are particularly prone to applying human-based gender expectations to machines when they express apparent gender through their voice.</p>
<p>This shift toward voice-based interaction raises important questions about how social expectations, particularly gender stereotypes, influence user experience and learning outcomes. Language learning apps, such as Duolingo, and computer programs, such as Rosetta Stone, are digital systems that users can use as a didactic tool for vocabulary building in a second language. Language learning apps are most commonly used for home learning (<xref ref-type="bibr" rid="B24">Godwin-Jones, 2011</xref>), but they are also being explored for use in the classroom (<xref ref-type="bibr" rid="B26">Guaqueta &amp; Castro-Garces, 2018</xref>). And there are projections that device-based language learning applications will increase in future years (<xref ref-type="bibr" rid="B25">Godwin-Jones, 2017</xref>). There is a growing body of work exploring whether people exhibit social behavior towards machines in the same ways that they do with other humans (e.g., <xref ref-type="bibr" rid="B22">Gambino &amp; Liu, 2022</xref>; <xref ref-type="bibr" rid="B40">Nass et al., 1997</xref>). For instance, in extending the classic Asch study of social conformity to peers, Brandstetter and colleagues (<xref ref-type="bibr" rid="B9">2014</xref>) found that while participants show conformity to inaccurate verbal judgements made by human peers, they do not show similar influence from robot &#8220;peers&#8221;. With respect to speech and language patterns, the literature is mixed. In some prior work, participants show perceptual adaptation of a vowel shift produced by an apparent voice-AI device to the same degree as they do for an apparent human speaker (<xref ref-type="bibr" rid="B67">Zellou et al., 2023</xref>). Yet other work has shown that college-age participants show little to no phonetic imitation of device voices (<xref ref-type="bibr" rid="B65">Zellou, Cohn, &amp; Ferenc Segedin, 2021</xref>; <xref ref-type="bibr" rid="B66">Zellou, Cohn, &amp; Kline, 2021</xref>). Thus, it remains an open question as to whether people learn, and how human-based social patterns will affect learning, when interacting with a voice-AI device. What role does voice gender play in learning via computer-based applications? The Nass et al. (<xref ref-type="bibr" rid="B40">1997</xref>) study found that incongruence between the gender of the voice and the semantic content of a computer-based tutoring session affects listeners&#8217; learning experience. Thus, for the present study, we predict that, even under the guise of a word learning computer app, gender biases in word learning will be found.</p>
<p>There are important societal implications of the growing ubiquity of voice-AI and an observation of gender associations impacting language learning when interacting with machines. For one, female-voiced digital assistants as the default option can lead to reinforcing and exacerbating gender asymmetries that exist in society. Digital devices are often used to assist, are subservient toward, and rarely disagree with or are unpleasant to, the user. Female-voiced AI systems send the message to users that women and girls should express themselves similarly (<xref ref-type="bibr" rid="B59">West et al., 2019</xref>, UNESCO report). Moreover, as mentioned earlier, the current state of speech technology is that most commercially available systems in use provide text-to-speech voices that reflect a gender binary. These choices can impact behavior during interactions with speech and language technology. More generally, such biases could reinforce and amplify the notion that gender is a binary, as well as societal biases around gender-nonconformity.</p>
<p>Examining the role of social factors on lexical acquisition in the context of human-computer interaction can also speak to theoretical accounts of attention and learning. Apparent gender in synthetic voices might activate gendered expectations about content and competence, which, in turn can shape attention during learning tasks. Device voices can act as sources of prior expectations, which can lead to differential allocations of attention. Within a predictive-processing account, incongruence of voice and semantic cues could trigger helpful surprise, boosting learning. In contrast, cue-reliability accounts of learning predict that cue mismatches impose processing costs that hinder learning; when voice-AI-based social cues are congruent with content, perceptual fluency is higher and learning is facilitated.</p>
</sec>
<sec>
<title>1.3 Current study</title>
<p>The current study investigates whether statistical associations between social properties of voices and semantic meaning together influence word learning. We tested this using a mock word learning app called WordBot. In general use, text-to-speech voices typically align stereotypical gender cues (f0, vocal tract length, prosody, source characteristics), which means that non-stereotypical combinations remain understudied in human-computer interaction.</p>
<p>In the training phase, listeners hear sixteen novel words in an invented language produced by two talkers. We manipulate the relationship between voice and semantic content across two conditions: gender-aligning (female voice produces female-associated words like <italic>lipstick, flower</italic>, male voice produces male-associated words like <italic>hammer, football</italic>) and gender-mismatching (voice-meaning pairings are reversed). In the test phase, listeners hear the same voices as in the training phase, producing the same words. Listeners need to identify associated objects.</p>
<p>Following the test phase, listeners enter a second test phase in which they hear the same words in two additional voices. The relationship between voice and semantic content is the same as with the training voices. This phase examines whether generalization differs across alignment conditions. Prior work on perceptual learning shows that generalization to new talkers is a hallmark of robust learning, yet it is often weaker than same-talker performance, because cross-talker variability introduces processing costs (<xref ref-type="bibr" rid="B19">Eisner &amp; McQueen, 2005</xref>; <xref ref-type="bibr" rid="B64">Xie &amp; Myers, 2017</xref>). In the present study, we predict that participants will generalize their learning to novel voices, but with reduced accuracy compared to the same-voice test block. This expectation is based on theories of attention and cue reliability: when social cues (voice&#8211;semantic associations) align with prior expectations, perceptual fluency should facilitate generalization; when mismatched, reduced cue reliability may hinder it. Thus, we anticipate successful generalization overall, but also expect that alignment versus mismatch will modulate the strength of generalization effects.</p>
<p>We test voice stereotypicality across two conditions, original text-to-speech voices (stereotypical associations between gendered speech patterns and gross acoustic cues) or flipped voices (gross acoustic patterns and fine-grained phonetic patterns incongruent, creating non-stereotypical male and female voices). For example, in the flipped condition, male voices would receive the average f0 and spectral cues typical of female voices, while maintaining the more subtle gendered cues associated with the male voices (e.g., intonation, phonological variation).</p>
<p>Following the word learning study, participants provide social evaluations of the voices: communicative competence ratings and assessment of whether they would use a language learning app with each voice.</p>
<p>We had the following hypotheses:</p>
<list list-type="bullet">
<list-item><p>H1. Participants will show better word learning performance in the gender association-aligning conditions, compared to gender-mismatching conditions.</p></list-item>
<list-item><p>H2. Participants will successfully generalize learning to novel voices, but accuracy will be lower than in the same-voice test block, and this reduction may interact with alignment conditions (gender-aligning vs. gender-mismatching).</p></list-item>
<list-item><p>H3. The mismatch effect will be weaker with the flipped AI voices than with the original voices.</p></list-item>
<list-item><p>H4. If explicit judgments of voices affect word learning, we predict that if participants rate the voices as less communicatively competent and less preferred, the participants will have lower learning outcomes.</p></list-item>
</list>
</sec>
</sec>
<sec>
<title>2. Methods</title>
<sec>
<title>2.1 Stimuli</title>
<p>In the current study, all participants learned 16 novel word-image pairings. The novel words were taken from a previous artificial language study (<xref ref-type="bibr" rid="B29">Kaushanskaya &amp; Marian, 2009</xref>). We took 16 of the &#8220;phonologically familiar&#8221; items, designed to be similar in structure to many English real words. They were constructed from 8 English phonemes: four vowels (/&#593;/, /&#603;/, /i/, and /u/) and four consonants (/f/, /n/, /t/, and /g/).</p>
<p>In order to identify what the meaning associations would be, we used a publicly available corpus of psycholinguistic ratings of over 5,500 English words (<xref ref-type="bibr" rid="B51">Scott et al., 2019</xref>). In that study, they had participants rate words on a number of scales, including gender. Participants were told &#8220;A word&#8217;s gender is how strongly its meaning is associated with male or female behavior. A word can be considered MASCULINE if it is linked to male behavior. Alternatively, a word can be considered FEMININE if it is linked to female behavior. Please indicate the gender associated with each word on a scale of VERY FEMININE (1) to VERY MASCULINE (7), with the midpoint being neuter (neither feminine nor masculine).&#8221;</p>
<p>From the Scott et al. (<xref ref-type="bibr" rid="B51">2019</xref>) ratings data, we selected only meanings that were highly imageable (rated 5.4 or higher on a 7 point scale of imageability). Then, we selected the top eight most female-associated words (gender-associated rating mean of 1.5 for the eight words) and the top eight most male-associated words (gender-associated rating mean of 6.06 for the eight words). Words were replaced with the next appropriate word if they referred to drugs/addictive substances (<italic>cigar</italic>), weapons (<italic>grenade, gun</italic>), a physical person (<italic>lady</italic>), or a body part (<italic>vagina, penis</italic>). The lexical items and meanings (with imageability and gender ratings) used in this study are provided in the Appendix.</p>
</sec>
<sec>
<title>2.2 Voices</title>
<p>Each of the 16 words was generated in 8 text-to-speech US-English voices (4 female: Danielle, Kendra, Kimberly, Ruth; 4 male: Gregory, Matthew, Joey, Stephen) using AWS (Amazon Web Services) Polly in neural text-to-speech. The items were generated using SSML tags with the IPA form of the word (e.g., &lt;speak&gt; &lt;phoneme alphabet=&#8221;ipa&#8221; ph=&#8221;f&#593;n&#603;t&#8221;&gt; &lt;/phoneme&gt; &lt;/speak&gt;). Words were downloaded from the AWS console and amplitude normalized to 60 dB.</p>
<p>All stimuli were resynthesized using the &#8216;Change gender&#8217; function in Praat. This function manipulates the sampling frequency to achieve a linear rescaling of the sound spectral envelope, thereby mimicking differences in speaker vocal-tract length (<xref ref-type="bibr" rid="B3">Barreda, 2020</xref>). Differences in f0 are implemented using overlap-add synthesis. In order to maintain uniformity in their processing, all stimuli were resynthesized using this function in two conditions: unmodified and flipped. For the unmodified condition, spectral envelopes were not manipulated, but male voices were modified to have a median f0 of 110 Hz, and female voices were modified to have a median f0 of 220 Hz. This allowed each speaker to maintain their unique pitch contour, but placed speakers of each gender in the same general range. In the flipped condition, male voices were given a median f0 of 220 Hz, and the spectral envelope was scaled up by a factor of 1.2; female voices were given a median f0 of 110 Hz, and the spectral envelope was scaled down by a factor of 1/1.2 (0.83). The result of these manipulations is a potential mismatch between gross gender cues (overall f0 and spectral cues to vocal tract length) and fine-grained gender cues (everything else: prosody, coarticulation, etc.) in the flipped condition. This mismatch will reduce the gender stereotypicality of flipped voices, and may affect the voices&#8217; perceived masculinity/femininity, even for voices all classified as female or male.</p>
</sec>
<sec>
<title>2.3 Participants</title>
<p>231 native English speakers took part in the study (self-reported: 174 female, 1 trans woman, 5 non-binary/gender non-conforming, 51 male; mean age = 19.7 years old). They completed the experiment online via a Qualtrics survey. Participants were recruited from the UC Davis subjects&#8217; pool and given partial course credit for their participation. This study was approved by the UC Davis Institutional Review Board and all participants completed informed consent. Participants were instructed to complete the experiment in a quiet room without distractions or noise, to silence their phones, and to wear headphones. None of the listeners reported having a hearing or language impairment.</p>
</sec>
<sec>
<title>2.4 Procedure</title>
<p>The study began with a pre-test of participant audio: participants heard one sentence presented auditorily (&#8220;She asked about the host&#8221;) and were asked to identify the sentence from three multiple choice options, each containing a phonologically close target word (<italic>host, toast, coast</italic>). All participants passed this audio check.</p>
<p>Participants then completed the experiment, which consisted of four phases: training phase, talker-specific testing block, generalization testing block, then, finally, a voice ratings block. <xref ref-type="fig" rid="F1">Figure 1</xref> provides a flowchart demonstrating the blocks and their ordering as completed by participants.</p>
<fig id="F1">
<label>Figure 1</label>
<caption>
<p>Flowchart of the experimental blocks.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g1.png"/>
</fig>
<p>The training phase consisted of a paradigm adapted from Wong and Perrachione (<xref ref-type="bibr" rid="B62">2007</xref>). In this phase, participants were trained to identify word meanings as depicted by drawings of concrete nouns. Each participant was trained on a vocabulary of 16 items. Every participant was trained on the same 16 item-meaning pairings (see the Supplementary Materials), with 8 of the items referring to a &#8220;female-associated&#8221; object and 8 of the items referring to a &#8220;male-associated&#8221; object.</p>
<p>The format of the training session is illustrated in <xref ref-type="fig" rid="F2">Figure 2</xref>. To facilitate learning, participants were trained on a group of 3 words at a time (i.e., a trial group of 3 items). Each of the 16 items was presented a total of 3 times, each item randomly assigned to one of 16 trial groups. The trial groups varied in having gender-mixed words and gender-consistent words (the example in <xref ref-type="fig" rid="F1">Figure 1</xref> happens to show a gender-mixed trial group). In a learning session, participants heard a word production and saw an image which they were told corresponded to the meaning of the word (<xref ref-type="fig" rid="F2">Figure 2.A</xref>). After three trials, participants were then given a mini-test on the three words they just learned. One of the three words was played and participants selected the correct corresponding image from among three images (<xref ref-type="fig" rid="F2">Figure 2.B</xref>). After making a selection, feedback was provided to facilitate recognition of the items and to correct if a mistake was made. Participants heard 48 items (16 words * 3 repetitions across different trials groups).</p>
<fig id="F2">
<label>Figure 2</label>
<caption>
<p>Example of a training session trial group. Panel A illustrates three separate trials from the same group, shown together for compactness; in the actual experiment, each word&#8211;image pair was presented one at a time. After hearing all three words with their corresponding images, participants were quizzed on those items (Panel B). During training, feedback was provided after each quiz response. Across the training phase, participants learned 16 word&#8211;image associations.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g2.png"/>
</fig>
<p>In the talker-specific testing block, which directly followed the training phase, participants were presented with each of the 16 images (see <xref ref-type="fig" rid="F1">Figure 1</xref>). On each trial, they heard one of the 16 learned words presented in the voice that they had heard for that item in the training phase. They were asked to select the correct image. Words were presented randomly, twice each. No feedback was provided. Participants completed 32 total testing trials (16 words * 2 repetitions).</p>
<p>Directly following the talker-specific (Same Talker) testing block, participants completed the generalization (Novel Talker) testing block. The trial design was identical to the previous block, except participants now heard two novel voices produce the items. The gender-meaning pairing of the novel voices was the same as that for exposure for each participant. For instance, if they had heard the female voice produce female-associated words in training and testing block 1, in the novel-talker testing block, they heard female-associated words in a novel female voice. Words were presented randomly, twice each. No feedback was provided. Participants completed 32 total testing trials (16 words * 2 repetitions).</p>
<p>Participants were randomly assigned to one of the four experimental conditions: 57 were assigned to gender-aligning conditions with original voices; 57 were assigned to gender-mismatching conditions with original voices; 57 were assigned to gender-aligning conditions with flipped voices; 60 were assigned to gender-mismatching conditions with flipped voices.</p>
<p>Finally, participants completed several ratings for each of the four voices they had heard in the ratings block. First, we asked participants to provide a gender categorization of each of the voices (&#8220;What do you think the gender of this voice is?&#8221;; binary choice: Female or Male). Next, we asked them to rate the voices on gender stereotypicality (two questions: &#8220;How feminine/masculine does this voice sound?&#8221;, Likert scale (1 = Totally feminine, 5 = Totally masculine); &#8220;How typical does this voice sound for its gender?&#8221;, Likert scale (1 = Very unusual, 5 = Very typical). Then, we asked them to provide ratings of two statements to assess the voice&#8217;s communicative competence. One statement assessed comprehension competence: &#8220;This WordBot will understand my voice and what I am saying&#8221;. One statement assessed production competence: &#8220;This WordBot speaks clearly so that I understand everything it says&#8221;. Next, we asked them to respond to two statements to assess the voice&#8217;s naturalness, which prior work has shown to vary in perception of robots that have congruent vs. conflicting cues (<xref ref-type="bibr" rid="B37">Moore, 2012</xref>): Naturalness: &#8220;This WordBot sounds naturalistic/human-like&#8221; and Creepiness: &#8220;This WordBot sounds creepy&#8221;. Finally, we asked them to provide an overall rating of the voice: &#8220;Overall, I would use this WordBot if it were available&#8221;. Responses were provided on a Likert scale (1 = Strongly disagree, 5 = Strongly agree).</p>
</sec>
<sec>
<title>2.5 Data analysis</title>
<p>We collected 14,784 responses to items in the test and generalization phases of the task from 231 participants. We also collected a set of ratings from each participant on each of the voices that participant encountered.</p>
<p>We used R, ggplot, and sjplot for data analysis and visualization (<xref ref-type="bibr" rid="B34">L&#252;decke &amp; L&#252;decke, 2015</xref>; <xref ref-type="bibr" rid="B46">R Core Team 2024</xref>; <xref ref-type="bibr" rid="B61">Wickham, 2016</xref>). We fit linear models using brms and rstan (<xref ref-type="bibr" rid="B10">B&#252;rkner, 2017</xref>; <xref ref-type="bibr" rid="B53">Stan Development Team, 2024</xref>). The models were fit using 5000 iterations on 4 chains and weakly informative priors. We used leave-one-out cross-validation to select the best model. Data analysis was pre-registered (<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://aspredicted.org/dpy9-6ffy.pdf">https://aspredicted.org/dpy9-6ffy.pdf</ext-link>) for the learning data. We performed post hoc tests on the rating data.</p>
<p>We fit a model on the rating data and one set of models on the learning data to assess participant accuracy across conditions. For the rating data, in order to see whether &#8220;natural&#8221; ratings correlate with our modulation of the voices, we fit a Bayesian ordinal regression model predicting the Likert scale ratings for naturalness from voice type (original/flipped), using weakly informative priors. For participant accuracy, we fit a logistic regression model on the learning data predicting whether a response to an item was accurate (= the participant gave the appropriate label for an object). The predictors included: alignment between meaning and voice (aligned = a female voice read the items with female associations, and a male voice read the items with male associations, mismatched = a male voice read the items with female associations, a female voice read the items with male associations), voice type (original AI voice, flipped voice modulated to have mismatched gender cues), block (same voice or novel voice, see 2.2, <xref ref-type="fig" rid="F1">Figure 1</xref>) and the participants&#8217; rating on whether the specific voice &#8220;sounds natural&#8221;. That is, we used the &#8220;natural&#8221; rating as a proxy for other ratings in our model of participant accuracy. The model had a grouping factor for participants and items. The best model has a three-way interaction between alignment, voice type, and block.</p>
<p>We deviated from the pre-registration in three ways. First, we did not remove participants or trials based on the exclusion criteria, since we deemed task accuracy sufficient to reliably assess participant attention. Also, since we did not put enough emphasis on response speed in participant instructions, some participants might have been fully paying attention and still have been extra slow. Test and generalization trials required the participant to select one out of 16 options, and the relative difficulty of this trial type likely increased variability in trial duration. Second, we ended up with more participants than the sample size put forward in the preregistration (231 participants vs. 200 pre-registered). We did not stop recruitment based on observing the data, and since the final sample size was only 115% of the preregistered size, we decided not to discard extra participants. Third, we added participant ratings of the voices&#8217; typicality as a predictor in the learning analysis to account for the participants&#8217; subjective response to the voices.</p>
</sec>
</sec>
<sec>
<title>3. Results</title>
<p>Because the interpretation of the results hinges on the perception of gender, we will begin by outlining gender perception across conditions. The original voices were identified as their intended gender in 100% of cases for female speakers, and in 96% of cases for male speakers. The voice flip condition succeeded in reversing these judgments: in this condition, female speakers were identified as female in only 5% of cases, while male speakers were identified as male in only 15% of cases. We were also interested in the graded measure of gender typicality. We fit an ordinal model predicting the Likert scale response on gender typicality, using the voice flip condition as the predictor. The ordinal model revealed a meaningful effect for the voice flip condition (est = &#8211;0.81, se = 0.22, 95% CI: [&#8211;1.26; &#8211;0.4]), where original voices were rated as more typical for their gender than flipped voices. To present this in terms of the response scale (1 = Very unusual, 5 = Very typical), for flipped voices, the predicted probabilities were highest for a rating of 4 (0.53) and 3 (0.29), with very few predicted responses of 5 (0.03). In contrast, for original voices, the model predicted a higher probability of a rating of 4 (0.69), a lower probability of a rating of 3 (0.14), and a much higher probability of 5 (0.14), indicating a notable shift toward less typical gender ratings in the flipped condition. The gender identity of participants is notably absent from this analysis &#8211; we return to this at the end of this section.</p>
<p>Raw participant means across conditions can be seen in <xref ref-type="fig" rid="F3">Figure 3</xref>. Participants were broadly more accurate in the aligned condition than in the mismatched condition, and this was particularly visible for test trials (as opposed to generalization trials) and participants training and testing with the original AI voices (as opposed to the flipped voices). However, aggregating over the raw data is a poor proxy for the signal in the experiment, because it masks two important facets of the data, namely, that (a) participants responded to individual items, which is not shown by the means, and (b) the subjective experience of participants was likely more relevant for their accuracy than our labels for the AI voices.</p>
<fig id="F3">
<label>Figure 3</label>
<caption>
<p>Participant means across experiment block (test or generalization), voice-item alignment (aligned or mismatched) and AI voice type (original or flipped). Each participant responded to both test and generalization trials (within-participant condition). Each participant encountered either aligned or mismatched and either original or flipped voices (across-participant conditions). The line segments mark individual participant accuracy in test and generalization. The dotted horizontal line shows chance level.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g3.png"/>
</fig>
<p>Participants rated voices using several Likert scales: sounds feminine or masculine, creepy sounding, natural sounding, speaks clearly, sounds typical for their gender, easy to understand, and would use in the future. <xref ref-type="fig" rid="F4">Figure 4</xref> shows the Pearson correlations between ratings across participants. The &#8220;stereotypical&#8221; rating is included for the reader&#8217;s convenience. It is based on the feminine/masculine rating applied to female or male voices. If the voice is originally female, stereotypicality refers to the female end of the rating scale; if the voice is originally male, the other way around. For simplicity&#8217;s sake, this includes both the original and flipped voices. Across rating scales, responses are moderately correlated, indicating that (a) participants evaluated the voices holistically, and (b) voice gender perception was related to the other dimensions of the rating scales: if a participant rates a voice as less gender-confirming, it also sounds less natural and harder to understand.</p>
<fig id="F4">
<label>Figure 4</label>
<caption>
<p>Pearson correlations between participant ratings of the voices used in the experiment.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g4.png"/>
</fig>
<p>Principal components analysis reveals a stronger natural/clear/non-creepy dimension and a weaker typical/gender-stereotypical dimension. These are visible in the correlations in <xref ref-type="fig" rid="F4">Figure 4</xref>, as well. We used the &#8220;sounds natural&#8221; rating as our benchmark (as opposed to the &#8220;creepy&#8221; or &#8220;would use&#8221; rating, etc.) because it shows the strongest correlation with the other ratings (though not &#8220;typical&#8221;) and it is relatively straightforward to interpret.</p>
<p>Once we establish that &#8220;sounds natural&#8221; works as a good benchmark of how participants perceive the voices, we can go on to look at a model of participant accuracy that incorporates this subjective dimension. The estimates of our best model of participant accuracy across conditions can be seen in <xref ref-type="fig" rid="F5">Figure 5</xref>. Treatment coding was used, so the intercept captures the model&#8217;s baseline, a response to an object matching the voice gender, with an original voice in the same-talker test block. The model uses the logit scale. Response accuracy in this group is between &#8211;0.18 and 0.49; this translates to an accuracy of 45 &#8211; 62%, which is much higher than the chance accuracy of 6.25% (since each trial had 16 possible response choices). The rest of the estimates show relative accuracy compared to the intercept. If the model&#8217;s estimate largely excludes zero (the dotted vertical line) and has a magnitude that is large enough to noticeably affect outcomes, the effect is likely meaningful. For example, if a participant rated a voice as more natural, the participant was also more accurate in the experiment. The rest of the estimates are harder to interpret in isolation, because they take part in a three-way interaction between object-voice alignment, experiment block, and AI voice type.</p>
<fig id="F5">
<label>Figure 5</label>
<caption>
<p>Posterior distributions from the best model fit on the learning data. The vertical lines show the median, the shaded areas, and the 50%, the outlines, the 95% credible intervals.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g5.png"/>
</fig>
<p>We can better understand this interaction if we look at <xref ref-type="fig" rid="F6">Figure 6</xref>, which shows model predictions. We see predicted values for participants training and testing with original (left panel) and flipped AI voices (right panel). The predictions are similar to the raw data in <xref ref-type="fig" rid="F3">Figure 3</xref>. We see what appears to be an interaction between alignment and mismatch in the original condition, but not in the flipped condition (right panel). In the flipped condition, we see additive effects for generalization and alignment, with no interaction between the two, such that the lines in the plot are nearly parallel. One way to interpret this is that word-voice alignment is easier than word-voice mismatch, but this is mostly seen in the test phase and for the original voices. The aligned/mismatched difference is much smaller for flipped voices. Generalization is harder than test across the board, though still above chance level (1/16).</p>
<fig id="F6">
<label>Figure 6</label>
<caption>
<p>Samples (draws) of expected values of the posterior predictive distributions from the model fit on the learning data. Points show medians. Line ranges show 95% credible intervals. Predictions transformed from log odds to percentages. The dotted horizontal line shows chance level.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g6.png"/>
</fig>
<p>We can explicitly tie our results to our hypotheses using this model, detailed above, and Bayesian hypothesis testing.</p>
<list list-type="bullet">
<list-item><p>H1. Participants were less accurate with word-voice mismatch than with word-voice alignment (est = &#8211;0.61, se = 0.23, 95% CI: [&#8211;0.99; &#8211;0.22]).</p></list-item>
<list-item><p>H2. Participants were less accurate in the generalization block than in the test block (est = &#8211;0.58, se = 0.08, 95% CI: [&#8211;0.71; &#8211;0.46]). However, participants were still above chance (p = 1/16 or lo = -2.71) in the generalization block (est = 2.28, se = 0.02, 95% CI: [1.87; 2.69]).</p></list-item>
<list-item><p>H3. The mismatch effect was weaker with flipped AI voices. If participants were trained with a flipped AI voice, there was a much smaller difference between the aligned and mismatched conditions (est = 0.42, se = 0.51, 95% CI: [&#8211;0.4;1.26]).</p></list-item>
<list-item><p>H4. If a participant found a voice more natural, their responses were more accurate (est = 0.18, se = 0.08, 95% CI: [0.04; 0.31]). Voice naturalness best captures the behaviour of other ratings and robustly correlates with voice type (original or flipped), so we used this as the proxy for the participant preference and competence judgements. At the same time, naturalness ratings explain additional variation beyond voice type. That is, the subjective experience of participants does not reduce to our labels for the voices.</p></list-item>
</list>
<p>Any discussion of how participants reacted to voice gender directly ties in with participant gender identity. However, 175 of our 231 participants are women, and 174 are cis women. As a result, we cannot provide a comprehensive account of the effects of listener gender identity across our conditions. However, participants were randomly assigned to conditions. When we look at participant random intercepts in our best model, which express an overall rate of accuracy for participants, above and beyond treatment effects, we do not see that ciswomen and the other groups in the sample cluster or pattern differently. This can be seen in <xref ref-type="fig" rid="F7">Figure 7</xref>. This strongly suggests that, at least in the sample, participant gender identity did not have a major, discernible effect on participant accuracy.</p>
<fig id="F7">
<label>Figure 7</label>
<caption>
<p>Listener random intercepts and listener gender identity in our best model.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-5-1-47021-g7.png"/>
</fig>
</sec>
<sec>
<title>4. General discussion</title>
<p>The current study investigated the impact of social associations between a speaker&#8217;s voice acoustics and lexical meaning on the learning of novel words, using a paradigm that simulates a language learning app. In the subsections below, we discuss our findings in detail, and interpret their implications for learning at the intersection of speech variation, society, and technology.</p>
<sec>
<title>4.1 Role of mismatching voice and semantic gender cues</title>
<p>Our key finding is that participants learned words less well when the gender associations of their meanings mismatched with the gross gender acoustics of the voices (H1). We observed this across two separate experiments in which voices contained stereotypical and non-stereotypical fine-grained acoustic features.</p>
<p>These results can contribute to our understanding of the role of socio-indexical information on language processing and learning. Prior work shows that the statistical distribution of real words across social groups affects the usage of novel pseudo-words (<xref ref-type="bibr" rid="B41">Needle &amp; Pierrehumbert, 2018</xref>) and that associations between language variation and speaker gender in an artificial language are learned (<xref ref-type="bibr" rid="B47">R&#225;cz et al., 2017</xref>). The present study extends this to the auditory domain and novel word learning, by finding that learning performance is lower when gendered associations of the meaning and the voice are mismatching. Why is this the case? Since prior work shows that the social distribution in which people experience words affects the processing and comprehension of their first language (<xref ref-type="bibr" rid="B30">Kim, 2016</xref>; <xref ref-type="bibr" rid="B36">McGowan, 2015</xref>; <xref ref-type="bibr" rid="B42">Niedzielski, 1999</xref>; <xref ref-type="bibr" rid="B50">Rubin, 1992</xref>; <xref ref-type="bibr" rid="B55">Strand, 1999</xref>; <xref ref-type="bibr" rid="B67">Zellou et al., 2023</xref>), such statistical patterns could be transferred when learning and processing new words. For instance, the real world statistical and episodic experience people have with words in their native language is proposed to form the representations from which listeners process language (<xref ref-type="bibr" rid="B28">Johnson &amp; Sjerps, 2021</xref>; <xref ref-type="bibr" rid="B43">Pierrehumbert, 2016</xref>). Such patterns might include voices that have female-presenting attributes (e.g., higher f0) producing words that evoke concepts such as those used in the present study (e.g., dresses).</p>
<p>Thus, the learning of new lexical forms can be supported when the gendered associations between the voice and meaning align with past experiences learners have from their first language. More broadly, this finding supports models of cognitive processing in which sensory inputs containing features that align with prior experience are easier to process and, therefore, easier to learn and remember (e.g., <xref ref-type="bibr" rid="B60">Westerman et al., 2002</xref>), as well as models of memory in which stimuli containing expected information are encoded and recognized better (e.g., <xref ref-type="bibr" rid="B7">Bein et al., 2015</xref>). This is consistent with prior work examining how the reliability of social cues in the input guides attention and learning: when social cues are incongruent, learners might find the cues less reliable and divert their attention to tracking multiple hypotheses about the sources of the uncertainty. This distributed attention can hinder learning performance. The results in the present study are also consistent with work showing that mismatches in expectations about social associations lead to reduced language comprehension (i.e., seeing an Asian face and hearing native-English accented speech in <xref ref-type="bibr" rid="B50">Rubin, 1992</xref>; cf. increased comprehension when there is alignment between an Asian face and hearing Mandarin-accented speech in <xref ref-type="bibr" rid="B36">McGowan, 2015</xref>). Thus, when multiple social cues align with listeners&#8217; prior expectation, this allows them to devote attentional resources more robustly to the encoding and storage of new word forms, as well.</p>
<p>Of course, explanations that hinge on statistical relations derived from previous experience rely on the existence of a veridical empirical relationship between, for example, &#8220;female&#8221; semantic content (e.g., <italic>flowers</italic>) and expected female speech acoustics. Another possibility is that listener expectations, even when these are only tenuously connected to the empirical facts, can affect perception. For example, Barreda and Predeck (<xref ref-type="bibr" rid="B5">2024</xref>) found that perceived gender affects the apparent size of speakers independently of speech acoustics. In a whispered speech condition, listeners rated the same tokens as produced by taller speakers when they identified the speakers as male, and shorter when they identified them as female. In fact, listeners identified all male speakers as taller than all female speakers, even when vocal-tract length cues were equivalent, and even when the female speakers were, in fact, taller than male speakers in the experiment. In this case, the general expectation that men are taller than women overrides the empirical facts of the matter, which indicate that speakers who produce approximately the same formant frequencies tend to be approximately the same size (<xref ref-type="bibr" rid="B57">Turner et al., 2009</xref>).</p>
</sec>
<sec>
<title>4.2 Generalization of learning</title>
<p>Our experiments also tested the extent to which participants generalized learning to new voices &#8211; a hallmark indicator of successful, productive language learning. Participant learning performance was above chance across all groups. Thus, a second key finding is that learning did generalize to different voices. Yet, performance was reduced in the generalization condition, which is consistent with prior work indicating that cross-talker learning can be weaker than same-talker learning (e.g., <xref ref-type="bibr" rid="B19">Eisner &amp; McQueen, 2005</xref>; <xref ref-type="bibr" rid="B63">Xie et al., 2021</xref>; <xref ref-type="bibr" rid="B64">Xie &amp; Myers, 2017</xref>).</p>
<p>With respect to theories about how attention shapes learning and generalization, we propose that adult word learning in voice-AI contexts depends on (a) prior expectations about voice-semantic content associations, (b) the reliability of social cues, and (c) processing fluency. Violations of expectations that increase informativity under reliable cueing (e.g., social cues in voices that align with prior expectations about distributions between social and semantic information) can boost encoding, leading to stronger learning and generalization, whereas violations that undermine cue reliability (e.g., incongruent voice&#8211;content pairs) can hinder learning.</p>
</sec>
<sec>
<title>4.3 How voice typicality modulates these effects</title>
<p>Another key finding is how voice typicality modulated learning across conditions. We find that these differences in learning performance are reduced if the acoustic gender cues of the text-to-speech voices are flipped (H2). More specifically, overall, the difference in accuracy was actually diminished in the flipped voices condition, contrary to our expectation that presenting non-stereotypical voices would make learning in the mismatched condition harder (H3, H4). This is likely because overall accuracy was lower in the flipped voices condition, so that smaller differences did not have the space to shine, and, possibly, because the alignment-mismatch distance was simply smaller for the gender-flipped voices, such that training with a female item and a male voice was less taxing if the voice sounded less stereotypically male. An additional, noteworthy aspect of our data is that flipped voices are hard to work with: participants find it difficult to generalize to new items when training and testing with a flipped voice. This suggests that voices that defy gendered expectations regarding gross acoustics (average f0 or formant frequencies) and more subtle aspects of gender in voice incur additional processing costs that generally make learning more difficult.</p>
<p>Thus, the degree to which a voice conforms to expectations regarding gross speech acoustics (e.g., average f0) and more subtle variation related to gender (e.g., specific intonation patterns) can influence the learning of novel words. Because of the arbitrariness of most gender-related speech variation, it is important to note that what is typical or atypical in the opinion of listeners is socially constructed and may be largely inconsistent across languages and cultures. As a result, the atypicality of some voices, and the associated processing costs, stem from the listener&#8217;s expectations regarding what male and female speakers &#8220;should&#8221; sound like, rather than from any inherent property held by these so-called &#8220;atypical&#8221; speakers. Thus, exploring how linguistic stereotyping, word learning, and human-computer interaction changes with a range of gender-diverse voices is important to understanding how listener expectations regarding gender shape language learning and processing.</p>
</sec>
<sec>
<title>4.4 The effect of the learning exposure condition on voice ratings</title>
<p>In general, our ratings study yielded mostly unsurprising results: participants used gross gender acoustics to categorize female voices as female and male voices as male. Participants made a clear difference between flipped and unmodified voices. If the gender cues were flipped, participants were more likely to classify female voices as male and give female voices higher masculinity ratings (and the other way around). Participants also found flipped voices less natural, which, overall, accounted for most of the ratings differences across voices.</p>
<p>An unexpected finding from our ratings was the effect that the <italic>alignment</italic> condition in the learning experiment had on participants&#8217; later ratings of the voices. When participants were trained on flipped voices, they provided lower naturalness ratings of the voices.</p>
</sec>
<sec>
<title>4.5 Implications for language learning via voice-AI</title>
<p>There are important implications for the role of gendered associations in language learning using devices. In the current study, the effect of voice gender on learning occurred in a context where participants were told they were using a word learning computer application, not actual human language teachers. Computers and applications do not have bodies or genders &#8211; they are machines. Yet, as demonstrated by Nass et al. (<xref ref-type="bibr" rid="B40">1997</xref>) and Ernst and Herm-Stapelberg (<xref ref-type="bibr" rid="B20">2020</xref>), people still apply human-based social patterns of behavior to computers when they display cues of humanness, such as using voices with apparent genders. There are societal implications of the observation that race and gender biases impact behavior during interactions with speech and language technology. More generally, such biases could reinforce and amplify societal stereotypes. For instance, for gender, having female voice assistant voices as the default could amplify the stereotype that women are the helpers, rather than the ones in charge (<xref ref-type="bibr" rid="B59">West et al., 2019</xref>). This can explicitly and implicitly send users the message that women and girls should communicate in the way voice assistants do &#8211; by being subservient, pleasant, and not questioning.</p>
<p>Moreover, our stereotypicality manipulation is also relevant for understanding speech communication with devices. Our unmodified experiment utilized widely-available text-to-speech voices. The current state of speech technology is skewed toward providing voices to users that reflect a binary. Unlike in the real world, where people experience human talkers who express a range of gender identities, the overwhelming majority of text-to-speech voices and voice-assistant personae are presented to reflect stereotypical expressions of binary gender identity: non-binary, gender non-conforming, and gender-neutral voices are not available in commercially-available systems (<xref ref-type="bibr" rid="B56">Sutton, 2020</xref>; though genderless voices have been developed by various researchers: e.g., Project Q; <xref ref-type="bibr" rid="B2">Assink, 2021</xref>; <xref ref-type="bibr" rid="B11">Carpenter, 2019</xref>; <xref ref-type="bibr" rid="B12">Chaves, 2021</xref>).</p>
<p>Our findings reveal that people extend social expectations from human speakers to synthetic voices; thus, design choices about apparent gender and cue reliability can affect word learning and evaluation in speech and language technologies. Systems that allow gender-diverse and less stereotyped voice options, and that align social cues with instructional content, may improve perceptual fluency without reinforcing harmful stereotypes.</p>
</sec>
</sec>
<sec>
<title>5. Conclusion</title>
<p>Language acquisition is challenging for adult learners. One reason for this is that the linguistic and cultural associations adults have, based on the statistical patterns they experience in their native language, make some words even harder to learn than others. Gendered associations between speaker and word frequency, for instance, create an expectation that women and men have distinct language usage patterns through which incoming experiences are filtered. Under that premise, the current study tested whether adults show differences in word learning performance of an artificial language in which lexical items are presented in either a male or a female voice. Participants in gender-aligning training conditions learned the novel words at higher rates &#8211; for both the same and other voice &#8211; than participants in gender-mismatching conditions. Ratings of the voices across conditions were not different, indicating that the gendered associations have an implicit effect on language learning. These results have significant practical and societal implications for the role of social cognition and biases in second language learning, speech and language technology, and human-computer interaction. Finally, our results suggest that &#8220;atypical&#8221; voices may incur increased processing costs, which can hinder learning by incurring additional processing costs. Though some may interpret this as motivation to avoid the use of such &#8220;atypical&#8221; voices, we suggest the opposite is the case. By increasing listeners&#8217; exposure to a wide range of voices, the range considered atypical may be narrowed, and the potential processing costs associated with the voices of such speakers may be minimized.</p>
</sec>
</body>
<back>
<sec>
<title>Appendix</title>
<table-wrap>
<table>
<tbody>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold>Novel Word (IPA)</bold></td>
<td align="left" valign="top"><bold>Image</bold></td>
<td align="left" valign="top"><bold>Mean Imageability Rating</bold></td>
<td align="left" valign="top"><bold>Mean Gender-association Rating</bold></td>
</tr>
<tr>
<td align="left" valign="top" rowspan="8">Female-associated items</td>
<td align="left" valign="top">f&#593;n&#603;t</td>
<td align="left" valign="top">lipstick</td>
<td align="left" valign="top">6.7</td>
<td align="left" valign="top">1.3</td>
</tr>
<tr>
<td align="left" valign="top">&#603;gun</td>
<td align="left" valign="top">skirt</td>
<td align="left" valign="top">6.5</td>
<td align="left" valign="top">1.4</td>
</tr>
<tr>
<td align="left" valign="top">gitu</td>
<td align="left" valign="top">handbag</td>
<td align="left" valign="top">6.8</td>
<td align="left" valign="top">1.4</td>
</tr>
<tr>
<td align="left" valign="top">ut&#593;f</td>
<td align="left" valign="top">dress</td>
<td align="left" valign="top">6.8</td>
<td align="left" valign="top">1.5</td>
</tr>
<tr>
<td align="left" valign="top">t&#593;fun</td>
<td align="left" valign="top">necklace</td>
<td align="left" valign="top">6.5</td>
<td align="left" valign="top">1.5</td>
</tr>
<tr>
<td align="left" valign="top">&#603;fit</td>
<td align="left" valign="top">flower</td>
<td align="left" valign="top">6.8</td>
<td align="left" valign="top">1.7</td>
</tr>
<tr>
<td align="left" valign="top">tugi</td>
<td align="left" valign="top">bracelet</td>
<td align="left" valign="top">6.5</td>
<td align="left" valign="top">1.7</td>
</tr>
<tr>
<td align="left" valign="top">itun</td>
<td align="left" valign="top">perfume</td>
<td align="left" valign="top">5.4</td>
<td align="left" valign="top">1.8</td>
</tr>
<tr>
<td align="left" valign="top" rowspan="8">Male-associated items</td>
<td align="left" valign="top">fun&#593;</td>
<td align="left" valign="top">motorbike</td>
<td align="left" valign="top">6.8</td>
<td align="left" valign="top">6.1</td>
</tr>
<tr>
<td align="left" valign="top">f&#603;ti</td>
<td align="left" valign="top">screwdriver</td>
<td align="left" valign="top">6.9</td>
<td align="left" valign="top">6.1</td>
</tr>
<tr>
<td align="left" valign="top">g&#593;fun</td>
<td align="left" valign="top">suit</td>
<td align="left" valign="top">6.5</td>
<td align="left" valign="top">6.1</td>
</tr>
<tr>
<td align="left" valign="top">&#603;tug</td>
<td align="left" valign="top">football</td>
<td align="left" valign="top">6.8</td>
<td align="left" valign="top">6.2</td>
</tr>
<tr>
<td align="left" valign="top">nig&#603;f</td>
<td align="left" valign="top">tie</td>
<td align="left" valign="top">6.6</td>
<td align="left" valign="top">6.2</td>
</tr>
<tr>
<td align="left" valign="top">n&#593;fit</td>
<td align="left" valign="top">steak</td>
<td align="left" valign="top">6.6</td>
<td align="left" valign="top">5.9</td>
</tr>
<tr>
<td align="left" valign="top">iguf</td>
<td align="left" valign="top">hammer</td>
<td align="left" valign="top">6.6</td>
<td align="left" valign="top">6</td>
</tr>
<tr>
<td align="left" valign="top">nutig</td>
<td align="left" valign="top">drill</td>
<td align="left" valign="top">6.5</td>
<td align="left" valign="top">6</td>
</tr>
</tbody>
</table>
</table-wrap>
</sec>
<sec>
<title>Data accessibility statement</title>
<p>All data and code for this article are available at: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://doi.org/10.5281/zenodo.18367222">https://doi.org/10.5281/zenodo.18367222</ext-link>.</p>
</sec>
<sec>
<title>Ethics and consent</title>
<p>Research was performed in accordance with the UCD IRB (IRB number 1328085).</p>
</sec>
<sec>
<title>Competing interests</title>
<p>The authors have no competing interests to declare.</p>
</sec>
<sec>
<title>Authors&#8217; contributions</title>
<p>GZ conceptualized the study. GZ, PR, and SB designed the experiment, and wrote and edited the paper. PR performed the statistical analysis. All authors contributed to the article and approved of the final version.</p>
</sec>
<sec>
<title>ORCiD IDs</title>
<p><bold>Georgia Zellou</bold>: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://orcid.org/0000-0001-9167-0744">https://orcid.org/0000-0001-9167-0744</ext-link></p>
<p><bold>P&#233;ter R&#225;cz</bold>: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://orcid.org/0000-0001-7896-4801">https://orcid.org/0000-0001-7896-4801</ext-link></p>
<p><bold>Santiago Barreda</bold>: <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://orcid.org/0000-0002-1552-083X">https://orcid.org/0000-0002-1552-083X</ext-link></p>
</sec>
<ref-list>
<ref id="B1"><mixed-citation publication-type="journal"><string-name><surname>Ammari</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Kaye</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Tsai</surname>, <given-names>J. Y.</given-names></string-name>, &amp; <string-name><surname>Bentley</surname>, <given-names>F.</given-names></string-name> (<year>2019</year>). <article-title>Music, search, and IoT: How people (really) use voice assistants</article-title>. <source>ACM Transactions on Computer-Human Interaction (TOCHI)</source>, <volume>26</volume>(<issue>3</issue>), <fpage>1</fpage>&#8211;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.1145/3311956</pub-id></mixed-citation></ref>
<ref id="B2"><mixed-citation publication-type="thesis"><string-name><surname>Assink</surname>, <given-names>L. M.</given-names></string-name> (<year>2021</year>). <source>Making the invisible visible: Exploring gender bias in AI voice assistants</source>. [Master&#8217;s thesis]. <publisher-name>University of Twente</publisher-name>.</mixed-citation></ref>
<ref id="B3"><mixed-citation publication-type="journal"><string-name><surname>Barreda</surname>, <given-names>S.</given-names></string-name> (<year>2020</year>). <article-title>Vowel normalization as perceptual constancy</article-title>. <source>Language</source>, <volume>96</volume>(<issue>2</issue>), <fpage>224</fpage>&#8211;<lpage>254</lpage>. <pub-id pub-id-type="doi">10.1353/lan.2020.0018</pub-id></mixed-citation></ref>
<ref id="B4"><mixed-citation publication-type="journal"><string-name><surname>Barreda</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Assmann</surname>, <given-names>P. F.</given-names></string-name> (<year>2021</year>). <article-title>Perception of gender in children&#8217;s voices</article-title>. <source>The Journal of the Acoustical Society of America</source>, <volume>150</volume>(<issue>5</issue>), <fpage>3949</fpage>&#8211;<lpage>3963</lpage>. <pub-id pub-id-type="doi">10.1121/10.0006785</pub-id></mixed-citation></ref>
<ref id="B5"><mixed-citation publication-type="journal"><string-name><surname>Barreda</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Predeck</surname>, <given-names>K.</given-names></string-name> (<year>2024</year>). <article-title>Inaccurate but predictable: Vocal-tract length estimation and gender stereotypes in height perception</article-title>. <source>Journal of Phonetics</source>, <volume>102</volume>, <elocation-id>101290</elocation-id>. <pub-id pub-id-type="doi">10.1016/j.wocn.2023.101290</pub-id></mixed-citation></ref>
<ref id="B6"><mixed-citation publication-type="book"><string-name><surname>Baugh</surname>, <given-names>J.</given-names></string-name> (<year>2016</year>). <chapter-title>Linguistic profiling and discrimination</chapter-title>. In <string-name><given-names>O</given-names> <surname>Garc&#237;a</surname></string-name>, <string-name><given-names>N</given-names> <surname>Flores</surname></string-name> &amp; <string-name><given-names>M</given-names> <surname>Spotti</surname></string-name> (Eds.), <source>The Oxford handbook of language and society</source> (pp. <fpage>349</fpage>&#8211;<lpage>368</lpage>). <publisher-name>Oxford University Press</publisher-name>. <pub-id pub-id-type="doi">10.1093/oxfordhb/9780190212896.013.13</pub-id></mixed-citation></ref>
<ref id="B7"><mixed-citation publication-type="journal"><string-name><surname>Bein</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Livneh</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Reggev</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Gilead</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Goshen-Gottstein</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name><surname>Maril</surname>, <given-names>A.</given-names></string-name> (<year>2015</year>). <article-title>Delineating the effect of semantic congruency on episodic memory: The role of integration and relatedness</article-title>. <source>PLoS One</source>, <volume>10</volume>(<issue>2</issue>), <elocation-id>e0115624</elocation-id>. <pub-id pub-id-type="doi">10.1371/journal.pone.0115624</pub-id></mixed-citation></ref>
<ref id="B8"><mixed-citation publication-type="journal"><string-name><surname>Bovolenta</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Marsden</surname>, <given-names>E.</given-names></string-name> (<year>2021</year>). <article-title>Expectation violation enhances the development of new abstract syntactic representations: Evidence from an artificial language learning study</article-title>. <source>Language Development Research</source>, <volume>1</volume>(<issue>1</issue>), <fpage>193</fpage>&#8211;<lpage>243</lpage>. <pub-id pub-id-type="doi">10.31219/osf.io/zyegf</pub-id></mixed-citation></ref>
<ref id="B9"><mixed-citation publication-type="book"><string-name><surname>Brandstetter</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>R&#225;cz</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Beckner</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Sandoval</surname>, <given-names>E. B.</given-names></string-name>, <string-name><surname>Hay</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Bartneck</surname>, <given-names>C.</given-names></string-name> (<year>2014</year>, <month>September</month>). <chapter-title>A peer pressure experiment: Recreation of the Asch conformity experiment with robots</chapter-title>. In <source>2014 IEEE/RSJ International Conference on Intelligent Robots and Systems</source> (pp. <fpage>1335</fpage>&#8211;<lpage>1340</lpage>). <publisher-name>IEEE</publisher-name>. <pub-id pub-id-type="doi">10.1109/IROS.2014.6942730</pub-id></mixed-citation></ref>
<ref id="B10"><mixed-citation publication-type="journal"><string-name><surname>B&#252;rkner</surname>, <given-names>P. C.</given-names></string-name> (<year>2017</year>). <article-title>brms: An R package for Bayesian multilevel models using Stan</article-title>. <source>Journal of Statistical Software</source>, <volume>80</volume>, <fpage>1</fpage>&#8211;<lpage>28</lpage>. <pub-id pub-id-type="doi">10.18637/jss.v080.i01</pub-id></mixed-citation></ref>
<ref id="B11"><mixed-citation publication-type="journal"><string-name><surname>Carpenter</surname>, <given-names>J.</given-names></string-name> (<year>2019</year>). <article-title>Why project Q is more than the world&#8217;s first nonbinary voice for technology</article-title>. <source>Interactions</source>, <volume>26</volume>(<issue>6</issue>), <fpage>56</fpage>&#8211;<lpage>59</lpage>. <pub-id pub-id-type="doi">10.1145/3358912</pub-id></mixed-citation></ref>
<ref id="B12"><mixed-citation publication-type="thesis"><string-name><surname>Chaves</surname>, <given-names>C.</given-names></string-name> (<year>2021</year>). <source>Voice as identity: Creating a genderless voice assistant</source>. [Doctoral dissertation]. <publisher-name>San Francisco State University</publisher-name>. <pub-id pub-id-type="doi">10.46569/20.500.12680/1g05fh80z</pub-id></mixed-citation></ref>
<ref id="B13"><mixed-citation publication-type="book"><string-name><surname>Cheshire</surname>, <given-names>J.</given-names></string-name> (<year>2004</year>). <chapter-title>Sex and gender in variationist research</chapter-title>. In <string-name><given-names>J. K.</given-names> <surname>Chambers</surname></string-name>, <string-name><given-names>P.</given-names> <surname>Trudgill</surname></string-name>, &amp; <string-name><given-names>N.</given-names> <surname>Schilling-Estes</surname></string-name> (Eds.), <source>The handbook of language variation and change</source> (pp. <fpage>423</fpage>&#8211;<lpage>443</lpage>). <publisher-name>Wiley</publisher-name>. <pub-id pub-id-type="doi">10.1002/9780470756591.ch17</pub-id></mixed-citation></ref>
<ref id="B14"><mixed-citation publication-type="book"><string-name><surname>Coates</surname>, <given-names>J.</given-names></string-name> (<year>2015</year>). <source>Women, men and language: A sociolinguistic account of gender differences in language</source>. <publisher-name>Routledge</publisher-name>. <pub-id pub-id-type="doi">10.4324/9781315835778</pub-id></mixed-citation></ref>
<ref id="B15"><mixed-citation publication-type="webpage"><string-name><surname>De Renesse</surname>, <given-names>R.</given-names></string-name> <year>2017</year>. <article-title>Virtual digital assistants to overtake world population by 2021</article-title>. <source>Ovum</source>, <month>May</month>, <day>17</day>. Retrieved November 26, 2020, from <uri>https://ovum.informa.com/resources/product-content/virtual-digital-assistants-to-overtakeworld-population-by-2021</uri></mixed-citation></ref>
<ref id="B16"><mixed-citation publication-type="book"><string-name><surname>Drager</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Kirtley</surname>, <given-names>M. J.</given-names></string-name> (<year>2016</year>). <chapter-title>Awareness, salience, and stereotypes in exemplar-based models of speech production and perception</chapter-title>. In <string-name><given-names>K.</given-names> <surname>Campbell-Kibler</surname></string-name> &amp; <string-name><given-names>A.</given-names> <surname>Babel</surname></string-name> (Eds.), <source>Awareness and control in sociolinguistic research</source> (pp. <fpage>1</fpage>&#8211;<lpage>24</lpage>). <publisher-name>Cambridge University Press</publisher-name>. <pub-id pub-id-type="doi">10.1017/CBO9781139680448.003</pub-id></mixed-citation></ref>
<ref id="B17"><mixed-citation publication-type="journal"><string-name><surname>Eckert</surname>, <given-names>P.</given-names></string-name> (<year>1989</year>). <article-title>The whole woman: Sex and gender differences in variation</article-title>. <source>Language Variation and Change</source>, <volume>1</volume>(<issue>3</issue>), <fpage>245</fpage>&#8211;<lpage>267</lpage>. <pub-id pub-id-type="doi">10.1017/S095439450000017X</pub-id></mixed-citation></ref>
<ref id="B18"><mixed-citation publication-type="book"><string-name><surname>Eckert</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>McConnell-Ginet</surname>, <given-names>S.</given-names></string-name> (<year>2013</year>). <source>Language and gender</source>. <publisher-name>Cambridge University Press</publisher-name>. <pub-id pub-id-type="doi">10.1017/CBO9781139245883</pub-id></mixed-citation></ref>
<ref id="B19"><mixed-citation publication-type="journal"><string-name><surname>Eisner</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>McQueen</surname>, <given-names>J. M.</given-names></string-name> (<year>2005</year>). <article-title>The specificity of perceptual learning in speech processing</article-title>. <source>Perception &amp; Psychophysics</source>, <volume>67</volume>(<issue>2</issue>), <fpage>224</fpage>&#8211;<lpage>238</lpage>. <pub-id pub-id-type="doi">10.3758/BF03206487</pub-id></mixed-citation></ref>
<ref id="B20"><mixed-citation publication-type="webpage"><string-name><surname>Ernst</surname>, <given-names>C. P. H.</given-names></string-name>, &amp; <string-name><surname>Herm-Stapelberg</surname>, <given-names>N.</given-names></string-name> (<year>2020</year>). <article-title>The impact of gender stereotyping on the perceived likability of virtual assistants</article-title>. In <source>AMCIS Proceedings 4</source>. <uri>https://aisel.aisnet.org/amcis2020/cognitive_in_is/cognitive_in_is/4</uri></mixed-citation></ref>
<ref id="B21"><mixed-citation publication-type="journal"><string-name><surname>Feher</surname>, <given-names>O.</given-names></string-name>, <string-name><surname>Kirby</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Smith</surname>, <given-names>K.</given-names></string-name> (<year>2014</year>). <article-title>Social influences on the regularization of unpredictable linguistic variation</article-title>. In <source>Proceedings of the Annual Meeting of the Cognitive Science Society</source> (Vol. <volume>36</volume>, No. <issue>36</issue>).</mixed-citation></ref>
<ref id="B22"><mixed-citation publication-type="journal"><string-name><surname>Gambino</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Liu</surname>, <given-names>B.</given-names></string-name> (<year>2022</year>). <article-title>Considering the context to build theory in HCI, HRI, and HMC: Explicating differences in processes of communication and socialization with social technologies</article-title>. <source>Human-Machine Communication</source>, <volume>4</volume>, <fpage>111</fpage>&#8211;<lpage>130</lpage>. <pub-id pub-id-type="doi">10.30658/hmc.4.6</pub-id></mixed-citation></ref>
<ref id="B23"><mixed-citation publication-type="journal"><string-name><surname>Girard</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Floccia</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Goslin</surname>, <given-names>J.</given-names></string-name> (<year>2008</year>). <article-title>Perception and awareness of accents in young children</article-title>. <source>British Journal of Developmental Psychology</source>, <volume>26</volume>(<issue>3</issue>), <fpage>409</fpage>&#8211;<lpage>433</lpage>. <pub-id pub-id-type="doi">10.1348/026151007X251712</pub-id></mixed-citation></ref>
<ref id="B24"><mixed-citation publication-type="webpage"><string-name><surname>Godwin-Jones</surname>, <given-names>R.</given-names></string-name> (<year>2011</year>). <article-title>Mobile apps for language learning</article-title>. <source>Language Learning &amp; Technology</source>, <volume>15</volume>(<issue>2</issue>), <fpage>2</fpage>&#8211;<lpage>11</lpage>. <uri>http://llt.msu.edu/issues/june2011/emerging.pdf</uri></mixed-citation></ref>
<ref id="B25"><mixed-citation publication-type="webpage"><string-name><surname>Godwin-Jones</surname>, <given-names>R.</given-names></string-name> (<year>2017</year>). <article-title>Smartphones and language learning</article-title>. <source>Language Learning &amp; Technology</source>, <volume>21</volume>(<issue>2</issue>), <fpage>3</fpage>&#8211;<lpage>17</lpage>. <uri>http://llt.msu.edu/issues/june2017/emerging.pdf</uri></mixed-citation></ref>
<ref id="B26"><mixed-citation publication-type="journal"><string-name><surname>Guaqueta</surname>, <given-names>C. A.</given-names></string-name>, &amp; <string-name><surname>Castro-Garces</surname>, <given-names>A. Y.</given-names></string-name> (<year>2018</year>). <article-title>The use of language learning apps as a didactic tool for EFL vocabulary building</article-title>. <source>English Language Teaching</source>, <volume>11</volume>(<issue>2</issue>), <fpage>61</fpage>&#8211;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.5539/elt.v11n2p61</pub-id></mixed-citation></ref>
<ref id="B27"><mixed-citation publication-type="journal"><string-name><surname>Hay</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Nolan</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Drager</surname>, <given-names>K.</given-names></string-name> (<year>2006</year>). <article-title>From fush to feesh: Exemplar priming in speech perception</article-title>. <source>Linguistic Review</source>, <volume>23</volume>(<issue>3</issue>), <fpage>351</fpage>&#8211;<lpage>379</lpage>. <pub-id pub-id-type="doi">10.1515/TLR.2006.014</pub-id></mixed-citation></ref>
<ref id="B28"><mixed-citation publication-type="book"><string-name><surname>Johnson</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Sjerps</surname>, <given-names>M. J.</given-names></string-name> (<year>2021</year>). <chapter-title>Speaker normalization in speech perception</chapter-title>. In <string-name><given-names>J. S.</given-names> <surname>Pardo</surname></string-name>, <string-name><given-names>L. C.</given-names> <surname>Nygaard</surname></string-name>, <string-name><given-names>R. E.</given-names> <surname>Remez</surname></string-name>, &amp; <string-name><given-names>D. B.</given-names> <surname>Pisoni</surname></string-name> (Eds.), <source>The handbook of speech perception</source> (pp. <fpage>145</fpage>&#8211;<lpage>176</lpage>). <publisher-name>Wiley</publisher-name>. <pub-id pub-id-type="doi">10.1002/9781119184096.ch6</pub-id></mixed-citation></ref>
<ref id="B29"><mixed-citation publication-type="journal"><string-name><surname>Kaushanskaya</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Marian</surname>, <given-names>V.</given-names></string-name> (<year>2009</year>). <article-title>Bilingualism reduces native-language interference during novel-word learning</article-title>. <source>Journal of Experimental Psychology: Learning, Memory, and Cognition</source>, <volume>35</volume>(<issue>3</issue>), <fpage>829</fpage>&#8211;<lpage>835</lpage>. <pub-id pub-id-type="doi">10.1037/a0015275</pub-id></mixed-citation></ref>
<ref id="B30"><mixed-citation publication-type="journal"><string-name><surname>Kim</surname>, <given-names>J.</given-names></string-name> (<year>2016</year>). <article-title>Perceptual associations between words and speaker age</article-title>. <source>Laboratory Phonology</source>, <volume>7</volume>(<issue>1</issue>). <pub-id pub-id-type="doi">10.5334/labphon.33</pub-id></mixed-citation></ref>
<ref id="B31"><mixed-citation publication-type="journal"><string-name><surname>Kuchenbrandt</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>H&#228;ring</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Eichberg</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Eyssel</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Andr&#233;</surname>, <given-names>E.</given-names></string-name> (<year>2014</year>). <article-title>Keep an eye on the task! How gender typicality of tasks influence human&#8211;robot interactions</article-title>. <source>International Journal of Social Robotics</source>, <volume>6</volume>, <fpage>417</fpage>&#8211;<lpage>427</lpage>. <pub-id pub-id-type="doi">10.1007/s12369-014-0244-0</pub-id></mixed-citation></ref>
<ref id="B32"><mixed-citation publication-type="book"><string-name><surname>Labov</surname>, <given-names>W.</given-names></string-name> (<year>2006</year>). <source>The social stratification of English in New York City</source>. <publisher-name>Cambridge University Press</publisher-name>. <pub-id pub-id-type="doi">10.1017/CBO9780511618208</pub-id></mixed-citation></ref>
<ref id="B33"><mixed-citation publication-type="journal"><string-name><surname>Lee</surname>, <given-names>C.</given-names></string-name>, &amp; <string-name><surname>Lew-Williams</surname>, <given-names>C.</given-names></string-name> (<year>2022</year>). <article-title>Speech and social cues combine at discourse boundaries to promote word learning</article-title>. <source>Cognitive Development</source>, <volume>64</volume>, <elocation-id>101254</elocation-id>. <pub-id pub-id-type="doi">10.1016/j.cogdev.2022.101254</pub-id></mixed-citation></ref>
<ref id="B34"><mixed-citation publication-type="journal"><string-name><surname>L&#252;decke</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>L&#252;decke</surname>, <given-names>M. D.</given-names></string-name> (<year>2015</year>). <article-title>Package &#8216;sjplot&#8217;</article-title>. <source>R package Version</source>, <volume>1</volume>(<issue>9</issue>), <fpage>1</fpage>&#8211;<lpage>106</lpage>.</mixed-citation></ref>
<ref id="B35"><mixed-citation publication-type="journal"><string-name><surname>MacDonald</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Yurovsky</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Frank</surname>, <given-names>M. C.</given-names></string-name> (<year>2017</year>). <article-title>Social cues modulate the representations underlying cross-situational learning</article-title>. <source>Cognitive Psychology</source>, <volume>94</volume>, <fpage>67</fpage>&#8211;<lpage>84</lpage>. <pub-id pub-id-type="doi">10.1016/j.cogpsych.2017.02.003</pub-id></mixed-citation></ref>
<ref id="B36"><mixed-citation publication-type="journal"><string-name><surname>McGowan</surname>, <given-names>K. B.</given-names></string-name> (<year>2015</year>). <article-title>Social expectation improves speech perception in noise</article-title>. <source>Language and Speech</source>, <volume>58</volume>(<issue>4</issue>), <fpage>502</fpage>&#8211;<lpage>521</lpage>. <pub-id pub-id-type="doi">10.1177/0023830914565191</pub-id></mixed-citation></ref>
<ref id="B37"><mixed-citation publication-type="journal"><string-name><surname>Moore</surname>, <given-names>R. K.</given-names></string-name> (<year>2012</year>). <article-title>A Bayesian explanation of the &#8216;Uncanny Valley&#8217; effect and related psychological phenomena</article-title>. <source>Scientific Reports</source>, <volume>2</volume>(<issue>1</issue>), <elocation-id>864</elocation-id>. <pub-id pub-id-type="doi">10.1038/srep00864</pub-id></mixed-citation></ref>
<ref id="B38"><mixed-citation publication-type="journal"><string-name><surname>Most</surname>, <given-names>S. B.</given-names></string-name>, <string-name><surname>Sorber</surname>, <given-names>A. V.</given-names></string-name>, &amp; <string-name><surname>Cunningham</surname>, <given-names>J. G.</given-names></string-name> (<year>2007</year>). <article-title>Auditory Stroop reveals implicit gender associations in adults and children</article-title>. <source>Journal of Experimental Social Psychology</source>, <volume>43</volume>(<issue>2</issue>), <fpage>287</fpage>&#8211;<lpage>294</lpage>. <pub-id pub-id-type="doi">10.1016/j.jesp.2006.02.002</pub-id></mixed-citation></ref>
<ref id="B39"><mixed-citation publication-type="journal"><string-name><surname>Munson</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Lackas</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Koeppe</surname>, <given-names>K.</given-names></string-name> (<year>2022</year>). <article-title>Individual differences in the development of gendered speech in preschool children: Evidence from a longitudinal study</article-title>. <source>Journal of Speech, Language, and Hearing Research</source>, <volume>65</volume>(<issue>4</issue>), <fpage>1311</fpage>&#8211;<lpage>1330</lpage>. <pub-id pub-id-type="doi">10.1044/2021_JSLHR-21-00465</pub-id></mixed-citation></ref>
<ref id="B40"><mixed-citation publication-type="journal"><string-name><surname>Nass</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Moon</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name><surname>Green</surname>, <given-names>N.</given-names></string-name> (<year>1997</year>). <article-title>Are machines gender neutral? Gender-stereotypic responses to computers with voices</article-title>. <source>Journal of Applied Social Psychology</source>, <volume>27</volume>(<issue>10</issue>), <fpage>864</fpage>&#8211;<lpage>876</lpage>. <pub-id pub-id-type="doi">10.1111/j.1559-1816.1997.tb00275.x</pub-id></mixed-citation></ref>
<ref id="B41"><mixed-citation publication-type="journal"><string-name><surname>Needle</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Pierrehumbert</surname>, <given-names>J.</given-names></string-name> (<year>2018</year>). <article-title>Gendered associations of English morphology</article-title>. <source>Laboratory Phonology</source>, <volume>9</volume>(<issue>1</issue>). <pub-id pub-id-type="doi">10.5334/labphon.134</pub-id></mixed-citation></ref>
<ref id="B42"><mixed-citation publication-type="journal"><string-name><surname>Niedzielski</surname>, <given-names>N.</given-names></string-name> (<year>1999</year>). <article-title>The effect of social information on the perception of sociolinguistic variables</article-title>. <source>Journal of Language and Social Psychology</source>, <volume>18</volume>(<issue>1</issue>), <fpage>62</fpage>&#8211;<lpage>85</lpage>. <pub-id pub-id-type="doi">10.1177/0261927X99018001005</pub-id></mixed-citation></ref>
<ref id="B43"><mixed-citation publication-type="journal"><string-name><surname>Pierrehumbert</surname>, <given-names>J. B.</given-names></string-name> (<year>2016</year>). <article-title>Phonological representation: Beyond abstract versus episodic</article-title>. <source>Annual Review of Linguistics</source>, <volume>2</volume>(<issue>1</issue>), <fpage>33</fpage>&#8211;<lpage>52</lpage>. <pub-id pub-id-type="doi">10.1146/annurev-linguistics-030514-125050</pub-id></mixed-citation></ref>
<ref id="B44"><mixed-citation publication-type="journal"><string-name><surname>Poon</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Ng</surname>, <given-names>M. L.</given-names></string-name> (<year>2015</year>). <article-title>The role of fundamental frequency and formants in voice gender identification</article-title>. <source>Speech, Language and Hearing</source>, <volume>18</volume>(<issue>3</issue>), <fpage>161</fpage>&#8211;<lpage>165</lpage>. <pub-id pub-id-type="doi">10.1179/2050572814Y.0000000058</pub-id></mixed-citation></ref>
<ref id="B45"><mixed-citation publication-type="journal"><string-name><surname>Prystawski</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Grant</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Nematzadeh</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Lee</surname>, <given-names>S. W.</given-names></string-name>, <string-name><surname>Stevenson</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Xu</surname>, <given-names>Y.</given-names></string-name> (<year>2020</year>). <article-title>Tracing the emergence of gendered language in childhood</article-title>. In <source>Proceedings of the Annual Meeting of the Cognitive Science Society</source> (Vol. <volume>42</volume>).</mixed-citation></ref>
<ref id="B46"><mixed-citation publication-type="webpage"><collab>R Core Team</collab>. (<year>2024</year>). <chapter-title>R: A language and environment for statistical computing</chapter-title>. <publisher-name>R Foundation for Statistical Computing</publisher-name>, <publisher-loc>Vienna, Austria</publisher-loc>. <uri>https://www.R-project.org/</uri>.</mixed-citation></ref>
<ref id="B47"><mixed-citation publication-type="journal"><string-name><surname>R&#225;cz</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Hay</surname>, <given-names>J. B.</given-names></string-name>, &amp; <string-name><surname>Pierrehumbert</surname>, <given-names>J. B.</given-names></string-name> (<year>2017</year>). <article-title>Social salience discriminates learnability of contextual cues in an artificial language</article-title>. <source>Frontiers in Psychology</source>, <volume>8</volume>, <elocation-id>207450</elocation-id>. <pub-id pub-id-type="doi">10.3389/fpsyg.2017.00051</pub-id></mixed-citation></ref>
<ref id="B48"><mixed-citation publication-type="journal"><string-name><surname>Roberts</surname>, <given-names>G.</given-names></string-name> (<year>2017</year>). <article-title>The linguist&#8217;s Drosophila: Experiments in language change</article-title>. <source>Linguistics Vanguard</source>, <volume>3</volume>(<issue>1</issue>), <elocation-id>20160086</elocation-id>. <pub-id pub-id-type="doi">10.1515/lingvan-2016-0086</pub-id></mixed-citation></ref>
<ref id="B49"><mixed-citation publication-type="journal"><string-name><surname>Robertson</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Murachver</surname>, <given-names>T.</given-names></string-name> (<year>2003</year>). <article-title>Children&#8217;s speech accommodation to gendered language styles</article-title>. <source>Journal of Language and Social Psychology</source>, <volume>22</volume>(<issue>3</issue>), <fpage>321</fpage>&#8211;<lpage>333</lpage>. <pub-id pub-id-type="doi">10.1177/0261927X03255304</pub-id></mixed-citation></ref>
<ref id="B50"><mixed-citation publication-type="journal"><string-name><surname>Rubin</surname>, <given-names>D. L.</given-names></string-name> (<year>1992</year>). <article-title>Nonlanguage factors affecting undergraduates&#8217; judgments of nonnative English-speaking teaching assistants</article-title>. <source>Research in Higher Education</source>, <volume>33</volume>, <fpage>511</fpage>&#8211;<lpage>531</lpage>. <pub-id pub-id-type="doi">10.1007/BF00973770</pub-id></mixed-citation></ref>
<ref id="B51"><mixed-citation publication-type="journal"><string-name><surname>Scott</surname>, <given-names>G. G.</given-names></string-name>, <string-name><surname>Keitel</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Becirspahic</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Yao</surname>, <given-names>B.</given-names></string-name>, &amp; <string-name><surname>Sereno</surname>, <given-names>S. C.</given-names></string-name> (<year>2019</year>). <article-title>The Glasgow Norms: Ratings of 5,500 words on nine scales</article-title>. <source>Behavior Research Methods</source>, <volume>51</volume>(<issue>3</issue>), <fpage>1258</fpage>&#8211;<lpage>1270</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-018-1099-3</pub-id></mixed-citation></ref>
<ref id="B52"><mixed-citation publication-type="journal"><string-name><surname>Smith</surname>, <given-names>J.</given-names></string-name>, <string-name><surname>Durham</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Fortune</surname>, <given-names>L.</given-names></string-name> (<year>2009</year>). <article-title>Universal and dialect-specific pathways of acquisition: Caregivers, children, and t/d deletion</article-title>. <source>Language Variation and Change</source>, <volume>21</volume>(<issue>1</issue>), <fpage>69</fpage>&#8211;<lpage>95</lpage>. <pub-id pub-id-type="doi">10.1017/S0954394509000039</pub-id></mixed-citation></ref>
<ref id="B53"><mixed-citation publication-type="webpage"><collab>Stan Development Team</collab>. (<year>2024</year>). <article-title>RStan: the R interface to Stan. R package version</article-title>. (<uri>https://mc-stan.org/rstan/</uri>).</mixed-citation></ref>
<ref id="B54"><mixed-citation publication-type="journal"><string-name><surname>Stennes</surname>, <given-names>L. M.</given-names></string-name>, <string-name><surname>Burch</surname>, <given-names>M. M.</given-names></string-name>, <string-name><surname>Sen</surname>, <given-names>M. G.</given-names></string-name>, &amp; <string-name><surname>Bauer</surname>, <given-names>P. J.</given-names></string-name> (<year>2005</year>). <article-title>A longitudinal study of gendered vocabulary and communicative action in young children</article-title>. <source>Developmental Psychology</source>, <volume>41</volume>(<issue>1</issue>), <fpage>75</fpage>&#8211;<lpage>88</lpage>. <pub-id pub-id-type="doi">10.1037/0012-1649.41.1.75</pub-id></mixed-citation></ref>
<ref id="B55"><mixed-citation publication-type="journal"><string-name><surname>Strand</surname>, <given-names>E. A.</given-names></string-name> (<year>1999</year>). <article-title>Uncovering the role of gender stereotypes in speech perception</article-title>. <source>Journal of Language and Social Psychology</source>, <volume>18</volume>(<issue>1</issue>), <fpage>86</fpage>&#8211;<lpage>100</lpage>. <pub-id pub-id-type="doi">10.1177/0261927X99018001006</pub-id></mixed-citation></ref>
<ref id="B56"><mixed-citation publication-type="journal"><string-name><surname>Sutton</surname>, <given-names>S. J.</given-names></string-name> (<year>2020</year>, <month>July</month>). <article-title>Gender ambiguous, not genderless: Designing gender in voice user interfaces (VUIs) with sensitivity</article-title>. In <source>Proceedings of the 2nd conference on conversational user interfaces</source> (pp. <fpage>1</fpage>&#8211;<lpage>8</lpage>). <pub-id pub-id-type="doi">10.1145/3405755.3406123</pub-id></mixed-citation></ref>
<ref id="B57"><mixed-citation publication-type="journal"><string-name><surname>Turner</surname>, <given-names>R. E.</given-names></string-name>, <string-name><surname>Walters</surname>, <given-names>T. C.</given-names></string-name>, <string-name><surname>Monaghan</surname>, <given-names>J. J.</given-names></string-name>, &amp; <string-name><surname>Patterson</surname>, <given-names>R. D.</given-names></string-name> (<year>2009</year>). <article-title>A statistical, formant-pattern model for segregating vowel type and vocal-tract length in developmental formant data</article-title>. <source>The Journal of the Acoustical Society of America</source>, <volume>125</volume>(<issue>4</issue>), <fpage>2374</fpage>&#8211;<lpage>2386</lpage>. <pub-id pub-id-type="doi">10.1121/1.3079772</pub-id></mixed-citation></ref>
<ref id="B58"><mixed-citation publication-type="journal"><string-name><surname>Waytz</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Cacioppo</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Epley</surname>, <given-names>N.</given-names></string-name> (<year>2010</year>). <article-title>Who sees human? The stability and importance of individual differences in anthropomorphism</article-title>. <source>Perspectives on Psychological Science</source>, <volume>5</volume>(<issue>3</issue>), <fpage>219</fpage>&#8211;<lpage>232</lpage>. <pub-id pub-id-type="doi">10.1177/1745691610369336</pub-id></mixed-citation></ref>
<ref id="B59"><mixed-citation publication-type="webpage"><string-name><surname>West</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Kraut</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Chew</surname>, <given-names>H. E.</given-names></string-name> (<year>2019</year>). <chapter-title>I&#8217;d blush if I could: closing gender divides in digital skills through education</chapter-title>. <uri>https://unesdoc.unesco.org/ark:/48223/pf0000367416</uri>, (<publisher-name>EQUALS/UNESCO</publisher-name>).</mixed-citation></ref>
<ref id="B60"><mixed-citation publication-type="journal"><string-name><surname>Westerman</surname>, <given-names>D. L.</given-names></string-name>, <string-name><surname>Lloyd</surname>, <given-names>M. E.</given-names></string-name>, &amp; <string-name><surname>Miller</surname>, <given-names>J. K.</given-names></string-name> (<year>2002</year>). <article-title>The attribution of perceptual fluency in recognition memory: The role of expectation</article-title>. <source>Journal of Memory and Language</source>, <volume>47</volume>(<issue>4</issue>), <fpage>607</fpage>&#8211;<lpage>617</lpage>. <pub-id pub-id-type="doi">10.1016/S0749-596X(02)00022-0</pub-id></mixed-citation></ref>
<ref id="B61"><mixed-citation publication-type="journal"><string-name><surname>Wickham</surname>, <given-names>M. H.</given-names></string-name> (<year>2016</year>). <article-title>Package &#8216;ggplot2&#8217;. Create elegant data visualisations using the grammar of graphics</article-title>. Version 2.1, <fpage>1</fpage>&#8211;<lpage>189</lpage>.</mixed-citation></ref>
<ref id="B62"><mixed-citation publication-type="journal"><string-name><surname>Wong</surname>, <given-names>P. C.</given-names></string-name>, &amp; <string-name><surname>Perrachione</surname>, <given-names>T. K.</given-names></string-name> (<year>2007</year>). <article-title>Learning pitch patterns in lexical identification by native English-speaking adults</article-title>. <source>Applied Psycholinguistics</source>, <volume>28</volume>(<issue>4</issue>), <fpage>565</fpage>&#8211;<lpage>585</lpage>. <pub-id pub-id-type="doi">10.1017/S0142716407070312</pub-id></mixed-citation></ref>
<ref id="B63"><mixed-citation publication-type="journal"><string-name><surname>Xie</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Liu</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name><surname>Jaeger</surname>, <given-names>T. F.</given-names></string-name> (<year>2021</year>). <article-title>Cross-talker generalization in the perception of nonnative speech: A large-scale replication</article-title>. <source>Journal of Experimental Psychology: General</source>, <volume>150</volume>(<issue>11</issue>), <elocation-id>e22</elocation-id>. <pub-id pub-id-type="doi">10.1037/xge0001039</pub-id></mixed-citation></ref>
<ref id="B64"><mixed-citation publication-type="journal"><string-name><surname>Xie</surname>, <given-names>X.</given-names></string-name>, &amp; <string-name><surname>Myers</surname>, <given-names>E. B.</given-names></string-name> (<year>2017</year>). <article-title>Learning a talker or learning an accent: Acoustic similarity constrains generalization of foreign accent adaptation to new talkers</article-title>. <source>Journal of Memory and Language</source>, <volume>97</volume>, <fpage>30</fpage>&#8211;<lpage>46</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2017.07.005</pub-id></mixed-citation></ref>
<ref id="B65"><mixed-citation publication-type="journal"><string-name><surname>Zellou</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Cohn</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Ferenc Segedin</surname>, <given-names>B.</given-names></string-name> (<year>2021</year>). <article-title>Age-and gender-related differences in speech alignment toward humans and voice-AI</article-title>. <source>Frontiers in Communication</source>, <volume>5</volume>, <elocation-id>600361</elocation-id>. <pub-id pub-id-type="doi">10.3389/fcomm.2020.600361</pub-id></mixed-citation></ref>
<ref id="B66"><mixed-citation publication-type="journal"><string-name><surname>Zellou</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Cohn</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Kline</surname>, <given-names>T.</given-names></string-name> (<year>2021</year>). <article-title>The influence of conversational role on phonetic alignment toward voice-AI and human interlocutors</article-title>. <source>Language, Cognition and Neuroscience</source>, <volume>36</volume>(<issue>10</issue>), <fpage>1298</fpage>&#8211;<lpage>1312</lpage>. <pub-id pub-id-type="doi">10.1080/23273798.2021.1931372</pub-id></mixed-citation></ref>
<ref id="B67"><mixed-citation publication-type="journal"><string-name><surname>Zellou</surname>, <given-names>G</given-names></string-name>, <string-name><surname>Cohn</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Pycha</surname>, <given-names>A.</given-names></string-name> (<year>2023</year>). <article-title>Listener beliefs and perceptual learning: Differences between device and human guises</article-title>. <source>Language</source>, <volume>99</volume>(<issue>4</issue>), <fpage>692</fpage>&#8211;<lpage>725</lpage>. <pub-id pub-id-type="doi">10.1353/lan.2023.a914191</pub-id></mixed-citation></ref>
</ref-list>
</back>
</article>