<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.2 20120330//EN" "http://jats.nlm.nih.gov/publishing/1.2/JATS-journalpublishing1.dtd">
<!--<?xml-stylesheet type="text/xsl" href="article.xsl"?>-->
<article article-type="research-article" dtd-version="1.2" xml:lang="en" xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id journal-id-type="issn">2767-0279</journal-id>
<journal-title-group>
<journal-title>Glossa Psycholinguistics</journal-title>
</journal-title-group>
<issn pub-type="epub">2767-0279</issn>
<publisher>
<publisher-name>eScholarship Publishing</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.5070/G601133588</article-id>
<article-categories>
<subj-group>
<subject>Regular article</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Canvassing the whole neighborhood: A large-scale view of neighbor network structure, and how it relates to lexical processing</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author" corresp="yes">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-3078-6855</contrib-id>
<name>
<surname>Carlson</surname>
<given-names>Matthew T.</given-names>
</name>
<email>mtc173@psu.edu</email>
<xref ref-type="aff" rid="aff-1">1</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-6517-3733</contrib-id>
<name>
<surname>DiMercurio</surname>
<given-names>Dominick</given-names>
<suffix>II</suffix>
</name>
<email>dimercurio@umbc.edu</email>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
<contrib contrib-type="author">
<name>
<surname>Gertel</surname>
<given-names>Victoria H.</given-names>
</name>
<email>victoriagertel@gmail.com</email>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0001-7263-1694</contrib-id>
<name>
<surname>Diaz</surname>
<given-names>Michele T.</given-names>
</name>
<email>mtd143@psu.edu</email>
<xref ref-type="aff" rid="aff-3">3</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-2147-661X</contrib-id>
<name>
<surname>Sandberg</surname>
<given-names>Chaleece W.</given-names>
</name>
<email>cws18@psu.edu</email>
<xref ref-type="aff" rid="aff-2">2</xref>
</contrib>
</contrib-group>
<aff id="aff-1"><label>1</label>Department of Spanish, Italian, and Portuguese and Center for Language Science, The Pennsylvania State University</aff>
<aff id="aff-2"><label>2</label>Department of Communication Sciences and Disorders and Center for Language Science, The Pennsylvania State University</aff>
<aff id="aff-3"><label>3</label>Department of Psychology and Center for Language Science, The Pennsylvania State University</aff>
<pub-date publication-format="electronic" date-type="pub" iso-8601-date="2025-06-18">
<day>18</day>
<month>06</month>
<year>2025</year>
</pub-date>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<volume>4</volume>
<issue>1</issue>
<elocation-id>16</elocation-id>
<permissions>
<copyright-statement>Copyright: &#x00A9; 2025 The Author(s)</copyright-statement>
<copyright-year>2025</copyright-year>
<license license-type="open-access" xlink:href="http://creativecommons.org/licenses/by/4.0/">
<license-p>This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International License (CC-BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. See <uri xlink:href="http://creativecommons.org/licenses/by/4.0/">http://creativecommons.org/licenses/by/4.0/</uri>.</license-p>
</license>
</permissions>
<self-uri xlink:href="https://glossapsycholinguistics.journalpub.escholarship.org/articles/10.5070/G601133588/"/>
<abstract>
<p>Lexical processing reflects patterns of phonological and/or orthographic similarity among words. One approach to explaining this is to conceive of the mental lexicon as being structured according to these similarity patterns, and to model that structure as a network, most commonly by connecting each word to its immediate neighbors. Evidence that lexical processing is related to structure beyond the size of words&#8217; immediate neighborhoods suggests that network analyses can capture psycholinguistically relevant structural patterns in the lexicon, but it remains unclear how that structure is represented in the mind and how it relates to the mechanisms used in the most prominent theoretical approaches to explaining lexical processing. To shed light on this issue, we use a latent variable approach to identify the underlying dimensions of phonological and orthographic structure in the lexicon by modeling multiple network-derived properties, and testing those dimensions against word recognition data from two mega-studies. Our results confirm the importance of network measures and show that their effects on behavior are captured by three latent constructs: how densely words are packed in the region surrounding a target word (not just immediate neighbors), the interconnectedness of words residing near a target, and target words&#8217; connectedness to multiple subregions of the network (cf. community structure). We propose that these latent constructs offer crucial guidance for interpreting the theoretical idea of structure in the mental lexicon, inviting new explanations for why words are distributed in this way, and for how structure relates to theoretical accounts of lexical processing.</p>
</abstract>
</article-meta>
</front>
<body>
<sec>
<title>1. Introduction</title>
<p>A cornerstone of lexical processing research for decades has been the idea that sets of words similar in form are activated together, and that the size and contents of these sets influence the course and results of processing. These sets are often referred to as <italic>neighborhoods</italic>, most commonly defined as the set of words that differ from a given target word by the substitution (<xref ref-type="bibr" rid="B34">Coltheart et al., 1977</xref>; <xref ref-type="bibr" rid="B58">Landauer &amp; Streeter, 1973</xref>), addition, or subtraction (<xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>) of one phoneme (phonological neighbors) or letter (orthographic neighbors).</p>
<p>There is now copious evidence that <italic>neighborhood density</italic> (ND, the number of neighbors a target word has) relates to auditory and visual word recognition (e.g., <xref ref-type="bibr" rid="B4">Andrews, 1997</xref>; <xref ref-type="bibr" rid="B114">Vitevitch &amp; Luce, 2016</xref>), and all current theories account for density in some way. It has, moreover, become nearly ubiquitous as a covariate or control variable. ND effects are also well-documented in spoken (<xref ref-type="bibr" rid="B114">Vitevitch &amp; Luce, 2016</xref>), signed (<xref ref-type="bibr" rid="B24">Caselli et al., 2021</xref>), and written (<xref ref-type="bibr" rid="B77">Roux &amp; Bonin, 2009</xref>) word production, in language development (<xref ref-type="bibr" rid="B30">Charles-Luce &amp; Luce, 1990</xref>, <xref ref-type="bibr" rid="B31">1995</xref>; <xref ref-type="bibr" rid="B33">Coady &amp; Aslin, 2003</xref>; <xref ref-type="bibr" rid="B102">Swingley &amp; Aslin, 2002</xref>, <xref ref-type="bibr" rid="B103">2007</xref>), in the errors made by healthy younger and older adults (<xref ref-type="bibr" rid="B47">Harley &amp; Brown, 1998</xref>; <xref ref-type="bibr" rid="B88">Sommers, 1996</xref>; <xref ref-type="bibr" rid="B94">Stemberger, 2004</xref>; <xref ref-type="bibr" rid="B106">Vitevitch, 1997</xref>, <xref ref-type="bibr" rid="B107">2002</xref>; <xref ref-type="bibr" rid="B118">Vitevitch &amp; Sommers, 2003</xref>), and by individuals with developmental disabilities or neurological impairments (<xref ref-type="bibr" rid="B40">Goldrick et al., 2010</xref>; <xref ref-type="bibr" rid="B96">Storkel, 2009</xref>).</p>
<p>Various refinements to ND have been proposed to address theoretical and practical concerns with this measure. These include attempts to incorporate a similarity gradient beyond an edit distance of one<xref ref-type="fn" rid="n1">1</xref> (<xref ref-type="bibr" rid="B52">Iverson et al., 1998</xref>; <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>; <xref ref-type="bibr" rid="B97">Strand, 2014</xref>; <xref ref-type="bibr" rid="B100">Su&#225;rez et al., 2011</xref>; <xref ref-type="bibr" rid="B122">Yarkoni et al., 2008</xref>), or to compute finer gradations of similarity by incorporating subsegmental features (<xref ref-type="bibr" rid="B10">Bailey &amp; Hahn, 2005</xref>; <xref ref-type="bibr" rid="B46">Hahn &amp; Bailey, 2005</xref>; <xref ref-type="bibr" rid="B97">Strand, 2014</xref>). Others have shown that ND effects depend on other properties of neighborhoods, such as the frequencies of the target word&#8217;s neighbors (<xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>; <xref ref-type="bibr" rid="B112">Vitevitch &amp; Luce, 1998</xref>), competition from neighbors at the same segment position (<xref ref-type="bibr" rid="B68">McClelland &amp; Rumelhart, 1981</xref>), and whether neighbors overlap early or later in their forms, or form a cohort (<xref ref-type="bibr" rid="B1">Allopenna et al., 1998</xref>; <xref ref-type="bibr" rid="B37">Fricke et al., 2016</xref>; <xref ref-type="bibr" rid="B65">Marslen-Wilson &amp; Tyler, 1980</xref>). ND may also be conditioned on the syntactic categories of the target and its neighbors (<xref ref-type="bibr" rid="B49">Heller &amp; Goldrick, 2014</xref>; <xref ref-type="bibr" rid="B98">Strand et al., 2014</xref>).</p>
<p>A different approach to examining word-form similarity structure in the lexicon builds on the idea that neighbor-to-neighbor links allow an entire lexicon to be represented as a graph (<xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>). A <italic>graph</italic> is a mathematical structure comprised of a set of <italic>nodes</italic> and a set of <italic>edges</italic> that connect the nodes based on some criteria. Network science studies the application of graph theory to real-world phenomena, including a recent burst of activity related to cognition (see <xref ref-type="bibr" rid="B12">Baronchelli et al., 2013</xref>; <xref ref-type="bibr" rid="B87">Siew et al., 2019</xref>, for recent reviews). In nearly all studies of phonological or orthographic networks, the <italic>nodes</italic> represent words, and <italic>edges</italic> connect words that have an edit distance of 1, i.e., neighbors. We refer to this kind of network as <italic>neighbor networks</italic> (following <xref ref-type="bibr" rid="B18">Brown et al., 2018</xref>).<xref ref-type="fn" rid="n2">2</xref> Networks could be built using other criteria, e.g., incorporating syllable structure (cf. <xref ref-type="bibr" rid="B101">Sun &amp; Poeppel, 2023</xref>, on the role or syllables in lexical organization, though they did not use a network), or phonotactics (e.g., via biphones or a bipartite network linking phonemes to words; <xref ref-type="bibr" rid="B116">Vitevitch et al., 2021</xref>), or placing edges based on human misperceptions or intuitions about phonological associations (<xref ref-type="bibr" rid="B27">Castro &amp; Vitevitch, 2022</xref>). However, useful alternatives to the edit distance = 1 criterion are potentially many and have not yet been systematically explored.</p>
<p>Neighbor networks go beyond addressing some of the weaknesses of ND (e.g., by incorporating gradient similarity by providing paths connecting words that are not immediate neighbors), but, more importantly, they implement the idea of phonological and orthographic structure in the lexicon. Network science offers a wealth of concepts and measures for quantifying the position of nodes in a network, which we will refer to as <italic>nodal properties</italic> (<xref ref-type="bibr" rid="B21">Bullmore &amp; Sporns, 2009</xref>; <xref ref-type="bibr" rid="B89">Sporns, 2011</xref>; <xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>). There is growing evidence in psycholinguistics that a variety of nodal properties capture useful information about the properties of words&#8217; immediate neighborhoods (micro-level structure), as well as complexity in the distribution and relationships among words at larger (meso- and macro-) scales and distances from the target (for a review, see <xref ref-type="bibr" rid="B109">Vitevitch, 2022</xref>).</p>
<p>The idea of word-form based structure in the lexicon has potential for opening new theoretical perspectives on how word-form similarity constrains lexical processing (<xref ref-type="bibr" rid="B25">Castro &amp; Siew, 2020</xref>; <xref ref-type="bibr" rid="B87">Siew et al., 2019</xref>; <xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>, <xref ref-type="bibr" rid="B109">2022</xref>), but the underlying theoretical nature of structure, as it may be measured in neighbor networks, is not yet clear, making it hard to explore that potential. There are three reasons for this. First, as noted above, useful alternatives to the edit distance = 1 criterion are potentially many and have not yet been systematically explored. Second, the nodal properties used to measure aspects of structure have precise mathematical definitions that depend on specific assumptions about how information flows in the network, given the phenomenon being modeled (<xref ref-type="bibr" rid="B16">Borgatti, 2005</xref>). These assumptions have guided the psycholinguistic study of new nodal properties, yielding important insights, but the paths in a lexical neighbor network are not as concrete as they are for other phenomena that may be modeled using networks, for example, person-to-person contact in a model of disease transmission. Finally, while networks suggest paths for activation to spread from word to word (<xref ref-type="bibr" rid="B83">Siew, 2019</xref>; <xref ref-type="bibr" rid="B109">Vitevitch, 2022</xref>), word recognition theories in general lack lateral spreading activation of this kind (e.g. <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>; <xref ref-type="bibr" rid="B65">Marslen-Wilson &amp; Tyler, 1980</xref>; <xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>; <xref ref-type="bibr" rid="B68">McClelland &amp; Rumelhart, 1981</xref>; <xref ref-type="bibr" rid="B73">Norris, 1994</xref>; <xref ref-type="bibr" rid="B74">Norris &amp; McQueen, 2008</xref>; see <xref ref-type="bibr" rid="B120">Weber &amp; Scharenborg, 2012</xref>, for a detailed review), and some models do not include word form representations (i.e., the network nodes) at all (<xref ref-type="bibr" rid="B8">Baayen et al., 2019</xref>; <xref ref-type="bibr" rid="B79">Shafaei-Bajestan et al., 2023</xref>).</p>
<p>In this article, we therefore take the documented effects of various nodal properties as evidence that the networks capture relevant aspects of phonological and orthographic structure in the mental lexicon, but we explore the usefulness of considering those measures as proxies of some set of underlying dimensions that serve to describe that structure. To gain insight into those dimensions, we apply a data-driven latent variable technique to relate a large set of nodal properties, including several that have not been explored previously, to auditory and visual word recognition data from recent mega-studies (<xref ref-type="bibr" rid="B11">Balota et al., 2007</xref>; <xref ref-type="bibr" rid="B105">Tucker et al., 2019</xref>).</p>
<p>In this way, we seek to shed light on what it means to say that the lexicon is structured based on phonological and orthographic word-form similarity. Theoretically, we hope that the insights gained will help narrow the gap between the notion of structure suggested by network models and word recognition theory. Methodologically, we aim to clarify the links between nodal properties and the underlying constructs they measure, contributing to our understanding of how networks can be used and interpreted in psycholinguistics.</p>
<sec>
<title>1.1 The lexicon as a neighbor network</title>
<p>The first studies of neighbor networks focused on properties of the entire network. Vitevitch (<xref ref-type="bibr" rid="B108">2008</xref>; see also <xref ref-type="bibr" rid="B5">Arbesman et al., 2010a</xref>, <xref ref-type="bibr" rid="B6">2010b</xref>) studied a phonological neighbor network comprising 19,340 words from the <italic>Merriam-Webster Pocket Dictionary</italic>. The largest connected component of the network, called the Giant Component (GC), contained 6,508 nodes, and there were many smaller components, including <italic>hermits</italic> (words with no neighbors). The GC itself had a high mean clustering coefficient and assortativity, and low average shortest path length. <italic>Clustering</italic> quantifies the extent to which each word&#8217;s neighbors are also neighbors of each other, and <italic>assortativity</italic> means that words and their neighbors tend to have similar degree (<italic>degree</italic> is the network science term for the number of neighbors a node has). High clustering and low average shortest path length are indicative of small-worldness (<xref ref-type="bibr" rid="B51">Humphries &amp; Gurney, 2008</xref>; <xref ref-type="bibr" rid="B119">Watts &amp; Strogatz, 1998</xref>).</p>
<p>Properties like this, found in many networks, including social, neural, and semantic networks, are thought to favor efficient flow of information (<xref ref-type="bibr" rid="B119">Watts &amp; Strogatz, 1998</xref>), leading to the suggestion that lexical structure evolves to favor efficient search (<xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>). However, many of the topological properties of neighbor networks in several languages, including the degree distribution, the distribution of connected component sizes, mean shortest path length, clustering, and assortativity, are well approximated in random pseudolexicons generated with minimal linguistic constraints &#8211; namely, with the same word length distribution and number of phonetic categories, but with no constraints on CV structure or phonotactics (<xref ref-type="bibr" rid="B18">Brown et al., 2018</xref>; <xref ref-type="bibr" rid="B45">Gruenenfelder &amp; Pisoni, 2009</xref>; but see <xref ref-type="bibr" rid="B93">Stella &amp; Brede, 2015</xref>). Moreover, adding such constraints has little impact on the properties of the resulting pseudolexicon networks. These topological properties may, thus, be an inevitable feature of networks in which the nodes represent strings constructed from a limited set of sub-nodal units, such as words encoded as strings of letters/phonemes. Global network structure, therefore, does not provide unambiguous evidence that lexicons have evolved to favor efficient search, but structure may, nevertheless, help explain the overall efficiency of lexical processing.</p>
<p>The other approach to studying network structure in lexical processing is to use properties of the nodes to quantify each word&#8217;s position in the network in various ways (see <xref ref-type="table" rid="T1">Table 1</xref>, below, for a list of the nodal properties we study, and their definitions). The entire literature on neighborhood density comprises the simplest way to do this, since ND is merely the most local, or micro-level, property of nodes in a network (called <italic>degree</italic>). Research has explored new nodal properties incrementally, based on their expected effects on the flow of information through the network. This includes meso- and macro-level properties (<xref ref-type="bibr" rid="B109">Vitevitch, 2022</xref>), which quantify structure in progressively larger subsets of words surrounding a target (e.g., clustering, community structure), ultimately encompassing the whole network or connected component (e.g., <italic>closeness centrality</italic>, which is the average shortest path length between each target word and all other words in the component).</p>
<table-wrap id="T1">
<caption>
<p><bold>Table 1:</bold> Network Measures and Descriptions.</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"><bold>Measure</bold></td>
<td align="left" valign="top"><bold>Description</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">Degree</td>
<td align="left" valign="top">The number of edges connected to a node, i.e., neighborhood density.</td>
</tr>
<tr>
<td align="left" valign="top">Average Neighbor Degree</td>
<td align="left" valign="top">The average degree of each node&#8217;s neighbors. This is related to degree assortativity (<xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>), which is defined as the correlation between a node&#8217;s degree and the degrees of its neighbors, across all pairs of neighbors in the network.</td>
</tr>
<tr>
<td align="left" valign="top">Average Neighbor Frequency</td>
<td align="left" valign="top">The average log frequency of each node&#8217;s neighbors. This is not, strictly speaking, a property of the network itself, but of certain nodes identified based on their relationship in the network. Neighborhood density is often weighted by neighbor frequencies in the literature (<xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>).</td>
</tr>
<tr>
<td align="left" valign="top">Clustering Coefficient</td>
<td align="left" valign="top">Proportion of pairs of a node&#8217;s neighbors that are also neighbors to each other (undefined for nodes with fewer than two neighbors).</td>
</tr>
<tr>
<td align="left" valign="top">Local Efficiency</td>
<td align="left" valign="top">A measure of the average efficiency of information transfer. Related to clustering coefficient, local efficiency of each node is the average of the inverse shortest path lengths in a subgraph comprised of the node&#8217;s neighbors (excluding the target node).</td>
</tr>
<tr>
<td align="left" valign="top">Coreness</td>
<td align="left" valign="top">The <italic>k</italic><sup>th</sup> core consists of those nodes that remain after recursively pruning all nodes with degree &lt; <italic>k</italic>, but are removed when recursively pruning all nodes with degree = <italic>k</italic>. Since pruning is recursive, coreness may be smaller than degree, but it cannot be greater.<break/>It measures longer-range integration or embeddedness within a network, in the sense that words with high coreness are interconnected with a set of words having equal or greater degree, such that those words survive the pruning procedure for lower values of <italic>k</italic>. Coreness in the phonological network predicts children&#8217;s vocabulary growth (<xref ref-type="bibr" rid="B23">Carlson et al., 2014</xref>).</td>
</tr>
<tr>
<td align="left" valign="top">Community Size</td>
<td align="left" valign="top">The number of nodes that are in a node&#8217;s community. A community is a subgraph selected to optimize modularity of the network (<xref ref-type="bibr" rid="B80">Siew, 2013</xref>) and consists of a set of nodes that are more densely interconnected with each other than with other (nearby) nodes.</td>
</tr>
<tr>
<td align="left" valign="top">Participation Coefficient</td>
<td align="left" valign="top">A measure of how many communities a node&#8217;s neighbors are connected to and how evenly distributed a node&#8217;s neighbors are across those communities. This has not been directly examined in relation to word recognition, but we include it based on recent work on the relevance of community structure more generally (<xref ref-type="bibr" rid="B80">Siew, 2013</xref>).<break/>Nodes with high participation serve as &#8220;connector hubs&#8221;, having neighbors in many communities outside of their own, and nodes with low participation tend to lack neighbors outside of their own community.</td>
</tr>
<tr>
<td align="left" valign="top">Within-module degree z-score</td>
<td align="left" valign="top">The z-score of each node&#8217;s degree, within its community. High values identify nodes that serve as &#8220;provincial hubs&#8221; within their own communities. This is also included based on its relevance to community structure.</td>
</tr>
<tr>
<td align="left" valign="top">Betweenness Centrality</td>
<td align="left" valign="top">The fraction of shortest paths between all possible pairs of nodes in the same connected component that pass through a given node.<break/>It measures the extent to which each node contributes to the efficient traversal of the network.</td>
</tr>
<tr>
<td align="left" valign="top">PageRank Centrality</td>
<td align="left" valign="top">A recursive measure that is the sum of a node&#8217;s neighbors&#8217; PageRank centralities divided by their out-degrees (<xref ref-type="bibr" rid="B17">Brin &amp; Page, 1998</xref>). Usually used for directed graphs, but also defined for undirected graphs, such as those analyzed here (<xref ref-type="bibr" rid="B44">Grolmusz, 2012</xref>).<break/>A measure of node importance, PageRank can be interpreted as the probability that a random exploration of the graph will lead to a given node. In semantic network research, it has been linked with spreading activation and random walk mechanisms for explaining letter fluency responses, and may reflect processes that shaped the prior probabilities of words (<xref ref-type="bibr" rid="B43">Griffiths et al., 2007</xref>). While this has not been studied in phonological and orthographic neighbor networks, we include it here, because spreading activation and random walks have been used to interpret the relationship between these networks and word recognition (<xref ref-type="bibr" rid="B109">Vitevitch, 2022</xref>).</td>
</tr>
<tr>
<td align="left" valign="top">Closeness Centrality</td>
<td align="left" valign="top">The reciprocal of the sum of the shortest distances between a node and all other nodes in the same connected component.<break/>Centrally located nodes are reachable from any point in the component via a minimum number of steps, but peripherally located nodes are more distant from some parts of the component. Reflects nodal importance for any process that can follow a shortest path, even if in parallel (<xref ref-type="bibr" rid="B16">Borgatti, 2005</xref>).</td>
</tr>
<tr>
<td align="left" valign="top">Eccentricity</td>
<td align="left" valign="top">The length of the shortest path between a node and the farthest node from that node, residing in the same connected component. Similar to closeness, nodes with low eccentricity are relatively close to every node in their connected component, computed via shortest paths.</td>
</tr>
</tbody>
</table>
</table-wrap>
<p>The first of these properties to attract attention was clustering, because of its contribution to the efficient transmission of information. Clustering in the phonological network is related to increased auditory (<xref ref-type="bibr" rid="B2">Altieri et al., 2010</xref>; <xref ref-type="bibr" rid="B28">Chan &amp; Vitevitch, 2009</xref>) and visual (<xref ref-type="bibr" rid="B123">Yates, 2013</xref>)<xref ref-type="fn" rid="n3">3</xref> lexical decision and naming response times, and to increased speech errors (<xref ref-type="bibr" rid="B29">Chan &amp; Vitevitch, 2010</xref>). According to Chan and Vitevitch, this occurs because, when clustering is low, activation from the neighbors either flows back to the target or dissipates to the rest of the network, making the target stand out. When clustering is high, neighbors pass activation among themselves, maintaining their higher activation relative to the target. They also successfully simulated the effect in a computational model with this kind of activation flow. Siew (<xref ref-type="bibr" rid="B81">2017</xref>) observed a similar effect on spoken word recognition, using a version of clustering that includes neighbors at a Levenshtein distance of 1 or 2 (2-hop neighborhoods). Using a slightly different way of constructing the network (words were considered neighbors only if both their phonological and orthographic representations were neighbors), Siew and Vitevitch (<xref ref-type="bibr" rid="B83">2019</xref>) found a negative effect of clustering coefficient, again on spoken word recognition latencies.</p>
<p>At a more macro level, faster auditory lexical decision responses (at least for frequent words) have been found for words with high <italic>closeness centrality</italic>, which measures how quickly a target may be reached, on average, from anywhere in the same connected component (<xref ref-type="bibr" rid="B42">Goldstein &amp; Vitevitch, 2017</xref>). Faster responses are also found for keywords in repetition, identification in noise, and auditory lexical decision (<xref ref-type="bibr" rid="B111">Vitevitch &amp; Goldstein, 2014</xref>). Keywords are words that link larger clusters of words together in the network, such that removing them would result in fragmentation of the network, disrupting the flow of information.</p>
<p>At a meso level, Siew (<xref ref-type="bibr" rid="B80">2013</xref>) studied <italic>communities</italic>, which are groups of words that are more tightly interconnected with each other than with other words in the same component. Words in large communities tended to be shorter, more frequent, acquired earlier, and have higher degree than words in smaller communities. Siew suggested that this grouping of words together may be relevant to lexical processing because the flow of information will tend to pool within one community, which may help explain phonotactic probability effects, because words in the same community tend to share similar sound sequences. Finally, at the scale of the whole network, Siew and Vitevitch (<xref ref-type="bibr" rid="B84">2016</xref>) found that spoken words located in <italic>islands</italic> (smaller connected components that are not linked to the GC) were recognized and recalled more quickly and accurately than words located in the GC.</p>
<p>Siew (<xref ref-type="bibr" rid="B82">2018</xref>) brought some of these measures together in a large-scale analysis of visual lexical decision and word naming data for 11,358 words from the English Lexicon Project (<xref ref-type="bibr" rid="B11">Balota et al., 2007</xref>). Siew found that degree was associated with decreased response latencies (contrary to many, though not all, published studies; see the review by <xref ref-type="bibr" rid="B114">Vitevitch &amp; Luce, 2016</xref>) and higher accuracy in both lexical decision and word naming. A high clustering coefficient was associated with longer response latencies and marginally lower accuracy in lexical decision, and high closeness centrality was associated with faster lexical decision responses, and with slower and less accurate naming responses. These three variables together accounted for an increase in R<sup>2</sup> of a little under 1%, over a model including lexical frequency and length in letters, phonemes, and syllables. Despite these apparently small effects, these results provide evidence that network-like structure in the mental lexicon at various scales, regarding phonological or orthographic word forms, influences processing.<xref ref-type="fn" rid="n4">4</xref></p>
<p>The link between these nodal properties and information flow has led to successful simulations of the clustering, giant component membership, and keyword effects reviewed above (<xref ref-type="bibr" rid="B28">Chan &amp; Vitevitch, 2009</xref>; <xref ref-type="bibr" rid="B84">Siew &amp; Vitevitch, 2016</xref>; <xref ref-type="bibr" rid="B111">Vitevitch &amp; Goldstein, 2014</xref>), using computational models in which activation spreads from word to word by traversing the edges between neighbors. These models can vary in how this takes place, e.g., whether some activation remains behind as it spreads (duplication), and whether a decay function is applied (<xref ref-type="bibr" rid="B83">Siew, 2019</xref>; <xref ref-type="bibr" rid="B110">Vitevitch et al., 2011</xref>; <xref ref-type="bibr" rid="B115">Vitevitch &amp; Mullin, 2021</xref>), and other processes that exploit the network&#8217;s edges, such as random or (semi-) directed search, could be substituted for spreading activation in variants of this model (<xref ref-type="bibr" rid="B109">Vitevitch, 2022</xref>). Other kinds of network flow, and other network structures, are currently being explored (<xref ref-type="bibr" rid="B27">Castro &amp; Vitevitch, 2022</xref>; <xref ref-type="bibr" rid="B116">Vitevitch et al., 2021</xref>), and the inclusion of processes operating outside (e.g., governing where search begins), as well as within, networks is also possible (<xref ref-type="bibr" rid="B50">Hills &amp; Kenett, 2022</xref>).</p>
</sec>
<sec>
<title>1.2 Interpreting nodal property effects on word recognition</title>
<p>The studies reviewed above support the conclusion that neighbor networks capture psycholinguistically important structural features of the mental lexicon, but it is not immediately clear how this should be interpreted. As noted above, the relevance of nodal properties depends on specific assumptions about information flow, given a specific kind of network structure (<xref ref-type="bibr" rid="B16">Borgatti, 2005</xref>). However, only a few simulations explicitly incorporate neighbor network structure of the kind used in the vast majority of the network science-based literature on word recognition, with information flowing via word-to-word activation spread over the network&#8217;s edges (<xref ref-type="bibr" rid="B28">Chan &amp; Vitevitch, 2009</xref>; <xref ref-type="bibr" rid="B84">Siew &amp; Vitevitch, 2016</xref>; <xref ref-type="bibr" rid="B111">Vitevitch &amp; Goldstein, 2014</xref>).</p>
<p>In most models, including models with distributed representations, words are activated via some feed-forward or interactive activation process that links phonetic and lexical units, often incrementally (e.g., <xref ref-type="bibr" rid="B61">Luce et al., 2000</xref>; NAM, <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>; TRACE, <xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>; IAM, <xref ref-type="bibr" rid="B68">McClelland &amp; Rumelhart, 1981</xref>; <xref ref-type="bibr" rid="B73">Norris, 1994</xref>). These models can also be viewed through a network science lens, with varying assumptions about networks&#8217; architecture and information flow. The bipartite network of Vitevitch et al. (<xref ref-type="bibr" rid="B116">2021</xref>) is one example, with edges between phonological units and words, rather than between words (which Vitevitch et al. suggest may help account for effects of phonotactic probability). The NAM (<xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>) also links different levels of representation. Word decision units monitor activation of acoustic-phonetic patterns, top-down information about word likelihoods, and the overall level of activity among decision units, with competition arising from the activation of decision units for multiple candidates at once. TRACE (<xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>) uses interactive links between phonological and lexical units (like Vitevitch et al.&#8217;s bipartite network), plus inhibitory links between competitors as a mechanism for resolving competition, where <italic>competitors</italic> are words that are simultaneously active, but need not be neighbors, strictly speaking. Some models, such as the more recent Shortlist-B (<xref ref-type="bibr" rid="B74">Norris &amp; McQueen, 2008</xref>) or LDL-AURIS (<xref ref-type="bibr" rid="B79">Shafaei-Bajestan et al., 2023</xref>), do not directly implement network-like structure, lacking not only direct edges between word-form representations, but also explicit word-form (e.g., phonolexical) representations themselves. In Shortlist-B, words are identified by evaluating their posterior probabilities, given the phonetic representation of the incoming speech, and in LDL-AURIS, semantic representations are computed directly from acoustic or other low-level phonetic representations.</p>
<p>This diversity of network structures and assumptions about information flow indicates a need for care when interpreting results obtained using specific network architectures and assumptions about information flow, such as are found in the psycholinguistic literature using neighbor networks. This leads us to ask whether structure at different scales in this kind of network should be thought of as reflecting variability in the configuration and evolution of sets of candidate words during processing, such that words&#8217; nodal properties in neighbor networks might be better interpreted as proxies for latent dimensions of structure. As an example of this thinking, Yarkoni et al. (<xref ref-type="bibr" rid="B122">2008</xref>) propose that findings of both inhibitory and facilitatory effects of ND in visual word recognition might be reconciled in a two-stage process in which a set of candidates become active quickly, based on the phonetic units being perceived, followed by a later inhibitory effect arising through competition between the target and specific, highly similar words.</p>
<p>Similarly, Yates (<xref ref-type="bibr" rid="B123">2013</xref>) interprets the effect of clustering in the phonological lexicon on visual word recognition through interactive activation between phonetic and lexical units (note that his study crossed modalities, although the interpretation would seem to apply regardless of modality). He argued that feedback to the phonetic units reinforces a target&#8217;s neighbors more strongly when clustering is high, because the neighbors overlap more consistently, with the differences occurring in limited positions &#8211; i.e., clustering is interpreted as a proxy measure of this consistency. They therefore remain more active relative to the target. This can explain the effects of clustering on visual word recognition, because strongly active neighbors are associated with inhibitory effects, and weakly active neighbors with facilitation (<xref ref-type="bibr" rid="B32">Chen &amp; Mirman, 2012</xref>).</p>
<p>Interestingly, Chan and Vitevitch (see also <xref ref-type="bibr" rid="B115">Vitevitch &amp; Mullin, 2021</xref>) failed to reproduce the effects of phonological clustering on auditory word recognition in simulations using computational implementations of models using interactive activation (TRACE, <xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>; jTRACE, <xref ref-type="bibr" rid="B99">Strauss et al., 2007</xref>) and bottom-up activation (Shortlist, <xref ref-type="bibr" rid="B73">Norris, 1994</xref>), but this could be due to peculiarities in the computational implementation of the models (e.g. the phonetic units, size of the lexicon, and lexicon copying procedure used in TRACE), and not necessarily to the underlying mechanisms.</p>
</sec>
<sec>
<title>1.3 The present study</title>
<p>In light of the foregoing, structure may be interpreted as a complex distribution of words in phonological or orthographic similarity space, leading to differences in how groups of related words participate over the course of word recognition. While networks provide many ways of measuring that structure, they do not tell us what underlying features or dimensions would suffice for describing that structure as it relates to word recognition. Our goal in the present study is to explore those underlying dimensions, shedding light on what it means to say that the mental lexicon is structured based on phonological and orthographic word forms, and what it means for that structure to influence processing.</p>
<p>Note that while our goal is to help work towards better explanatory models by refining our understanding of structure in the lexicon and its relationship to processing, our methods have much in common with predictive modeling and machine learning, in that we apply dimension-reduction techniques to a large pool of variables, including several nodal properties that have not yet been studied directly. The link between our methods and our longer-term explanatory goals is the theoretical construct of structure in the lexicon, and the use of neighbor networks as models of that structure. Based on this view, our latent variable approach provides a data-driven way to determine not only what measures of structure are predictive of performance in psycholinguistic tasks, but, principally, what underlying, latent constructs are useful for describing phonological and orthographic structure in the lexicon. Our work thus complements the prior literature, which has sought to determine which measures are useful in psycholinguistics by choosing nodal properties that reasonably match existing constructs (like neighborhoods) and testing them experimentally.</p>
</sec>
</sec>
<sec>
<title>2. Methods</title>
<sec>
<title>2.1 Network construction</title>
<p>To construct our networks, we began with the 79,672 real words in the English Lexicon Project database (ELP; <xref ref-type="bibr" rid="B11">Balota et al., 2007</xref>), minus words lacking phonetic transcriptions and retaining only words with HAL or SUBTLEX frequencies greater than zero (which eliminated many acronyms, misspellings, and other anomalous items). This left 70,741 real words. Within this lexicon, the ELP identifies a subset of 40,481 words (40,398 with phonetic transcriptions) based on further screening for obvious nonwords, misspellings, alternate spellings, and so forth, for which lexical decision and naming data are available (see <xref ref-type="bibr" rid="B11">Balota et al., 2007</xref>, for details). This subset is approximately the basis for the orthographic network used by Siew (<xref ref-type="bibr" rid="B82">2018</xref>), who considered it an adequate reflection of the vocabulary of an average English speaker (about 15,000 word families; <xref ref-type="bibr" rid="B20">Brysbaert et al., 2016</xref>).</p>
<p>For the present study, we follow Siew (<xref ref-type="bibr" rid="B82">2018</xref>), constructing networks using the 40,398 words for which behavioral data and phonetic transcriptions are available in the ELP.<xref ref-type="fn" rid="n5">5</xref> However, since this subset excluded many, mainly derived, words that would be familiar to an educated English speaker, we repeated our analyses using networks incorporating the larger set of 70,741 words with frequencies greater than zero. The results did not differ substantially, and the code, networks, data, and results from the larger networks are available in the online materials.</p>
<p>We note that both of these lexicons include many inflected forms, among them plural nouns, past tense verbs, along with participles, and so forth. While we are by no means the first to include inflected forms (e.g., <xref ref-type="bibr" rid="B82">Siew, 2018</xref>), many of the earlier studies used dictionaries, which lack inflected forms (<xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>). Though inflection clearly impacts neighbor network structure (at least in English, where many inflections are also neighbors), the issue of morphological relationships in phonological and orthographic networks has not yet been examined closely. We consider the inclusion of inflected words to be reasonable, based on substantial data and theory supporting the storage of regular inflected forms in memory (e.g., <xref ref-type="bibr" rid="B22">Bybee, 2006</xref>; <xref ref-type="bibr" rid="B48">Hay &amp; Baayen, 2005</xref>; <xref ref-type="bibr" rid="B53">Jackendoff &amp; Audring, 2020</xref>). Eventually, of course, a more sophisticated network structure, such as multilayer networks (see Section 4), would be needed to capture both phonological/orthographic and morphological structure. For now, since our focus is on spoken and written word recognition, we include inflected word forms.</p>
<p>To reflect both the auditory and visual modalities, we built two different networks, with the nodes representing the phonological and orthographic word forms, respectively. We use separate networks for two reasons. First, the literature has generally studied auditory and visual word recognition separately, and there is good reason to think that word recognition unfolds somewhat differently in each modality. Second, while there is also value to a more generalized view, there are many ways to incorporate both modalities into the same network (e.g., using multilayer networks, or drawing edges only between words that are neighbors in both modalities), and the literature to date offers no guidance for how to do this (the only example we are aware of is <xref ref-type="bibr" rid="B85">Siew &amp; Vitevitch, 2019</xref>, who take the latter strategy). We therefore focus on the latent dimensionality of structure in unimodal networks, leaving the question of representing multi-modal structure for future research. Nevertheless, as described below, we incorporate both modalities into the same statistical analysis.</p>
<p>The phonological forms were taken from the ELP, which uses a transcription like X-SAMPA (<xref ref-type="bibr" rid="B121">Wells, 1997</xref>; <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="www.phon.ucl.ac.uk/home/sampa">www.phon.ucl.ac.uk/home/sampa</ext-link>). To prepare these phonological transcriptions for network construction, stress and syllable boundaries were removed, and digraphs were replaced to ensure a 1:1 relationship between characters and phonemes. After collapsing homophones into a single node (homographs were already represented only once in the ELP, but some homophones with different spellings were present) the unique pronunciations/letter strings served as the labels for the nodes in the phonological and orthographic networks, respectively. This procedure resulted in a phonological network containing 38,982 nodes, and an orthographic network containing 40,398 nodes.</p>
<p>In both networks, edges were placed between two nodes if the Levenshtein edit distance (<xref ref-type="bibr" rid="B59">Levenshtein, 1966</xref>) between their labels equaled 1, i.e., the insertion, deletion, or substitution of a single letter or phoneme. Consistent with previous research, this led to a Giant Component (GC) and several smaller island components, along with many hermits. The GC in the phonological network comprised 13,396 words (34% of the network), and 11,358 words (28%) in the orthographic network.</p>
</sec>
<sec>
<title>2.2 Network measures</title>
<p>We analyzed 13 nodal properties, listed in <xref ref-type="table" rid="T1">Table 1</xref> with brief definitions. For purposes of exposition, we organize the table roughly from the most local to the most global properties (cf. micro, meso, and macro level structure; <xref ref-type="bibr" rid="B109">Vitevitch, 2022</xref>), although this should not be taken as a principled ordering. We selected these variables based on their use in prior research on lexical networks (degree, clustering, coreness, closeness centrality, and community structure) and for their potential relevance given spreading activation, random walk, or other processes suggested in the literature (such as the idea of dense attractor basins; <xref ref-type="bibr" rid="B122">Yarkoni et al., 2008</xref>). Our criteria for inclusion were liberal, based on our holistic approach to networks as models of structure in the lexicon, with the expectation that measures with little relevance to word recognition will show little relationship to the behavioral data. Where specific measures have not yet been used in word recognition research, we comment briefly in the table about their expected relevance.</p>
<p>The MATLAB (<xref ref-type="bibr" rid="B66">MATLAB, 2010</xref>) native function &#8220;centrality&#8221; was used to calculate degree, betweenness centrality, closeness centrality, and PageRank centrality. Average neighbor degree and frequency were computed in MATLAB. The MATLAB native function &#8220;conncomp&#8221; was used to determine each node&#8217;s connected component and its corresponding component size. The Brain Connectivity Toolbox (BCT; <xref ref-type="bibr" rid="B78">Rubinov &amp; Sporns, 2010</xref>) was used to calculate eccentricity, coreness, local efficiency, and clustering coefficient. The Louvain community detection method (<xref ref-type="bibr" rid="B15">Blondel et al., 2008</xref>) was applied in a custom Python program that employed the NetworkX (<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://pypi.org/project/networkx/">https://pypi.org/project/networkx/</ext-link>) and python-Louvain (<ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://github.com/taynaud/python-louvain">https://github.com/taynaud/python-louvain</ext-link>) libraries to determine the modules in the phonological network (30,783 modules, with modularity <italic>Q</italic> = 0.86) and the orthographic network (34,165 modules; <italic>Q</italic> = 0.81). Using the module information, the participation coefficient and within-module degree z-score of each node were calculated using a BCT script. The bivariate relationships of all nodal properties with the dependent variables were assessed visually, and only betweenness showed any evidence of nonlinearity or potentially influential extreme values. To remedy this, betweenness was log-transformed.</p>
<p>Before proceeding further, we explored the correlations among the nodal properties, both as a first approximation to how they may group together, and to assess multicollinearity. This was done using the subset of 6,457 words used in the analysis of the behavioral data described below. The bivariate correlations for nodal properties in the phonological network are shown in <xref ref-type="table" rid="T2">Table 2</xref>. The pattern of correlations in the orthographic network is similar (see the online materials).</p>
<table-wrap id="T2">
<caption>
<p><bold>Table 2:</bold> Bivariate correlations between nodal properties (in the phonological network) in the sample of words to be analyzed in the SCGLR.</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold>degree</bold></td>
<td align="left" valign="top"><bold>coreness</bold></td>
<td align="left" valign="top"><bold>av Neigh Degree</bold></td>
<td align="left" valign="top"><bold>av Neigh Frq</bold></td>
<td align="left" valign="top"><bold>clustering</bold></td>
<td align="left" valign="top"><bold>efficiency</bold></td>
<td align="left" valign="top"><bold>closeness</bold></td>
<td align="left" valign="top"><bold>between</bold></td>
<td align="left" valign="top"><bold>PageRank</bold></td>
<td align="left" valign="top"><bold>eccentricity</bold></td>
<td align="left" valign="top"><bold>community Size</bold></td>
<td align="left" valign="top"><bold>participation</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top">coreness</td>
<td align="left" valign="top">.93</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">av Neigh Degree</td>
<td align="left" valign="top">.87</td>
<td align="left" valign="top">.93</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">av Neigh Frq</td>
<td align="left" valign="top">.44</td>
<td align="left" valign="top">.50</td>
<td align="left" valign="top">.55</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">clustering</td>
<td align="left" valign="top">&#8211;.03</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">efficiency</td>
<td align="left" valign="top">.04</td>
<td align="left" valign="top">.15</td>
<td align="left" valign="top">.16</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">.98</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">closeness</td>
<td align="left" valign="top">.77</td>
<td align="left" valign="top">.79</td>
<td align="left" valign="top">.86</td>
<td align="left" valign="top">.50</td>
<td align="left" valign="top">&#8211;.07</td>
<td align="left" valign="top">&#8211;.02</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">between</td>
<td align="left" valign="top">.53</td>
<td align="left" valign="top">.44</td>
<td align="left" valign="top">.39</td>
<td align="left" valign="top">.17</td>
<td align="left" valign="top">&#8211;.56</td>
<td align="left" valign="top">&#8211;.50</td>
<td align="left" valign="top">.47</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">PageRank</td>
<td align="left" valign="top">.45</td>
<td align="left" valign="top">.27</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">&#8211;.01</td>
<td align="left" valign="top">&#8211;.27</td>
<td align="left" valign="top">&#8211;.23</td>
<td align="left" valign="top">.16</td>
<td align="left" valign="top">.54</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">eccentricity</td>
<td align="left" valign="top">&#8211;.62</td>
<td align="left" valign="top">&#8211;.65</td>
<td align="left" valign="top">&#8211;.73</td>
<td align="left" valign="top">&#8211;.44</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">.03</td>
<td align="left" valign="top">&#8211;.91</td>
<td align="left" valign="top">&#8211;.41</td>
<td align="left" valign="top">&#8211;.11</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">community Size</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">&#8211;.02</td>
<td align="left" valign="top">&#8211;.02</td>
<td align="left" valign="top">&#8211;.01</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.05</td>
<td align="left" valign="top">&#8211;.07</td>
<td align="left" valign="top"></td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">participation</td>
<td align="left" valign="top">.40</td>
<td align="left" valign="top">.39</td>
<td align="left" valign="top">.45</td>
<td align="left" valign="top">.24</td>
<td align="left" valign="top">&#8211;.25</td>
<td align="left" valign="top">&#8211;.23</td>
<td align="left" valign="top">.59</td>
<td align="left" valign="top">.34</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">&#8211;.53</td>
<td align="left" valign="top">&#8211;.10</td>
<td align="left" valign="top"></td>
</tr>
<tr>
<td align="left" valign="top">wMDZscore</td>
<td align="left" valign="top">.82</td>
<td align="left" valign="top">.75</td>
<td align="left" valign="top">.63</td>
<td align="left" valign="top">.31</td>
<td align="left" valign="top">.03</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">.48</td>
<td align="left" valign="top">.51</td>
<td align="left" valign="top">.54</td>
<td align="left" valign="top">&#8211;.38</td>
<td align="left" valign="top">.00</td>
<td align="left" valign="top">.07</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn><p><italic>Note</italic>. Correlations were computed over only the 6,457 words included for analysis of the behavioral data. Abbreviations: avNeighFrq: average neighbor frequency, avNeighDegree: average neighbor degree, between: betweenness centrality, closeness: closeness centrality, clustering: clustering coefficient, efficiency: local efficiency, PageRank: PageRank centrality, participation: participation coefficient, wMDZscore: within-module degree z-score. Significance is not noted (nearly all were significant), because of the large sample size. We focus on the strength of the correlations instead.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>The most extensive relationships among these predictors are the strong positive correlations between degree, coreness, average neighbor degree, and closeness centrality, which are all negatively related to eccentricity. Degree and coreness are related by definition (such that coreness cannot exceed degree), and the positive correlation between degree and average neighbor degree corresponds to the definition of assortative mixing by degree, which is a well-documented property of phonological networks (<xref ref-type="bibr" rid="B6">Arbesman et al., 2010b</xref>). All three of these variables, in turn, are positively related to closeness and negatively related to eccentricity: words with many and well-connected neighbors tend to be reachable from anywhere within the connected component via comparatively short paths, nodes on the periphery are not. Other patterns are observable here, but they are described in detail with the results of the SCGLR modeling below. For the moment, these observations suggest that the relationships among nodal properties are consistent with the idea that the position of a node in the network may be describable using a smaller number of dimensions, although individual nodal properties may also make unique contributions.</p>
</sec>
<sec>
<title>2.3 Behavioral data and dependent measures of lexical processing</title>
<p>We tested the nodal properties and covariates (described below) against behavioral data from the ELP (<xref ref-type="bibr" rid="B11">Balota et al., 2007</xref>) and the MALD (<xref ref-type="bibr" rid="B105">Tucker et al., 2019</xref>) databases, representing visual lexical decision (ELP), word naming (ELP), and auditory lexical decision (MALD).<xref ref-type="fn" rid="n6">6</xref> We used the by-item mean response times from these mega-studies, inverse-transformed (and multiplied by &#8211;1000) to approximate normal distributions, as confirmed by visual inspection of quantile-quantile plots. While Siew (<xref ref-type="bibr" rid="B82">2018</xref>) also found effects on response accuracy in the ELP, accuracy was generally near ceiling, with low variability, so we omitted it. These mega-studies provide a large data set against which to test how effects of network-derived properties compare across different task conditions, and how they scale up to a more representative lexicon, in contrast to the small and more uniform word sets used in many psycholinguistic experiments.</p>
<p>While the phonological and orthographic networks included 38,982 and 40,398 words, respectively, analysis of the behavioral data was performed on only a subset of these data, comprising 6,457 words selected based on the following criteria. First, since several nodal properties are undefined when computed across disconnected components (e.g., closeness centrality cannot include shortest paths for pairs of nodes between which no path exists), only words residing in the GC of both networks were included (reducing the target word list to 7,026), consistent with most studies of lexical networks (but see <xref ref-type="bibr" rid="B84">Siew &amp; Vitevitch, 2016</xref>). Second, only words with more than one neighbor, in both networks, were included (leaving 6,461 words), because clustering is undefined for words with only one neighbor (the denominator is the number of pairs of neighbors, but where there is only one neighbor, there are zero pairs). Finally, we included only words with RTs available in both the ELP and MALD (<xref ref-type="bibr" rid="B105">Tucker et al., 2019</xref>) and for which all covariates were available.</p>
</sec>
<sec>
<title>2.4 Supervised Component Generalized Linear Regression (SCGLR)</title>
<p>SCGLR (<xref ref-type="bibr" rid="B19">Bry et al., 2013</xref>; <xref ref-type="bibr" rid="B72">Mortier et al., 2015</xref>; <xref ref-type="bibr" rid="B104">Tomaschek et al., 2018</xref>) is a regression technique well-suited to data sets with many correlated predictors. It works by reducing the data to a smaller number of orthogonal components, like other latent variable techniques, but unlike other techniques, it seeks an optimal partitioning of the predictors with respect to their relationships with one or more dependent variables (DV). SCGLR, therefore, resembles Principal Components Regression, a two-step technique in which the predictors are first reduced to a smaller number of principal components based only on shared variance among the predictors (using Principal Components Analysis), and then the selected components are used to predict the dependent variable (using multiple regression). SCGLR combines these two steps into one, jointly modeling the predictors and one or more dependent variables in one step by maximizing a function based both on the shared variance among the predictors (like PCA) and on the relationship of the components with one or more dependent variables (like regression).</p>
<p>Since we are not merely interested in the interrelatedness of words&#8217; network-based properties, but rather in how measures of network structure relate to lexical processing, these features make SCGLR ideally suited to our goals. Where several predictors bear a strong joint relationship to the dependent variables, groups of predictors and outcomes will load together on selected components, and where individual predictors relate uniquely to dependent variables, selected components will have loadings from those predictors. The importance of individual predictors is also estimated through standardized regression coefficients that are shrunk to account for multicollinearity, but we stress here that our goal is not to &#8220;deal with&#8221; collinearity for the purpose of identifying predictors&#8217; unique contributions. We take a different approach to multicollinearity, allowing for the possibility that it may be more enlightening to focus on what the predictors have in common than to seek what independent contributions they may make (<xref ref-type="bibr" rid="B71">Morrissey &amp; Ruxton, 2018</xref>). When we sidestep collinearity, either through stimulus design or through stepwise regression techniques,<xref ref-type="fn" rid="n7">7</xref> we focus on the portion of variance that one variable does not share with the correlated variable(s), but it may not be clear whether these unique contributions should be interpreted as evidence of an independent underlying construct, or whether it indicates different strengths and weaknesses of correlated predictors as measures of the same underlying, latent construct. An additional advantage of our approach, applied to large-scale data, is that it avoids several difficulties with using controlled subsets of items for testing. Specifically, controlling for many variables (if this can even be achieved; <xref ref-type="bibr" rid="B35">Cutler, 1981</xref>) is unlikely to yield representative samples of words, compromising ecological validity, and the matching of small item lists based on inherently continuous variables can lead to inconsistent and sample-dependent results (<xref ref-type="bibr" rid="B60">Liben-Nowell et al., 2019</xref>).</p>
<p>By estimating the joint variance components among the predictors that are most useful for predicting behavioral data, SCGLR is, thus, consistent with our goal of exploring what underlying dimensions of phonological and orthographic structure in the mental lexicon are psychologically relevant, and how well those dimensions are measured by the various nodal properties made available through using neighbor networks as models of structure. This data-driven approach complements the existing literature by helping to identify the underlying constructs that can link network models to theories of word recognition, helping to reconcile these mathematical models (networks) with the theoretical construct of structure in the mental lexicon.</p>
</sec>
<sec>
<title>2.5 Data analysis</title>
<p>SCGLR models were fit to the 6,457 words residing in the GC of both networks, and meeting the other criteria outlined above, using the SCGLR package in R (<xref ref-type="bibr" rid="B72">Mortier et al., 2015</xref>; <xref ref-type="bibr" rid="B75">R Core Team, 2023</xref>). Since SCGLR supports the simultaneous fitting of multiple dependent variables, all three DVs were included (visual lexical decision and naming latencies from the ELP, and auditory lexical decision latencies from the MALD, inverse-transformed, as described above). We included both modalities in the same analysis in order to gain insight into both the commonalities and differences in how phonological and orthographic structure in the lexicon relates to word recognition. Since SCGLR partitions the common and unique variance among the dependent variables and relates this to the partitioning of the predictors, relationships that are common to both modalities, as well as those unique to one or the other, should be captured in the fitted model. Some of the covariates we include are expected to confirm this. For example, the initial phoneme is unlikely to affect RTs to written stimuli made by button press, but it should be related to RTs measured by voice key (naming), and possibly also RTs to auditory stimuli. This expectation was borne out in the results (see below and the online materials), confirming that the model is indeed capturing this kind of subtlety. We also validated this approach by fitting separate models to visual lexical decision (with predictors from the orthographic network) and auditory lexical decision (with predictors from the phonological network). The results were very similar to the patterns reported here in the combined model, and they are presented and discussed in the online materials.</p>
<p>The network-derived predictors (see <xref ref-type="table" rid="T1">Table 1</xref>), and their treatment, as described above, were drawn from both the phonological and the orthographic networks. We also included several covariates, based on their well-known effects on behavioral measures of lexical processing. Frequency of the orthographic wordform (<xref ref-type="bibr" rid="B64">Lund &amp; Burgess, 1996</xref>),<xref ref-type="fn" rid="n8">8</xref> length in phonemes, letters, and syllables, phonotactic probability (mean bigram frequency), and semantic neighborhood density (semND) were retrieved from the ELP database. Orthographic and phonological <italic>uniqueness points</italic> (UPs) were retrieved from the Massive Auditory Lexical Decision database (MALD; <xref ref-type="bibr" rid="B105">Tucker et al., 2019</xref>), along with the durations, in milliseconds, of the stimuli used in the MALD mega-study. UPs in the MALD are defined as the letter/phoneme position at which each word is distinguishable from all others, with the first letter/phoneme being at position 1. Words that cannot be uniquely identified at the last letter/phoneme (e.g. <italic>cat</italic>, which cannot be distinguished at position 3 from <italic>category)</italic> have UPs one more than the word length, (i.e., the word boundary). Thus, the uniqueness point for <italic>cat</italic> would be 4. To approximate normal distributions, frequency and the measures of word length were log-transformed, and the auditory stimulus durations were inverse-transformed; the transformations were selected based on visual inspection of quantile-quantile plots. Finally, we added a set of five dummy-coded variables encoding the initial phoneme, because amplitude-triggered voice keys, such as that used in the ELP megastudy, are affected by the acoustic properties of the initial sound. Based on the results of prior investigations into this issue (<xref ref-type="bibr" rid="B55">Kessler et al., 2002</xref>; <xref ref-type="bibr" rid="B76">Rastle &amp; Davis, 2002</xref>; <xref ref-type="bibr" rid="B112">Vitevitch &amp; Luce, 1998</xref>), these variables encoded whether the initial phoneme was (a) a vowel, (b) an obstruent, (c) a voiced obstruent, (d) a posterior consonant (/g, k, h/), and (e) an /s/ followed by another obstruent (note that sonorants are identified as phonemes that are neither vowels nor obstruents). All variables, including the dependent variables, were standardized, other than the categorical, dummy-coded predictors. All code and data, and additional details are available in the online materials.</p>
<p>Several steps taken to establish the validity of the results are reported in the online materials. In these, we compared the fitted response times from our model with published by-item mean RTs from two prior studies (<xref ref-type="bibr" rid="B3">Andrews, 1992</xref>; <xref ref-type="bibr" rid="B84">Siew &amp; Vitevitch, 2016</xref>), and we refit the model using various subsets of the lexicon that more closely resemble the words used in psycholinguistic studies on network-based lexical properties, including short words, low-frequency words, and the published item lists from Vitevitch and Luce (<xref ref-type="bibr" rid="B112">1998</xref>, <xref ref-type="bibr" rid="B113">1999</xref>), Siew and Vitevitch (<xref ref-type="bibr" rid="B84">2016</xref>), and Andrews (<xref ref-type="bibr" rid="B3">1992</xref>). Overall, these steps supported the validity of our model.</p>
</sec>
</sec>
<sec>
<title>3. Results</title>
<sec>
<title>3.1 Supervised components</title>
<p>SCGLR, like other latent variable techniques, reduces the dimensionality of a set of related variables, with the added feature of incorporating the relationships between latent dimensions (supervised components, sc) and a set of DVs. The number of components needed to effectively model the data was determined by minimizing the Bayesian Information Criterion using 10-fold cross-validation (<xref ref-type="bibr" rid="B72">Mortier et al., 2015</xref>), indicating that 8 components were sufficient. We then re-fit the SCGLR model using eight components. This model accounted for 38% of the variance (<italic>r</italic><sup>2</sup>) in visual LDT latencies, 32% in word naming latencies, and 13% in auditory LDT latencies.</p>
<p>We begin by exploring how the variance in the predictors and DVs is partitioned across the 8 components via the squared correlation loadings, shown in <xref ref-type="table" rid="T3">Table 3</xref>. The squared correlation loadings represent the proportion of variance that is shared between each variable and each component, showing which groups of predictors are most related to each component. The best plane indicates the pair of components most strongly related to each variable, and best value is the proportion of variance in each variable accounted for by the best plane.</p>
<table-wrap id="T3">
<caption>
<p><bold>Table 3:</bold> Squared Correlation Loadings of Predictors and Dependent Variables.</p>
</caption>
<table>
<thead>
<tr>
<td align="left" valign="top"></td>
<td align="left" valign="top"><bold>sc1</bold></td>
<td align="left" valign="top"><bold>sc2</bold></td>
<td align="left" valign="top"><bold>sc3</bold></td>
<td align="left" valign="top"><bold>sc4</bold></td>
<td align="left" valign="top"><bold>sc5</bold></td>
<td align="left" valign="top"><bold>sc6</bold></td>
<td align="left" valign="top"><bold>sc7</bold></td>
<td align="left" valign="top"><bold>sc8</bold></td>
<td align="left" valign="top"><bold>best plane</bold></td>
<td align="left" valign="top"><bold>best value</bold></td>
</tr>
</thead>
<tbody>
<tr>
<td align="left" valign="top" colspan="11">Dependent Variables</td>
</tr>
<tr>
<td align="left" valign="top">ldt_ELP</td>
<td align="left" valign="top">.38</td>
<td align="left" valign="top">.12</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">.28</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.66</td>
</tr>
<tr>
<td align="left" valign="top">nmg_ELP</td>
<td align="left" valign="top">.58</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.15</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.73</td>
</tr>
<tr>
<td align="left" valign="top">ldt_MALD</td>
<td align="left" valign="top">.20</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">.11</td>
<td align="left" valign="top">.17</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.27</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">1/7</td>
<td align="left" valign="top">.47</td>
</tr>
<tr>
<td align="left" valign="top" colspan="11">Predictors</td>
</tr>
<tr>
<td align="left" valign="top">ph_degree</td>
<td align="left" valign="top">.70</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.12</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.83</td>
</tr>
<tr>
<td align="left" valign="top">ph_avNeighDegree</td>
<td align="left" valign="top">.71</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">1/8</td>
<td align="left" valign="top">.79</td>
</tr>
<tr>
<td align="left" valign="top">ph_coreness</td>
<td align="left" valign="top">.71</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.78</td>
</tr>
<tr>
<td align="left" valign="top">ph_closeness</td>
<td align="left" valign="top">.69</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.77</td>
</tr>
<tr>
<td align="left" valign="top">or_length</td>
<td align="left" valign="top">.69</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.74</td>
</tr>
<tr>
<td align="left" valign="top">or_degree</td>
<td align="left" valign="top">.65</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.74</td>
</tr>
<tr>
<td align="left" valign="top">or_avNeighDegree</td>
<td align="left" valign="top">.64</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/2</td>
<td align="left" valign="top">.71</td>
</tr>
<tr>
<td align="left" valign="top">or_closeness</td>
<td align="left" valign="top">.63</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/5</td>
<td align="left" valign="top">.71</td>
</tr>
<tr>
<td align="left" valign="top">or_coreness</td>
<td align="left" valign="top">.64</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.70</td>
</tr>
<tr>
<td align="left" valign="top">ph_length</td>
<td align="left" valign="top">.66</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/6</td>
<td align="left" valign="top">.70</td>
</tr>
<tr>
<td align="left" valign="top">obstruent_initial</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.61</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">3/6</td>
<td align="left" valign="top">.64</td>
</tr>
<tr>
<td align="left" valign="top">ph_eccentricity</td>
<td align="left" valign="top">.51</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.11</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.62</td>
</tr>
<tr>
<td align="left" valign="top">ph_clustering</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.43</td>
<td align="left" valign="top">.15</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.18</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">2/5</td>
<td align="left" valign="top">.62</td>
</tr>
<tr>
<td align="left" valign="top">or_clustering</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.42</td>
<td align="left" valign="top">.19</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.17</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">2/3</td>
<td align="left" valign="top">.61</td>
</tr>
<tr>
<td align="left" valign="top">or_efficiency</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.41</td>
<td align="left" valign="top">.19</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.16</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">2/3</td>
<td align="left" valign="top">.61</td>
</tr>
<tr>
<td align="left" valign="top">ph_efficiency</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.42</td>
<td align="left" valign="top">.15</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.18</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">2/5</td>
<td align="left" valign="top">.60</td>
</tr>
<tr>
<td align="left" valign="top">ph_wMDZscore</td>
<td align="left" valign="top">.42</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">.16</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.58</td>
</tr>
<tr>
<td align="left" valign="top">freq</td>
<td align="left" valign="top">.28</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.16</td>
<td align="left" valign="top">.27</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.55</td>
</tr>
<tr>
<td align="left" valign="top">or_UP</td>
<td align="left" valign="top">.44</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/2</td>
<td align="left" valign="top">.54</td>
</tr>
<tr>
<td align="left" valign="top">V_initial</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.44</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">4/6</td>
<td align="left" valign="top">.53</td>
</tr>
<tr>
<td align="left" valign="top">ph_UP</td>
<td align="left" valign="top">.42</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/2</td>
<td align="left" valign="top">.51</td>
</tr>
<tr>
<td align="left" valign="top">or_wMDZscore</td>
<td align="left" valign="top">.32</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.17</td>
<td align="left" valign="top">.19</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.51</td>
</tr>
<tr>
<td align="left" valign="top">or_between</td>
<td align="left" valign="top">.24</td>
<td align="left" valign="top">.26</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.22</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/2</td>
<td align="left" valign="top">.50</td>
</tr>
<tr>
<td align="left" valign="top">nSyll</td>
<td align="left" valign="top">.38</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.12</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.50</td>
</tr>
<tr>
<td align="left" valign="top">ph_between</td>
<td align="left" valign="top">.27</td>
<td align="left" valign="top">.22</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.13</td>
<td align="left" valign="top">.12</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/2</td>
<td align="left" valign="top">.49</td>
</tr>
<tr>
<td align="left" valign="top">or_eccentricity</td>
<td align="left" valign="top">.39</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.49</td>
</tr>
<tr>
<td align="left" valign="top">semND</td>
<td align="left" valign="top">.19</td>
<td align="left" valign="top">.12</td>
<td align="left" valign="top">.17</td>
<td align="left" valign="top">.30</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.49</td>
</tr>
<tr>
<td align="left" valign="top">ph_participation</td>
<td align="left" valign="top">.22</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.20</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.42</td>
</tr>
<tr>
<td align="left" valign="top">ph_avNeighFrq</td>
<td align="left" valign="top">.35</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">1/8</td>
<td align="left" valign="top">.42</td>
</tr>
<tr>
<td align="left" valign="top">sObstruent_initial</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">.29</td>
<td align="left" valign="top">.13</td>
<td align="left" valign="top">7/8</td>
<td align="left" valign="top">.42</td>
</tr>
<tr>
<td align="left" valign="top">dur_MALD</td>
<td align="left" valign="top">.17</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.21</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/7</td>
<td align="left" valign="top">.38</td>
</tr>
<tr>
<td align="left" valign="top">or_avNeighFrq</td>
<td align="left" valign="top">.27</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/4</td>
<td align="left" valign="top">.36</td>
</tr>
<tr>
<td align="left" valign="top">ph_PageRank</td>
<td align="left" valign="top">.09</td>
<td align="left" valign="top">.21</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.14</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">2/4</td>
<td align="left" valign="top">.34</td>
</tr>
<tr>
<td align="left" valign="top">or_PageRank</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">.22</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/2</td>
<td align="left" valign="top">.30</td>
</tr>
<tr>
<td align="left" valign="top">or_participation</td>
<td align="left" valign="top">.19</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.10</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.30</td>
</tr>
<tr>
<td align="left" valign="top">voicedObs_initial</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.15</td>
<td align="left" valign="top">.15</td>
<td align="left" valign="top">.11</td>
<td align="left" valign="top">6/7</td>
<td align="left" valign="top">.30</td>
</tr>
<tr>
<td align="left" valign="top">posterior_initial</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.14</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">6/7</td>
<td align="left" valign="top">.22</td>
</tr>
<tr>
<td align="left" valign="top">meanBG</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.14</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/3</td>
<td align="left" valign="top">.20</td>
</tr>
<tr>
<td align="left" valign="top">ph_communitySize</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.07</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">.08</td>
<td align="left" valign="top">4/8</td>
<td align="left" valign="top">.15</td>
</tr>
<tr>
<td align="left" valign="top">or_communitySize</td>
<td align="left" valign="top">.06</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">&#8212;</td>
<td align="left" valign="top">1/7</td>
<td align="left" valign="top">.12</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn><p><italic>Note</italic>. Variables are sorted in descending order by their best value. Empty cells (&#8220;&#8212;&#8221;) indicate correlation loadings less than .05.</p></fn>
</table-wrap-foot>
</table-wrap>
<p>The sc1/sc4 plane, thus, captures the bulk of the relationship between the DVs and the nodal properties. This is the best plane for visual LDT and naming RTs. The best plane for auditory LDT is sc1/sc7, but its next strongest loading is on sc4, and sc7 primarily captures the effects of covariates that are most relevant to this particular task (duration of the MALD stimuli and their initial phonemes). The squared correlation loadings of the DVs on the remaining components are much smaller.</p>
<p>Among the predictors, most of the nodal properties and several covariates load substantially on sc1, including degree, coreness, average neighbor degree, closeness, and eccentricity, along with length-related variables and uniqueness points. This component accounted for 30% of the common variance (i.e., inertia) among the predictors (more than three times higher than the inertia of the next component). Most of these share sc1/sc4 as their best plane, like the DVs, but only frequency and semND load substantially on sc4. A largely different set of nodal properties loads more strongly on sc2, sc3, and sc5, comprised primarily of clustering, efficiency, and PageRank, but also including betweenness, participation and wMDZscore, which also load on sc1. The only variables loading substantially on sc6-sc8 are those encoding the initial phoneme and the duration of the MALD stimuli, so we do not comment further on these components.</p>
<p>To examine how these groups of predictors relate to the DVs, we plot the correlation loadings for several planes in <xref ref-type="fig" rid="F1">Figures 1</xref> and <xref ref-type="fig" rid="F2">2</xref>, respectively. The correlation loadings on the two components in these planes are the coordinates of each vector in these plots, and they are not squared (as in <xref ref-type="table" rid="T3">Table 3</xref>), so that the directions of the effects may be seen. The length of each vector represents the strength of the joint correlation between one predictor and the two components represented on the x and y axes (only predictors with a joint correlation loading with absolute value greater than .5 are shown). Vectors that are nearly parallel represent variables that are strongly correlated with each other on the plane. Variables that are negatively correlated with each other point in opposite directions. Other planes are available in the Plots folder in the online materials, and the code and data are also available there for further exploration.</p>
<fig id="F1">
<caption>
<p><bold>Figure 1:</bold> Correlation Loadings on the sc1/sc4 Plane. The solid circle indicates the maximum correlation (i.e., &#177;1), and the dotted circle indicates a correlation of &#177;.5. Dependent variables are labeled in red, and predictors in blue. The numbers in parentheses along each axis represent the proportion of common predictor variance accounted for by each component (i.e., inertia).</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-4-1-33588-g1.png"/>
</fig>
<fig id="F2">
<caption>
<p><bold>Figure 2:</bold> Correlation Loadings on the sc2/sc5 and sc3/sc5 Planes.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-4-1-33588-g2.png"/>
</fig>
<p>To begin interpreting the sc1/sc4 plane (<xref ref-type="fig" rid="F1">Figure 1</xref>), note that all three DVs align closely with each other in the third quadrant, with frequency, semND, and the average neighbor frequency measures aligned more or less opposite the DVs, in the first quadrant. This indicates that this plane captures a negative relationship between shared variance among these four predictors, and shared variance among all three DVs. That is, frequent words with many semantic neighbors, and with frequent phonological and orthographic neighbors, are responded to more quickly on these tasks than infrequent words with few semantic neighbors and infrequent phonological and orthographic neighbors.</p>
<p>The bundle of predictors that primarily load on sc1 is also clearly seen here. Pointing to the left are length in phonemes, letters, and syllables, UPs, and eccentricity, and pointing to the right are degree, coreness, average neighbor degree, closeness, betweenness, and wMDZscore (the latter two loading somewhat more negatively on sc4). The angle between this bundle and the DVs shows that this complex of variables also explains variance in the behavioral data, with short words residing in dense areas of the lexicon associated with faster RTs.</p>
<p>As noted above, betweenness, wMDZscore, and participation differ from the other predictors loading on sc1 in that they also load on sc2, sc3, and/or sc5. Sc2 differentiates betweenness from the other variables in this bundle, where it shares variance with PageRank and semND, together relating negatively to the DVs. Sc3 differentiates wMDZscore and participation from the main sc1 bundle, with participation loading positively, along with the DVs on sc3, and wMDZscore loading negatively (planes sc1/sc2, and sc1/sc3, showing these patterns more directly, are available in the online materials).</p>
<p>Clustering, efficiency, and PageRank load on sc2, sc3, and sc5, but not on sc1. <xref ref-type="fig" rid="F2">Figure 2</xref> shows the sc2/sc5 and sc3/sc5 planes. On these planes, we can see that sc2 captures common variance involving clustering, efficiency, betweenness, and PageRank (phonological PageRank loads similarly to betweenness on sc2/sc5, but more weakly), and sc3 captures common variance in clustering, efficiency, and participation (orthographic participation loads positively on sc3 and negatively on sc5, but relatively weakly, and is not shown). Note that participation loads opposite wMDZscore on sc3, though weakly. Sc5 differentiates these variables, based on the modality of the network. Interestingly, these planes show a positive relationship between the DVs and phonological clustering/efficiency, and a negative relationship with orthographic clustering/efficiency (the lengths of the DV vectors show that these relationships are relatively weak).</p>
</sec>
<sec>
<title>3.2 Regression coefficients</title>
<p>To explore the overall contributions of individual predictors, we turn our attention to their standardized regression coefficients, for each DV. Unlike ordinary and mixed effects regression, where multicollinearity can lead to unreliable coefficient estimates, SCGLR shrinks the coefficients to account for the collinearity, making them interpretable (<xref ref-type="bibr" rid="B104">Tomaschek et al., 2018</xref>). Note, however, that they must still be interpreted in the context of the whole model, including their loadings on the supervised components, as we discuss below. Perhaps more importantly, the standardized regression coefficients allow comparison of the relative effect sizes to gauge predictors&#8217; importances. A caterpillar plot of the standardized regression coefficients is given in <xref ref-type="fig" rid="F3">Figure 3</xref>.</p>
<fig id="F3">
<caption>
<p><bold>Figure 3:</bold> Standardized Regression Coefficients. The predictors are ordered, top to bottom, from the largest negative to the largest positive effect on visual LDT response times, which was the DV having the highest R<sup>2</sup> in the model. Error bars show 95% CIs, based on 1000 bootstrap samples. Network-based properties are bolded.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="glossapx-4-1-33588-g3.png"/>
</fig>
<p>The standardized regression coefficients for the nodal properties are considerably smaller than the effects of frequency and semND, as well as duration and the initial phoneme for the naming and auditory LDT tasks. The directions of these larger effects of covariates are as expected.</p>
<p>The effects for many, though not all, nodal properties are also generally consistent with what was observed in the supervised components and with previous studies, Interestingly, some properties exhibit different effects based on network modality and/or the behavioral task. Orthographic degree and closeness are negatively related to visual LDT and naming (the two tasks which involved reading written words; cf. <xref ref-type="bibr" rid="B3">Andrews, 1992</xref>), but phonological closeness is positively related to the DVs, and the CI for phonological degree overlaps with zero for all three tasks. A similar pattern of results is observed for the effects of coreness, closeness (cf. <xref ref-type="bibr" rid="B42">Goldstein &amp; Vitevitch, 2017</xref>), and average neighbor degree, along with eccentricity in the opposite direction, all of which loaded primarily on sc1, like degree. As phonological clustering and efficiency increased, reaction time increased for naming and visual, though not auditory, LDT (cf. <xref ref-type="bibr" rid="B28">Chan &amp; Vitevitch, 2009</xref>; <xref ref-type="bibr" rid="B123">Yates, 2013</xref>), and the CIs for orthographic clustering and efficiency overlapped with zero. As PageRank centrality increased, reaction time decreased on all tasks. Related to community structure, community size and participation had mainly positive effects.</p>
</sec>
</sec>
<sec>
<title>4. Discussion</title>
<p>To summarize the results, the components identified in the SCGLR partitioned the common variance among predictors in ways that reflect the conceptual and mathematical relationships among nodal properties, and in some cases, network modality, and in ways that related to the behavioral data. Roughly the same latent variable structure and pattern of importances was found in SCGLR models fit to previously published item lists (and RT data) carefully matched for several lexical properties (see the online materials). The effect directions shown by the standardized regression coefficients are largely, though not completely, consistent with prior findings, but most effect sizes, particularly for the nodal properties, were small. Effect sizes for degree and other nodal properties tend to be small in general (cf. <xref ref-type="bibr" rid="B82">Siew, 2018</xref>), so this is not necessarily surprising, and it is also consistent with the shrinkage that results from the multicollinearity among these variables. Since our primary goal was to shed light on the underlying dimensions of phonological and orthographic structure in the mental lexicon, we focus first on interpreting the supervised components, and then comment on the predictors&#8217; importances and (shrunken) estimated effects.</p>
<sec>
<title>4.1 Underlying dimensions of structure</title>
<p>Where nodal properties are concerned, the decomposition into supervised components yielded three main dimensions, accounting for varying amounts of shared variance in the DVs. We interpret these as reflecting generalized constructs of density, interconnectedness, and community structure. The first candidate construct identified in our analysis, captured primarily in sc1, encompassed the bulk of the shared variance among length, UP, degree, coreness, average neighbor degree, betweenness, closeness, and eccentricity. This component accounted for the largest portion of the common variance in the DVs (see <xref ref-type="table" rid="T3">Table 3</xref>). The correlations among these variables are no surprise. Short words have earlier UPs (because UP, as defined in the MALD database, can never be more than one greater than length), more neighbors, and their neighbors (being short also) tend to have more neighbors, leading to high coreness. Due to their dense neighborhoods, they also lie on many paths, making high closeness and low eccentricity likely. Betweenness and within module degree Z-score (wMDZscore) also loaded, to some extent, on sc1, correlating positively with degree, coreness, etc., but since they loaded on other components as well, it is less clear how to interpret their role in this dimension. It may simply be that words with many neighbors are more likely to be provincial hubs (wMDZscore), and to lie on shortest paths between more distant areas of the network (betweenness). We will comment further on these nodal properties below.</p>
<p>This grouping of variables suggests a generalized concept of centrality or density in the network, with faster responses to words embedded in areas of greater density. Interestingly, this parallels, to some extent, the finding of Brown et al. (<xref ref-type="bibr" rid="B18">2018</xref>) that the length distribution of their random lexicons was the primary driver of differences in overall network structure, including assortativity, average path length, degree distribution, and clustering. However, this component does not seem to be driven by length alone, because the same group of variables patterned together in SCGLR analyses of monosyllabic words (see the online materials). Rather, it seems that much of the effect of word length in phonemes, syllables, or letters may have less to do with the size of words than it does with the greater likelihood that the forms of short words will overlap substantially.<xref ref-type="fn" rid="n9">9</xref></p>
<p>Neighborhood density/degree (<xref ref-type="bibr" rid="B34">Coltheart et al., 1977</xref>; <xref ref-type="bibr" rid="B58">Landauer &amp; Streeter, 1973</xref>; <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>) was, of course, originally proposed as a first approximation of something like this, offering a way to quantify competition arising from the activation of similar words. This is expected, in some sense, under every theory of word recognition to date, regardless of how competitors become activated. Sc1 offers a more gradient way of capturing this idea (cf. earlier proposals by <xref ref-type="bibr" rid="B46">Hahn &amp; Bailey, 2005</xref>; <xref ref-type="bibr" rid="B122">Yarkoni et al., 2008</xref>), incorporating measures of density that vary in how locally they are focused around a target word. We note that different nodal properties capturing variation in density at different distances from the target, such as closeness (<xref ref-type="bibr" rid="B42">Goldstein &amp; Vitevitch, 2017</xref>) and two-hop degree (<xref ref-type="bibr" rid="B81">Siew, 2017</xref>), account for unique variance in word recognition, as does coreness for children&#8217;s vocabulary development (<xref ref-type="bibr" rid="B23">Carlson et al., 2014</xref>). However, accounting for unique variance does not necessarily mean that density at different distances from the target impacts word recognition in different ways. Rather, since none of the variables loaded independently on any supervised component, and the standardized regression coefficients were small, it seems better to interpret these variables as measures of a single latent construct.</p>
<p>The other latent constructs suggested by our results involve clustering, betweenness, PageRank, participation, and wMDZscore, which loaded on sc2, sc3, and sc5. Apart from clustering and a general approach to community structure/modularity, these have not received much attention in psycholinguistics, so our interpretation is tentative. The planes shown in <xref ref-type="fig" rid="F2">Figure 2</xref> suggest that these variables relate to two latent dimensions, both of which depend on the network modality, and that each account for a relatively small portion of the shared variance in the DVs. In one of these, clustering and efficiency are negatively related to PageRank and betweenness. We suggest that this dimension reflects a more generalized version of clustering, or the extent to which the words surrounding a target, including words that are not immediate neighbors, are connected to each other. Where this interconnectedness is high, many paths will cross that region, such that there are many short or likely paths. Clustering/efficiency of any given node are, thus, expected to be high, and betweenness and PageRank are expected to be low. This interpretation would be consistent with the positive effects of clustering on RT documented in prior studies (<xref ref-type="bibr" rid="B2">Altieri et al., 2010</xref>; <xref ref-type="bibr" rid="B28">Chan &amp; Vitevitch, 2009</xref>; <xref ref-type="bibr" rid="B123">Yates, 2013</xref>).</p>
<p>In the other dimension, clustering and efficiency are related negatively to participation. Since participation measures the extent to which words are neighbors to words residing in different communities from themselves, we suggest that this dimension may reflect aspects of community structure (cf. <xref ref-type="bibr" rid="B80">Siew, 2013</xref>). WMDZscore, which measures the connectedness of a word to the other words in its own community, also loads on sc3, where it is positively related to participation. Interestingly, these variables are related to words&#8217; positions in their communities, but not the size of the communities, which was the only nodal property that did not load strongly on any component (though the standardized regression coefficients suggest that it has a positive overall effect on RTs). The loading of clustering and efficiency on these components may reflect a relationship between clustering and modularity. Areas of lower clustering may contribute to increased modularity, because if two neighbors of a target node reside in different communities (as when participation is high), those neighbors are less likely to be neighbors of each other. Taking these two proposed latent constructs together, this suggests that regions exhibiting generally high interconnectedness may vary in the presence of boundaries between one or more communities, such that some words lie on those boundaries, whereas others do not. The alignment of the DVs with this dimension suggests slower responses to words in regions where many communities intersect.</p>
<p>One additional feature of these latter two constructs merits further investigation. On the sc2/sc5 plane (<xref ref-type="fig" rid="F3">Figure 3</xref>, left panel), which we interpret as revealing a generalized concept of interconnectedness (or clustering), the DVs are positively related to clustering/efficiency, and negatively related to betweenness (and PageRank, which had a joint loading on this plane of less than .5, but was nearly parallel to betweenness), but only based on the phonological network. The same variables measured in the orthographic network were orthogonal (perpendicular) to the DVs. Conversely, on the sc3/sc5 plane, which we have interpreted as proximity to the intersections among communities, the DVs were negatively related to clustering and positively to participation (again, loading more weakly on this plane, but parallel to clustering and efficiency), and were orthogonal to the phonological versions of these variables. If our interpretations of these latent constructs are correct, then the opposite loadings of clustering relative to the DVs may be due to the different reasons why clustering/efficiency are related to these constructs. More puzzling is the dependence of these effects on network modality. This is not without precedent (note the facilitatory effect of phonographic clustering reported by <xref ref-type="bibr" rid="B85">Siew &amp; Vitevitch, 2019</xref>) and contrasting effects of neighborhood density have long been reported in the literature. This may reflect the complex relationship between phonology and orthography in English, but we leave this for future investigation.</p>
</sec>
<sec>
<title>4.2 Importance of nodal properties</title>
<p>For the most part, as we have already pointed out, the standardized regression coefficients are consistent with the prior evidence, and the mostly small effect sizes are not surprising, especially in light of the shrinkage due to multicollinearity. We focus here on interpreting these coefficients in light of the latent structure revealed in the supervised components, and on observations that either depart from established findings or have little precedent.</p>
<p>First, however, observe that the effects of the duration of the auditory stimuli from the MALD and the initial phonemes of words yielded much larger effects on the tasks involving auditory stimuli or speech production, as expected. Notwithstanding some small effects on visual lexical decision, these variables loaded primarily on components that did not correlate with visual LDT. For example, the loadings on sc7 (<xref ref-type="table" rid="T3">Table 3</xref>) show strong loadings only for dur_MALD, [s]+obstruent onsets, and auditory LDT latencies. This supports the validity of the SCGLR approach, in that it distinguished when predictors related primarily to one or two DVs, vs. to what all three DVs shared in common.</p>
<p>In general, though, the standardized coefficients must be interpreted in light of the fact that they are shrunken, due to the multicollinearity captured in the supervised components. Thus, while most of the nodal properties contribute to the latent dimensions we have identified, their small and relatively uniform effects show that no one measure stands out as being substantially more important, such that we might otherwise interpret it as more central to the underlying construct. Thus, while stepwise regression might reveal unique contributions to explaining the behavioral data, this alone does not provide strong evidence that they represent different constructs, and we take the present data to favor interpreting nodal properties as providing different ways of measuring the same underlying constructs, as decomposed in the supervised components, each measure being subject to its own strengths and weaknesses.</p>
<p>Interestingly, (phonological) PageRank and the measures of community size stand out, in that their effects were relatively large (though still small) among the nodal properties, including better studied ones like degree and clustering. We attribute this partly to the fact that PageRank and community size did not strongly associate with the other groupings of variables (see <xref ref-type="table" rid="T3">Table 3</xref>), leading to less shrinkage of the standardized coefficients. Note also that our interpretation of the third dimension of structure identified above refers to location relative to the boundaries between communities, but not to the size of the communities. It is not yet clear whether these variables might be understood as revealing further subtleties in the structure of the lexicon, but if so, their weak representation in the supervised components, compared to the three constructs we have identified, suggests that any role they play is likely to be minor.</p>
<p>There are a few instances where the standardized coefficients do not seem to reflect previously established findings. First, we note that the very small negative relationship between length in phonemes and visual LDT, in contrast to all other effects of length-related variables. This could be an artifact of measuring length in different units in a language with relatively nontransparent orthography, and we note that all length measures loaded primarily on what we have interpreted as the generalized density construct. We interpret this to mean that their effects are not entirely reducible to the idea that longer words take longer to process (in fact, the relation between ldt_MALD and dur_MALD captured on sc7 is the only obvious effect of this nature).</p>
<p>Regarding the measures of structure, we note that while the coefficients of phonological degree are positive, as is well established (<xref ref-type="bibr" rid="B38">Goh et al., 2009</xref>; <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>; <xref ref-type="bibr" rid="B113">Vitevitch &amp; Luce, 1999</xref>), they are very small, and while positive effects of clustering have been consistently reported (<xref ref-type="bibr" rid="B2">Altieri et al., 2010</xref>; <xref ref-type="bibr" rid="B28">Chan &amp; Vitevitch, 2009</xref>; <xref ref-type="bibr" rid="B123">Yates, 2013</xref>), the coefficients here only show this effect for visual lexical decision and naming. The absence of a positive relationship between degree and auditory word recognition response times here could have to do with task strategies or list effects (<xref ref-type="bibr" rid="B4">Andrews, 1997</xref>; <xref ref-type="bibr" rid="B18">Brown et al., 2018</xref>) specific to the MALD, and we note that a positive effect of degree was found to be significant by Castro and Vitevitch (<xref ref-type="bibr" rid="B109">2022</xref>) in the Auditory ELP data (<xref ref-type="bibr" rid="B39">Goh et al., 2020</xref>). On the other hand, the literature on neighbor networks, the present study included, provides evidence that neighborhoods differ not only in size, but also in the overall density and interconnectedness of the surrounding region, properties that may modulate the activation of words in that region, given a specific target. Since the direction of ND effects on word recognition response latencies depends on the strength of activation of the neighbors (<xref ref-type="bibr" rid="B32">Chen &amp; Mirman, 2012</xref>), this may account for the discrepancy here. However, we find it more consistent with our overall analysis to attribute these effects to the status of these nodal properties as proxies for more general latent dimensions of structure. Coupled with the strong relationships between degree, clustering, and the dimensions revealed in the supervised components, we have interpreted small or nonsignificant regression coefficients to indicate that the associated nodal properties do not dominate those components, relative to the other measures.</p>
<p>Concerning the coefficients for clustering (and efficiency), we have already discussed how these properties&#8217; loadings on the supervised components related to two different dimensions of structure, suggesting that they relate to the behavioral data in multiple ways. In a model, such as ours, that accounts for this multiplicity, shrinking the regression coefficients accordingly, those coefficients are not as straightforwardly interpretable.</p>
</sec>
<sec>
<title>4.3 Structure and processing</title>
<p>The present results offer a new perspective on the theoretical construct of (phonological/orthographic) structure in the mental lexicon, by suggesting that its psycholinguistically important features are describable using three dimensions: the density of words packed into the region surrounding a target word, the interconnectedness of those words, and the location of the target word relative to the margins between subregions of greater interconnectedness. This leads to three related questions. Where does this structure come from, how does it relate to word recognition, and what are the best ways to measure it?</p>
<p>Theories of lexical and phonological development suggest that structure emerges as the need to distinguish one word from among similar words drives sensitivity to phonetic or orthographic detail. This is captured in the Lexical Restructuring Hypothesis (<xref ref-type="bibr" rid="B69">Metsala, 1997</xref>; <xref ref-type="bibr" rid="B70">Metsala &amp; Walley, 1998</xref>; for extension of this idea to network science, see <xref ref-type="bibr" rid="B23">Carlson et al., 2014</xref>; <xref ref-type="bibr" rid="B108">Vitevitch, 2008</xref>), and in discriminative learning models (<xref ref-type="bibr" rid="B8">Baayen et al., 2019</xref>; <xref ref-type="bibr" rid="B54">Kapatsinski, 2018</xref>). Learned sensitivity to detail would reflect both the presence of similar words and the specific ways those words relate to the target, leading to the observed effects in word recognition. Supporting this developmental view, discriminative learning models that lack lexical representations entirely have simulated effects of ND (<xref ref-type="bibr" rid="B7">Baayen, 2010</xref>).</p>
<p>Similar to these ideas, Vitevitch and Sale (<xref ref-type="bibr" rid="B117">2023</xref>) suggested that words whose phonological properties are most useful for distinguishing words form a <italic>backbone</italic> of words that serve to maintain the overall, macro-level structure of the phonological network. Intriguingly, using a network simplification procedure, they found that words in this backbone had properties that load positively on sc1 in our analysis, namely, short words with high degree and closeness (and high frequency as well), and that the backbone GC also had higher modularity.</p>
<p>Work inspired by the Lexical Restructuring Hypothesis usually operationalizes words&#8217; similarity at the segment level, i.e., through minimal pairs, corresponding closely to basing network structure on neighbor relationships. But broader structure in neighbor networks, e.g., clustering and modularity, has been linked to phonotactic distributions (<xref ref-type="bibr" rid="B80">Siew, 2013</xref>; <xref ref-type="bibr" rid="B116">Vitevitch et al., 2021</xref>), suggesting that structure in the lexicon also reflects larger scale units, including consonant clusters, syllables, and possibly also morphology. Taking a developmental view of structure as reflecting the details that allow words to be discriminated reliably, the picture of structure emerging here draws attention to these units of different sizes and types as impacting the similarity of word forms simultaneously on different scales, in ways that can be summarized by the constructs of density, interconnectedness, and proximity to community boundaries.</p>
<p>How, then, do these three constructs (rather than the nodal properties we might use to measure them) connect to theories of word recognition, or lexical processing more generally? Theories of word recognition universally incorporate the activation of sets of related words during processing. The density with which words occur within a certain range of phonological and/or orthographic similarity has long been incorporated into psycholinguistic theories, whether in models relying on bottom-up or interactive activation (e.g., <xref ref-type="bibr" rid="B61">Luce et al., 2000</xref>; <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>; <xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>; <xref ref-type="bibr" rid="B68">McClelland &amp; Rumelhart, 1981</xref>; <xref ref-type="bibr" rid="B73">Norris, 1994</xref>; <xref ref-type="bibr" rid="B74">Norris &amp; McQueen, 2008</xref>), including models incorporating incremental processing (e.g., <xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>; <xref ref-type="bibr" rid="B74">Norris &amp; McQueen, 2008</xref>). Understanding structure as a reflection of the information that serves to distinguish words (cf. the model of discriminative learning cited above) also provides a link to models involving distributed, rather than localist, representations. While this is most commonly quantified via ND, its long acknowledged weaknesses (discussed above) show that there is nothing special about immediate neighbors, consistent with the gradience evident in the latent density variable we have identified.</p>
<p>Other aspects of neighbor network structure have not usually been explicitly addressed in the literature on lexical processing, except for the network science-based literature. Nevertheless, there is good reason to think that these constructs are compatible. Yates (<xref ref-type="bibr" rid="B123">2013</xref>) provides a relatively direct example, explaining clustering effects through bottom-up and/or interactive activation of a target&#8217;s neighbors, because clustering reflects the extent to which those neighbors overlap with the target in the same way, and this reasoning would seem to apply to our second latent variable, which extends the idea of clustering to capture the interconnectedness of words beyond the circle of immediate neighbors.</p>
<p>The third latent dimension, which we have tentatively related to community structure, represents a way of conceptualizing words&#8217; relationships to the rest of the lexicon that has only recently gained attention in psycholinguistics (<xref ref-type="bibr" rid="B80">Siew, 2013</xref>), but there are nonetheless some points of contact. One interpretation of community structure involves higher resting activation (or a lower activation threshold) for words that link several communities, because their relatedness to a more diverse set of words results in more frequent activation (cf. <xref ref-type="bibr" rid="B111">Vitevitch &amp; Goldstein, 2014</xref>). Alternatively, since community structure is thought to capture shared phonotactic structure (<xref ref-type="bibr" rid="B80">Siew, 2013</xref>), words at the intersections of several communities may be relatively unique among their competitors, because their competitors divide into subgroups defined by their community membership.</p>
<p>We thus argue that there is good reason to think that most prominent theoretical models are compatible to some extent with this richer view of structure. One obvious way to evaluate this compatibility is to study the conditions influencing modularity in neighbor networks. Phonotactics provides one obvious way to begin pursuing this, and we suggest that prosodic structure and morphology may also yield insight into more psycholinguistically articulated interpretations of community structure.</p>
<p>New computational modeling will also have an important role to play. Documented effects of several nodal properties are captured by computational models implementing activation spread or search over neighbor networks (<xref ref-type="bibr" rid="B83">Siew, 2019</xref>; <xref ref-type="bibr" rid="B110">Vitevitch et al., 2011</xref>; <xref ref-type="bibr" rid="B115">Vitevitch &amp; Mullin, 2021</xref>), but not by computational implementations of two earlier models that instead use either interactive or bottom-up activation (TRACE; <xref ref-type="bibr" rid="B67">McClelland &amp; Elman, 1986</xref>; <xref ref-type="bibr" rid="B99">Strauss et al., 2007</xref>; Shortlist; <xref ref-type="bibr" rid="B73">Norris, 1994</xref>), plus word-to-word inhibition, in the case of TRACE. This could indicate a need to incorporate lateral activation spread into theories of word recognition, but given the well-known limitations of TRACE and Shortlist, it would be useful to explore the latent constructs emerging from the present analysis using more recent models, such as Shortlist-B (<xref ref-type="bibr" rid="B74">Norris &amp; McQueen, 2008</xref>) or LDL-AURIS (<xref ref-type="bibr" rid="B79">Shafaei-Bajestan et al., 2023</xref>).</p>
<p>The third question we consider here is how best to measure word-form based structure in the lexicon. We have identified three underlying dimensions of structure, using 12 different nodal properties measured in phonological and orthographic neighbor networks. But are all of these measures necessary? And are traditional neighbor networks the best way to approach the measurement of structure? Relatedly, a weakness of neighbor networks of the kind used here (i.e., where edges are placed between neighbors only) is that many nodal properties are only defined within a single connected component or for words with degree of at least 2. As a result, we followed the common practice in the network science literature, analyzing behavioral data only for words residing in the GC of our networks. Since this limitation is inherent to neighbor networks as implemented here, generalizing the present findings to the whole lexicon would require identifying alternative network structures in which to quantify the latent dimensions we have described.</p>
<p>Practically, if one&#8217;s goal is to select or control words for use in experiments, or in diagnosis or treatment of language-related disorders, words&#8217; scores on supervised components could be used (ours are available in the online materials). On the other hand, since none of the nodal properties emerged as better encapsulating the underlying dimensions, and since the first dimension (generalized density) was more strongly associated with the behavioral data than the other two constructs, interconnectedness or relationship to communities, it may be enough to use degree as a proxy, as has been done for decades. If all three constructs are in view, clustering, betweenness, and/or participation might also be used as proxies. Using only a small number of proxy measures in this way would not be subject to the shrinkage reflected in our regression coefficients, and we would expect them to adequately capture the desired effects, provided their interpretation is contextualized with respect to the theoretical construct of structure in the lexicon.</p>
<p>If, however, our goal is to probe the notion of phonological and orthographic structure more deeply, then it would be useful to use insights gained from the underlying dimensions of structure to move beyond defining networks through the traditional understanding of neighbors. Maintaining the general idea of networks, this might involve building networks in ways that better match our current understanding of the sources of structure. For example, Vitevitch et al. (<xref ref-type="bibr" rid="B116">2021</xref>) sought to incorporate phonotactics more directly by placing network edges based on shared biphones, or using a bipartite network linking phonetic units to word nodes, but lacking links between word nodes. A different approach would use participant responses to infer neighborhood structure (<xref ref-type="bibr" rid="B40">Goldrick et al., 2010</xref>; <xref ref-type="bibr" rid="B62">Luce &amp; Large, 2001</xref>), rather than a priori critiera, similar to common practice in research on semantic networks (e.g., <xref ref-type="bibr" rid="B14">Benedek et al., 2017</xref>). Castro and Vitevitch (<xref ref-type="bibr" rid="B109">2022</xref>) pursued this approach, building phonological networks using phonological association and misperception data collected from human participants and obtaining somewhat better prediction of behavioral data. This avoids some potential limitations of specific similarity metrics, as human associations likely incorporate positional and other information beyond the segmental content of words, and some insight can be gained into what information respondents use to identify associates by exploring what features contribute to perceived phonological similarity (e.g., phonetic features; <xref ref-type="bibr" rid="B46">Hahn &amp; Bailey, 2005</xref>). It also appears to help with network fragmentation, as their GC comprised well over 90% of the network. However, there are practical challenges with obtaining enough data to reasonably approximate the contents of the lexicon.</p>
<p>Another approach would be to build on graded views of neighborhood structure, e.g., by weighting edit distance by phoneme similarity and frequency (<xref ref-type="bibr" rid="B9">Bailey &amp; Hahn, 2001</xref>; <xref ref-type="bibr" rid="B46">Hahn &amp; Bailey, 2005</xref>; <xref ref-type="bibr" rid="B63">Luce &amp; Pisoni, 1998</xref>). The use of 2-hop neighborhoods is one way to incorporate gradedness (<xref ref-type="bibr" rid="B81">Siew, 2017</xref>). Another is to use a fully connected network, in which each word is connected to all other words in the lexicon, with the edges of the network weighted according to some similarity metric. The OLD20 metric is one nodal property of such an orthographic network, where similarity is measured as edit distance in letters, OLD20 being the average weight of edges to the 20 nearest neighbors (<xref ref-type="bibr" rid="B122">Yarkoni et al., 2008</xref>). Measuring nodal properties in a fully-weighted lexical network would be highly processor-intensive with existing algorithms, and no attempts have been made to use this kind of network in psycholinguistics.</p>
<p>More recently, multilayer networks have attracted attention as techniques for incorporating multiple levels of representation (e.g., phonetic features, phonological and prosodic units, morphemes, lexical items, etc.) to model structure in the lexicon. Multilayer networks arise from the mathematical concept of a multigraph, where pairs of nodes can be connected by more than one edge; we can think of multilayer networks as edge-labeled multigraphs. For instance, both orthography and phonology may be used in the same network (<xref ref-type="bibr" rid="B85">Siew &amp; Vitevitch, 2019</xref>), words may be connected via similarity of both form and meaning (<xref ref-type="bibr" rid="B90">Stella, 2020</xref>; <xref ref-type="bibr" rid="B92">Stella et al., 2018</xref>), and other levels of representation can also be incorporated, including features, bigrams, syllables, and even morphemes. While definitions vary, multiplex networks are often described as a type of multilayer network where each layer has the same set of nodes but a different set of edges, allowing the network to represent unique types of interactions in each layer (<xref ref-type="bibr" rid="B56">Kinsley et al., 2020</xref>). Recent work by Castro, Stella, and Siew (<xref ref-type="bibr" rid="B26">2020</xref>) demonstrates the utility of a multiplex lexical network in psycholinguistics: their multiplex network combines semantic and phonological information to help understand the nature of clinical impairments, specifically, word retrieval errors in aphasia. While our analysis finds latent components in phonological and orthographic neighbors in separate single-layer networks, future studies may combine both into a multilayer network, providing for a more comprehensive model of the structure of the lexicon and potentially uncovering interlayer factors that could be absent when analyzing each layer on its own.</p>
</sec>
</sec>
<sec>
<title>5. Conclusion</title>
<p>The novel contribution of our work is that the relationship between nodal properties, measured in phonological and orthographic neighbor networks, and word recognition data can be summarized in three dimensions, reflecting generalized and gradient conceptualizations of density, interconnectedness, and community structure. Regarding density, short words with higher values on most centrality measures led to faster LDT and naming responses. Concerning interconnectedness, words in regions where most words are neighbors of most other words led to slower responses. And in community structure, words linked to multiple, distinct areas of greater interconnectivity were responded to more slowly.</p>
<p>These three dimensions point to a relatively simple notion of (phonological and/or orthographic) structure in the mental lexicon as capturing density and different kinds of unevenness in the distribution of words in similarity space. This helps to contextualize the diverse tools and concepts available in network science within an interpretation of structure that is sensitive to how psycholinguistic phenomena like word recognition interact with that structure. We consider the interpretation of structure that our analysis offers to be, in principle, compatible with prominent theories of word recognition, indicating a need for more modeling work to help clarify this relationship. In particular, the results concerning interconnectedness and community structure suggest that there are advances to be made in exploring why words are unevenly distributed in phonological/orthographic similarity space. We have suggested that this may be due to morphology, which has long been a target of investigation in psycholinguistics, as well as to phonotactics and prosodic structure, which have not been approached with as much sophistication. It is our hope that the underlying dimensions of structure that we have identified will serve to guide this work, leading to a better understanding of what it means for the mental lexicon to be structured, and how structure influences processing.</p>
</sec>
</body>
<back>
<sec>
<title>Abbreviations</title>
<p>BG: mean bigram frequency</p>
<p>DV: dependent variable</p>
<p>ELP: English Lexicon Project (<xref ref-type="bibr" rid="B11">Balota et al., 2007</xref>)</p>
<p>GC: Giant Component</p>
<p>LDT: lexical decision task</p>
<p>MALD: Massive Auditory Lexical Decision (<xref ref-type="bibr" rid="B105">Tucker et al., 2019</xref>)</p>
<p>ND: neighborhood density</p>
<p>RT: response time</p>
<p>SCGLR: Supervised Component Generalized Linear Regression</p>
<p>sc1, sc2, etc.: Supervised Component 1, 2, etc.</p>
<p>UP: uniqueness point</p>
</sec>
<sec>
<title>Data accessibility statement</title>
<p>All data, code, and additional materials are available at <ext-link ext-link-type="uri" xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="https://osf.io/mhjvx/">https://osf.io/mhjvx/</ext-link>.</p>
</sec>
<sec>
<title>Ethics and consent</title>
<p>The data used were from publicly available databases, so additional consent procedures were unnecessary.</p>
</sec>
<sec>
<title>Acknowledgments</title>
<p>An earlier version of the data/analyses was presented at the 2019 Annual Meeting of the Psychonomic Society. DD was funded through a 2020&#8211;2021 William Orr Dingwall Foundation Dissertation Fellowship in the Cognitive, Clinical and Neural Foundations of Language.</p>
</sec>
<sec>
<title>Competing interests</title>
<p>The authors have no competing interests to declare.</p>
</sec>
<sec>
<title>Author contributions</title>
<p>All authors participated in conceptualizing the project, interpreting the results, and revising the manuscript. DD wrote the code for constructing the networks and extracting nodal properties and MTC had primary responsibility for performing the SCGLR analyses, and took the lead in writing and compiling the manuscript.</p>
</sec>
<fn-group>
<fn id="n1"><p>The insertion, deletion, or substitution of a single letter or phoneme (<xref ref-type="bibr" rid="B59">Levenshtein, 1966</xref>).</p></fn>
<fn id="n2"><p>Networks have also been used to explore other domains of linguistic structure, such as semantic or syntactic features or associations (e.g., <xref ref-type="bibr" rid="B13">Beckage &amp; Colunga, 2019</xref>; <xref ref-type="bibr" rid="B36">Ferrer i Cancho et al., 2004</xref>; <xref ref-type="bibr" rid="B57">Kumar et al., 2021</xref>; <xref ref-type="bibr" rid="B95">Steyvers &amp; Tenenbaum, 2005</xref>), or used multiple kinds of edges, i.e., multiplex networks (e.g., phonological and semantic; <xref ref-type="bibr" rid="B91">Stella et al., 2017</xref>). We focus here on word-form networks, but acknowledge that a complete model must include form, meaning, and other levels of linguistic structure.</p></fn>
<fn id="n3"><p>Curiously, Yates&#8217; study crosses phonological network structure with visual word recognition, although his findings are consistent with within-modality results.</p></fn>
<fn id="n4"><p>Additional research explores how network structure emerges from word-learning biases, particularly in children, (<xref ref-type="bibr" rid="B23">Carlson et al., 2014</xref>; <xref ref-type="bibr" rid="B41">Goldstein &amp; Vitevitch, 2014</xref>; <xref ref-type="bibr" rid="B86">Siew &amp; Vitevitch, 2020</xref>), cf. similar work on semantic networks (e.g. <xref ref-type="bibr" rid="B13">Beckage &amp; Colunga, 2019</xref>; <xref ref-type="bibr" rid="B95">Steyvers &amp; Tenenbaum, 2005</xref>).</p></fn>
<fn id="n5"><p>Note that the analysis of behavioral data reported below uses a much smaller sample of words, as described below. This is because not all nodal properties are defined for all words in a neighbor network of the kind we use here. 40,398, thus, refers to the set of words used to construct the network, not to the number of words for which behavioral data will be modeled. We comment on this shortcoming of neighbor networks in Section 4.</p></fn>
<fn id="n6"><p>We note that the MALD stimuli were recorded by a single talker, and, therefore, generalizing the results depends on the assumption that listeners respond to this talker similarly to other talkers. This weakness is common in psycholinguistic research on auditory word recognition, which we acknowledge here.</p></fn>
<fn id="n7"><p>In any case, hierarchical regression, the strategy adopted most frequently in the psycholinguistic literature on neighbor networks (e.g., <xref ref-type="bibr" rid="B23">Carlson et al., 2014</xref>; <xref ref-type="bibr" rid="B82">Siew, 2018</xref>), is not only impractical, given the number of variables we consider here, but it would require many decisions about the stepwise procedure for which there is currently little theoretical basis.</p></fn>
<fn id="n8"><p>We note that frequency is not directly represented in word recognition models, and itself probably reflects the joint effects of several variables. It is, nonetheless, universally controlled for in psycholinguistic studies.</p></fn>
<fn id="n9"><p>Note that a different measure of word length, the duration of the MALD auditory stimuli, loaded on a separate component, sc7, where it accounted for variance in the auditory LDT and naming RTs.</p></fn>
</fn-group>
<ref-list>
<ref id="B1"><mixed-citation publication-type="journal"><string-name><surname>Allopenna</surname>, <given-names>P. D.</given-names></string-name>, <string-name><surname>Magnuson</surname>, <given-names>J. S.</given-names></string-name>, &amp; <string-name><surname>Tanenhaus</surname>, <given-names>M. K.</given-names></string-name> (<year>1998</year>). <article-title>Tracking the time course of spoken word recognition using eye movements: Evidence for continuous mapping models</article-title>. <source>Journal of Memory and Language</source>, <volume>38</volume>(<issue>4</issue>), <fpage>419</fpage>&#8211;<lpage>439</lpage>. <pub-id pub-id-type="doi">10.1006/jmla.1997.2558</pub-id></mixed-citation></ref>
<ref id="B2"><mixed-citation publication-type="journal"><string-name><surname>Altieri</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Gruenenfelder</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Pisoni</surname>, <given-names>D. B.</given-names></string-name> (<year>2010</year>). <article-title>Clustering coefficients of lexical neighborhoods: Does neighborhood structure matter in spoken word recognition?</article-title> <source>The Mental Lexicon</source>, <volume>5</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>21</lpage>. <pub-id pub-id-type="doi">10.1075/ml.5.1.01alt</pub-id></mixed-citation></ref>
<ref id="B3"><mixed-citation publication-type="journal"><string-name><surname>Andrews</surname>, <given-names>S.</given-names></string-name> (<year>1992</year>). <article-title>Frequency and neighborhood effects on lexical access: Lexical similarity or orthographic redundancy?</article-title> <source>Journal of Experimental Psychology: Learning, Memory and Cognition</source>, <volume>18</volume>(<issue>2</issue>), <fpage>234</fpage>&#8211;<lpage>254</lpage>. <pub-id pub-id-type="doi">10.1037/0278-7393.18.2.234</pub-id></mixed-citation></ref>
<ref id="B4"><mixed-citation publication-type="journal"><string-name><surname>Andrews</surname>, <given-names>S.</given-names></string-name> (<year>1997</year>). <article-title>The effect of orthographic similarity on lexical retrieval: Resolving neighborhood conflicts</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>4</volume>(<issue>4</issue>), <fpage>439</fpage>&#8211;<lpage>461</lpage>. <pub-id pub-id-type="doi">10.3758/BF03214334</pub-id></mixed-citation></ref>
<ref id="B5"><mixed-citation publication-type="journal"><string-name><surname>Arbesman</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Strogatz</surname>, <given-names>S. H.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2010a</year>). <article-title>Comparative analysis of networks of phonologically similar words in English and Spanish</article-title>. <source>Entropy</source>, <volume>12</volume>(<issue>3</issue>), <fpage>327</fpage>&#8211;<lpage>337</lpage>. <pub-id pub-id-type="doi">10.3390/e12030327</pub-id></mixed-citation></ref>
<ref id="B6"><mixed-citation publication-type="journal"><string-name><surname>Arbesman</surname>, <given-names>S.</given-names></string-name>, <string-name><surname>Strogatz</surname>, <given-names>S. H.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2010b</year>). <article-title>The structure of phonological networks across multiple languages</article-title>. <source>International Journal of Bifurcations and Chaos</source>, <volume>20</volume>(<issue>3</issue>), <fpage>679</fpage>&#8211;<lpage>685</lpage>. <pub-id pub-id-type="doi">10.1142/S021812741002596X</pub-id></mixed-citation></ref>
<ref id="B7"><mixed-citation publication-type="journal"><string-name><surname>Baayen</surname>, <given-names>R. H.</given-names></string-name> (<year>2010</year>). <article-title>Demythologizing the word frequency effect: A discriminative learning perspective</article-title>. <source>The Mental Lexicon</source>, <volume>5</volume>(<issue>3</issue>), <fpage>436</fpage>&#8211;<lpage>461</lpage>. <pub-id pub-id-type="doi">10.1075/ml.5.3.10baa</pub-id></mixed-citation></ref>
<ref id="B8"><mixed-citation publication-type="journal"><string-name><surname>Baayen</surname>, <given-names>R. H.</given-names></string-name>, <string-name><surname>Chuang</surname>, <given-names>Y.-Y.</given-names></string-name>, <string-name><surname>Shafaei-Bajestan</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Blevins</surname>, <given-names>J. P.</given-names></string-name> (<year>2019</year>). <article-title>The discriminative lexicon: A unified computational model for the lexicon and lexical processing in comprehension and production grounded not in (de)composition but in Linear Discriminative Learning</article-title>. <source>Complexity</source>, <volume>2019</volume>, <elocation-id>e4895891</elocation-id>. <pub-id pub-id-type="doi">10.1155/2019/4895891</pub-id></mixed-citation></ref>
<ref id="B9"><mixed-citation publication-type="journal"><string-name><surname>Bailey</surname>, <given-names>T. M.</given-names></string-name>, &amp; <string-name><surname>Hahn</surname>, <given-names>U.</given-names></string-name> (<year>2001</year>). <article-title>Determinants of wordlikeness: Phonotactics or lexical neighborhoods?</article-title> <source>Journal of Memory and Language</source>, <volume>44</volume>(<issue>4</issue>), <fpage>568</fpage>&#8211;<lpage>591</lpage>. <pub-id pub-id-type="doi">10.1006/jmla.2000.2756</pub-id></mixed-citation></ref>
<ref id="B10"><mixed-citation publication-type="journal"><string-name><surname>Bailey</surname>, <given-names>T. M.</given-names></string-name>, &amp; <string-name><surname>Hahn</surname>, <given-names>U.</given-names></string-name> (<year>2005</year>). <article-title>Phoneme similarity and confusability</article-title>. <source>Journal of Memory and Language</source>, <volume>52</volume>(<issue>3</issue>), <fpage>339</fpage>&#8211;<lpage>362</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2004.12.003</pub-id></mixed-citation></ref>
<ref id="B11"><mixed-citation publication-type="journal"><string-name><surname>Balota</surname>, <given-names>D. A.</given-names></string-name>, <string-name><surname>Yap</surname>, <given-names>M. J.</given-names></string-name>, <string-name><surname>Hutchison</surname>, <given-names>K. A.</given-names></string-name>, <string-name><surname>Cortese</surname>, <given-names>M. J.</given-names></string-name>, <string-name><surname>Kessler</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Loftis</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Neely</surname>, <given-names>J. H.</given-names></string-name>, <string-name><surname>Nelson</surname>, <given-names>D. L.</given-names></string-name>, <string-name><surname>Simpson</surname>, <given-names>G. B.</given-names></string-name>, &amp; <string-name><surname>Treiman</surname>, <given-names>R.</given-names></string-name> (<year>2007</year>). <article-title>The English Lexicon Project</article-title>. <source>Behavior Research Methods</source>, <volume>39</volume>(<issue>3</issue>), <fpage>445</fpage>&#8211;<lpage>459</lpage>. <pub-id pub-id-type="doi">10.3758/BF03193014</pub-id></mixed-citation></ref>
<ref id="B12"><mixed-citation publication-type="journal"><string-name><surname>Baronchelli</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Ferrer-i-Cancho</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Pastor-Satorras</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Chater</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Christiansen</surname>, <given-names>M. H.</given-names></string-name> (<year>2013</year>). <article-title>Networks in cognitive science</article-title>. <source>Trends in Cognitive Sciences</source>, <volume>17</volume>(<issue>7</issue>), <fpage>348</fpage>&#8211;<lpage>360</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2013.04.010</pub-id></mixed-citation></ref>
<ref id="B13"><mixed-citation publication-type="journal"><string-name><surname>Beckage</surname>, <given-names>N. M.</given-names></string-name>, &amp; <string-name><surname>Colunga</surname>, <given-names>E.</given-names></string-name> (<year>2019</year>). <article-title>Network growth modeling to capture individual lexical learning</article-title>. <source>Complexity</source>, <volume>2019</volume>, <fpage>1</fpage>&#8211;<lpage>17</lpage>. <pub-id pub-id-type="doi">10.1155/2019/7690869</pub-id></mixed-citation></ref>
<ref id="B14"><mixed-citation publication-type="journal"><string-name><surname>Benedek</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Kenett</surname>, <given-names>Y. N.</given-names></string-name>, <string-name><surname>Umdasch</surname>, <given-names>K.</given-names></string-name>, <string-name><surname>Anaki</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Faust</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Neubauer</surname>, <given-names>A. C.</given-names></string-name> (<year>2017</year>). <article-title>How semantic memory structure and intelligence contribute to creative thought: A network science approach</article-title>. <source>Thinking &amp; Reasoning</source>, <volume>23</volume>(<issue>2</issue>), <fpage>158</fpage>&#8211;<lpage>183</lpage>. <pub-id pub-id-type="doi">10.1080/13546783.2016.1278034</pub-id></mixed-citation></ref>
<ref id="B15"><mixed-citation publication-type="journal"><string-name><surname>Blondel</surname>, <given-names>V. D.</given-names></string-name>, <string-name><surname>Guillaume</surname>, <given-names>J.-L.</given-names></string-name>, <string-name><surname>Lambiotte</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Lefebvre</surname>, <given-names>E.</given-names></string-name> (<year>2008</year>). <article-title>Fast unfolding of communities in large networks</article-title>. <source>Journal of Statistical Mechanics: Theory and Experiment</source>, <volume>2008</volume>(<issue>10</issue>), <elocation-id>P10008</elocation-id>. <pub-id pub-id-type="doi">10.1088/1742-5468/2008/10/p10008</pub-id></mixed-citation></ref>
<ref id="B16"><mixed-citation publication-type="journal"><string-name><surname>Borgatti</surname>, <given-names>S. P.</given-names></string-name> (<year>2005</year>). <article-title>Centrality and network flow</article-title>. <source>Social Networks</source>, <volume>27</volume>(<issue>1</issue>), <fpage>55</fpage>&#8211;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1016/j.socnet.2004.11.008</pub-id></mixed-citation></ref>
<ref id="B17"><mixed-citation publication-type="journal"><string-name><surname>Brin</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Page</surname>, <given-names>L.</given-names></string-name> (<year>1998</year>). <article-title>The anatomy of a large-scale hypertextual web search engine</article-title>. <source>Computer Networks</source>, <volume>30</volume>, <fpage>107</fpage>&#8211;<lpage>117</lpage>. <pub-id pub-id-type="doi">10.1016/S0169-7552(98)00110-X</pub-id></mixed-citation></ref>
<ref id="B18"><mixed-citation publication-type="journal"><string-name><surname>Brown</surname>, <given-names>K. S.</given-names></string-name>, <string-name><surname>Allopenna</surname>, <given-names>P. D.</given-names></string-name>, <string-name><surname>Hunt</surname>, <given-names>W. R.</given-names></string-name>, <string-name><surname>Steiner</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Saltzman</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>McRae</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Magnuson</surname>, <given-names>J. S.</given-names></string-name> (<year>2018</year>). <article-title>Universal features in phonological neighbor networks</article-title>. <source>Entropy</source>, <volume>20</volume>(<issue>7</issue>), Article <elocation-id>7</elocation-id>. <pub-id pub-id-type="doi">10.3390/e20070526</pub-id></mixed-citation></ref>
<ref id="B19"><mixed-citation publication-type="journal"><string-name><surname>Bry</surname>, <given-names>X.</given-names></string-name>, <string-name><surname>Trottier</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Verron</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Mortier</surname>, <given-names>F.</given-names></string-name> (<year>2013</year>). <article-title>Supervised component generalized linear regression using a PLS-extension of the Fisher scoring algorithm</article-title>. <source>Journal of Multivariate Analysis</source>, <volume>119</volume>, <fpage>47</fpage>&#8211;<lpage>60</lpage>. <pub-id pub-id-type="doi">10.1016/j.jmva.2013.03.013</pub-id></mixed-citation></ref>
<ref id="B20"><mixed-citation publication-type="journal"><string-name><surname>Brysbaert</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Stevens</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Mandera</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Keuleers</surname>, <given-names>E.</given-names></string-name> (<year>2016</year>). <article-title>How many words do we know? Practical estimates of vocabulary size dependent on word definition, the degree of language input and the participant&#8217;s age</article-title>. <source>Frontiers in Psychology</source>, <volume>7</volume>, <elocation-id>1116</elocation-id>. <pub-id pub-id-type="doi">10.3389/fpsyg.2016.01116</pub-id></mixed-citation></ref>
<ref id="B21"><mixed-citation publication-type="journal"><string-name><surname>Bullmore</surname>, <given-names>E.</given-names></string-name>, &amp; <string-name><surname>Sporns</surname>, <given-names>O.</given-names></string-name> (<year>2009</year>). <article-title>Complex brain networks: Graph theoretical analysis of structural and functional systems</article-title>. <source>Nature Reviews Neuroscience</source>, <volume>10</volume>(<issue>3</issue>), <fpage>186</fpage>&#8211;<lpage>198</lpage>. <pub-id pub-id-type="doi">10.1038/nrn2575</pub-id></mixed-citation></ref>
<ref id="B22"><mixed-citation publication-type="book"><string-name><surname>Bybee</surname>, <given-names>J.</given-names></string-name> (<year>2006</year>). <chapter-title>From usage to grammar: The mind&#8217;s response to repetition</chapter-title>. <source>Language</source>, <volume>82</volume>(<issue>4</issue>), <fpage>711</fpage>&#8211;<lpage>733</lpage>. <publisher-name>JSTOR</publisher-name>. <pub-id pub-id-type="doi">10.1353/lan.2006.0186</pub-id></mixed-citation></ref>
<ref id="B23"><mixed-citation publication-type="journal"><string-name><surname>Carlson</surname>, <given-names>M. T.</given-names></string-name>, <string-name><surname>Sonderegger</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Bane</surname>, <given-names>M.</given-names></string-name> (<year>2014</year>). <article-title>How children explore the phonological network in child-directed speech: A survival analysis of children&#8217;s first word productions</article-title>. <source>Journal of Memory and Language</source>, <volume>75</volume>, <fpage>159</fpage>&#8211;<lpage>180</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2014.05.005</pub-id></mixed-citation></ref>
<ref id="B24"><mixed-citation publication-type="journal"><string-name><surname>Caselli</surname>, <given-names>N. K.</given-names></string-name>, <string-name><surname>Emmorey</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Cohen-Goldberg</surname>, <given-names>A. M.</given-names></string-name> (<year>2021</year>). <article-title>The signed mental lexicon: Effects of phonological neighborhood density, iconicity, and childhood language experience</article-title>. <source>Journal of Memory and Language</source>, <volume>121</volume>, <elocation-id>104282</elocation-id>. <pub-id pub-id-type="doi">10.1016/j.jml.2021.104282</pub-id></mixed-citation></ref>
<ref id="B25"><mixed-citation publication-type="journal"><string-name><surname>Castro</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name> (<year>2020</year>). <article-title>Contributions of modern network science to the cognitive sciences: Revisiting research spirals of representation and process</article-title>. <source>Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences</source>, <volume>476</volume>(<issue>2238</issue>), <elocation-id>20190825</elocation-id>. <pub-id pub-id-type="doi">10.1098/rspa.2019.0825</pub-id></mixed-citation></ref>
<ref id="B26"><mixed-citation publication-type="journal"><string-name><surname>Castro</surname>, <given-names>N.</given-names></string-name>, <string-name><surname>Stella</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name> (<year>2020</year>). <article-title>Quantifying the interplay of semantics and phonology during failures of word retrieval by people with aphasia using a multiplex lexical network</article-title>. <source>Cognitive Science</source>, <volume>44</volume>(<issue>9</issue>). <pub-id pub-id-type="doi">10.1111/cogs.12881</pub-id></mixed-citation></ref>
<ref id="B27"><mixed-citation publication-type="journal"><string-name><surname>Castro</surname>, <given-names>N.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2022</year>). <article-title>Using network science and psycholinguistic megastudies to examine the dimensions of phonological similarity</article-title>. <source>Language and Speech</source>, <italic>32</italic>. <pub-id pub-id-type="doi">10.1177/00238309221095455</pub-id></mixed-citation></ref>
<ref id="B28"><mixed-citation publication-type="journal"><string-name><surname>Chan</surname>, <given-names>K. Y.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2009</year>). <article-title>The influence of the phonological neighborhood clustering coefficient on spoken word recognition</article-title>. <source>Journal of Experimental Psychology. Human Perception and Performance</source>, <volume>35</volume>, <fpage>1934</fpage>&#8211;<lpage>1949</lpage>. <pub-id pub-id-type="doi">10.1037/a0016902</pub-id></mixed-citation></ref>
<ref id="B29"><mixed-citation publication-type="journal"><string-name><surname>Chan</surname>, <given-names>K. Y.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2010</year>). <article-title>Network structure influences speech production</article-title>. <source>Cognitive Science</source>, <volume>34</volume>(<issue>4</issue>), <fpage>685</fpage>&#8211;<lpage>697</lpage>. <pub-id pub-id-type="doi">10.1111/j.1551-6709.2010.01100.x</pub-id></mixed-citation></ref>
<ref id="B30"><mixed-citation publication-type="journal"><string-name><surname>Charles-Luce</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name> (<year>1990</year>). <article-title>Similarity neighbourhoods of words in young children&#8217;s lexicons</article-title>. <source>Journal of Child Language</source>, <volume>17</volume>(<issue>1</issue>), <fpage>205</fpage>&#8211;<lpage>215</lpage>. <pub-id pub-id-type="doi">10.1017/S0305000900013180</pub-id></mixed-citation></ref>
<ref id="B31"><mixed-citation publication-type="journal"><string-name><surname>Charles-Luce</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name> (<year>1995</year>). <article-title>An examination of similarity neighbourhoods in young children&#8217;s receptive vocabularies</article-title>. <source>Journal of Child Language</source>, <volume>22</volume>(<issue>3</issue>), <fpage>727</fpage>&#8211;<lpage>735</lpage>. <pub-id pub-id-type="doi">10.1017/S0305000900010023</pub-id></mixed-citation></ref>
<ref id="B32"><mixed-citation publication-type="journal"><string-name><surname>Chen</surname>, <given-names>Q.</given-names></string-name>, &amp; <string-name><surname>Mirman</surname>, <given-names>D.</given-names></string-name> (<year>2012</year>). <article-title>Competition and cooperation among similar representations: Toward a unified account of facilitative and inhibitory effects of lexical neighbors</article-title>. <source>Psychological Review</source>, <volume>119</volume>(<issue>2</issue>), <fpage>417</fpage>&#8211;<lpage>430</lpage>. <pub-id pub-id-type="doi">10.1037/a0027175</pub-id></mixed-citation></ref>
<ref id="B33"><mixed-citation publication-type="journal"><string-name><surname>Coady</surname>, <given-names>J. A.</given-names></string-name>, &amp; <string-name><surname>Aslin</surname>, <given-names>R. N.</given-names></string-name> (<year>2003</year>). <article-title>Phonological neighborhoods in the developing lexicon</article-title>. <source>Journal of Child Language</source>, <volume>30</volume>(<issue>2</issue>), <fpage>441</fpage>&#8211;<lpage>469</lpage>. <pub-id pub-id-type="doi">10.1017/S0305000903005579</pub-id></mixed-citation></ref>
<ref id="B34"><mixed-citation publication-type="book"><string-name><surname>Coltheart</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Davelaar</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Jonasson</surname>, <given-names>J. T.</given-names></string-name>, &amp; <string-name><surname>Besner</surname>, <given-names>D.</given-names></string-name> (<year>1977</year>). <chapter-title>Access to the internal lexicon</chapter-title>. In <string-name><given-names>S.</given-names> <surname>Dornic&#774;</surname></string-name> (Ed.), <source>Attention and performance VI: Proceedings of the Sixth International Symposium on Attention and Performance, Stockholm, Sweden, July 28&#8211;August 1, 1975</source> (pp. <fpage>535</fpage>&#8211;<lpage>555</lpage>). <publisher-name>Routledge</publisher-name>. <pub-id pub-id-type="doi">10.4324/9781003309734-29</pub-id></mixed-citation></ref>
<ref id="B35"><mixed-citation publication-type="journal"><string-name><surname>Cutler</surname>, <given-names>A.</given-names></string-name> (<year>1981</year>). <article-title>Making up materials is a confounded nuisance, or: Will we able to run any psycholinguistic experiments at all in 1990?</article-title> <source>Cognition</source>, <volume>10</volume>(<issue>1</issue>), <fpage>65</fpage>&#8211;<lpage>70</lpage>. <pub-id pub-id-type="doi">10.1016/0010-0277(81)90026-3</pub-id></mixed-citation></ref>
<ref id="B36"><mixed-citation publication-type="journal"><string-name><surname>Ferrer i Cancho</surname>, <given-names>R.</given-names></string-name>, <string-name><surname>Sol&#233;</surname>, <given-names>R. V.</given-names></string-name>, &amp; <string-name><surname>K&#246;hler</surname>, <given-names>R.</given-names></string-name> (<year>2004</year>). <article-title>Patterns in syntactic dependency networks</article-title>. <source>Physical Review E</source>, <volume>69</volume>(<issue>5</issue>), <elocation-id>051915</elocation-id>. <pub-id pub-id-type="doi">10.1103/PhysRevE.69.051915</pub-id></mixed-citation></ref>
<ref id="B37"><mixed-citation publication-type="book"><string-name><surname>Fricke</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Baese-Berk</surname>, <given-names>M. M.</given-names></string-name>, &amp; <string-name><surname>Goldrick</surname>, <given-names>M.</given-names></string-name> (<year>2016</year>). <chapter-title>Dimensions of similarity in the mental lexicon</chapter-title>. In <source>Language, Cognition, and Neuroscience</source> <volume>31</volume>(<issue>5</issue>), <fpage>639</fpage>&#8211;<lpage>645</lpage>. <pub-id pub-id-type="doi">10.1080/23273798.2015.1130234</pub-id></mixed-citation></ref>
<ref id="B38"><mixed-citation publication-type="journal"><string-name><surname>Goh</surname>, <given-names>W. D.</given-names></string-name>, <string-name><surname>Su&#225;rez</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Yap</surname>, <given-names>M. J.</given-names></string-name>, &amp; <string-name><surname>Tan</surname>, <given-names>S. H.</given-names></string-name> (<year>2009</year>). <article-title>Distributional analyses in auditory lexical decision: Neighborhood density and word-frequency effects</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>16</volume>(<issue>5</issue>), <elocation-id>882</elocation-id>. <pub-id pub-id-type="doi">10.3758/PBR.16.5.882</pub-id></mixed-citation></ref>
<ref id="B39"><mixed-citation publication-type="journal"><string-name><surname>Goh</surname>, <given-names>W. D.</given-names></string-name>, <string-name><surname>Yap</surname>, <given-names>M. J.</given-names></string-name>, &amp; <string-name><surname>Chee</surname>, <given-names>Q. W.</given-names></string-name> (<year>2020</year>). <article-title>The Auditory English Lexicon Project: A multi-talker, multi-region psycholinguistic database of 10,170 spoken words and nonwords</article-title>. <source>Behavior Research Methods</source>, <volume>52</volume>(<issue>5</issue>), <fpage>2202</fpage>&#8211;<lpage>2231</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-020-01352-0</pub-id></mixed-citation></ref>
<ref id="B40"><mixed-citation publication-type="journal"><string-name><surname>Goldrick</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Folk</surname>, <given-names>J. R.</given-names></string-name>, &amp; <string-name><surname>Rapp</surname>, <given-names>B.</given-names></string-name> (<year>2010</year>). <article-title>Mrs. Malaprop&#8217;s neighborhood: Using word errors to reveal neighborhood structure</article-title>. <source>Journal of Memory and Language</source>, <volume>62</volume>(<issue>2</issue>), <fpage>113</fpage>&#8211;<lpage>134</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2009.11.008</pub-id></mixed-citation></ref>
<ref id="B41"><mixed-citation publication-type="book"><string-name><surname>Goldstein</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2014</year>). <chapter-title>The influence of clustering coefficient on word-learning: How groups of similar sounding words facilitate acquisition</chapter-title>. <source>Frontiers in Psychology</source>, <volume>5</volume>, <fpage>1307</fpage>&#8211;<lpage>1307</lpage>. PubMed. <pub-id pub-id-type="doi">10.3389/fpsyg.2014.01307</pub-id></mixed-citation></ref>
<ref id="B42"><mixed-citation publication-type="journal"><string-name><surname>Goldstein</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2017</year>). <article-title>The influence of closeness centrality on lexical processing</article-title>. <source>Frontiers in Psychology</source>, <volume>8</volume>. <pub-id pub-id-type="doi">10.3389/fpsyg.2017.01683</pub-id></mixed-citation></ref>
<ref id="B43"><mixed-citation publication-type="journal"><string-name><surname>Griffiths</surname>, <given-names>T. L.</given-names></string-name>, <string-name><surname>Steyvers</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Firl</surname>, <given-names>A.</given-names></string-name> (<year>2007</year>). <article-title>Google and the mind: Predicting fluency with PageRank</article-title>. <source>Psychological Science</source>, <volume>18</volume>(<issue>12</issue>), <fpage>1069</fpage>&#8211;<lpage>1076</lpage>. <pub-id pub-id-type="doi">10.1111/j.1467-9280.2007.02027.x</pub-id></mixed-citation></ref>
<ref id="B44"><mixed-citation publication-type="journal"><string-name><surname>Grolmusz</surname>, <given-names>V.</given-names></string-name> (<year>2012</year>). <article-title>A note on the PageRank of undirected graphs</article-title>. <source>arXiv:1205.1960 [Cs]</source>. <pub-id pub-id-type="doi">10.1016/j.ipl.2015.02.015</pub-id></mixed-citation></ref>
<ref id="B45"><mixed-citation publication-type="journal"><string-name><surname>Gruenenfelder</surname>, <given-names>T. M.</given-names></string-name>, &amp; <string-name><surname>Pisoni</surname>, <given-names>D. B.</given-names></string-name> (<year>2009</year>). <article-title>The Lexical Restructuring Hypothesis and graph theoretic analyses of networks based on random lexicons</article-title>. <source>Journal of Speech, Language, and Hearing Research</source>, <volume>52</volume>(<issue>3</issue>), <fpage>596</fpage>&#8211;<lpage>609</lpage>. <pub-id pub-id-type="doi">10.1044/1092-4388(2009/08-0004)</pub-id></mixed-citation></ref>
<ref id="B46"><mixed-citation publication-type="journal"><string-name><surname>Hahn</surname>, <given-names>U.</given-names></string-name>, &amp; <string-name><surname>Bailey</surname>, <given-names>T. M.</given-names></string-name> (<year>2005</year>). <article-title>What makes words sound similar?</article-title> <source>Cognition</source>, <volume>97</volume>(<issue>3</issue>), <fpage>227</fpage>&#8211;<lpage>267</lpage>. <pub-id pub-id-type="doi">10.1016/j.cognition.2004.09.006</pub-id></mixed-citation></ref>
<ref id="B47"><mixed-citation publication-type="book"><string-name><surname>Harley</surname>, <given-names>T. A.</given-names></string-name>, &amp; <string-name><surname>Brown</surname>, <given-names>H. E.</given-names></string-name> (<year>1998</year>). <chapter-title>What causes a tip-of-the-tongue state? Evidence for lexical neighbourhood effects in speech production</chapter-title>. <source>British Journal of Psychology</source>, <volume>89</volume>(<issue>1</issue>), <elocation-id>151</elocation-id>. <publisher-name>Periodicals Archive Online; Periodicals Index Online</publisher-name>. <pub-id pub-id-type="doi">10.1111/j.2044-8295.1998.tb02677.x</pub-id></mixed-citation></ref>
<ref id="B48"><mixed-citation publication-type="journal"><string-name><surname>Hay</surname>, <given-names>J.</given-names></string-name>, &amp; <string-name><surname>Baayen</surname>, <given-names>R. H.</given-names></string-name> (<year>2005</year>). <article-title>Shifting paradigms: Gradient structure in morphology</article-title>. <source>Trends in Cognitive Sciences</source>, <volume>9</volume>(<issue>7</issue>), <fpage>342</fpage>&#8211;<lpage>348</lpage>. <pub-id pub-id-type="doi">10.1016/j.tics.2005.04.002</pub-id></mixed-citation></ref>
<ref id="B49"><mixed-citation publication-type="journal"><string-name><surname>Heller</surname>, <given-names>J. R.</given-names></string-name>, &amp; <string-name><surname>Goldrick</surname>, <given-names>M.</given-names></string-name> (<year>2014</year>). <article-title>Grammatical constraints on phonological encoding in speech production</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>21</volume>(<issue>6</issue>), <fpage>1576</fpage>&#8211;<lpage>1582</lpage>. <pub-id pub-id-type="doi">10.3758/s13423-014-0616-3</pub-id></mixed-citation></ref>
<ref id="B50"><mixed-citation publication-type="journal"><string-name><surname>Hills</surname>, <given-names>T. T.</given-names></string-name>, &amp; <string-name><surname>Kenett</surname>, <given-names>Y. N.</given-names></string-name> (<year>2022</year>). <article-title>Is the mind a network? Maps, vehicles, and skyhooks in cognitive network science</article-title>. <source>Topics in Cognitive Science</source>, <volume>14</volume>(<issue>1</issue>), <fpage>189</fpage>&#8211;<lpage>208</lpage>. <pub-id pub-id-type="doi">10.1111/tops.12570</pub-id></mixed-citation></ref>
<ref id="B51"><mixed-citation publication-type="journal"><string-name><surname>Humphries</surname>, <given-names>M. D.</given-names></string-name>, &amp; <string-name><surname>Gurney</surname>, <given-names>K.</given-names></string-name> (<year>2008</year>). <article-title>Network &#8216;small-world-ness&#8217;: A quantitative method for determining canonical network equivalence</article-title>. <source>PLOS ONE</source>, <volume>3</volume>(<issue>4</issue>), <elocation-id>e0002051</elocation-id>. <pub-id pub-id-type="doi">10.1371/journal.pone.0002051</pub-id></mixed-citation></ref>
<ref id="B52"><mixed-citation publication-type="journal"><string-name><surname>Iverson</surname>, <given-names>P.</given-names></string-name>, <string-name><surname>Bernstein</surname>, <given-names>L. E.</given-names></string-name>, &amp; <string-name><surname>Auer</surname> <suffix>Jr</suffix>, <given-names>E. T.</given-names></string-name> (<year>1998</year>). <article-title>Modeling the interaction of phonemic intelligibility and lexical structure in audiovisual word recognition</article-title>. <source>Speech Communication</source>, <volume>26</volume>(<issue>1&#8211;2</issue>), <fpage>45</fpage>&#8211;<lpage>63</lpage>. <pub-id pub-id-type="doi">10.1016/S0167-6393(98)00049-1</pub-id></mixed-citation></ref>
<ref id="B53"><mixed-citation publication-type="book"><string-name><surname>Jackendoff</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Audring</surname>, <given-names>J.</given-names></string-name> (<year>2020</year>). <source>The texture of the lexicon: Relational morphology and the parallel architecture</source>. <publisher-name>Oxford University Press</publisher-name>. <pub-id pub-id-type="doi">10.1093/oso/9780198827900.001.0001</pub-id></mixed-citation></ref>
<ref id="B54"><mixed-citation publication-type="book"><string-name><surname>Kapatsinski</surname>, <given-names>V.</given-names></string-name> (<year>2018</year>). <source>Changing minds changing tools: From learning theory to language acquisition to language change</source>. <publisher-name>MIT Press</publisher-name>. <pub-id pub-id-type="doi">10.7551/mitpress/9780262037860.001.0001</pub-id></mixed-citation></ref>
<ref id="B55"><mixed-citation publication-type="journal"><string-name><surname>Kessler</surname>, <given-names>B.</given-names></string-name>, <string-name><surname>Treiman</surname>, <given-names>R.</given-names></string-name>, &amp; <string-name><surname>Mullennix</surname>, <given-names>J.</given-names></string-name> (<year>2002</year>). <article-title>Phonetic biases in voice key response time measurements</article-title>. <source>Journal of Memory and Language</source>, <volume>47</volume>(<issue>1</issue>), <fpage>145</fpage>&#8211;<lpage>171</lpage>. <pub-id pub-id-type="doi">10.1006/jmla.2001.2835</pub-id></mixed-citation></ref>
<ref id="B56"><mixed-citation publication-type="journal"><string-name><surname>Kinsley</surname>, <given-names>A. C.</given-names></string-name>, <string-name><surname>Rossi</surname>, <given-names>G.</given-names></string-name>, <string-name><surname>Silk</surname>, <given-names>M. J.</given-names></string-name>, &amp; <string-name><surname>VanderWaal</surname>, <given-names>K.</given-names></string-name> (<year>2020</year>). <article-title>Multilayer and multiplex networks: An introduction to their use in veterinary epidemiology</article-title>. <source>Frontiers in Veterinary Science</source>, <volume>7</volume>. <pub-id pub-id-type="doi">10.3389/fvets.2020.00596</pub-id></mixed-citation></ref>
<ref id="B57"><mixed-citation publication-type="book"><string-name><surname>Kumar</surname>, <given-names>A. A.</given-names></string-name>, <string-name><surname>Steyvers</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Balota</surname>, <given-names>D. A.</given-names></string-name> (<year>2021</year>). <chapter-title>A critical review of network-based and distributional approaches to semantic memory structure and processes</chapter-title>. <source>Topics in Cognitive Science</source>, <publisher-name>tops.12548</publisher-name>. <pub-id pub-id-type="doi">10.1111/tops.12548</pub-id></mixed-citation></ref>
<ref id="B58"><mixed-citation publication-type="journal"><string-name><surname>Landauer</surname>, <given-names>T. K.</given-names></string-name>, &amp; <string-name><surname>Streeter</surname>, <given-names>L. A.</given-names></string-name> (<year>1973</year>). <article-title>Structural differences between common and rare words: Failure of equivalence assumptions for theories of word recognition</article-title>. <source>Journal of Verbal Learning and Verbal Behavior</source>, <volume>12</volume>(<issue>2</issue>), <fpage>119</fpage>&#8211;<lpage>131</lpage>. <pub-id pub-id-type="doi">10.1016/S0022-5371(73)80001-5</pub-id></mixed-citation></ref>
<ref id="B59"><mixed-citation publication-type="journal"><string-name><surname>Levenshtein</surname>, <given-names>V. I.</given-names></string-name> (<year>1966</year>). <article-title>Binary codes capable of correcting deletions, insertions, and reversals</article-title>. <source>Soviet Physics Doklady</source>, <volume>10</volume>(<issue>8</issue>), <fpage>707</fpage>&#8211;<lpage>710</lpage>.</mixed-citation></ref>
<ref id="B60"><mixed-citation publication-type="journal"><string-name><surname>Liben-Nowell</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Strand</surname>, <given-names>J. F.</given-names></string-name>, <string-name><surname>Sharp</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Wexler</surname>, <given-names>T.</given-names></string-name>, &amp; <string-name><surname>Woods</surname>, <given-names>K.</given-names></string-name> (<year>2019</year>). <article-title>The danger of testing by selecting controlled subsets, with applications to spoken-word recognition</article-title>. <source>Journal of Cognition</source>, <volume>2</volume>(<issue>1</issue>), Article <elocation-id>1</elocation-id>. <pub-id pub-id-type="doi">10.5334/joc.51</pub-id></mixed-citation></ref>
<ref id="B61"><mixed-citation publication-type="journal"><string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name>, <string-name><surname>Goldinger</surname>, <given-names>S. D.</given-names></string-name>, <string-name><surname>Auer</surname>, <given-names>E. T.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2000</year>). <article-title>Phonetic priming, neighborhood activation, and PARSYN</article-title>. <source>Perception &amp; Psychophysics</source>, <volume>62</volume>(<issue>3</issue>), <fpage>615</fpage>&#8211;<lpage>625</lpage>. <pub-id pub-id-type="doi">10.3758/BF03212113</pub-id></mixed-citation></ref>
<ref id="B62"><mixed-citation publication-type="journal"><string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name>, &amp; <string-name><surname>Large</surname>, <given-names>N. R.</given-names></string-name> (<year>2001</year>). <article-title>Phonotactics, density, and entropy in spoken word recognition</article-title>. <source>Language and Cognitive Processes</source>, <volume>16</volume>(<issue>5/6</issue>), <fpage>565</fpage>&#8211;<lpage>581</lpage>. <pub-id pub-id-type="doi">10.1080/01690960143000137</pub-id></mixed-citation></ref>
<ref id="B63"><mixed-citation publication-type="journal"><string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name>, &amp; <string-name><surname>Pisoni</surname>, <given-names>D. B.</given-names></string-name> (<year>1998</year>). <article-title>Recognizing spoken words: The Neighborhood Activation Model</article-title>. <source>Ear and Hearing</source>, <volume>19</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>36</lpage>. <pub-id pub-id-type="doi">10.1097/00003446-199802000-00001</pub-id></mixed-citation></ref>
<ref id="B64"><mixed-citation publication-type="journal"><string-name><surname>Lund</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Burgess</surname>, <given-names>C.</given-names></string-name> (<year>1996</year>). <article-title>Producing high-dimensional semantic spaces from lexical co-occurrence</article-title>. <source>Behavior Research Methods, Instruments, &amp; Computers</source>, <volume>28</volume>(<issue>2</issue>), <fpage>203</fpage>&#8211;<lpage>208</lpage>. <pub-id pub-id-type="doi">10.3758/BF03204766</pub-id></mixed-citation></ref>
<ref id="B65"><mixed-citation publication-type="journal"><string-name><surname>Marslen-Wilson</surname>, <given-names>W. D.</given-names></string-name>, &amp; <string-name><surname>Tyler</surname>, <given-names>L. K.</given-names></string-name> (<year>1980</year>). <article-title>The temporal structure of spoken language understanding</article-title>. <source>Cognition</source>, <volume>8</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>71</lpage>. <pub-id pub-id-type="doi">10.1016/0010-0277(80)90015-3</pub-id></mixed-citation></ref>
<ref id="B66"><mixed-citation publication-type="book"><collab>MATLAB</collab>. (<year>2010</year>). <source>MATLAB</source> (Version 7.10.0 (R2010a)) [Computer software]. <publisher-name>The MathWorks, Inc</publisher-name>.</mixed-citation></ref>
<ref id="B67"><mixed-citation publication-type="journal"><string-name><surname>McClelland</surname>, <given-names>J. L.</given-names></string-name>, &amp; <string-name><surname>Elman</surname>, <given-names>J. L.</given-names></string-name> (<year>1986</year>). <article-title>The TRACE model of speech perception</article-title>. <source>Cognitive Psychology</source>, <volume>18</volume>(<issue>1</issue>), <fpage>1</fpage>&#8211;<lpage>86</lpage>. <pub-id pub-id-type="doi">10.1016/0010-0285(86)90015-0</pub-id></mixed-citation></ref>
<ref id="B68"><mixed-citation publication-type="journal"><string-name><surname>McClelland</surname>, <given-names>J. L.</given-names></string-name>, &amp; <string-name><surname>Rumelhart</surname>, <given-names>D. E.</given-names></string-name> (<year>1981</year>). <article-title>An interactive activation model of context effects in letter perception: I. An account of basic findings</article-title>. <source>Psychological Review</source>, <volume>88</volume>(<issue>5</issue>), <fpage>375</fpage>&#8211;<lpage>407</lpage>. <pub-id pub-id-type="doi">10.1037/0033-295X.88.5.375</pub-id></mixed-citation></ref>
<ref id="B69"><mixed-citation publication-type="journal"><string-name><surname>Metsala</surname>, <given-names>J. L.</given-names></string-name> (<year>1997</year>). <article-title>An examination of word frequency and neighborhood density in the development of spoken-word recognition</article-title>. <source>Memory &amp; Cognition</source>, <volume>25</volume>(<issue>1</issue>), <fpage>47</fpage>&#8211;<lpage>56</lpage>. <pub-id pub-id-type="doi">10.3758/BF03197284</pub-id></mixed-citation></ref>
<ref id="B70"><mixed-citation publication-type="book"><string-name><surname>Metsala</surname>, <given-names>J. L.</given-names></string-name>, &amp; <string-name><surname>Walley</surname>, <given-names>A. C.</given-names></string-name> (<year>1998</year>). <chapter-title>Spoken vocabulary growth and the segmental restructuring of lexical representations: Precursors to phonemic awareness and early reading ability</chapter-title>. In <string-name><given-names>J. L.</given-names> <surname>Metsala</surname></string-name> &amp; <string-name><given-names>L. C.</given-names> <surname>Ehri</surname></string-name> (Eds.), <source>Word recognition in beginning literacy</source> (pp. <fpage>89</fpage>&#8211;<lpage>120</lpage>). <publisher-name>Lawrence Erlbaum Associates</publisher-name>.</mixed-citation></ref>
<ref id="B71"><mixed-citation publication-type="journal"><string-name><surname>Morrissey</surname>, <given-names>M. B.</given-names></string-name>, &amp; <string-name><surname>Ruxton</surname>, <given-names>G. D.</given-names></string-name> (<year>2018</year>). <article-title>Multiple regression is not multiple regressions: The meaning of multiple regression and the non-problem of collinearity</article-title>. <source>Philosophy, Theory, and Practice in Biology</source>, <volume>10</volume>(<issue>3</issue>). <pub-id pub-id-type="doi">10.3998/ptpbio.16039257.0010.003</pub-id></mixed-citation></ref>
<ref id="B72"><mixed-citation publication-type="journal"><string-name><surname>Mortier</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Trottier</surname>, <given-names>C.</given-names></string-name>, <string-name><surname>Cornu</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Bry</surname>, <given-names>X.</given-names></string-name> (<year>2015</year>). <source>SCGLR &#8211; An R package for Supervised Component Generalized Linear Regression</source>, <elocation-id>13</elocation-id>.</mixed-citation></ref>
<ref id="B73"><mixed-citation publication-type="journal"><string-name><surname>Norris</surname>, <given-names>D.</given-names></string-name> (<year>1994</year>). <article-title>Shortlist: A connectionist model of continuous speech recognition</article-title>. <source>Cognition</source>, <volume>52</volume>(<issue>3</issue>), <fpage>189</fpage>&#8211;<lpage>234</lpage>. <pub-id pub-id-type="doi">10.1016/0010-0277(94)90043-4</pub-id></mixed-citation></ref>
<ref id="B74"><mixed-citation publication-type="journal"><string-name><surname>Norris</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>McQueen</surname>, <given-names>J. M.</given-names></string-name> (<year>2008</year>). <article-title>Shortlist B: A Bayesian model of continuous speech recognition</article-title>. <source>Psychological Review</source>, <volume>115</volume>(<issue>2</issue>), <fpage>357</fpage>&#8211;<lpage>395</lpage>. <pub-id pub-id-type="doi">10.1037/0033-295X.115.2.357</pub-id></mixed-citation></ref>
<ref id="B75"><mixed-citation publication-type="webpage"><collab>R Core Team</collab>. (<year>2023</year>). <source>R: A language and environment for statistical computing</source> [Computer software]. <publisher-name>R Foundation for Statistical Computing</publisher-name>. URL <uri>http://www.R-project.org/</uri></mixed-citation></ref>
<ref id="B76"><mixed-citation publication-type="journal"><string-name><surname>Rastle</surname>, <given-names>K.</given-names></string-name>, &amp; <string-name><surname>Davis</surname>, <given-names>M. H.</given-names></string-name> (<year>2002</year>). <article-title>On the complexities of measuring naming</article-title>. <source>Journal of Experimental Psychology: Human Perception and Performance</source>, <volume>28</volume>(<issue>2</issue>), <fpage>307</fpage>&#8211;<lpage>314</lpage>. <pub-id pub-id-type="doi">10.1037/0096-1523.28.2.307</pub-id></mixed-citation></ref>
<ref id="B77"><mixed-citation publication-type="journal"><string-name><surname>Roux</surname>, <given-names>S.</given-names></string-name>, &amp; <string-name><surname>Bonin</surname>, <given-names>P.</given-names></string-name> (<year>2009</year>). <article-title>Neighborhood effects in spelling in adults</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>16</volume>(<issue>2</issue>), <fpage>369</fpage>&#8211;<lpage>373</lpage>. <pub-id pub-id-type="doi">10.3758/PBR.16.2.369</pub-id></mixed-citation></ref>
<ref id="B78"><mixed-citation publication-type="journal"><string-name><surname>Rubinov</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Sporns</surname>, <given-names>O.</given-names></string-name> (<year>2010</year>). <article-title>Complex network measures of brain connectivity: Uses and interpretations</article-title>. <source>Neuroimage</source>, <volume>52</volume>(<issue>3</issue>), <fpage>1059</fpage>&#8211;<lpage>1069</lpage>. <pub-id pub-id-type="doi">10.1016/j.neuroimage.2009.10.003</pub-id></mixed-citation></ref>
<ref id="B79"><mixed-citation publication-type="journal"><string-name><surname>Shafaei-Bajestan</surname>, <given-names>E.</given-names></string-name>, <string-name><surname>Moradipour-Tari</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Uhrig</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Baayen</surname>, <given-names>R. H.</given-names></string-name> (<year>2023</year>). <article-title>LDL-AURIS: A computational model, grounded in error-driven learning, for the comprehension of single spoken words</article-title>. <source>Language, Cognition and Neuroscience</source>, <volume>38</volume>(<issue>4</issue>), <fpage>509</fpage>&#8211;<lpage>536</lpage>. <pub-id pub-id-type="doi">10.1080/23273798.2021.1954207</pub-id></mixed-citation></ref>
<ref id="B80"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name> (<year>2013</year>). <article-title>Community structure in the phonological network</article-title>. <source>Frontiers in Psychology</source>, <volume>4</volume>. <pub-id pub-id-type="doi">10.3389/fpsyg.2013.00553</pub-id></mixed-citation></ref>
<ref id="B81"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name> (<year>2017</year>). <article-title>The influence of 2-hop network density on spoken word recognition</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>24</volume>(<issue>2</issue>), <fpage>496</fpage>&#8211;<lpage>502</lpage>. <pub-id pub-id-type="doi">10.3758/s13423-016-1103-9</pub-id></mixed-citation></ref>
<ref id="B82"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name> (<year>2018</year>). <article-title>The orthographic similarity structure of English words: Insights from network science</article-title>. <source>Applied Network Science</source>, <volume>3</volume>(<issue>1</issue>), <elocation-id>13</elocation-id>. <pub-id pub-id-type="doi">10.1007/s41109-018-0068-1</pub-id></mixed-citation></ref>
<ref id="B83"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name> (<year>2019</year>). <article-title>spreadr: An R package to simulate spreading activation in a network</article-title>. <source>Behavior Research Methods</source>, <volume>51</volume>(<issue>2</issue>), <fpage>910</fpage>&#8211;<lpage>929</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-018-1186-5</pub-id></mixed-citation></ref>
<ref id="B84"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2016</year>). <article-title>Spoken word recognition and serial recall of words from components in the phonological network</article-title>. <source>Journal of Experimental Psychology: Learning, Memory, and Cognition</source>, <volume>42</volume>(<issue>3</issue>), <fpage>394</fpage>&#8211;<lpage>410</lpage>. <pub-id pub-id-type="doi">10.1037/xlm0000139</pub-id></mixed-citation></ref>
<ref id="B85"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2019</year>). <article-title>The phonographic language network: Using network science to investigate the phonological and orthographic similarity structure of language</article-title>. <source>Journal of Experimental Psychology: General</source>, <volume>148</volume>(<issue>3</issue>), <fpage>475</fpage>&#8211;<lpage>500</lpage>. <pub-id pub-id-type="doi">10.1037/xge0000575</pub-id></mixed-citation></ref>
<ref id="B86"><mixed-citation publication-type="webpage"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name>, &amp; <string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2020</year>). <article-title>An investigation of network growth principles in the phonological language network</article-title>. <source>Journal of Experimental Psychology: General</source>, <volume>149</volume>(<issue>12</issue>), <fpage>2376</fpage>&#8211;<lpage>2394</lpage>. <uri>https://psycnet.apa.org/doi/10.1037/xge0000876</uri></mixed-citation></ref>
<ref id="B87"><mixed-citation publication-type="journal"><string-name><surname>Siew</surname>, <given-names>C. S. Q.</given-names></string-name>, <string-name><surname>Wulff</surname>, <given-names>D. U.</given-names></string-name>, <string-name><surname>Beckage</surname>, <given-names>N. M.</given-names></string-name>, <string-name><surname>Kenett</surname>, <given-names>Y. N.</given-names></string-name>, &amp; <string-name><surname>Me&#353;trovi&#263;</surname>, <given-names>A.</given-names></string-name> (<year>2019</year>). <article-title>Cognitive network science: A review of research on cognition through the lens of network representations, processes, and dynamics</article-title>. <source>Complexity</source>, <volume>2019</volume>, <fpage>1</fpage>&#8211;<lpage>24</lpage>. <pub-id pub-id-type="doi">10.1155/2019/2108423</pub-id></mixed-citation></ref>
<ref id="B88"><mixed-citation publication-type="journal"><string-name><surname>Sommers</surname>, <given-names>M. S.</given-names></string-name> (<year>1996</year>). <article-title>The structural organization of the mental lexicon and its contribution to age-related declines in spoken-word recognition</article-title>. <source>Psychology and Aging</source>, <volume>11</volume>(<issue>2</issue>), <fpage>333</fpage>&#8211;<lpage>341</lpage>. <pub-id pub-id-type="doi">10.1037/0882-7974.11.2.333</pub-id></mixed-citation></ref>
<ref id="B89"><mixed-citation publication-type="book"><string-name><surname>Sporns</surname>, <given-names>O.</given-names></string-name> (<year>2011</year>). <source>Networks of the brain</source>. <publisher-name>MIT Press</publisher-name>. <pub-id pub-id-type="doi">10.7551/mitpress/8476.001.0001</pub-id></mixed-citation></ref>
<ref id="B90"><mixed-citation publication-type="journal"><string-name><surname>Stella</surname>, <given-names>M.</given-names></string-name> (<year>2020</year>). <article-title>Multiplex networks quantify robustness of the mental lexicon to catastrophic concept failures, aphasic degradation and ageing</article-title>. <source>Physica A: Statistical Mechanics and Its Applications</source>, <volume>554</volume>, <elocation-id>124382</elocation-id>. <pub-id pub-id-type="doi">10.1016/j.physa.2020.124382</pub-id></mixed-citation></ref>
<ref id="B91"><mixed-citation publication-type="journal"><string-name><surname>Stella</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Beckage</surname>, <given-names>N. M.</given-names></string-name>, &amp; <string-name><surname>Brede</surname>, <given-names>M.</given-names></string-name> (<year>2017</year>). <article-title>Multiplex lexical networks reveal patterns in early word acquisition in children</article-title>. <source>Scientific Reports</source>, <volume>7</volume>, <elocation-id>46730</elocation-id>. <pub-id pub-id-type="doi">10.1038/srep46730</pub-id></mixed-citation></ref>
<ref id="B92"><mixed-citation publication-type="journal"><string-name><surname>Stella</surname>, <given-names>M.</given-names></string-name>, <string-name><surname>Beckage</surname>, <given-names>N. M.</given-names></string-name>, <string-name><surname>Brede</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>De Domenico</surname>, <given-names>M.</given-names></string-name> (<year>2018</year>). <article-title>Multiplex model of mental lexicon reveals explosive learning in humans</article-title>. <source>Scientific Reports</source>, <volume>8</volume>(<issue>1</issue>), <elocation-id>2259</elocation-id>. <pub-id pub-id-type="doi">10.1038/s41598-018-20730-5</pub-id></mixed-citation></ref>
<ref id="B93"><mixed-citation publication-type="journal"><string-name><surname>Stella</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Brede</surname>, <given-names>M.</given-names></string-name> (<year>2015</year>). <article-title>Patterns in the English language: Phonological networks, percolation and assembly models</article-title>. <source>Journal of Statistical Mechanics: Theory and Experiment</source>, <volume>2015</volume>(<issue>5</issue>), <elocation-id>P05006</elocation-id>. <pub-id pub-id-type="doi">10.1088/1742-5468/2015/05/P05006</pub-id></mixed-citation></ref>
<ref id="B94"><mixed-citation publication-type="journal"><string-name><surname>Stemberger</surname>, <given-names>J. P.</given-names></string-name> (<year>2004</year>). <article-title>Neighbourhood effects on error rates in speech production</article-title>. <source>Third International Conference on the Mental Lexicon</source>, <volume>90</volume>(<issue>1</issue>), <fpage>413</fpage>&#8211;<lpage>422</lpage>. <pub-id pub-id-type="doi">10.1016/S0093-934X(03)00452-8</pub-id></mixed-citation></ref>
<ref id="B95"><mixed-citation publication-type="journal"><string-name><surname>Steyvers</surname>, <given-names>M.</given-names></string-name>, &amp; <string-name><surname>Tenenbaum</surname>, <given-names>J. B.</given-names></string-name> (<year>2005</year>). <article-title>The large-scale structure of semantic networks: Statistical analyses and a model of semantic growth</article-title>. <source>Cognitive Science</source>, <volume>29</volume>(<issue>1</issue>), <fpage>41</fpage>&#8211;<lpage>78</lpage>. <pub-id pub-id-type="doi">10.1207/s15516709cog2901_3</pub-id></mixed-citation></ref>
<ref id="B96"><mixed-citation publication-type="journal"><string-name><surname>Storkel</surname>, <given-names>H. L.</given-names></string-name> (<year>2009</year>). <article-title>Developmental differences in the effects of phonological, lexical and semantic variables on word learning by infants</article-title>. <source>Journal of Child Language</source>, <volume>36</volume>(<issue>2</issue>), <fpage>291</fpage>&#8211;<lpage>321</lpage>. <pub-id pub-id-type="doi">10.1017/S030500090800891X</pub-id></mixed-citation></ref>
<ref id="B97"><mixed-citation publication-type="journal"><string-name><surname>Strand</surname>, <given-names>J. F.</given-names></string-name> (<year>2014</year>). <article-title>Phi-square lexical competition database (Phi-Lex): An online tool for quantifying auditory and visual lexical competition</article-title>. <source>Behavior Research Methods</source>, <volume>46</volume>(<issue>1</issue>), <fpage>148</fpage>&#8211;<lpage>158</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-013-0356-8</pub-id></mixed-citation></ref>
<ref id="B98"><mixed-citation publication-type="journal"><string-name><surname>Strand</surname>, <given-names>J. F.</given-names></string-name>, <string-name><surname>Simenstad</surname>, <given-names>A.</given-names></string-name>, <string-name><surname>Cooperman</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Rowe</surname>, <given-names>J.</given-names></string-name> (<year>2014</year>). <article-title>Grammatical context constrains lexical competition in spoken word recognition</article-title>. <source>Memory &amp; Cognition</source>, <volume>42</volume>(<issue>4</issue>), <fpage>676</fpage>&#8211;<lpage>687</lpage>. <pub-id pub-id-type="doi">10.3758/s13421-013-0378-6</pub-id></mixed-citation></ref>
<ref id="B99"><mixed-citation publication-type="journal"><string-name><surname>Strauss</surname>, <given-names>T. J.</given-names></string-name>, <string-name><surname>Harris</surname>, <given-names>H. D.</given-names></string-name>, &amp; <string-name><surname>Magnuson</surname>, <given-names>J. S.</given-names></string-name> (<year>2007</year>). <article-title>jTRACE: A reimplementation and extension of the TRACE model of speech perception and spoken word recognition</article-title>. <source>Behavior Research Methods</source>, <volume>39</volume>(<issue>1</issue>), <fpage>19</fpage>&#8211;<lpage>30</lpage>. <pub-id pub-id-type="doi">10.3758/BF03192840</pub-id></mixed-citation></ref>
<ref id="B100"><mixed-citation publication-type="journal"><string-name><surname>Su&#225;rez</surname>, <given-names>L.</given-names></string-name>, <string-name><surname>Tan</surname>, <given-names>S. H.</given-names></string-name>, <string-name><surname>Yap</surname>, <given-names>M. J.</given-names></string-name>, &amp; <string-name><surname>Goh</surname>, <given-names>W. D.</given-names></string-name> (<year>2011</year>). <article-title>Observing neighborhood effects without neighbors</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>18</volume>(<issue>3</issue>), <fpage>605</fpage>&#8211;<lpage>611</lpage>. <pub-id pub-id-type="doi">10.3758/s13423-011-0078-9</pub-id></mixed-citation></ref>
<ref id="B101"><mixed-citation publication-type="journal"><string-name><surname>Sun</surname>, <given-names>Y.</given-names></string-name>, &amp; <string-name><surname>Poeppel</surname>, <given-names>D.</given-names></string-name> (<year>2023</year>). <article-title>Syllables and their beginnings have a special role in the mental lexicon</article-title>. <source>Proceedings of the National Academy of Sciences</source>, <volume>120</volume>(<issue>36</issue>), <elocation-id>e2215710120</elocation-id>. <pub-id pub-id-type="doi">10.1073/pnas.2215710120</pub-id></mixed-citation></ref>
<ref id="B102"><mixed-citation publication-type="journal"><string-name><surname>Swingley</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Aslin</surname>, <given-names>R. N.</given-names></string-name> (<year>2002</year>). <article-title>Lexical neighborhoods and the word-form representations of 14-month-olds</article-title>. <source>Psychological Science</source>, <volume>13</volume>(<issue>5</issue>), <fpage>480</fpage>&#8211;<lpage>484</lpage>. <pub-id pub-id-type="doi">10.1111/1467-9280.00485</pub-id></mixed-citation></ref>
<ref id="B103"><mixed-citation publication-type="journal"><string-name><surname>Swingley</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Aslin</surname>, <given-names>R. N.</given-names></string-name> (<year>2007</year>). <article-title>Lexical competition in young children&#8217;s word learning</article-title>. <source>Cognitive Psychology</source>, <volume>54</volume>(<issue>2</issue>), <fpage>99</fpage>&#8211;<lpage>132</lpage>. <pub-id pub-id-type="doi">10.1016/j.cogpsych.2006.05.001</pub-id></mixed-citation></ref>
<ref id="B104"><mixed-citation publication-type="journal"><string-name><surname>Tomaschek</surname>, <given-names>F.</given-names></string-name>, <string-name><surname>Hendrix</surname>, <given-names>P.</given-names></string-name>, &amp; <string-name><surname>Baayen</surname>, <given-names>R. H.</given-names></string-name> (<year>2018</year>). <article-title>Strategies for addressing collinearity in multivariate linguistic data</article-title>. <source>Journal of Phonetics</source>, <volume>71</volume>, <fpage>249</fpage>&#8211;<lpage>267</lpage>. <pub-id pub-id-type="doi">10.1016/j.wocn.2018.09.004</pub-id></mixed-citation></ref>
<ref id="B105"><mixed-citation publication-type="journal"><string-name><surname>Tucker</surname>, <given-names>B. V.</given-names></string-name>, <string-name><surname>Brenner</surname>, <given-names>D.</given-names></string-name>, <string-name><surname>Danielson</surname>, <given-names>D. K.</given-names></string-name>, <string-name><surname>Kelley</surname>, <given-names>M. C.</given-names></string-name>, <string-name><surname>Nenadi&#263;</surname>, <given-names>F.</given-names></string-name>, &amp; <string-name><surname>Sims</surname>, <given-names>M.</given-names></string-name> (<year>2019</year>). <article-title>The Massive Auditory Lexical Decision (MALD) database</article-title>. <source>Behavior Research Methods</source>, <volume>51</volume>(<issue>3</issue>), <fpage>1187</fpage>&#8211;<lpage>1204</lpage>. <pub-id pub-id-type="doi">10.3758/s13428-018-1056-1</pub-id></mixed-citation></ref>
<ref id="B106"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>1997</year>). <article-title>The neighborhood characteristics of malapropisms</article-title>. <source>Language and Speech</source>, <volume>40</volume>(<issue>3</issue>), <fpage>211</fpage>&#8211;<lpage>228</lpage>. <pub-id pub-id-type="doi">10.1177/002383099704000301</pub-id></mixed-citation></ref>
<ref id="B107"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2002</year>). <article-title>The influence of phonological similarity neighborhoods on speech production</article-title>. <source>Journal of Experimental Psychology: Learning, Memory, and Cognition</source>, <volume>28</volume>(<issue>4</issue>), <fpage>735</fpage>&#8211;<lpage>747</lpage>. <pub-id pub-id-type="doi">10.1037/0278-7393.28.4.735</pub-id></mixed-citation></ref>
<ref id="B108"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2008</year>). <article-title>What can graph theory tell us about word learning and lexical retrieval?</article-title> <source>Journal of Speech, Language, and Hearing Research</source>, <volume>51</volume>(<issue>2</issue>), <fpage>408</fpage>&#8211;<lpage>422</lpage>. <pub-id pub-id-type="doi">10.1044/1092-4388(2008/030)</pub-id></mixed-citation></ref>
<ref id="B109"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name> (<year>2022</year>). <article-title>What can network science tell us about phonology and language processing?</article-title> <source>Topics in Cognitive Science</source>, <volume>14</volume>(<issue>1</issue>), <fpage>127</fpage>&#8211;<lpage>142</lpage>. <pub-id pub-id-type="doi">10.1111/tops.12532</pub-id></mixed-citation></ref>
<ref id="B110"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, <string-name><surname>Ercal</surname>, <given-names>G.</given-names></string-name>, &amp; <string-name><surname>Adagarla</surname>, <given-names>B.</given-names></string-name> (<year>2011</year>). <article-title>Simulating retrieval from a highly clustered network: Implications for spoken word recognition</article-title>. <source>Frontiers in Psychology</source>, <volume>2</volume>. <pub-id pub-id-type="doi">10.3389/fpsyg.2011.00369</pub-id></mixed-citation></ref>
<ref id="B111"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Goldstein</surname>, <given-names>R.</given-names></string-name> (<year>2014</year>). <article-title>Keywords in the mental lexicon</article-title>. <source>Journal of Memory and Language</source>, <volume>73</volume>, <fpage>131</fpage>&#8211;<lpage>147</lpage>. <pub-id pub-id-type="doi">10.1016/j.jml.2014.03.005</pub-id></mixed-citation></ref>
<ref id="B112"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name> (<year>1998</year>). <article-title>When words compete: Levels of processing in perception of spoken words</article-title>. <source>Psychological Science</source>, <volume>9</volume>(<issue>4</issue>), <fpage>325</fpage>&#8211;<lpage>329</lpage>. <pub-id pub-id-type="doi">10.1111/1467-9280.00064</pub-id></mixed-citation></ref>
<ref id="B113"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name> (<year>1999</year>). <article-title>Probabilistic phonotactics and neighborhood activation in spoken word recognition</article-title>. <source>Journal of Memory and Language</source>, <volume>40</volume>(<issue>3</issue>), <fpage>374</fpage>&#8211;<lpage>408</lpage>. <pub-id pub-id-type="doi">10.1006/jmla.1998.2618</pub-id></mixed-citation></ref>
<ref id="B114"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Luce</surname>, <given-names>P. A.</given-names></string-name> (<year>2016</year>). <article-title>Phonological neighborhood effects in spoken word perception and production</article-title>. <source>Annual Review of Linguistics</source>, <volume>2</volume>(<issue>1</issue>), <fpage>75</fpage>&#8211;<lpage>94</lpage>. <pub-id pub-id-type="doi">10.1146/annurev-linguistics-030514-124832</pub-id></mixed-citation></ref>
<ref id="B115"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Mullin</surname>, <given-names>G. J. D.</given-names></string-name> (<year>2021</year>). <article-title>What do cognitive networks do? Simulations of spoken word recognition using the cognitive network science approach</article-title>. <source>Brain Sciences</source>, <volume>11</volume>(<issue>12</issue>), Article <elocation-id>12</elocation-id>. <pub-id pub-id-type="doi">10.3390/brainsci11121628</pub-id></mixed-citation></ref>
<ref id="B116"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, <string-name><surname>Niehorster-Cook</surname>, <given-names>L.</given-names></string-name>, &amp; <string-name><surname>Niehorster-Cook</surname>, <given-names>S.</given-names></string-name> (<year>2021</year>). <article-title>Exploring how phonotactic knowledge can be represented in cognitive networks</article-title>. <source>Big Data and Cognitive Computing</source>, <volume>5</volume>(<issue>4</issue>), <elocation-id>47</elocation-id>. <pub-id pub-id-type="doi">10.3390/bdcc5040047</pub-id></mixed-citation></ref>
<ref id="B117"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Sale</surname>, <given-names>M.</given-names></string-name> (<year>2023</year>). <article-title>Identifying the phonological backbone in the mental lexicon</article-title>. <source>PLOS ONE</source>, <volume>18</volume>(<issue>6</issue>), <elocation-id>e0287197</elocation-id>. <pub-id pub-id-type="doi">10.1371/journal.pone.0287197</pub-id></mixed-citation></ref>
<ref id="B118"><mixed-citation publication-type="journal"><string-name><surname>Vitevitch</surname>, <given-names>M. S.</given-names></string-name>, &amp; <string-name><surname>Sommers</surname>, <given-names>M. S.</given-names></string-name> (<year>2003</year>). <article-title>The facilitative influence of phonological similarity and neighborhood frequency in speech production in younger and older adults</article-title>. <source>Memory &amp; Cognition</source>, <volume>31</volume>(<issue>4</issue>), <fpage>491</fpage>&#8211;<lpage>504</lpage>. PubMed. <pub-id pub-id-type="doi">10.3758/bf03196091</pub-id></mixed-citation></ref>
<ref id="B119"><mixed-citation publication-type="journal"><string-name><surname>Watts</surname>, <given-names>D. J.</given-names></string-name>, &amp; <string-name><surname>Strogatz</surname>, <given-names>S. H.</given-names></string-name> (<year>1998</year>). <article-title>Collective dynamics of &#8220;small-world&#8221; networks</article-title>. <source>Nature</source>, <volume>393</volume>, <fpage>440</fpage>&#8211;<lpage>442</lpage>. <pub-id pub-id-type="doi">10.1038/30918</pub-id></mixed-citation></ref>
<ref id="B120"><mixed-citation publication-type="journal"><string-name><surname>Weber</surname>, <given-names>A.</given-names></string-name>, &amp; <string-name><surname>Scharenborg</surname>, <given-names>O.</given-names></string-name> (<year>2012</year>). <article-title>Models of spoken-word recognition: Models of word recognition</article-title>. <source>Wiley Interdisciplinary Reviews: Cognitive Science</source>, <volume>3</volume>(<issue>3</issue>), <fpage>387</fpage>&#8211;<lpage>401</lpage>. <pub-id pub-id-type="doi">10.1002/wcs.1178</pub-id></mixed-citation></ref>
<ref id="B121"><mixed-citation publication-type="book"><string-name><surname>Wells</surname>, <given-names>J. C.</given-names></string-name> (<year>1997</year>). <chapter-title>SAMPA computer readable phonetic alphabet</chapter-title>. In <string-name><given-names>D.</given-names> <surname>Gibbon</surname></string-name>, <string-name><given-names>R.</given-names> <surname>Moore</surname></string-name>, &amp; <string-name><given-names>R.</given-names> <surname>Winski</surname></string-name> (Eds.), <source>Handbook of standards and resources for spoken language systems</source>. <publisher-name>Mouton de Gruyter</publisher-name>.</mixed-citation></ref>
<ref id="B122"><mixed-citation publication-type="journal"><string-name><surname>Yarkoni</surname>, <given-names>T.</given-names></string-name>, <string-name><surname>Balota</surname>, <given-names>D.</given-names></string-name>, &amp; <string-name><surname>Yap</surname>, <given-names>M.</given-names></string-name> (<year>2008</year>). <article-title>Moving beyond Coltheart&#8217;s N: A new measure of orthographic similarity</article-title>. <source>Psychonomic Bulletin &amp; Review</source>, <volume>15</volume>(<issue>5</issue>), <fpage>971</fpage>&#8211;<lpage>979</lpage>. <pub-id pub-id-type="doi">10.3758/PBR.15.5.971</pub-id></mixed-citation></ref>
<ref id="B123"><mixed-citation publication-type="journal"><string-name><surname>Yates</surname>, <given-names>M.</given-names></string-name> (<year>2013</year>). <article-title>How the clustering of phonological neighbors affects visual word recognition</article-title>. <source>Journal of Experimental Psychology: Learning, Memory, and Cognition</source>, <volume>39</volume>(<issue>5</issue>), <fpage>1649</fpage>&#8211;<lpage>1656</lpage>. <pub-id pub-id-type="doi">10.1037/a0032422</pub-id></mixed-citation></ref>
</ref-list>
</back>
</article>