- Home
- e-Journals
- The Mental Lexicon
- Previous Issues
- Volume 18, Issue 3, 2023
The Mental Lexicon - Volume 18, Issue 3, 2023
Volume 18, Issue 3, 2023
-
Lexical and contextual emotional valence in foreign language vocabulary retention
Author(s): Yu Kanazawapp.: 339–365 (27)More LessAbstractStudies suggest that not only words per se but also surrounding contexts play significant roles in foreign language vocabulary retention. This study investigated how different conditions of lexical and contextual emotional valence differently affect foreign language vocabulary retention. The target words were either positive (LexVal+), neutral (LexVal=), or negative (LexVal−) in meaning. Each visually enhanced target word was embedded in a sentence either positive (CtxVal+), neutral (CtxVal=), or negative (CtxVal−) in meaning. Sentences with different combinations of lexical valence and contextual valence were presented in the Study Session, which were later incidentally recalled in the Test Session. It was revealed that positive and negative words were remembered more often than neutral words and that negative contexts resulted in better retention of the embedded words than neutral contexts. These findings are in accordance with the predictions of the emotionally enhanced memory, the Emotion-Involved Processing Hypothesis, and the role of affect in the Modular Cognition Framework. Further, even if the target item is not emotional, embedding it in emotional context may result in better retention, a finding with potential pedagogical implications. Interestingly, words embedded in emotionally congruent contexts were not learned better than those in incongruent contexts, a finding contrary to expectation. The result may be explainable via the Deep Epistemic Emotion Hypothesis, calling for more empirical study.
-
Eye-voice and finger-voice spans in adults’ oral reading of connected texts*
Author(s): Andrea Nadalini, Claudia Marzi, Marcello Ferro, Loukia Taxitari, Alessandro Lento, Davide Crepaldi and Vito Pirrellipp.: 366–400 (35)More LessAbstractThe present paper investigates the interaction between eye movements, voice articulation and the movements of the index finger dynamically pointing to a text line in oral finger-point reading of Italian. During finger-point reading, the finger appears to be ahead of the voice most of the times, by a margin that is significantly modulated by the distribution of phrasal and prosodic units in the reading text. Eye movements replicate the same effects on a different time scale. The eye is ahead of both voice and finger by a wide margin (confirming evidence observed for English and German sentence reading), while showing a tendency to re-synchronise with voice articulation at the right edge of strong prosodic units (sentence boundaries). Our evidence suggests a multicomponent view of the time span between the eye/finger and the voice. The span is shown to be the dynamic outcome of an optimally adaptive reading strategy, resulting from the interaction between individual decoding skills, the reader’s phonological buffer capacity, and the structural complexity of a reading text. Proficient readers modulate their span to compensate for the different timing between word fixation and word articulation, read faster, and dynamically adjust their processing window to the meaningful, prosodic units of a text.
-
Bilingual and monolingual adults’ lexical choice in storytelling
Author(s): Elena Nicoladis, Danat Tewelde and Valin Zeschukpp.: 401–416 (16)More LessAbstractBilinguals often have a harder time accessing words for production than monolinguals, perhaps because they have less exposure to words from each language (the weaker-links hypothesis). This research on lexical access mostly comes from studies of words in isolation. The purpose of the present study was to test whether bilinguals also show greater lexical access difficulties than monolinguals when telling a story. In the context of a narrative, we predicted that bilinguals would produce fewer different words and words of higher frequency than monolinguals, in order to make lexical access easier. For the same reason, we also predicted that bilinguals would use more cognates than monolinguals. English monolinguals, French monolinguals, and highly proficient French-English bilinguals watched a cartoon and told the story back. We coded the words they used for frequency and cognate status. In English, the results showed little difference between bilinguals and monolinguals on word frequency and cognate status. In French, first-language-English bilinguals used higher frequency words than first-language-French bilinguals. These results support predictions from the weaker-links hypothesis in lexical access for storytelling, albeit only for French.
-
German nominal number interpretation in an impaired mental lexicon
Author(s): Ingo Plag, Maria Heitmeier and Frank Domahspp.: 417–445 (29)More LessAbstractThere is an ongoing debate on how speakers and listeners process and interpret information in a morphological system that is very complex and not very transparent. A well-known test case is the German nominal number system. In this paper we employ discriminative learning (e.g., Ramscar & Yarlett, 2007; Baayen et al., 2011, 2019) to test whether discriminative learning networks can be used to better understand the processing of German number. We analyse behavioral data obtained from a patient with primary progressive aphasia (Domahs et al., 2017), and the unimpaired system. We test a model that implements the traditional cues borrowed from the schema approach (Köpcke, 1988, 1993; Köpcke et al., 2021), and compare it to a model that uses segmental and phonotactic information only. Our results for the unimpaired system demonstrate that a model based on only biphones as cues is better able to predict the number of a given word-form than a model using structural phonological cues. We also test whether a discriminative learning model can predict the number decisions by the aphasic patient. The results demonstrate that a biphone-based discriminative model trained on the patient’s responses is superior to a structure-based model in approximating the patient’s behavior.
-
Adult L2 learners’ morphological sensitivity in a morphosyllabic language
Author(s): Sihui (Echo) Ke, Rui Jin and Keiko Kodapp.: 446–471 (26)More LessAbstractThis study examined adult L2 learners’ morphological sensitivity in Chinese, a morphosyllabic language, and explored whether there is any modulating effects of L2 proficiency. Two word naming experiments (segment shifting and standard word naming) were administered to three participant groups, including native Chinese speakers, higher L2 Chinese proficiency learners, and lower L2 Chinese proficiency learners. In both experiments, reaction times (RTs) displayed only main effects of Chinese proficiency group and word type. This suggests that the morphological processing of L2 learners did not differ from that of native speakers, although the RTs of L2 learners were longer and exhibited more variability. Concerning error rates, both experiments showed that learners with higher and lower L2 proficiency had significantly higher error rates for words with unreliable morphological cues compared to those with reliable cues. Taken together, these findings indicate that L2 learners developed sensitivity to intraword morphological structure and employed decompositional strategies when reading Chinese words, irrespective of their L2 proficiency levels.
-
How direct is the link between words and images?
pp.: 472–511 (40)More LessAbstractGünther et al. (2022) investigated the relationship between words and images in which they concluded the possibility of a direct link between words and embodied experience. In their study, participants were presented with a target noun and a pair of images, one chosen by their model and another chosen randomly. Participants were asked to select the image that best matched the target noun. Building upon their work, we addressed the following questions. 1. Apart from utilizing visually embodied simulation, what other strategies subjects might have used? How much does this setup rely on visual information? Can it be solved using textual representations? 2. Do current visually-grounded embeddings explain subjects’ selection behavior better than textual embeddings? 3. Does visual grounding improve the representations of both concrete and abstract words? For this aim, we designed novel experiments based on pre-trained word embeddings. Our experiments reveal that subjects’ selection behavior is explained to a large extend on text-based embeddings and word-based similarities. Visually grounded embeddings offered modest advantages over textual embeddings in certain cases. These findings indicate that the experiment by Günther et al. (2022) may not be well suited for tapping into the perceptual experience of participants, and the extent to which it measures visually grounded knowledge is unclear.