Displaying 1 - 37 of 37
-
Fairs, A., Bögels, S., & Meyer, A. S. (2017). Dual-tasking in language: Concurrent production and comprehension interfere at the phonological level. Poster presented at Psycholinguistics in Flanders (PiF 2017), Leuven, Belgium.
-
Fairs, A., Bögels, S., & Meyer, A. S. (2017). Dual-tasking in language: Concurrent production and comprehension interfere at the phonological level. Poster presented at the Experimental Psychology Society Belfast Meeting, Belfast, UK.
Abstract
Conversation often involves simultaneous production and comprehension, yet little research has investigated whether these two processes interfere with one another. We tested participants’ ability to dual-task with production and comprehension tasks. Task one (production task) was picture naming. Task two (comprehension task) was either syllable identification (linguistic condition) or tone identification (non-linguistic condition). The two identification tasks were matched for difficulty. Three SOAs (50ms, 300ms, and 1800ms) resulted in different amounts of overlap between the production and comprehension tasks. We hypothesized that as production and comprehension use similar resources there would be greater interference with concurrent linguistic than non-linguistic tasks.
At the 50ms SOA, picture naming latencies were slower in the linguistic compared to the non-linguistic condition, suggesting that the resources required for production and comprehension overlap more in the linguistic condition. As the syllables were non-words without lexical representations, this interference likely occurs primarily at the phonological level. Across all SOAs, identification RTs were longer in the linguistic condition, showing that such phonological interference percolates through to the comprehension task, regardless of SOA. In sum, these results demonstrate that concurrent access to the phonological level in production and comprehension results in measurable interference in both speaking and comprehending.
-
Fairs, A., Bögels, S., & Meyer, A. S. (2017). Serial or parallel dual-task language processing: Production planning and comprehension are not carried out in parallel. Talk presented at Psycholinguistics in Flanders (PiF 2017). Leuven, Belgium. 2017-05-29 - 2017-05-30.
-
Fairs, A., Bögels, S., & Meyer, A. S. (2017). Serial or parallel dual-task language processing: Production planning and comprehension are not carried out in parallel. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2017), Lancaster, UK.
-
Hoedemaker, R. S., & Meyer, A. S. (2017). Coordination and preparation of utterances in a joint-naming task. Talk presented at the Experimental Psychology Society Belfast Meeting. Belfast, UK. 2017-04-10 - 2017-04-12.
-
Iacozza, S., Meyer, A. S., & Lev-Ari, S. (2017). “That’s a spatelhouder!”: How source memory is influenced by speakers’ social categories in a word-learning paradigm. Talk presented at Psycholinguistics in Flanders (PiF 2017). Leuven, Belgium. 2017-05-29 - 2017-05-30.
-
Iacozza, S., Meyer, A. S., & Lev-Ari, S. (2017). Speakers' social identity affects source memory for novel words. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2017), Lancaster, UK.
-
Jongman, S. R., Meyer, A. S., & Piai, V. (2017). Brain signature of planning for production: An EEG study. Talk presented at the Workshop 'Revising formal semantic and pragmatic theories from a neurocognitive perspective'. Bochum, Germany. 2017-06-19 - 2017-06-20.
-
Jongman, S. R., & Meyer, A. S. (2017). Simultaneous listening and planning for production: Full or partial comprehension?. Poster presented at the 30th Annual CUNY Conference on Human Sentence Processing, Cambridge, MA, USA.
-
Maslowski, M., Meyer, A. S., & Bosker, H. R. (2017). When slow speech sounds fast: How the speech rate of one talker influences perception of another talker. Talk presented at the IPS workshop: Abstraction, Diversity, and Speech Dynamics. Herrsching am Ammersee, Germany. 2017-05-03 - 2017-05-05.
Abstract
Listeners are continuously exposed to a broad range of speech rates. Earlier work has shown that listeners perceive phonetic category boundaries relative to contextual speech rate. This process of rate-dependent speech perception has been suggested to occur across talker changes, with the speech rate of talker A influencing perception of talker B. This study tested whether a ‘global’ speech rate calculated over multiple talkers and over a longer period of time affected perception of the temporal Dutch vowel contrast /ɑ/-/a:/. First, Experiment 1 demonstrated that listeners more often reported hearing long /a:/ in fast contexts than in ‘neutral rate’ contexts, replicating earlier findings. Then, in Experiment 2, one participant group was exposed to ‘neutral’ speech from talker A intermixed with slow speech from talker B. Another group listened to the same ‘neutral’ speech from talker A, but to fast speech from talker B. Between-group comparison in the ‘neutral’ condition revealed that Group 1 reported more long /a:/ than Group 2, indicating that A’s ‘neutral’ speech sounded faster when B was slower. Finally, Experiment 3 tested whether talking at slow or fast rates oneself elicits the same ‘global’ rate effects. However, no evidence was found that self-produced speech modulated perception of talker A. This study corroborates the idea that ‘global’ rate-dependent effects occur across talkers, but are insensitive to one’s own speech rate. Results are interpreted in light of the general auditory mechanisms thought to underlie rate normalization, with implications for our understanding of dialogue.Additional information
http://www.phonetik.uni-muenchen.de/institut/veranstaltungen/abstraction-divers… -
Maslowski, M., Meyer, A. S., & Bosker, H. R. (2017). Whether long-term tracking of speech affects perception depends on who is talking. Poster presented at the Donders Poster Sessions, Nijmegen, The Netherlands.
Abstract
Speech rate is known to modulate perception of temporally ambiguous speech sounds. For instance, a vowel may be perceived as short when the immediate speech context is slow, but as long when the context is fast. Yet, effects of long-term tracking of speech rate are largely unexplored. Two experiments tested whether long-term tracking of rate influences perception of the temporal Dutch vowel contrast /A/-/a:/. In Experiment 1, one low-rate group listened to ‘neutral’ rate speech from talker A and to slow speech from talker B. Another high-rate group was exposed to the same neutral speech from A, but to fast speech from B. Between-group comparison of the ‘neutral’ trials revealed that the low-rate group reported a higher proportion of /a:/ in A’s ‘neutral’ speech, indicating that A sounded faster when B was slow. Experiment 2 tested whether one’s own speech rate also contributes to effects of long-term tracking of rate. Here, talker B’s speech was replaced by playback of participants’ own fast or slow speech. No evidence was found that one’s own voice affected perception of talker A in larger speech contexts. These results carry implications for our understanding of the mechanisms involved in rate-dependent speech perception and of dialogue. -
Maslowski, M., Meyer, A. S., & Bosker, H. R. (2017). Whether long-term tracking of speech rate affects perception depends on who is talking. Poster presented at Interspeech 2017, Stockholm, Sweden.
Abstract
Speech rate is known to modulate perception of temporally ambiguous speech sounds. For instance, a vowel may be perceived as short when the immediate speech context is slow, but as long when the context is fast. Yet, effects of long-term tracking of speech rate are largely unexplored. Two experiments tested whether long-term tracking of rate influences perception of the temporal Dutch vowel contrast /ɑ/-/a:/. In Experiment 1, one low-rate group listened to 'neutral' rate speech from talker A and to slow speech from talker B. Another high-rate group was exposed to the same neutral speech from A, but to fast speech from B. Between-group comparison of the 'neutral' trials revealed that the low-rate group reported a higher proportion of /a:/ in A's 'neutral' speech, indicating that A sounded faster when B was slow. Experiment 2 tested whether one's own speech rate also contributes to effects of long-term tracking of rate. Here, talker B's speech was replaced by playback of participants' own fast or slow speech. No evidence was found that one's own voice affected perception of talker A in larger speech contexts. These results carry implications for our understanding of the mechanisms involved in rate-dependent speech perception and of dialogue. -
Meyer, A. S., Decuyper, C., & Coopmans, C. W. (2017). Distribution of attention in question-answer sequences: Evidence for limited parallel processing. Talk presented at the Experimental Psychology Society London Meeting. London, UK. 2017-01-03 - 2017-01-06.
-
Meyer, A. S. (2017). Towards understanding conversation: A psycholinguist's perspective. Talk presented at Psycholinguistics in Flanders (PiF 2017). Leuven, Belgium. 2017-05-29 - 2017-05-30.
-
Raviv, L., Meyer, A. S., & Lev-Ari, S. (2017). Compositional structure can emerge without generational transmission. Talk presented at the Inaugural Cultural Evolution Society Conference (CESC 2017). Jena, Germany. 2017-09-13 - 2017-09-15.
-
Raviv, L., Meyer, A. S., & Lev-Ari, S. (2017). Compositional structure can emerge without generational transmission. Talk presented at the 30th Annual CUNY Conference on Human Sentence Processing. Cambridge, MA, USA. 2017-03-30 - 2017-04-01.
-
Raviv, L., Meyer, A. S., & Lev-Ari, S. (2017). The role of community size in the emergence of linguistic structure. Talk presented at XLanS: Triggers of language change in the Language Sciences. Lyon, France. 2017-10-11 - 2017-10-13.
-
Rodd, J., Bosker, H. R., Ernestus, M., Ten Bosch, L., & Meyer, A. S. (2017). How we regulate speech rate: Phonetic evidence for a 'gain strategy' in speech planning. Poster presented at the Abstraction, Diversity and Speech Dynamics Workshop, Herrsching, Germany.
-
Rodd, J., Bosker, H. R., Ernestus, M., Meyer, A. S., & Ten Bosch, L. (2017). Simulating speaking rate control: A spreading activation model of syllable timing. Poster presented at the Workshop Conversational speech and lexical representations, Nijmegen, The Netherlands.
Abstract
Speech can be produced at different rates. The ability to produce faster or slower speech may be thought to result from executive control processes enlisted to modulate lexical selection and phonological encoding stages of speech planning.
This study used simulations of the model of serial order in language by Dell, Burger and Svec (1997, DBS) to characterise the strategies adopted by speakers when naming pictures at fast, medium and slow prescribed rates. Our new implementation of DBS was able to produce activation patterns that correlated strongly with observed syllable-level timing of disyllabic words from this task.
For each participant, different speaking rates were associated with different regions of the DBS parameter space. The precise placement of the speaking rates in the parameter space differed markedly between participants. Participants applied broadly the same parameter manipulation to accelerate their speech. This was however not the case for deceleration. Hierarchical clustering revealed two distinct patterns of parameter adjustment employed to decelerate speech, suggesting that deceleration is not necessarily achieved by the inverse process of acceleration. In addition, potential refinements to the DBS model are discussed. -
Shao, Z., & Meyer, A. S. (2017). How word and phrase frequencies affect noun phrase production. Poster presented at the 30th Annual CUNY Conference on Human Sentence Processing, Cambridge, MA, USA.
-
Tromp, J., Peeters, D., Meyer, A. S., & Hagoort, P. (2017). Combining Virtual Reality and EEG to study semantic and pragmatic processing in a naturalistic environment. Talk presented at the workshop 'Revising formal Semantic and Pragmatic theories from a Neurocognitive Perspective' (NeuroPragSem, 2017). Bochum, Germany. 2017-06-19 - 2017-06-20.
-
Van Paridon, J., Roelofs, A., & Meyer, A. S. (2017). Coordinating simultaneous comprehension and production: Behavioral and modelling findings from shadowing and simultaneous interpreting. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2017), Lancaster, UK.
-
Weber, K., Meyer, A. S., & Hagoort, P. (2017). Learning lexical-syntactic biases: An fMRI study on how we connect words and structures. Poster presented at the 13th International Conference for Cognitive Neuroscience (ICON), Amsterdam, The Netherlands.
-
Zormpa, E., Hoedemaker, R. S., Brehm, L., & Meyer, A. S. (2017). The production and generation effect in picture naming: How lexical access and articulation influence memory. Poster presented at the Donders Posters Session, Nijmegen, The Netherlands.
-
Araújo, S., Huettig, F., & Meyer, A. S. (2016). What's the nature of the deficit underlying impaired naming? An eye-tracking study with dyslexic readers. Talk presented at IWORDD - International Workshop on Reading and Developmental Dyslexia. Bilbao, Spain. 2016-05-05 - 2016-05-07.
Abstract
Serial naming deficits have been identified as core symptoms of developmental dyslexia. A prominent hypothesis is that naming delays are due to inefficient phonological encoding, yet the exact nature of this underlying impairment remains largely underspecified. Here we used recordings of eye movements and word onset latencies to examine at what processing level the dyslexic naming deficit emerges: localized at an early stage of lexical encoding or rather later at the level of phonetic or motor planning. 23 dyslexic and 25 control adult readers were tested on a serial object naming task for 30 items and an analogous reading task, where phonological neighborhood density and word-frequency were manipulated. Results showed that both word properties influenced early stages of phonological activation (first fixation and first-pass duration) equally in both groups of participants. Moreover, in the control group any difficulty appeared to be resolved early in the reading process, while for dyslexic readers a processing disadvantage for low-frequency words and for words with sparse neighborhood also emerged in a measure that included late stages of output planning (eye-voice span). Thus, our findings suggest suboptimal phonetic and/or articulatory planning in dyslexia. -
Hoedemaker, R. S., Ernst, J., Meyer, A. S., & Belke, E. (2016). Language production in a shared task: Cumulative semantic interference from self- and other-produced context words. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2016), Bilbao, Spain.
-
Hoedemaker, R. S., Ernst, J., Meyer, A. S., & Belke, E. (2016). Language production in a shared task: Cumulative semantic interference from self- and other-produced context words. Talk presented at Psycholinguistics in Flanders (PiF 2016). Antwerp, Belgium. 2016-05-25 - 2016-05-27.
-
Kösem, A., Bosker, H. R., Meyer, A. S., Jensen, O., & Hagoort, P. (2016). Neural entrainment reflects temporal predictions guiding speech comprehension. Poster presented at the Eighth Annual Meeting of the Society for the Neurobiology of Language (SNL 2016), London, UK.
Abstract
Speech segmentation requires flexible mechanisms to remain robust to features such as speech rate and pronunciation. Recent hypotheses suggest that low-frequency neural oscillations entrain to ongoing syllabic and phrasal rates, and that neural entrainment provides a speech-rate invariant means to discretize linguistic tokens from the acoustic signal. How this mechanism functionally operates remains unclear. Here, we test the hypothesis that neural entrainment reflects temporal predictive mechanisms. It implies that neural entrainment is built on the dynamics of past speech information: the brain would internalize the rhythm of preceding speech to parse the ongoing acoustic signal at optimal time points. A direct prediction is that ongoing neural oscillatory activity should match the rate of preceding speech even if the stimulation changes, for instance when the speech rate suddenly increases or decreases. Crucially, the persistence of neural entrainment to past speech rate should modulate speech perception. We performed an MEG experiment in which native Dutch speakers listened to sentences with varying speech rates. The beginning of the sentence (carrier window) was either presented at a fast or a slow speech rate, while the last three words (target window) were displayed at an intermediate rate across trials. Participants had to report the perception of the last word of the sentence, which was ambiguous with regards to its vowel duration (short vowel /ɑ/ – long vowel /aː/ contrast). MEG data was analyzed in source space using beamformer methods. Consistent with previous behavioral reports, the perception of the ambiguous target word was influenced by the past speech rate; participants reported more /aː/ percepts after a fast speech rate, and more /ɑ/ after a slow speech rate. During the carrier window, neural oscillations efficiently tracked the dynamics of the speech envelope. During the target window, we observed oscillatory activity that corresponded in frequency to the preceding speech rate. Traces of neural entrainment to the past speech rate were significantly observed in medial prefrontal areas. Right superior temporal cortex also showed persisting oscillatory activity which correlated with the observed perceptual biases: participants whose perception was more influenced by the manipulation in speech rate also showed stronger remaining neural oscillatory patterns. The results show that neural entrainment lasts after rhythmic stimulation. The findings further provide empirical support for oscillatory models of speech processing, suggesting that neural oscillations actively encode temporal predictions for speech comprehension. -
Kösem, A., Bosker, H. R., Meyer, A. S., Jensen, O., & Hagoort, P. (2016). Neural entrainment to speech rhythms reflects temporal predictions and influences word comprehension. Poster presented at the 20th International Conference on Biomagnetism (BioMag 2016), Seoul, South Korea.
-
Mainz, N., Shao, Z., Brysbaert, M., & Meyer, A. S. (2016). The contribution of vocabulary size to language processing: Evidence from lexical decision and picture-word interference. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2016), Bilbao, Spain.
Abstract
Previous research indicates that general cognitive abilities, such as attention or executive control, contribute to language processing (Hartsuiker & Barkhuysen, 2006; Jongman et al., 2014; Shao et al., 2013). Potential effects of language-specific abilities, such as vocabulary, on language processing in adult native speakers have been examined less extensively. Goals: a) develop and assess measures of vocabulary size in Dutch native speakers, and b) investigate the relationship between individual differences in vocabulary and language processing. -
Maslowski, M., Bosker, H. R., & Meyer, A. S. (2016). Slow speech can sound fast: How the speech rate of one talker affects perception of another talker. Talk presented at the Donders Discussions 2016. Nijmegen, The Netherlands. 2016-11-24 - 2016-11-25.
-
Maslowski, M., Bosker, H. R., & Meyer, A. S. (2016). Slow speech can sound fast: How the speech rate of one talker has a contrastive effect on the perception of another talker. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2016), Bilbao, Spain.
Abstract
Listeners are continuously exposed to a broad range of speech rates. Earlier work has shown that listeners perceive phonetic category boundaries relative to contextual speech rate. It has been suggested that this process of speech rate normalization occurs across talker changes. This would predict that the speech rate of talker A influences perception of the rate of another talker B. We assessed this hypothesis by testing effects of speech rate on the perception on the Dutch vowel continuum /A/-/a:/. One participant group was exposed to 'neutral' speech from talker A intermixed with fast speech from talker B. Another group listened to the same speech from talker A, but to slow speech from talker B. We observed a difference in perception of talker A depending on the speech rate of talker B: A's 'neutral' speech was perceived as slow when B spoke faster. These findings corroborate the idea that speech rate normalization occurs across talkers, but they challenge the assumption that listeners average over speech rates from multiple talkers. Instead, they suggest that listeners contrast talker-specific rates. -
Maslowski, M., Meyer, A. S., & Bosker, H. R. (2016). Slow speech can sound fast: How the speech rate of one talker has a contrastive effect on the perception of another talker. Talk presented at MPI Proudly Presents. Nijmegen, The Netherlands. 2016-06-01.
-
McQueen, J. M., & Meyer, A. S. (2016). Cognitive architectures [Session Chair]. Talk presented at the Language in Interaction Summerschool on Human Language: From Genes and Brains to Behavior. Berg en Dal, The Netherlands. 2016-07-03 - 2016-07-14.
-
Meyer, A. S. (2016). Utterance planning and resource allocation in dialogue. Talk presented at the Psychology Department, University of Geneva. Geneva, Italy. 2016-05-09.
-
Meyer, A. S. (2016). Utterance planning and resource allocation in dialogue. Talk presented at the International Workshop on Language Production (IWLP 2016). La Jolla, CA, USA. 2016-07-25 - 2016-07-27.
Abstract
Natural conversations are characterized by smooth transitions of turns between interlocutors. For instance, speakers often respond to questions or requests within half a second. As planning the first word of an utterance can easily take a second or more, this suggests that utterance planning often overlaps with listening to the preceding speaker's utterance. A specific proposal concerning the temporal coordination of listening and speech planning has recently been made by Levinson and Torreira (2016, Frontiers in Psychology; Levinson, 2016, Trends in Cognitive Sciences). They propose that speakers initiate their speech planning as soon as they have understood the speech act and gist of the preceding utterance. However, direct evidence for simultaneous listening and speech planning is scarce. I will first review studies demonstrating that both comprehending spoken utterances and planning them require processing capacity and that these processes can substantially interfere with each other. These data suggest that concurrent speech planning and listening should be cognitively quite challenging. In the second part of the talk I will turn to studies examining directly when utterance planning in dialogue begins. These studies indicate that (regrettably) there are probably no hard-and-fast rules for the temporal coordination of listening and speech planning. I will argue that (regrettably again) we need models that are far more complex than Levinson and Torreira's proposal to understand how listening and speech planning are coordinated in conversation -
Weber, K., Meyer, A. S., & Hagoort, P. (2016). The acquisition of verb-argument and verb-noun category biases in a novel word learning task. Poster presented at Architectures and Mechanisms for Language Processing (AMLaP 2016), Bilbao, Spain.
Abstract
We show that language users readily learn the probabilities of novel lexical cues to syntactic information (verbs biasing towards a prepositional object dative vs. double-object dative and words biasing towards a verb vs. noun reading) and use these biases in a subsequent production task. In a one-hour exposure phase participants read 12 novel lexical items, embedded in 30 sentence contexts each, in their native language. The items were either strongly (100%) biased towards one grammatical frame or syntactic category assignment or unbiased (50%). The next day participants produced sentences with the newly learned lexical items. They were given the sentence beginning up to the novel lexical item. Their output showed that they were highly sensitive to the biases introduced in the exposure phase.
Given this rapid learning and use of novel lexical cues, this paradigm opens up new avenues to test sentence processing theories. Thus, with close control on the biases participants are acquiring, competition between different frames or category assignments can be investigated using reaction times or neuroimaging methods.
Generally, these results show that language users adapt to the statistics of the linguistic input, even to subtle lexically-driven cues to syntactic information.
Share this page