Deviant processing of audiovisual stimuli in aphasia D¨orte Hessler∗, Roel Jonkers and Roelien Bastiaanse School of Behavioral and Cognitive Neurosciences, Center for Language and Cognition Groningen, University of Groningen

Introduction During speech perception not only auditory but also visual information (seen speech) is processed (Rosenblum, 2008). Already in 1976 McGurk and MacDonald reported on the role of visual information in auditory processing. In their study subjects watched dubbed videos with non-matching auditory and visual information and had to report what they perceived. Instead of answering with the auditory (/ba/) or the visual (/ga/) component of the video, they mostly reported a fusion of both (/da/). This effect, called the ‘McGurk effect’, has been a strong argument for primary multimodal processing, which means that processing of both auditory and visual information occurs under all circumstances rather than visual information being a mere fall-back mechanism that only applies when needed. The McGurk effect has been previously investigated in aphasic populations (Campbell et al., 1990; Klitsch, 2008). Klitsch (2008) compared the performance of a group of aphasic subjects to a group of age-matched non-brain-damaged control subjects. Participants had to watch a video and choose between three answer options: the McGurktype answer, the auditory component or the visual component of the video. The response pattern (mostly McGurk type answers, followed by auditory and visual answers) did not differ significantly between the groups. However this task was an offline measure and therefore provided limited information. In the current study we investigated audiovisual processing in aphasic and healthy listeners by combining offline scores with online reaction times. The aim was to find out whether audiovisual processing really does not differ between healthy and aphasic processing.

Procedure The participants of the current study were three aphasic subjects and a group of fourteen non-brain-damaged control subjects within the same age-range. The aphasic subjects WB (male, 57 years, Wernicke’s aphasia, 148 months post onset), EK (male, 48 years, amnestic aphasia, 16 months post onset) and JH (female, 51 years, mixed aphasia, 44 months post ∗

e-mail: [email protected]

1

onset) all suffered from a deficit in the auditory analysis of speech, as measured with the nonword discrimination task from the Dutch PALPA Battery (Bastiaanse, Bosje, & Visch-Brink, 1995). The current study employed a nonword identification task. The subjects watched videos of a speaker pronouncing a syllable. Afterwards they had to choose which of three written syllables matched the video. This task was carried out in three conditions: ‘auditory only’ (subjects could only hear the speaker), ‘audiovisual’ (the speaker could be heard and seen) and ‘McGurk’ (auditory /p/ dubbed onto visual /k/, expected percept /t/). The material consisted of CVC(C) syllables that formed strings phonologically possible but non-existent in Dutch. All stimuli started with either /k/, /p/, or /t/. The written answer choices consisted of the target and two distractors (same rhyme as the target but different initial phoneme). They were presented simultaneously in a fixed arrangement, starting with /k/ on top, /p/ in the middle and /t/ at the bottom of the screen. Next to the answer patterns also the reaction times were recorded for all participants.

Results The data of the aphasic subjects were analyzed individually and compared to the group of non-brain-damaged speakers. Table 1 provides an overview of the results. Generally the reaction times of all aphasic subjects were longer than those of the controls. Auditory Only Initials WB EK JH

correct

RT

Audiovisual correct

RT

McGurk (per answer type) McGurk (/t/) Auditory (/p/) Visual (/k/) Amount RT Amount RT Amount RT

53% 59% 55%

2176ms 2718ms 2755ms

73% 76% 89%

1674ms 2516ms 2353ms

50% 18% 39%

2120ms 2119ms 2653ms

23% 46% 39%

2316ms 2020ms 2718ms

27% 36% 22%

2195ms 2297ms 2693ms

Controls 99% (mean)

1462ms

100%

1422ms

22%

2077ms

33%

1866ms

45%

1870ms

Table 1: Overview of the results

For the ‘McGurk condition’, it was analyzed which answer-type was dominant: the auditory component, the visual component or the fusion (McGurk) type. A Friedman Anova revealed significant differences between answer types for the non-brain-damaged control group (X²= 30.964, df = 2, p <.001). A post-hoc analysis showed the following hierarchy for the non-brain-damaged control group: most answers are of the visual, fewer of the auditory and the fewest of the fusion (McGurk) type. For none of the aphasic subjects, however, a difference between answer types was found. The performance in the ‘McGurk condition’ was further analyzed with regard to the reaction times for each answer type chosen: There was a significant increase in reaction time for the control group whenever a McGurk type answer was chosen, while there were no influences of the chosen answer type on the reaction times of the aphasic subjects. These results are depicted in Figure 1. 2

Figure 1: Reactiontime per answertype *: Mann-Whitney test, p<.05

Discussion In the current study we showed that the processing of audiovisual information was not only slower in three aphasic subjects but also differed qualitatively from the non-braindamaged control subjects. For the control subjects reaction times differed depending on what answer type was chosen in the ‘McGurk condition’, while this was not true for either of the aphasic subjects. The reaction time of the non-brain-damaged subjects was the largest when they gave a McGurk type response. This could reflect additional resources necessary when a fusion of auditory and visual information is perceived. According to Rosenblum (2008) subjects often recognize the mismatch and still experience the McGurk effect. This recognition of the mismatch could be slowing down the processing. It is possible that the aphasic subjects relied solely on automatic processing, not recognizing the mismatch, and therefore not showing a slow-down in processing compared to the other conditions. This hypothesis however still needs to be tested as in the current study we did not record whether subjects were aware of mismatches.

References Bastiaanse, R., Bosje, M., & Visch-Brink, E. (1995). Psycholinguistic Assessment of Language Processing in Aphasia: The Dutch version. Hove, UK: Lawrence Erlbaum Associates Ltd. Campbell, R., Garwood, J., Franklin, S., Howard, D., Landis, T., & Regard, M. (1990). Neuropsychological studies of auditory-visual fusion illusions. four case studies and their implications. Neuropsychologia, 28 (8), 787-802. Klitsch, J. (2008). Open your eyes and listen carefully. Auditory and audiovisual speech perception and the McGurk effect in Dutch speakers with and without

3

aphasia (Doctoral dissertation, University of Groningen). Groningen Dissertations in Linguistics (GRODIL), 67 . McGurk, H., & MacDonald, J. (1976). Hearing lips and seeing voices. Nature, 264 (5588), 746-748. Rosenblum, L. D. (2008). Speech perception as a multimodal phenomenon. Current Directions in Psychological Science, 17 (6), 405-409.

4

Deviant processing of audiovisual stimuli in aphasia

of the video, they mostly reported a fusion of both (/da/). This effect, called the 'McGurk effect', has been a strong argument for primary multimodal processing, ...

119KB Sizes 1 Downloads 223 Views

Recommend Documents

Syntactic Processing in Aphasia - Language and Cognitive ...
That noted, we hasten to add that the Wernicke's patients are not likely to be entirely ..... All are variations on the same theme, namely, that syntactic limitations.

Perspectives-On-Audiovisual-Translation-Lodz-Studies-In ...
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Perspectives-On-Audiovisual-Translation-Lodz-Studies-In-Language.pdf. Perspectives-On-Audiovisual-Translatio

The perception of self-produced sensory stimuli in ...
sensation (caused by a piece of soft foam) on the palm of their left hand ..... Impaired central mismatch error-correcting behavior in schizo- phrenia. Archives of ...

Representation of Concurrent Stimuli by Population Activity in Visual ...
Dec 23, 2009 - How do neuronal populations represent concurrent stimuli? .... data set (Figures 1D and 1E) and an average of 98.5% of the variance in all nine ...

Audiovisual Celebrity Recognition in Unconstrained Web Videos
To the best of our knowl- edge, this ... method for recognizing celebrity faces in unconstrained web ... [10]. Densities within each cluster are approximated using a single Gaussian with a full .... King and Bill O'Reilly) are popular talk show hosts

Grammaticality Judgment in Aphasia
color maps that depict the degree of behavioral in- volvement .... same networks that subserve performance of the task. (Karni et ... in a computer-controlled procedure in which potential effects of .... Administration Medical Centers and newspaper a

audiovisual services -
Meeting Room Projector Package - $630 ... From helping small gatherings create a great impact to amplifying a stage experience for hundreds of attendees,.

Juvenile Deviant Behavior in an Immigrant Bangladeshi ...
1 Department of Criminal Justice Kutztown University of Pennsylvania, USA ... This study reviewed the official data of an English city for three years to see.

Prey capture behavior evoked by simple visual stimuli in ... - Frontiers
Dec 16, 2011 - ior in larval zebrafish, we developed “virtual reality” assays in which precisely ...... Visual stimuli were designed using custom software written.

Linguistics and Aphasia
then I call-INF to Maud sister. It is not the case, however, that .... controls, initially activate the most frequent meaning of an ambiguous word. However, for normal .... Paper presented at the conference onlinguistic theory, speech and language ..

Preference for Audiovisual Speech Congruency in ... - Floris de Lange
Auditory speech perception can be altered by concurrent visual information. .... METHODS .... analysis in an independent data set in which participants listened ...

Preference for Audiovisual Speech Congruency in ... - Floris de Lange
multisensory fusion (see Figure 1). ... fected by perceptual fusion of incongruent auditory and ..... Brugge, J. F., Volkov, I. O., Garell, P. C., Reale, R. A., &. Howard ...

Expectations accelerate entry of visual stimuli into ...
Jun 26, 2015 - three different types of cues: 3a, explicitly predictive word cues (e.g., the word ..... alternative accounts including response bias, response priming, or pre-existing .... The free-energy principle: A unified brain theory? Nature ...

Prey capture behavior evoked by simple visual stimuli in ... - Frontiers
Dec 16, 2011 - fitting an ellipse to each eye and recording the angle between the long axis of the ... Visual stimuli were designed using custom software written.

Audiovisual tools as integrated part of information transfer ... - EFITA
and to create a lasting commitment, a dedicated communication platform such as a special news program ´Beet-TV´ is most effective. By means of such a platform with different audiovisual elements a more personalized, durable, comfortable, and perhap

EPUB Sociology of Deviant Behavior - Marshall Clinard ...
That is to say, from the two opposites the World is born Sociology of Deviant Behavior in the Image of its Creator. See Sociology of Deviant Behavior their job pages for medical and non-medical field vacancies. ... corporate crime; and the importance

Julia Dalcin - SOCIAL MEDIA AND AUDIOVISUAL ...
hftLo, If]qLo, eflifs, wfld{s, n}+lus lje]b / ;a} k|sf/sf hftLo 5'jf5"tsf]. cGTo u/L ... Retrying... Julia Dalcin - SOCIAL MEDIA AND AUDIOVISUAL COMMUNICATIONS.pdf.