Next Article in Journal
Exploring Causes of Depression and Anxiety Health Disparities (HD) by Examining Differences between 1:1 Matched Individuals
Previous Article in Journal
On the Design of Broad-Based Neuropsychological Test Batteries to Assess the Cognitive Abilities of Individuals with Down Syndrome in the Context of Clinical Trials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Audiovisual Lexical Retrieval Deficits Following Left Hemisphere Stroke

1
Department of Diagnostic Radiology & Nuclear Medicine, University of Maryland School of Medicine, Baltimore, MD 21201, USA
2
Department of Clinical Research for Rehabilitation, Korea National Rehabilitation Research Institute, Seoul 01022, Korea
3
Kansas City University of Medicine & Biosciences, Kansas City, MO 64106, USA
4
Department of Psychology, University of Maryland Baltimore County (Shady Grove), Rockville, MD 20850, USA
*
Author to whom correspondence should be addressed.
Brain Sci. 2018, 8(12), 206; https://doi.org/10.3390/brainsci8120206
Submission received: 19 October 2018 / Revised: 18 November 2018 / Accepted: 27 November 2018 / Published: 28 November 2018

Abstract

:
Binding sensory features of multiple modalities of what we hear and see allows formation of a coherent percept to access semantics. Previous work on object naming has focused on visual confrontation naming with limited research in nonverbal auditory or multisensory processing. To investigate neural substrates and sensory effects of lexical retrieval, we evaluated healthy adults (n = 118) and left hemisphere stroke patients (LHD, n = 42) in naming manipulable objects across auditory (sound), visual (picture), and multisensory (audiovisual) conditions. LHD patients were divided into cortical, cortical–subcortical, or subcortical lesions (CO, CO–SC, SC), and specific lesion location investigated in a predictive model. Subjects produced lower accuracy in auditory naming relative to other conditions. Controls demonstrated greater naming accuracy and faster reaction times across all conditions compared to LHD patients. Naming across conditions was most severely impaired in CO patients. Both auditory and visual naming accuracy were impacted by temporal lobe involvement, although auditory naming was sensitive to lesions extending subcortically. Only controls demonstrated significant improvement over visual naming with the addition of auditory cues (i.e., multisensory condition). Results support overlapping neural networks for visual and auditory modalities related to semantic integration in lexical retrieval and temporal lobe involvement, while multisensory integration was impacted by both occipital and temporal lobe lesion involvement. The findings support modality specificity in naming and suggest that auditory naming is mediated by a distributed cortical–subcortical network overlapping with networks mediating spatiotemporal aspects of skilled movements producing sound.

1. Introduction

The formation of a coherent percept requires binding sensory features related to our perception of what we hear (auditory) and see (visual) with semantics. This is distinguished from multisensory integration based on a cognitive decision to integrate auditory–visual information [1]. Integration of sensory information can improve performance, response speed and localization [2]. Although there are few investigations focused on auditory–visual integration during object recognition and naming, there is evidence multisensory presentation may facilitate lexical retrieval in aging and speech recognition [3,4].
Visual object tasks have predominated investigations of neural mechanisms mediating naming and have implicated the left posterior and temporoparietal regions as critical for word retrieval [5,6,7]. Conversely, auditory naming localization has not been well investigated with limited understanding of neural processing pathways for nonverbal sound recognition. Auditory description naming has been associated with disruption to the anterior temporal lobe in epilepsy patients, while both auditory and visual naming have demonstrated impairment following posterior temporal lobe direct cortical stimulation and postsurgical resection [8,9,10].
Language perception involves sensory systems including auditory and visual modalities, but previous studies in language systems have focused on semantics with minimal consideration of sensory systems [11]. The neural circuits underlying language functions involve the inferior frontal cortex (including Broca’s area), the superior temporal cortex (including Wernicke’s area), and the interplay between these regions. Viewing and naming pictures of objects and tools associated with skilled movements (as opposed to other object categories) may selectively activate left ventral premotor and posterior parietal cortex reflecting ventral (“what”) and dorsal (“where”) visual processing streams [12]. Despite evidence suggesting specialized auditory networks processing “what” and “where”, the functional organization of sound recognition has been difficult to characterize in terms of distinct networks [13]. Sound object recognition requires several steps in segregating spatial, temporal, and synchronous cues, and also in semantic identification of the sounds associated with actual objects.
Lexical retrieval deficits are common and persistent following left hemisphere stroke, but the anatomical factors predicting recovery and compensatory mechanisms remain unclear. In addition to age-related changes in the efficiency of lexical retrieval, naming is frequently impaired in aphasia following a vascular event to the left hemisphere that disrupts cortical–subcortical networks [14]. Voxel-based lesion symptom studies have predicted lexical retrieval deficits based on damage to a cortical–subcortical network involving the left mid-posterior inferior temporal cortex and underlying connections, especially the longitudinal fasciculus [15].
Previous studies have typically relied on visual confrontation naming (i.e., picture-naming) to assess lexical retrieval failures. It is possible that there are differences in visual and auditory naming as well as multisensory integration based on lesion distribution. There is evidence that in addition to damage to the left cortex, the underlying connections contribute to persistent naming deficits, supporting both cortical and subcortical contributions to lexical retrieval [15]. In healthy older adults, auditory naming proved more sensitive to detecting naming failures, but the differences in cortical–subcortical contributions to visual and auditory naming have not been previously investigated [3]. Therefore, it remains unclear how nonverbal processing is impacted post-stroke and whether select patients may benefit from multisensory cues similar to what has been demonstrated in normal aging.
Findings of numerous studies argue for modality specificity, although the critical regions mediating audiovisual integration in naming are not well established. Conversely, there are numerous investigations focused on audiovisual integration in speech processing that have revealed critical contributions from the auditory cortex, the posterior superior temporal sulcus (STS), and the posterior inferior frontal gyrus [16,17,18]. In addition to these regions, lesion studies have also implicated lateral occipital visual cortex in addition to multisensory regions in the STS as critical to multisensory speech integration [4]. Unlike multisensory speech integration, the overlapping network mediating auditory verbal and nonverbal (sound) naming has not been well-characterized. Thus, the brain regions impacting audiovisual integration in lexical retrieval warrant further elucidation to better characterize lexical retrieval and develop potential intervention models.
Furthermore, it remains unclear how the auditory network overlaps with visual processing networks related to biological motion and visual recognition of tools associated with multisensory processing of object-associated action knowledge [19]. Determination of the more effective unisensory response (e.g., to visual or auditory) in comparison to simultaneous presentation of two sensory stimuli (audiovisual), allows for estimation of gains attributable to multisensory interactions (visual–auditory presentation). It is conceivable that integrating information across modalities may improve behavioral outcomes and response speed [2]. Study findings will help determine whether multisensory integration is a useful approach to facilitate naming in impaired populations such as those with stroke and aphasia.
The objective of this study was to extend our previous work in auditory and audiovisual naming in aging to left hemisphere stroke. We evaluated how lesions in the dominant hemisphere for language (left hemisphere stroke) impact naming to visual, auditory, and multisensory cues to determine which patients benefit from sensory cues and multisensory information to facilitate lexical retrieval. Related to degree of language deficits following stroke and lesion location, we predicted there would be significant variability across naming conditions elucidating neural mechanisms mediating modality specificity and multisensory integration. These findings may be capable of assisting prediction of which patients will display persisting language deficits following stroke, and allow realistic goals for rehabilitation efforts to reduce disability.

2. Materials and Methods

The protocol conformed to the Declaration of Helsinki and was approved by the Institutional Review Boards (IRB) for the University of Kansas Medical Center (KUMC), Emory University School of Medicine, and the University of Maryland School of Medicine (UMB HP-00060452). All subjects gave written informed consent prior to entering the study, and caregivers were consulted for all stroke patients regarding the details of the experiment prior to their enrollment.

2.1. Participants

One hundred and sixty older participants comprised of 42 left hemisphere ischemic stroke patients (LHD; 27 males and 15 females) and 118 healthy controls (53 males and 65 females) were recruited from both KUMC and Emory medical centers, and data was processed at UMB. Left hemisphere stroke patients were characterized based on involvement of cortical (CO, n = 17), cortical–subcortical (CO–SC, n = 10), and subcortical lesion locations (SC, n = 15) derived from clinical neuroimaging results (see Table 1 for detailed lesion location based on clinical MRI or CT). Neuroimaging was not obtained on healthy controls. The average age of healthy controls (63.1 years) was slightly higher than stroke patients (60.4 years; see Table 2), but was not statistically significant. Age between lesion groups (SC = 60.9; CO = 63; CO–SC = 57.1 years of age) revealed lower age for CO–SC relative to CO patients. The average educational level of the healthy participants (16.5 years) was higher than stroke patients (13.4 years), but education was not different between lesion groups (SC = 13.1; CO = 13.4; CO–SC = 13.8 years). Age and education were both utilized in the statistical analyses as covariates.
All subjects were native English speakers, and healthy controls were free of dementia and language impairments (Table 1). Subjects were screened for cognitive deficits with the Mini-Mental Status Examination [21], and also received a comprehensive neuropsychological examination to ensure that healthy controls were free of significant language and cognitive deficits. Handedness was determined by the Edinburgh Handedness Inventory [22], and most participants were right hand dominant. However, five controls and 8 stroke patients were left hand dominant, but the results did not differ between right and left-handed subjects. General exclusion criteria for normal controls and stroke included: (1) chronic, serious medical conditions, (2) other neurological disease, (3) psychiatric disease including current depression and anxiety untreated, or (4) significant substance abuse/dependence. Stroke exclusion criteria included: (1) evidence of bilateral or right hemisphere stroke, or (2) significant confluent white matter abnormalities. Left hemisphere stroke patients were characterized based on involvement of cortical (CO, n = 17), cortical–subcortical (CO–SC, n = 10), and subcortical lesion locations (SC, n = 15) derived from clinical neuroimaging results (see Table 2 for detailed lesion location based on clinical MRI or CT). Lesion stroke interval was calculated based on the number of months that elapsed between the date of the stroke to language examination. The stroke-testing interval range spanned from 4.5 months to 6 years for the majority of the LHD patients, with two patients with intervals extending outside this range (see Table 2 for characteristics of patients enrolled). Stroke interval was utilized as a predictor in the regression models.

2.2. Language, Neuropsychological, and Audiological Assessments

Hearing and Audiology Screening. Subjects were screened with the Hearing Handicap Inventory for the Elderly (HHIE-S) [23,24], and an audiometer screening to identify hearing loss that may interfere with processing of auditory stimuli. The hearing screening test was conducted with an audiometer (MA27) at 25 dB at 250, 500, 1000, 2000, 4000, and 8000 Hz bilaterally. This was followed by a threshold test conducted for each ear using the same frequency levels. Eighteen of the 160 enrolled subjects did not receive an audiology evaluation (nine normal controls and nine stroke patients), although the majority of these participants did alternatively complete the HHIE-S (11 of 18) and most (seven of the 11) and responses were consistent with no hearing handicap. Participants were evaluated based on differences between the ears with a goal of obtaining less than 35 dB difference at any frequency level set similar to criteria in previous studies [3,25]. All enrolled controls with the exception of three subjects met this criterion, and since they exceeded at only one frequency and obtained adequate scores on auditory naming, they were not excluded from the study. Audiological assessment was considered in a between groups fashion in the analysis.
Language Assessment. Participants were initially screened for the presence of anomia with the Boston Naming Test and verbal and semantic fluency [26,27]. Language was also characterized by the first part of the Western Aphasia Battery [20] to derive an aphasia quotient (see WAB-AQ Table 1 and Table 2). Of the 42 LHD stroke patients included in the study, 24 stroke patients were classified into one of the aphasia subtypes based on WAB criteria [20,28]. The majority of the aphasic patients were classified as Anomic (n = 18), Broca’s aphasia (n = 2), Transcortical Motor aphasia (n = 2), followed by the remaining other aphasia subtypes (see Table 2). Stroke participants demonstrated adequate auditory comprehension based on the WAB (AC > 60%; see Table 1 for means). Naming performance on the WAB are also provided in Table 2 for each individual stroke patient by lesion subgroups. Three stroke patients with significant auditory comprehension disturbance (< 60 auditory comprehension) and associated severe naming deficits were eliminated as outliers. Control participants were required to have normal language and cognitive performance and adequate performance on experimental tasks for inclusion into the analysis.

2.3. Experimental Conditions

The participants were administered an experimental object naming battery developed and described in a study investigating naming performance of younger and older adults [3]. The same common man-made objects with manipulable properties (e.g., toothbrush, saw, zipper, and telephone) were utilized in this study, and contained identical normative properties of verbal fluency, name agreement, alternate response ranges, and familiarity. The properties of the naming battery can be obtained from the original publication and include verbal frequency, familiarity, written frequency and other details. The naming battery presented 25 objects across the following conditions in a counterbalanced order across subjects: (1) Auditory (A)—digitized sounds of objects normalized for loudness equivalence; (2) Visual (V)—static color photographs of objects without images of motion; and (3) Audiovisual (AV)—sounds and photographs of the same objects presented simultaneously. Auditory naming was retested following a brief delay to determine improvement following the experiment in comparison to the first auditory presentation. Auditory2 naming was completed in all LHD patients and a subset of the control group (n = 67) 20–30 minutes following Auditory 1 consistent with delayed recall in standardized tests of memory in neuropsychological assessment and our previous paper [3].
Presentation stimulus delivery software program (Neurobehavioral systems, version 10.1) via a Dell laptop and headphones provided 4 second stimulus presentation. Subject responses preceding 4 second stimulus presentation completion were recorded for reaction time by the experimenter. Subjects were allowed a total of 25 seconds to produce a naming response even though the stimulus presentation was only 4 seconds. General or nonspecific responses (e.g., naming a general semantic class of object) were followed-up by the examiner, and self-corrected responses were marked correct when provided in the 25 second time frame. Reaction times were measured from stimulus presentation to first response generation and recorded by the experimenter utilizing the Presentation program developed for the experiment. Response accuracy and reaction times in milliseconds were collected by the experimenter on the laptop.

2.4. Analysis

A multivariate mixed design was conducted with age and education as covariates allowing for inclusion of a within-subjects factor for sensory condition (auditory, visual, and auditory–visual), a between-subjects factor (control versus stroke groups), and the interaction between the two in one model. This model is conservative and controls for multiple comparisons with several dependent measures (accuracy and reaction time) and is not dependent upon sphericity. Univariate values were examined when the multivariate model was significant, followed by post-hoc analyses with Bonferroni correction when univariates were significant. Additional multiple regression analyses were conducted to explore the impact of specific lesion location, aphasia severity, stroke interval, and cognitive dysfunction.
Between-group factors allowed for the following group comparisons: (i) LHD stroke to normal controls, (ii) LHD lesion locations (CO = Cortical, CO–SC = Cortical–Subcortical, and SC = Subcortical) relative to normal controls. Finally, several stepwise regression models evaluated cognitive, language and lesion variables as predictors of naming performance by condition.

3. Results

3.1. LHD versus Controls

A multivariate model compared naming accuracy and reaction times in LHD stroke patients (n = 42) relative to controls (n = 118) across conditions. A multivariate model utilized age and education as covariates, group (controls, LHD) as the between-subjects factor and sensory condition (auditory, visual, and multisensory) as the within-subjects or repeated factor. The multivariate model was significant for age, F(2,155) = 20.6, p < 0.0001, but not education, F(2,155) = 0.04, p < 0.96. The multivariate model revealed significant between-subjects effects for group, F(2,155) = 18.5, p < 0.0001, within-subjects effects revealing changes in performance across sensory conditions, F(4,153) = 2.5, p = 0.042, and a sensory by group interaction, F(4,153) = 5.3, p < 0.0001.

3.1.1. Group

Univariate tests revealed a significant between-subjects effect of group for naming accuracy, F(1,156) = 33.4, p < 0.0001, and correct reaction time, F(1,156) = 19.6, p < 0.0001. Controls had greater naming accuracy (mean difference = 10.9) and faster reaction times for correctly named items (mean difference = 0.77 s) across all conditions relative to LHD patients p < 0.0001 (Table 3).

3.1.2. Sensory Condition

Univariate tests displayed significant within-subjects effects for sensory condition on naming accuracy, F(2,312) = 3.3, p = 0.038, but not for correct reaction times, F(2,312) = 1.05, p = 0.351. The A condition displayed lower accuracy than the V (mean difference = −24.2) and AV conditions (mean difference = −25.2), p < 0.0001. However, the V and AV conditions did not differ overall in terms of accuracy (mean difference = −0.974, p = 0.206; Table 3).

3.1.3. Sensory Condition by Group Interaction

There was also a sensory condition by group interaction for naming accuracy, F(2,312) = 14.1, p < 0.0001, and for correct reaction times, F(2,312) = 3.2, p = 0.043.
Naming Accuracy. Pairwise comparisons with Bonferroni correction revealed that LHD patients had lower accuracy for each of the sensory conditions relative to controls (mean difference A = −18.6, mean difference V = −6.8, and mean difference AV = −7.5), p < 0.001. However, LHD patients had greater differences between auditory naming accuracy relative to visual and multisensory accuracy (mean differences presented in Table 3 and Table 4). Also, while normal controls had significantly higher naming accuracy in the multisensory compared to the visual condition, p = 0.045, LHD patients did not differ between the visual and multisensory conditions, p = 1.0 (see Table 3 and Table 4).
Reaction Times for Correctly Named Items. Pairwise comparisons with Bonferroni correction revealed that LHD patients had slower reaction times for each of the sensory conditions relative to controls (mean difference A = 1.06, mean difference V = 0.469, and mean difference AV = 0.786), p < 0.001. However, LHD patients’ reaction times improved more significantly than controls when visual cues were provided in place of auditory cues, or with the addition to auditory cues in a multisensory condition (see Table 4 for mean differences).

3.2. Lesion Group

A multivariate model compared naming accuracy and reaction times in LHD stroke patients with CO lesions (n = 17), CO–SC lesions (n = 10), and SC lesions (n = 15) relative to controls (n = 118) across conditions. The multivariate model was significant for age, F(2,153) = 19.9, p < 0.0001, but not education, F(2,153) = 0.007, p < 0.993. The multivariate model revealed significant between-subjects effects for lesion group, F(6, 308) = 8.5, p < 0.0001, within-subjects effects of sensory conditions, F(4,151) = 2.65, p = 0.035 and a sensory by lesion group interaction, F(12,459) = 2.95, p = 0.001.

3.2.1. Group

Univariate tests revealed significant between-subjects effects for lesion group on naming accuracy, F(3,154) = 16.1, p < 0.0001, and correct reaction time, F(3,154) = 11.4, p < 0.0001.
Naming Accuracy. Controls had higher naming accuracy than patients with CO (mean difference = 16.1, p < 0.0001) and CO–SC lesion locations (mean difference = 10.6, p = 0.004), but were not significantly different than SC patients (mean difference = 5.1, p = 0.347). SC patients had higher naming than CO patients (mean difference = 11.1, p = 0.003), but were not significantly different from CO–SC patients (Figure 1A–C).
Reaction Times for Correctly Named Items. Controls produced faster reaction times than CO and CO-CS patients (CO mean difference = −1.2, p < 0.0001; CO–SC mean difference = −0.984, p < 0.005), but were not significantly different from the SC group (mean difference = −0.15, p = 1.0). The SC group produced faster times than the CO group (mean difference = −1.01, p < 0.005), but not the CO-CS group (mean difference = 0.83, p < 0.08).

3.2.2. Sensory Condition

Univariate within-subjects effects were not significant for sensory condition on naming accuracy, F(2,308) = 2.7, p = 0.07 or for correct reaction time, F(2,308) = 1.4, p = 0.26 (Table 5).

3.3.3. Sensory Condition by Lesion Group Interaction

There was a significant within-subjects effects for sensory condition by lesion location interaction for naming accuracy, F(6,308) = 5.5, p < 0.0001, and correct reaction time, F(6, 308) = 1.4, p = 0.041.
Naming Accuracy. Controls displayed higher naming accuracy compared with CO group across all conditions, although the differences were largest in the auditory condition. Controls were only significantly different from patients with subcortical involvement on the auditory condition. SC patients had higher naming accuracy for auditory and multisensory conditions relative to the CO patients. None of the lesion groups had significantly different performances in the auditory condition. See Table 5 for mean differences across groups.
Reaction Times for Correctly Named Items. Controls had faster reaction times for correctly named items across conditions relative to patients with cortical involvement (with the exception of visual naming for CO–SC patients). Reaction times were not significantly different between controls and SC patients for any of the conditions. However, the SC group displayed faster reaction times than the CO group for both V and AV conditions, but they were not faster in auditory naming. The CO-CS group did not differ from the other lesion groups across conditions (Table 5 for significance and mean differences).

3.3. Stepwise Multiple Regression

To investigate the effect of stroke associated variables (stroke interval, language, and cognitive dysfunction) on naming accuracy, three separate regression analyses were conducted for each of the sensory conditions. Predictor variables (age, education, stroke interval, MMSE, WAB-AQ, verbal fluency, and LNS—Letter Number Sequencing) were examined in stepwise regression models predicting auditory, visual, and multisensory naming accuracy. Age accounted for a significant amount of variance in auditory naming accuracy (43.5% of the variance, p < 0.001), with the addition of semantic fluency (8.9% variance), LNS (3.7% variance), and MMSE (7.6%) accounting for additional variance (63.7% total variance explained; Table 6). The MMSE explained 74.3% of the variance in visual naming accuracy (p < 0.0001), with the addition of LNS accounting for an additional 1.9% variance (76.2% total variance explained). Multisensory naming accuracy was explained by aphasia severity (WAB-AQ, p < 0.001), with the MMSE accounting for additional variance (3.9%; 71.1% total variance explained). See Model 1 Table 6.
In a second set of stepwise regression analyses, the impact of lesion location (temporal, parietal, occipital, frontal lobes, basal ganglia, thalamic, and white matter involvement, Figure 2) were evaluated on naming accuracy across conditions, while controlling for age and education. In Model 2, age accounted for significant variance in both auditory and visual naming accuracy (44.7% and 34.7% variance, respectively), followed by additional variance accounted for by temporal lobe involvement (7.7% and 15.6% variance, respectively). Conversely, multisensory naming accuracy was best predicted by occipital lobe involvement explaining 27% of the variance, with additional variance explained by age (14.5%) and temporal lobe involvement (16.1%; 57.6% total variance explained). See model 2 Table 6.

Control Analyses

Control analyses were conducted on a repeated auditory condition to assess learning, improvement in auditory naming and recognition (i.e., Auditory 2), to assess for impact of peripheral hearing impairment and randomization order on the test results.

3.4. Auditory2 Naming

All LHD patients and a subset of the control subjects (n = 68) repeated the A naming condition following a 20 to 30 min delay at the end of the experiment. The multivariate mixed model with age and education as covariates, and naming accuracy and correct reaction times as dependent variables evaluated changes in A accuracy across times 1 and 2. Age was the only covariate to significantly adjust the variance between groups, F(2,104) = 22.79, p < 0.0001. There was a significant between-subjects effect for group, F(2,104) = 10.75, p < 0.0001, but the within-subjects effects for change in naming from initial to delay, F(2,104) = 1.97, p = 0.15, and condition by group interaction, F(2, 104) = 2.46, p = 0.09, were not significant. Univariate tests revealed a significant between-subjects effect of group for A naming accuracy F(1,105) = 21.17, p < 0.001, and correct reaction time, F(1,105) = 9.27, p < 0.005. The results are consistent with lower A accuracy for LHD subjects (mean = 69.19) relative to normal controls (82.99) and longer reaction times (LHD = 4.58; controls = 3.64 s) across both initial and delayed conditions.

3.5. Audiometric Analyses

Differences between the left and right ears were computed for each subject at each frequency levels of 250, 500, 1000, 2000, and 4000 Hz. Group comparisons did not reveal differences between controls and LHD patients, F(5, 150) = 1.8, p = ns. Group comparisons based on stroke location were not significant at any of the frequency levels and thus were not utilized in adjustment of the variance in auditory naming performance.

3.6. Accuracy Analyses for the First Block of Stimulus Items

To evaluate the influence of counterbalancing, we analyzed the first block of each condition as a control analysis. A naming performance was higher when the first block was V (76.6%) or AV (80.3%) compared to A first block presentation (60.7%). V accuracy was higher when the first block was A (94.8), or AV (94.5%) compared to V (89.7%). The first block condition did not significantly impact AV accuracy. However, LHD CO patients had lower A naming accuracy irrespective of the first block order.

4. Discussion

This study evaluated lexical retrieval performance across auditory (nonverbal) and visual modalities in older healthy adults relative to left hemisphere stroke patients. We predicted there would be significant variability across naming conditions related to degree of language deficits following stroke and based on lesion location. Variability across patients clarify neural mechanisms mediating modality specificity as well as multisensory integration in lexical retrieval. Investigation of how language deficits and lesion location influence lexical retrieval under different sensory conditions provides insights into post-stroke intervention potential. Because of variability in deficits in cognition and language and lesion location, future therapeutic interventions should be tailored to patient characteristics to obtain optimal treatment outcomes [29].
Healthy adults produced more accurate and more rapid responses in naming compared to patients with left hemisphere stroke. There was also a sensory condition effect, with all subjects producing less accurate and slower naming responses in the auditory condition compared to the visual and multisensory conditions. Across groups, there was substantial performance improvement in accuracy when naming to visual cues compared to auditory cues. Past studies have typically relied on visual confrontation naming to assess persisting naming deficits. Consistent with these previous results, our study revealed that patients with cortical lesions were more impaired than those with subcortical lesions on both visual and multisensory naming, although they did not differ from patients with cortical−subcortical lesions. Conversely, auditory naming was the most sensitive to detecting deficits and revealed lower performance than other conditions for all subjects. Cortical and cortical−subcortical patients performed worse than healthy controls on auditory naming, suggesting auditory naming may provide increased sensitivity in detecting persisting lexical retrieval deficits across lesion groups. Although patients with cortical lesions had lower naming accuracy on visual and multisensory naming, they did not have significantly lower performance on auditory naming.

4.1. Lesion Location and Lexical Retrieval

Cerebrovascular insults often affect the middle cerebral artery constraining the distribution of affected perisylvian regions impacting language dysfunction. Comparison to other types of stroke such as those limited to subcortical regions can help elucidate critical regions involved in lexical retrieval. Left hemisphere stroke is frequently associated with language impairments, and there is evidence that damage to the mid-posterior inferior temporal cortex and underlying connections result in lexical retrieval deficits [15]. Extent of disconnection from language regions in the inferior frontal region has demonstrated predictive validity for degree of naming deficits and may account for persisting naming deficits following subcortical stroke [30]. Similar to results of lexical retrieval following neurosurgical removal of glioma, our findings support a role for the subcortical network in mediating lexical retrieval although the contribution is less robust and may depend on the degree of underlying white matter disconnection [15].
Our findings are similar to previous studies and support the critical role of the temporal lobe in naming irrespective of sensory modality, and suggest that visual confrontation naming may rely on different neural mechanisms than either auditory naming or audiovisual integration. The finding of significantly lower scores on visual confrontation naming exclusively in patients with cortical lesions, argues that visual naming is mediated by a less distributed network. Thus, patients with left hemisphere stroke with milder lexical-retrieval declines may go undetected on typical tests of visual confrontation naming. Conversely, auditory confrontation naming may be more sensitive to milder deficits and capable of detecting naming deficits in other lesion locations. Although cognitive functioning and temporal lobe involvement predicted naming accuracy across conditions, auditory cues required additional processing demands. Auditory naming was more impacted by age effects and semantic capacity compared with the other conditions. Our findings support left cortical localization of object semantics unrelated to discrimination of acoustically related sounds [31]. Future investigations are needed to delineate the distributed network for naming in healthy individuals, although our findings confirm the critical role of the temporal lobe in lexical retrieval irrespective of modality of elicitation.
Evidence from fMRI and lesion symptom mapping in aphasia suggest that verbal and nonverbal auditory processing relies on overlapping cortical regions within the left hemisphere, whereas there is greater activation in the right hemisphere for processing environmental sounds [32,33]. However, the distribution of auditory verbal and auditory nonverbal (sound) neural networks has not been well characterized. Studies of patients with lateralized lesions have displayed dissociation between auditory recognition and localization, and motion perception suggesting a lateral recognition pathway and a medial and posterior spatial pathway [34]. Furthermore, the association between nonverbal auditory recognition and linguistic processing has not been clearly established, although it has been considered that auditory recognition proceeds in parallel as opposed to hierarchical processing [35].
A distributed frontal, parietal, and temporal activation network has been identified for unimodal auditory and visual stimuli, while posterior superior temporal sulcus/middle temporal gyrus, dorsal lateral prefrontal cortex, and ventral temporal cortex in previous studies associated to audiovisual stimuli [17]. In our study, integration of simultaneous congruent information from a second modality (auditory added to visual) relied on intact left cortical functioning since efficient multisensory lexical retrieval was only apparent for healthy older adults. Furthermore, multisensory integration in LHD patients was predicted by severity of language impairment, and lesion involvement of the temporal and occipital lobes. Language processing can be conceptualized in two different systems (sensory−conceptual and sensory−motor) that rely on auditory cortices interacting with projections to the temporal lobe and with the motor system to temporal parietal regions (dorsal stream) [36,37]. Similarly, our findings support the premise that the visual system interfaces with conceptual representations of objects through ventral (occipitotemporal) and dorsal object recognition streams (occipitoparietal) [38]. Our results of predictors of multisensory performance in LHD stroke support findings from neuropsychological lesion studies and fMRI investigations providing evidence that the processing of visual object categories is mediated by an occipitotemporal network [17,39,40,41].
There are numerous investigations focused on audiovisual integration in speech processing that have revealed critical contributions from the auditory cortex, the posterior superior temporal sulcus (STS), and the posterior inferior frontal gyrus [16,17,18]. In addition to these regions, lesion studies have also implicated lateral occipital visual cortex in addition to multisensory regions in the STS as critical to multisensory speech integration [4]. Unlike multisensory speech integration, the overlapping network mediating auditory verbal and nonverbal (sound) naming has not been well characterized. Thus, the brain regions impacting audiovisual integration in lexical retrieval warrant further elucidation to better characterize lexical retrieval and develop potential intervention models.

4.2. Reaction Times in Naming

Simultaneous processing of audiovisual cues increases reaction times, even when accounting for accuracy. This is consistent with increased cognitive demands required for simultaneous processing of audiovisual information, and variance in multisensory naming was predicted by general cognitive capacity as well as degree of language severity. These findings are not supportive of integration across modalities, but demonstrate an increased cognitive load required for dual processing of sensory data. This suggests that the left cortex is critically involved in multisensory processing required for improved efficiency, and is reflected in the more significant increase in reaction time for LHD patients. There were longer reaction times for patients with cortical stroke across conditions relative to controls and subcortical stroke patients, even when accounting for reaction times for correct items. Although cortical patients had longer reaction times than subcortical patients in the auditory condition, this did not reach statistical significance likely related to the small sample sizes and variability in performance. The cortical patients displayed increased reaction times across experimental conditions related to increased information processing demands for accessing semantics unrelated to auditory processing demands. Therefore, additional information from a second modality was not integrated effectively and there were no gains or multisensory enhancement in lexical retrieval following left cortical damage. The reaction times were not significantly different between any of the other groups irrespective of experimental condition.

4.3. Lesion Size & Characteristics

Differences in lesion size contribute to the severity of cognitive and language deficits across lesion groups since cortical strokes are larger than subcortical strokes. Previous investigations focused on post-stroke aphasia have documented clear impact of lesion size, although many of these investigations have relied on gross estimates of size (e.g., small, medium, and large) as opposed to precise volumetric measurements [30,42]. The full extent of brain regions affected following stroke is likely underappreciated by visual inspection alone, and is influenced by other factors including disconnection and hemodynamic changes [43,44]. Therefore, covarying for lesion size may be insufficient to fully account for functional differences in language, especially since size and location are correlated and could obscure the impact of lesion location on pattern of deficits.
Comparison of cortical lesions to cortical−subcortical and subcortical lesions provides some estimation of how lesion size and location impact severity and naming profiles. Our study findings revealed that both larger lesions (cortical−subcortical) as well as smaller lesions with subcortical involvement had greater accuracy in visual confrontation naming compared to patients with lesions confined to the cortex (larger lesions). These findings argue against lesion size alone as being able to account for naming accuracy, and the most severe deficits resulted from disruption of critical language regions in the left cortex. We also recognize that the stroke testing interval impacts degree of recovery and severity of both cognitive and language deficits. While stroke interval was not predictive in any of regression models, degree of language and cognitive deficits were predictive and may reflect some of the overlapping variance in this predictor variables. Similarly, longer stroke-testing intervals have less influence on the severity of deficits in larger lesions than smaller lesions, recognizing the interplay between lesion size, chronicity, and persisting deficits. Nonetheless, we acknowledge that lesion size contributes to the severity of language deficits, and thus one study limitation is the absence of detailed volumetric and morphometric measurements. Therefore, our study findings should be substantiated with future investigations carefully controlling for both detailed anatomical distribution of lesions, lesion size as well as functional imaging measurements.

4.4. Aging & Multisensory Processing

Age-associated auditory processing difficulties are common, although complex auditory processing deficits are distinct from hearing insensitivity. This is supported by equivocal hearing sensitivity across frequencies between left hemisphere stroke patients and healthy controls, and the amount of variance age accounted for in auditory accuracy. However, age accounted for significant variance in naming accuracy in most of our predictive models and not just for auditory naming. Therefore, hearing sensitivity is unable to account for differences in auditory naming between groups, and emphasizes the higher level of auditory processing required for successful lexical retrieval.
The ability to integrate in a multisensory fashion may be a critical marker of successful aging [45]. This can be related to variable aging including extent of brain atrophy, or degree of plasticity that can facilitate integration of audiovisual information that can access semantics and/or lexical representations [45,46]. Age-related changes in morphology implicate a critical role of the left superior temporal cortex volume in predicting semantic performance over time in healthy older adults [47]. There is some indication both temporal acuity and multisensory integration decline with aging [48]. This is supported by age-related changes in width and depth of cortical sulci demonstrating more profound age-related changes in multisensory as opposed to unisensory cortical regions [49]. Contrary to this evidence for declines in temporal acuity in aging, our current findings and previous work demonstrate improvements with multisensory naming evident only in older adults [3]. In summary, integration of multiple sensory environmental signals (audiovisual) into one percept relies heavily on the left superior temporal region which is susceptible to increased atrophy in aging or neurodegeneration as well as the occipital region for processing visual information [47,49,50].

5. Conclusions

Processing of environmental sounds may share overlapping neural networks with language related to the semantic categorization and skilled movements in the left hemisphere that also have a strong temporal component. There is evidence that correctly categorizing tool sounds activates a distributed left hemisphere cortical network that overlaps with tool manipulations that require integration of spatial and temporal features, suggesting that sound recognition may involve not only a “what” pathway, but also a “how” pathway associated with the motor actions that produce the sound [19,51]. It remains unclear whether this auditory network partially overlaps with visual processing networks related to biological motion and visual recognition of tools associated with action, which might presume multisensory processing of object semantics [19]. Although our results are capable of confirming a critical role of the left cortex in multisensory integration, stroke survivors with left cortical involvement and aphasia are less likely to benefit from multisensory cues in terms of naming efficiency or accuracy.
Multisensory cues that result in behavioral advantages require overlap of both spatial and temporal information that optimize integration across the senses. Study results indicate that outcomes can improve with multisensory information with respect to detection, localization as well as speed of responding if critical regions in the left cortex are spared [52,53,54]. Our results support involvement of a distributed left hemisphere network for auditory naming that overlaps with a tool-use distributed cortical−subcortical network mediating spatiotemporal features of action associated with environmental sounds (how pathway) [51,55].

Author Contributions

Conceptualization, B.H.P.; methodology, B.H.P. & H.C.; software, H.C.; formal analysis, B.H.P., H.C. & S.H.; investigation, B.H.P., H.C. & B.H.; resources, B.H.P.; data curation, B.H.P., H.C., B.H. & S.H.; writing—original draft preparation, B.H.P.; writing—review and editing, B.H.P., H.C., B.H. & S.H.; supervision, B.H.P.; project administration, B.H.P., H.C. & B.H.; funding acquisition, B.H.P.

Funding

The author(s) received the following financial support for the research, authorship, and/ or publication of this article. This work was supported in part by National Institutes of Health grant K01HD060759 (BHP) and the Landon Center on Aging Endowment from the University of Kansas Medical Center.

Acknowledgments

This work was presented, in part, to the 27th Annual Association of Psychological Science (APS) meeting for presentation that took place between May 21 and 24, 2015 in New York, NY. A small subset of the control data was published in a previous paper, Psychology and Aging, Vol 25(3), Sep 2010, 691−696 http://dx.doi.org/10.1037/a0019455. The authors would like to thank Alicia MacKay for assistance with recruitment of stroke patients and data collection.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bizley, J.K.; Maddox, R.K.; Lee, A.K. Defining Auditory–Visual Objects: Behavioral Tests and Physiological Mechanisms. Trends Neurosci. 2016, 39, 74–85. [Google Scholar] [CrossRef] [PubMed]
  2. Stevenson, R.A.; Ghose, D.; Fister, J.K.; Sarko, D.K.; Altieri, N.A.; Nidiffer, A.R.; Kurela, L.R.; Siemann, J.K.; James, T.W.; Wallace, M.T. Identifying and Quantifying Multisensory Integration: A Tutorial Review. Brain Topogr. 2014, 27, 707–730. [Google Scholar] [CrossRef] [PubMed]
  3. Hanna-Pladdy, B.; Choi, H. Age–Related Deficits in Auditory Confrontation Naming. Psychol. Aging 2010, 25, 691–696. [Google Scholar] [CrossRef] [PubMed]
  4. Hickok, G.; Rogalsky, C.; Matchin, W.; Basilakos, A.; Cai, J.; Pillay, S.; Ferrill, M.; Mickelsen, S.; Anderson, S.W.; Love, T.; et al. Neural Networks Supporting Audiovisual Integration for Speech: A Large–Scale Lesion Study. Cortex 2018, 103, 360–371. [Google Scholar] [CrossRef] [PubMed]
  5. Moore, C.J.; Price, C.J. Three Distinct Ventral Occipitotemporal Regions for Reading and Object Naming. Neuroimage 1999, 10, 181–192. [Google Scholar] [CrossRef] [PubMed]
  6. Abrahams, S.; Goldstein, L.H.; Simmons, A.; Brammer, M.J.; Williams, S.C.; Giampietro, V.P.; Andrew, C.M.; Leigh, P.N. Functional Magnetic Resonance Imaging of Verbal Fluency and Confrontation Naming Using Compressed Image Acquisition to Permit Overt Responses. Hum. Brain Mapp. 2003, 20, 29–40. [Google Scholar] [CrossRef] [PubMed]
  7. Vigneau, M.; Beaucousin, V.; Herve, P.Y.; Duffau, H.; Crivello, F.; Houde, O.; Mazoyer, B.; Tzourio-Mazoyer, N. Meta–Analyzing Left Hemisphere Language Areas: Phonology, Semantics, and Sentence Processing. Neuroimage 2006, 30, 1414–1432. [Google Scholar] [CrossRef] [PubMed]
  8. Hamberger, M.J.; Seidel, W.T.; Mckhann, G.M.; Perrine, K.; Goodman, R.R. Brain Stimulation Reveals Critical Auditory Naming Cortex. Brain 2005, 128 Pt 11, 2742–2749. [Google Scholar] [CrossRef]
  9. Hamberger, M.J.; Goodman, R.R.; Perrine, K.; Tamny, T. Anatomic Dissociation of Auditory and Visual Naming in the Lateral Temporal Cortex. Neurology 2001, 56, 56–61. [Google Scholar] [CrossRef] [PubMed]
  10. Hamberger, M.J.; Habeck, C.G.; Pantazatos, S.P.; Williams, A.C.; Hirsch, J. Shared Space, Separate Processes: Neural Activation Patterns for Auditory Description and Visual Object Naming in Healthy Adults. Hum. Brain Mapp. 2014, 35, 2507–2520. [Google Scholar] [CrossRef] [PubMed]
  11. Friederici, A.D.; Gierhan, S.M. The Language Network. Curr.Opin. Neurobiol. 2013, 23, 250–254. [Google Scholar] [CrossRef] [PubMed]
  12. Chao, L.L.; Martin, A. Representation of Manipulable Man–Made Objects in the Dorsal Stream. Neuroimage 2000, 12, 478–484. [Google Scholar] [CrossRef] [PubMed]
  13. Adriani, M.; Maeder, P.; Meuli, R.; Thiran, A.B.; Frischknecht, R.; Villemure, J.G.; Mayer, J.; Annoni, J.M.; Bogousslavsky, J.; Fornari, E.; et al. Sound Recognition and Localization in Man: Specialized Cortical Networks and Effects of Acute Circumscribed Lesions. Exp. Brain Res. 2003, 153, 591–604. [Google Scholar] [CrossRef] [PubMed]
  14. Thompson, C.K.; Lukic, S.; King, M.C.; Mesulam, M.M.; Weintraub, S. Verb and Noun Deficits in Stroke–Induced and Primary Progressive Aphasia: The Northwestern Naming Battery. Aphasiology 2012, 26, 632–655. [Google Scholar] [CrossRef] [PubMed]
  15. Herbet, G.; Moritz-Gasser, S.; Boiseau, M.; Duvaux, S.; Cochereau, J.; Duffau, H. Converging Evidence for a Cortico–Subcortical Network Mediating Lexical Retrieval. Brain 2016, 136, 3007–3021. [Google Scholar] [CrossRef] [PubMed]
  16. Beauchamp, M.S.; Argall, B.D.; Bodurka, J.; Duyn, J.H.; Martin, A. Unraveling Multisensory Integration: Patchy Organization within Human Sts Multisensory Cortex. Nat. Neurosci. 2004, 7, 1190–1192. [Google Scholar] [CrossRef] [PubMed]
  17. Beauchamp, M.S.; Lee, K.E.; Argall, B.D.; Martin, A. Integration of Auditory and Visual Information About Objects in Superior Temporal Sulcus. Neuron 2004, 41, 809–823. [Google Scholar] [CrossRef]
  18. Venezia, J.H.; Vaden, K.I., Jr.; Rong, F.; Maddox, D.; Saberi, K.; Hickok, G. Auditory, Visual and Audiovisual Speech Processing Streams in Superior Temporal Sulcus. Front. Hum. Neurosci. 2017, 11, 174. [Google Scholar] [CrossRef] [PubMed]
  19. Lewis, J.W.; Wightman, F.L.; Brefczynski, J.A.; Phinney, R.E.; Binder, J.R.; DeYoe, E.A. Human Brain Regions Involved in Recognizing Environmental Sounds. Cereb. Cortex 2004, 14, 1008–1021. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Kertesz, A. The Western Aphasia Battery–Revised; Grune & Stratton: New York, NY, USA, 2007. [Google Scholar]
  21. Folstein, M.F.; Folstein, S.E.; McHugh, P.R. Mini–Mental State. A Practical Method for Grading the Cognitive State of Patients for the Clinician. J. Psychiatr. Res. 1975, 12, 189–198. [Google Scholar] [CrossRef]
  22. Oldfield, R.C. The Assessment and Analysis of Handedness: The Edinburgh Inventory. Neuropsychologia 1971, 9, 97–113. [Google Scholar] [CrossRef]
  23. Ventry, I.M.; Weinstein, B.E. The Hearing Handicap Inventory for the Elderly: A New Tool. Ear Hear. 1982, 3, 128–134. [Google Scholar] [CrossRef] [PubMed]
  24. Weinstein, B.E.; Ventry, I.M. Audiometric Correlates of the Hearing Handicap Inventory for the Elderly. J. Speech Hear. Disord. 1983, 48, 379–384. [Google Scholar] [CrossRef] [PubMed]
  25. Fabiani, M.; Kazmerski, V.A.; Cycowicz, Y.M.; Friedman, D. Naming Norms for Brief Environmental Sounds: Effects of Age and Dementia. Psychophysiology 1996, 33, 462–475. [Google Scholar] [CrossRef] [PubMed]
  26. Kaplan, E.; Goodglass, H.; Weintraub, S. Boston Naming Test; Lee & Febiger: Philadelphia, PA, USA, 1983. [Google Scholar]
  27. Spreen, O.; Strauss, E.A. A Compendium of Neuropsychological Tests: Administration, Norms and Commentary; Oxford University Press: New York, NY, USA, 1998. [Google Scholar]
  28. Shewan, C.; Kertesz, A. Reliability and Validity Characteristics of the Western Aphaisa Battery (Wab). J. Speech Hear. Disord. 1980, 45, 308–324. [Google Scholar] [CrossRef] [PubMed]
  29. Duncan, E.S.; Schmah, T.; Small, S.L. Performance Variability as a Predictor of Response to Aphasia Treatment. Neurorehabil. Neural Repair 2016, 30, 876–882. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Gialanella, B. Aphasia Assessment and Functional Outcome Prediction in Patients with Aphasia after Stroke. J. Neurol. 2011, 258, 343–349. [Google Scholar] [CrossRef] [PubMed]
  31. Schnider, A.; Benson, D.F.; Alexander, D.N.; Schnider-Klaus, A. Non–Verbal Environmental Sound Recognition after Unilateral Hemispheric Stroke. Brain 1994, 117 Pt 2, 281–287. [Google Scholar] [CrossRef]
  32. Dick, F.; Saygin, A.P.; Galati, G.; Pitzalis, S.; Bentrovato, S.; D’Amico, S.; Wilson, S.; Bates, E.; Pizzamiglio, L. What Is Involved and What Is Necessary for Complex Linguistic and Nonlinguistic Auditory Processing: Evidence from Functional Magnetic Resonance Imaging and Lesion Data. J. Cogn. Neurosci. 2007, 19, 799–816. [Google Scholar] [CrossRef] [PubMed]
  33. Saygin, A.P.; Dick, F.; Wilson, S.M.; Dronkers, N.F.; Bates, E. Neural Resources for Processing Language and Environmental Sounds: Evidence from Aphasia. Brain 2003, 126 Pt 4, 928–945. [Google Scholar] [CrossRef]
  34. Clarke, S.; Bellmann, A.; Meuli, R.A.; Assal, G.; Steck, A.J. Auditory Agnosia and Auditory Spatial Deficits Following Left Hemispheric Lesions: Evidence for Distinct Processing Pathways. Neuropsychologia 2000, 38, 797–807. [Google Scholar] [CrossRef]
  35. Clarke, S.; Bellmann, A.; de Ribaupierre, F.; Assal, G. Non–Verbal Auditory Recognition in Normal Subjects and Brain–Damaged Patients: Evidence for Parallel Processing. Neuropsychologia 1996, 34, 587–603. [Google Scholar] [CrossRef]
  36. Brauer, J.; Anwander, A.; Perani, D.; Friederici, A.D. Dorsal and Ventral Pathways in Language Development. Brain Lang. 2013, 127, 289–295. [Google Scholar] [CrossRef] [PubMed]
  37. Saur, D.; Kreher, B.W.; Schnell, S.; Kümmerer, D.; Kellmeyer, P.; Vry, M.S.; Umarova, R.; Musso, M.; Glauche, V.; Abel, S.; et al. Ventral and Dorsal Pathways for Language. Proc. Natl. Acad. Sci. USA 2008, 105, 18035–18040. [Google Scholar] [CrossRef] [PubMed]
  38. Goodale, M.A.; Milner, A.D. Separate Visual Pathways for Perception and Action. Trends Neurosci. 1992, 15, 20–25. [Google Scholar] [CrossRef]
  39. Beauchamp, M.S.; Lee, K.E.; Haxby, J.V.; Martin, A. Parallel Visual Motion Processing Streams for Manipulable Objects and Human Movements. Neuron 2002, 34, 149–159. [Google Scholar] [CrossRef] [Green Version]
  40. Chao, L.L.; Haxby, J.V.; Martin, A. Attribute–Based Neural Substrates in Temporal Cortex for Perceiving and Knowing About Objects. Nat. Neurosci. 1999, 2, 913–919. [Google Scholar] [CrossRef] [PubMed]
  41. Haxby, J.V.; Ungerleider, L.G.; Clark, V.P.; Schouten, J.L.; Hoffman, E.A.; Martin, A. The Effect of Face Inversion on Activity in Human Neural Systems for Face and Object Perception. Neuron 1999, 22, 189–199. [Google Scholar] [CrossRef]
  42. Hartman-Maeir, A.; Soroker, N.; Ring, H.; Katz, N. Awareness of Deficits in Stroke Rehabilitation. J. Rehabil. Med. 2002, 34, 158–164. [Google Scholar] [CrossRef] [PubMed]
  43. Bonakdarpour, B.; Beeson, P.M.; DeMarco, A.T.; Rapcsak, S.Z. Variability in Blood Oxygen Level Dependent (Bold) Signal in Patients with Stroke–Induced and Primary Progressive Aphasia. Neuroimage Clin. 2015, 8, 87–94. [Google Scholar] [CrossRef] [PubMed]
  44. Bonilha, L.; Rorden, C.; Fridriksson, J. Assessing the Clinical Effect of Residual Cortical Disconnection after Ischemic Strokes. Stroke 2014, 45, 988–993. [Google Scholar] [CrossRef] [PubMed]
  45. De Dieuleveult, A.L.; Siemonsma, P.C.; van Erp, J.B.; Brouwer, A.M. Effects of Aging in Multisensory Integration: A Systematic Review. Front. Aging Neurosci. 2017, 9, 80. [Google Scholar] [CrossRef] [PubMed]
  46. Diederich, A.; Colonius, H.; Schomburg, A. Assessing Age–Related Multisensory Enhancement with the Time–Window–of–Integration Model. Neuropsychologia 2008, 46, 2556–2562. [Google Scholar] [CrossRef] [PubMed]
  47. Cardenas, V.A.; Chao, L.L.; Studholme, C.; Yaffe, K.; Miller, B.L.; Madison, C.; Buckley, S.T.; Mungas, D.; Schuff, N.; Weine, M.W. Brain Atrophy Associated with Baseline and Longitudinal Measures of Cognition. Neurobiol. Aging 2011, 32, 572–580. [Google Scholar] [CrossRef] [PubMed]
  48. Stevenson, R.A.; Baum, S.H.; Krueger, J.; Newhouse, P.A.; Wallace, M.T. Links between Temporal Acuity and Multisensory Integration across Life Span. J. Exp. Psychol. Hum. Percept. Perform. 2018, 44, 106–116. [Google Scholar] [CrossRef] [PubMed]
  49. Kochunov, P.; Mangin, J.F.; Coyle, T.; Lancaster, J.; Thompson, P.; Riviere, D.; Cointepas, Y.; Regis, J.; Schlosser, A.; Royall, D.R.; et al. Age–Related Morphology Trends of Cortical Sulci. Hum. Brain Mapp. 2005, 26, 210–220. [Google Scholar] [CrossRef] [PubMed]
  50. Maass, A.; Landau, S.; Baker, S.L.; Horng, A.; Lockhart, S.N.; La Joie, R.; Rabinovici, G.D.; Jagust, W.J. Initiative Alzheimer’s Disease Neuroimaging. Comparison of Multiple Tau–Pet Measures as Biomarkers in Aging and Alzheimer’s Disease. Neuroimage 2017, 157, 448–463. [Google Scholar] [CrossRef] [PubMed]
  51. Lewis, J.W.; Brefczynski, J.A.; Phinney, R.E.; Janik, J.J.; DeYoe, E.A. Distinct Cortical Pathways for Processing Tool Versus Animal Sounds. J. Neurosci. 2005, 25, 5148–5158. [Google Scholar] [CrossRef] [PubMed]
  52. Lovelace, C.T.; Stein, B.E.; Wallace, M.T. An Irrelevant Light Enhances Auditory Detection in Humans: A Psychophysical Analysis of Multisensory Integration in Stimulus Detection. Brain Res. Cogn. Brain Res. 2003, 7, 447–453. [Google Scholar] [CrossRef]
  53. Stein, B.E.; Wallace, M.T. Comparisons of Cross–Modality Integration in Midbrain and Cortex. Prog. Brain Res. 1996, 112, 289–299. [Google Scholar] [PubMed]
  54. Diederich, A.; Colonius, H. Bimodal and Trimodal Multisensory Enhancement: Effects of Stimulus Onset and Intensity on Reaction Time. Percept. Psychophys. 2004, 66, 1388–1404. [Google Scholar] [CrossRef] [PubMed]
  55. Hanna-Pladdy, B.; Heilman, K.M.; Foundas, A.L. Cortical and Subcortical Contributions to Ideomotor Apraxia: Analysis of Task Demands and Error Types. Brain 2001, 124, 2513–2527. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Naming accuracy (% Correct) by group and condition. Box plots depicting median and quartiles, and subject numbers for values extending outside range. A: Auditory naming accuracy by LHD Lesion Group. B: Visual naming accuracy by LHD Lesion Group. C: Multisensory naming accuracy by LHD Lesion Group.
Figure 1. Naming accuracy (% Correct) by group and condition. Box plots depicting median and quartiles, and subject numbers for values extending outside range. A: Auditory naming accuracy by LHD Lesion Group. B: Visual naming accuracy by LHD Lesion Group. C: Multisensory naming accuracy by LHD Lesion Group.
Brainsci 08 00206 g001
Figure 2. Percent distribution of lesion location for LHD subgroups. Based on the total LHD subgroup (Cortical, Cortical−Subcortical, and Subcortical) with lesion involvement to specific anatomical regions.
Figure 2. Percent distribution of lesion location for LHD subgroups. Based on the total LHD subgroup (Cortical, Cortical−Subcortical, and Subcortical) with lesion involvement to specific anatomical regions.
Brainsci 08 00206 g002
Table 1. Characteristics of left hemisphere stroke patients.
Table 1. Characteristics of left hemisphere stroke patients.
#AgeSexStroke IntervalWAB AQAphasia TypeWAB
Naming
WAB
Aud. Comp.
Lesion Location
CO
175Male32.9594.4 9.5100Frontal and parietal lobes and deep white matter
277Male20.1478.7Anomic7.480Frontal, temporal, and parietal lobes
374Male16.0386.9Anomic9.899Middle and posterior frontal gyri, precentral, postcentral, and posterior parietal gyrus
450Male20.376.1Transcor. Motor8.396Superior temporal gyrus extending to the perirolandic region
550Male19.8794.9 9.899Anterior parietal lobe, extending anteriorly and inferiorly including Wernicke’s area
667Female5.8180.8Anomic9.091Temporal lobe
740Male59.5373.5Conduction8.783Middle cerebral artery territory
850Female9.4981.2Anomic8.387Insula, operculum, and punctate foci within the frontal and parietal lobes
977Female7.2975.1Transcor. Motor7.493Posterior parietal and superior temporal lobes
1078Male27.6636.8Broca’s4.067Posterior parietal lobe extending into the medial temporal lobe/subinsular cortex
1178Male28.4191.4Anomic9.497Insular cortex, superior frontal, and lateral parietal lobes
1247Male14.0984.4Anomic8.182Posterior parietal and occipital lobes
1352Male41.3995.6 9.8100Posterior frontal lobe
1485Male41.3958.5Wernicke’s5.065Occipital and posterolateral aspect of temproparietal lobes
1562Male18.3997.0 9.294Posterior territory of the left middle cerebral artery; lacunar infarct
1656Female17.4746.0Broca’s4.681Temporal and parietal occipital lobes
1744Male4.1396.1 10 Distal middle cerebral and left anterior cerebral artery; lateral ventricle in posterior horn
CO–SC
1852Female14.5892.1Anomic8.798Insula, subinsular white matter, perisylvian frontal & temporal lobe, and body of caudate
1956Male38.0185.0Anomic8.089Frontal and anterior superior temporal lobes, insula, and corona radiata
2038Female11.1797.2 9.8100Caudate head, internal capsule, putamen, perisylvian cortex, deep white matter, and parenchyma adjacent to the atria of lateral ventricle
2157Female21.1287.8Anomic9.596Posterior putamen extending to the corona radiata. Small focus of restricted diffusion noted in the periatrial region of temporal lobe
2252Male20.0784.5Anomic8.796Multiple small foci of restricted diffusion within MCA, ACA, and deep watershed territories. Acute basal ganglia infarct.
2344Male44.1294.3 9.794Caudate head, basal ganglia, posterior frontal, parietal, and temporal lobes
2455Female7.0395.5 9.696Frontal lobe and corona radiata
2574Female7.3278.2Anomic8.392Frontal lobe, and subinsular region, putamen, and external capsule
2664Female14.4891.2Anomic8.997Sylvian and perisylvian cortex, precentral and postcentral gyrus, posterior aspects of superior and middle frontal gyrus, caudate head and body, putamen, and external capsule
2769Male28.0583.1Anomic9.398Subinsular region, posterior temporal lobe, centrum semiovale, corona radiata, and parietal lobe
SC
2860Male38.3493.2Anomic9.6100Basal ganglia extending into corona radiata
2965Male25.4995.7 8.799Posterior limb of the internal capsule, inferior putamen, and anterior periatrial white matter
3064Female26.2596.0 9.2100Posterior limb of the internal capsule
3165Male145.8185.2Anomic9.698Corona radiata
3251Male7.8585.5Anomic9.296Caudate body, corona radiata, and putamen
3364Male12.7193.8 9.693Paramedian pons
3463Male26.9483.9Anomic 8.996Posterior limb of the internal capsule
3548Female31.2795.4 9.8100Thalamic lacunar infarct
3670Male12.099.0 9.5100Anterior limb of internal capsule
3770Male4.5981.4Anomic7.691Posterior limb of the internal capsule
3864Female39.5598.3 9.799Basal ganglia and centrum semiovale
3962Male33.897.2 9.894Pons
4063Male1.698.8 9.8100Medullary infarct
4147Female61.6399.6 9.8100Posterior putamen
4258Female19.5497.1 9.896Pontomedullary junction
Note. CO = Cortical Stroke location; CO–SC = Cortical–subcortical stroke location; SC = Subcortical stroke location; Stroke Interval = Interval in months from stroke to evaluation; WAB Quotient = Western Aphasia Battery Aphasia Quotient in percentage points; WAB AQ < 93.7 presence of aphasia; WAB Aud. Comp. = Western Aphasia Battery Auditory Comprehension in percentage points [20]; Aphasia = Aphasia Subtypes from the Western Aphasia Battery; WAB Naming = Score out of possible 10; MMSE = Mini-Mental State Examination; BNT = Boston Naming Test; FAS = Controlled Oral Word Association.
Table 2. Subject characteristics and screening measures (means).
Table 2. Subject characteristics and screening measures (means).
HC
(n = 118)
LHD
(n = 42)
FSig.
p
Age63.1 (9.5)60.4 (11.5)2.30.133
Education16.5 (2.5)13.4 (2.4)51.30.0001
MMSE29.2 (0.9)25.5 (4.8)66.40.0001
WAB AQ98.4 (2.3)86.6 (13.5)83.90.0001
Boston Naming51.8 (11.9))40.5 (14.1)25.70.0001
Phonemic Fluency49.5 (9.5)33.5 (13.3)70.70.0001
Semantic Fluency52.1 (9.5)40.7 (13.3)35.50.0001
Note. HC = Healthy Controls; LHD = Left Hemisphere Damage; Means (SDs); MMSE = Mini-Mental State Examination; WAB AQ = Western Aphasia Battery Aphasia Quotient (out of 100); Boston Naming = Boston Naming Test (T score); Phonemic (FAS); Semantic (animals) Fluency (T scores); F = F statistical value; Sig. p = Significance as a probability.
Table 3. Mean (SDs) of naming accuracy (%) and reaction time (seconds) by condition.
Table 3. Mean (SDs) of naming accuracy (%) and reaction time (seconds) by condition.
Accuracy ReactionTime
AuditoryVisualMultisensoryAuditoryVisualMultisensory
Controls75.294.6963.932.032.40
(n = 118)(13.1)(7.3)(5.1)(1.4)(0.52)(0.72)
LHD60.787.988.44.862.573.14
(n = 42)(24.5)(15.5)(15.5)(1.8)(0.73)(1.1)
Cortical56.581.780.55.193.043.56
(n = 17)(29.6)(21.5)(20.9)(1.9)(0.84)(1.5)
Cortical–Subcortical62.790.5905.522.493.17
(n = 10)(14.8)(8.3)(7.8)(1.8)(0.41)(0.91)
Subcortical64.393.396.34.032.072.65
(n = 15)(23.9)(6.9)(4.7)(1.3)(0.33)(0.57)
Naming accuracy based on percent correct. Reaction time in seconds for correctly named items.
Table 4. Mean differences for naming accuracy and correct reaction time across conditions.
Table 4. Mean differences for naming accuracy and correct reaction time across conditions.
V vs. AMS vs. AMS vs. V
ACCURACY
Controls18.3 ***19.7 ***1.3 *
LHD30.1 ***30.7 ***0.60
REACTION TIME
Controls−1.8 ***−1.5 ***0.32 ***
LHD−2.4 ***−1.8 ***0.63 ***
Based on estimated marginal means with age and education as covariates. Adjustment for multiple comparisons Bonferroni. Note: Bold values with * significance p < 0.05; ** significance p < 0.01; *** significance p < 0.001 denoting significance between groups. A = Condition; V = Visual; MS = Multisensory. Naming accuracy based on percent correct. Reaction time in seconds for correctly named items.
Table 5. Mean group differences for naming accuracy and correct reaction time across conditions.
Table 5. Mean group differences for naming accuracy and correct reaction time across conditions.
HC vs. COHC vs. CO–SCHC vs. SCSC vs. COSC vs. CO–SCCO vs. CO–SC
ACCURACY
Auditory21.2 ***19.5 **14.8 **6.54.7−1.8
Visual12.3 ***5.61.011.3 **4.6−6.7
Multisensory14.8 ***6.8−0.7315.5 ***7.6−7.9
REACTION TIME
Auditory−1.4 **−1.7 **−0.22−1.15−1.51−0.36
Visual−0.93 ***−0.41−0.04−0.97 ***−0.4540.52
Multisensory−1.2 ***−0.81 *−0.28−0.91 **−0.5320.38
Based on estimated marginal means with age and education as covariates. Adjustment for multiple comparisons Bonferroni. Note: * significance p < 0.05; ** significance p < 0.01; *** significance p < 0.001 denoting significance between groups in bold. SC = Subcortical; CO–SC = Cortical–Subcortical; CO = Cortical. Naming accuracy based on percent correct. Reaction time in seconds for correctly named items.
Table 6. Stepwise multiple regression results of predictors of naming accuracy.
Table 6. Stepwise multiple regression results of predictors of naming accuracy.
Dependent VariablePredictorBetaSEStand. BetatpR2Adj.
R2
MODEL 1
Auditory
Accuracy 1  Age−0.9250.245−0.431−3.780.0011   0.4501   0.435
2   Semantic Fluency1.130.4610.3092.460.0192   0.5482   0.524
3   WAIS−IV LNS−4.291.29−0.407−3.310.0023   0.5953   0.561
4   MMSE2.150.7340.4312.930.0064   0.6744   0.637
Visual
Accuracy1   MMSE3.180.3150.98610.070.00011   0.7491   0.743
2   WAIS−IV LNS−1.360.668−0.200−2.040.0482   0.7752   0.762
Multisensory
Accuracy1   WAB-AQ0.6190.1660.5373.730.0011   0.6801   0.672
2   MMSE1.150.4640.3592.490.0172   0.7262   0.711
MODEL 2
Auditory
Accuracy1   Age−1.470.233−0.690−5.780.00011   0.4611   0.447
2   Temporal Lobe−14.65.41−0.295−2.690.012   0.5482   0.524
Visual
Accuracy1   Age−0.6400.173−0.474−3.710.0011   0.2111   0.191
2   Temporal Lobe−12.94.01−0.411−3.220.0032   0.380 2   0.347
Multisensory
Accuracy1   Occipital Lobe−27.146.19−0.457−4.390.00011   0.2881   0.270
2   Age−0.5580.139−0.413−4.010.00012   0.4442   0.415
3   Temporal Lobe−12.813.27−0.408−3.920.00013   0.6073   0.576
Dependent Variables: Auditory Naming Accuracy (% Correct); Visual Naming Accuracy (% Correct); Multisensory Naming Accuracy (% Correct). Model 1 Predictor Variables: Age; Years of Education; Stroke−test Interval in Months; MMSE = Mini-Mental State Examination; WAB−AQ = Western Aphasia Battery Aphasia Quotient; FAS Verbal Fluency; Semantic Fluency (animals); WAIS−IV Letter Number Sequencing. Model 2 Predictor Variables: Age; Years of Education; Temporal Lobe, Parietal Lobe, Occipital Lobe, Frontal Lobe, Basal Ganglia, Thalamus, White Matter Involvement.

Share and Cite

MDPI and ACS Style

Hanna-Pladdy, B.; Choi, H.; Herman, B.; Haffey, S. Audiovisual Lexical Retrieval Deficits Following Left Hemisphere Stroke. Brain Sci. 2018, 8, 206. https://doi.org/10.3390/brainsci8120206

AMA Style

Hanna-Pladdy B, Choi H, Herman B, Haffey S. Audiovisual Lexical Retrieval Deficits Following Left Hemisphere Stroke. Brain Sciences. 2018; 8(12):206. https://doi.org/10.3390/brainsci8120206

Chicago/Turabian Style

Hanna-Pladdy, Brenda, Hyun Choi, Brian Herman, and Spenser Haffey. 2018. "Audiovisual Lexical Retrieval Deficits Following Left Hemisphere Stroke" Brain Sciences 8, no. 12: 206. https://doi.org/10.3390/brainsci8120206

APA Style

Hanna-Pladdy, B., Choi, H., Herman, B., & Haffey, S. (2018). Audiovisual Lexical Retrieval Deficits Following Left Hemisphere Stroke. Brain Sciences, 8(12), 206. https://doi.org/10.3390/brainsci8120206

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop