Evidence from functional magnetic resonance imaging of crossmodal binding in the human heteromodal cortex. (33/1550)

BACKGROUND: Integrating information from the different senses markedly enhances the detection and identification of external stimuli. Compared with unimodal inputs, semantically and/or spatially congruent multisensory cues speed discrimination and improve reaction times. Discordant inputs have the opposite effect, reducing performance and slowing responses. These behavioural features of crossmodal processing appear to have parallels in the response properties of multisensory cells in the superior colliculi and cerebral cortex of non-human mammals. Although spatially concordant multisensory inputs can produce a dramatic, often multiplicative, increase in cellular activity, spatially disparate cues tend to induce a profound response depression. RESULTS: Using functional magnetic resonance imaging (fMRI), we investigated whether similar indices of crossmodal integration are detectable in human cerebral cortex, and for the synthesis of complex inputs relating to stimulus identity. Ten human subjects were exposed to varying epochs of semantically congruent and incongruent audio-visual speech and to each modality in isolation. Brain activations to matched and mismatched audio-visual inputs were contrasted with the combined response to both unimodal conditions. This strategy identified an area of heteromodal cortex in the left superior temporal sulcus that exhibited significant supra-additive response enhancement to matched audio-visual inputs and a corresponding sub-additive response to mismatched inputs. CONCLUSIONS: The data provide fMRI evidence of crossmodal binding by convergence in the human heteromodal cortex. They further suggest that response enhancement and depression may be a general property of multisensory integration operating at different levels of the neuroaxis and irrespective of the purpose for which sensory inputs are combined.  (+info)

The implementation of speech recognition in an electronic radiology practice. (34/1550)

For both efficiency and economic reasons, our practice (200,000 examinations) has converted all remote dictation to speech recognition transcription (PowerScribe, L & H, Burlington, MA). The design criteria included complete automation to the existing radiology information system (RIS), with full RIS capabilities immediately available following dictation. All dictations for computed tomography, magnetic resonance imaging, ultrasound, and nuclear medicine were converted from remote transcription to speech recognition over a 2-week period (following a 4-week installation phase and 8 days of training). The average turnaround time for these reports decreased from approximately 2 hours to less than 1 minute. Reports are then sent to the institutional Electronic Medical Record and are available throughout all facilities in a nominal 2 minutes. Speech recognition rates were surprisingly high, although certain phrases caused consistent difficulties and certain staff required retraining. This presents our analysis of both successful and problematic areas during our design and implementation, as well as statistical performance analyses.  (+info)

Human temporal lobe activation by speech and nonspeech sounds. (35/1550)

Functional organization of the lateral temporal cortex in humans is not well understood. We recorded blood oxygenation signals from the temporal lobes of normal volunteers using functional magnetic resonance imaging during stimulation with unstructured noise, frequency-modulated (FM) tones, reversed speech, pseudowords and words. For all conditions, subjects performed a material-nonspecific detection response when a train of stimuli began or ceased. Dorsal areas surrounding Heschl's gyrus bilaterally, particularly the planum temporale and dorsolateral superior temporal gyrus, were more strongly activated by FM tones than by noise, suggesting a role in processing simple temporally encoded auditory information. Distinct from these dorsolateral areas, regions centered in the superior temporal sulcus bilaterally were more activated by speech stimuli than by FM tones. Identical results were obtained in this region using words, pseudowords and reversed speech, suggesting that the speech-tones activation difference is due to acoustic rather than linguistic factors. In contrast, previous comparisons between word and nonword speech sounds showed left-lateralized activation differences in more ventral temporal and temporoparietal regions that are likely involved in processing lexical-semantic or syntactic information associated with words. The results indicate functional subdivision of the human lateral temporal cortex and provide a preliminary framework for understanding the cortical processing of speech sounds.  (+info)

Brain-wave representation of words by superposition of a few sine waves. (36/1550)

Data from three previous experiments were analyzed to test the hypothesis that brain waves of spoken or written words can be represented by the superposition of a few sine waves. First, we averaged the data over trials and a set of subjects, and, in one case, over experimental conditions as well. Next we applied a Fourier transform to the averaged data and selected those frequencies with high energy, in no case more than nine in number. The superpositions of these selected sine waves were taken as prototypes. The averaged unfiltered data were the test samples. The prototypes were used to classify the test samples according to a least-squares criterion of fit. The results were seven of seven correct classifications for the first experiment using only three frequencies, six of eight for the second experiment using nine frequencies, and eight of eight for the third experiment using five frequencies.  (+info)

Comparative evaluation of three continuous speech recognition software packages in the generation of medical reports. (37/1550)

OBJECTIVE: To compare out-of-box performance of three commercially available continuous speech recognition software packages: IBM ViaVoice 98 with General Medicine Vocabulary; Dragon Systems NaturallySpeaking Medical Suite, version 3.0; and L&H Voice Xpress for Medicine, General Medicine Edition, version 1.2. DESIGN: Twelve physicians completed minimal training with each software package and then dictated a medical progress note and discharge summary drawn from actual records. MEASUREMENTS: Errors in recognition of medical vocabulary, medical abbreviations, and general English vocabulary were compared across packages using a rigorous, standardized approach to scoring. RESULTS: The IBM software was found to have the lowest mean error rate for vocabulary recognition (7.0 to 9.1 percent) followed by the L&H software (13.4 to 15.1 percent) and then Dragon software (14.1 to 15.2 percent). The IBM software was found to perform better than both the Dragon and the L&H software in the recognition of general English vocabulary and medical abbreviations. CONCLUSION: This study is one of a few attempts at a robust evaluation of the performance of continuous speech recognition software. Results of this study suggest that with minimal training, the IBM software outperforms the other products in the domain of general medicine; however, results may vary with domain. Additional training is likely to improve the out-of-box performance of all three products. Although the IBM software was found to have the lowest overall error rate, successive generations of speech recognition software are likely to surpass the accuracy rates found in this investigation.  (+info)

Modeling and perception of 'gesture reduction'. (38/1550)

The phenomenon of vowel reduction is investigated by modeling 'gesture reduction' with the use of the Distinctive Region Model (DRM). First, a definition is proposed for the term gesture, i.e. an acoustically efficient command aimed at deforming, in the time domain, the area function of the vocal tract. Second, tests are reported on the perception of vowel-to-vowel transitions obtained with reduced gestures. These tests show that a dual representation of formant transitions is required to explain the reduction phenomenon: the trajectory in the F(1)-F(2) plane and the time course of the formant changes. The results also suggest that time-domain integration of the trajectories constitutes an integral part of the auditory processing of transitions. Perceptual results are also discussed in terms of the acoustic traces of DRM gestures.  (+info)

Brain correlates of stuttering and syllable production. A PET performance-correlation analysis. (39/1550)

To distinguish the neural systems of normal speech from those of stuttering, PET images of brain blood flow were probed (correlated voxel-wise) with per-trial speech-behaviour scores obtained during PET imaging. Two cohorts were studied: 10 right-handed men who stuttered and 10 right-handed, age- and sex-matched non-stuttering controls. Ninety PET blood flow images were obtained in each cohort (nine per subject as three trials of each of three conditions) from which r-value statistical parametric images (SPI inverted question markr inverted question mark) were computed. Brain correlates of stutter rate and syllable rate showed striking differences in both laterality and sign (i.e. positive or negative correlations). Stutter-rate correlates, both positive and negative, were strongly lateralized to the right cerebral and left cerebellar hemispheres. Syllable correlates in both cohorts were bilateral, with a bias towards the left cerebral and right cerebellar hemispheres, in keeping with the left-cerebral dominance for language and motor skills typical of right-handed subjects. For both stutters and syllables, the brain regions that were correlated positively were those of speech production: the mouth representation in the primary motor cortex; the supplementary motor area; the inferior lateral premotor cortex (Broca's area); the anterior insula; and the cerebellum. The principal difference between syllable-rate and stutter-rate positive correlates was hemispheric laterality. A notable exception to this rule was that cerebellar positive correlates for syllable rate were far more extensive in the stuttering cohort than in the control cohort, which suggests a specific role for the cerebellum in enabling fluent utterances in persons who stutter. Stutters were negatively correlated with right-cerebral regions (superior and middle temporal gyrus) associated with auditory perception and processing, regions which were positively correlated with syllables in both the stuttering and control cohorts. These findings support long-held theories that the brain correlates of stuttering are the speech-motor regions of the non-dominant (right) cerebral hemisphere, and extend this theory to include the non-dominant (left) cerebellar hemisphere. The present findings also indicate a specific role of the cerebellum in the fluent utterances of persons who stutter. Support is also offered for theories that implicate auditory processing problems in stuttering.  (+info)

Neuropsychological consequences of chronic bilateral stimulation of the subthalamic nucleus in Parkinson's disease. (40/1550)

The aim of this study was to examine possible neuropsychological changes in patients with advanced idiopathic Parkinson's disease treated with bilateral deep brain stimulation (DBS) of the subthalamic nucleus (STN). Eleven patients (age = 67 +/- 8 years, years with Parkinson's disease = 15 +/- 3, verbal IQ = 114 +/- 12) were evaluated (in their best 'on state') with tests assessing processes reliant on the functional integrity of frontal striatal circuitry, prior to the procedure (n = 11), at 3-6 months (n = 11) and at 9-12 months (n =10) post-operatively. Six of these patients were older than 69 years. Despite clinical motor benefits at 3-6 months post-operative, significant declines were noted in working memory, speed of mental processing, bimanual motor speed and co-ordination, set switching, phonemic fluency, long-term consolidation of verbal material and the encoding of visuospatial material. Declines were more consistently observed in patients who were older than 69 years, leading to a mental state comparable with progressive supranuclear palsy. 'Frontal' behavioural dyscontrol without the benefit of insight was also reported by half (three of six) of the caregivers of the elderly subgroup. At 9-12 months postoperative, only learning based on multiple trials had recovered. Tasks reliant on the integrity of frontal striatal circuitry either did not recover or gradually worsened over time. Bilateral STN DBS can have a negative impact on various aspects of frontal executive functioning, especially in patients older than 69 years. Future studies will evaluate a larger group of patients and examine the possible reversibility of these effects by turning the DBS off.  (+info)