Strength of German accent under altered auditory feedback. (41/2052)

Borden's (1979, 1980) hypothesis that speakers with vulnerable speech systems rely more heavily on feedback monitoring than do speakers with less vulnerable systems was investigated. The second language (L2) of a speaker is vulnerable, in comparison with the native language, so alteration to feedback should have a detrimental effect on it, according to this hypothesis. Here, we specifically examined whether altered auditory feedback has an effect on accent strength when speakers speak L2. There were three stages in the experiment. First, 6 German speakers who were fluent in English (their L2) were recorded under six conditions--normal listening, amplified voice level, voice shifted in frequency, delayed auditory feedback, and slowed and accelerated speech rate conditions. Second, judges were trained to rate accent strength. Training was assessed by whether it was successful in separating German speakers speaking English from native English speakers, also speaking English. In the final stage, the judges ranked recordings of each speaker from the first stage as to increasing strength of German accent. The results show that accents were more pronounced under frequency-shifted and delayed auditory feedback conditions than under normal or amplified feedback conditions. Control tests were done to ensure that listeners were judging accent, rather than fluency changes caused by altered auditory feedback. The findings are discussed in terms of Borden's hypothesis and other accounts about why altered auditory feedback disrupts speech control.  (+info)

Cross-modal plasticity underpins language recovery after cochlear implantation. (42/2052)

Postlingually deaf subjects learn the meaning of sounds after cochlear implantation by forming new associations between sounds and their sources. Implants generate coarse frequency responses, preventing place-coding fine enough to discriminate sounds with similar temporal characteristics, e.g., buck/duck. This limitation imposes a dependency on visual cues, e.g., lipreading. We hypothesized that cross-modal facilitation results from engagement of the visual cortex by purely auditory tasks. In four functional neuroimaging experiments, we show recruitment of early visual cortex (V1/V2) when cochlear implant users listen to sounds with eyes closed. Activity in visual cortex evolved in a stimulus-specific manner as a function of time from implantation reflecting experience-dependent adaptations in the postimplant phase.  (+info)

Long-range synchrony in the gamma band: role in music perception. (43/2052)

Synchronization seems to be a central mechanism for neuronal information processing within and between multiple brain areas. Furthermore, synchronization in the gamma band has been shown to play an important role in higher cognitive functions, especially by binding the necessary spatial and temporal information in different cortical areas to build a coherent perception. Specific task-induced (evoked) gamma oscillations have often been taken as an indication of synchrony, but the presence of long-range synchrony cannot be inferred from spectral power in the gamma range. We studied the usefulness of a relatively new measure, called similarity index to detect asymmetric interdependency between two brain regions. Spontaneous EEG from two groups-musicians and non-musicians-were recorded during several states: listening to music, listening to text, and at rest (eyes closed and eyes open). While listening to music, degrees of the gamma band synchrony over distributed cortical areas were found to be significantly higher in musicians than non-musicians. Yet no differences between these two groups were found at resting conditions and while listening to a neutral text. In contrast to the degree of long-range synchrony, spectral power in the gamma band was higher in non-musicians. The degree of spatial synchrony, a measure of signal complexity based on eigen-decomposition method, was also significantly increased in musicians while listening to music. As compared with non-musicians, the finding of increased long-range synchrony in musicians independent of spectral power is interpreted as a manifestation of a more advanced musical memory of musicians in binding together several features of the intrinsic complexity of music in a dynamical way.  (+info)

Consequences of neural asynchrony: a case of auditory neuropathy. (44/2052)

The neural representation of sensory events depends upon neural synchrony. Auditory neuropathy, a disorder of stimulus-timing-related neural synchrony, provides a model for studying the role of synchrony in auditory perception. This article presents electrophysiological and behavioral data from a rare case of auditory neuropathy in a woman with normal hearing thresholds, making it possible to separate audibility from neuropathy. The experimental results, which encompass a wide range of auditory perceptual abilities and neurophysiologic responses to sound, provide new information linking neural synchrony with auditory perception. Findings illustrate that optimal eighth nerve and auditory brainstem synchrony do not appear to be essential for understanding speech in quiet listening situations. However, synchrony is critical for understanding speech in the presence of noise.  (+info)

Effects of electrode configuration and place of stimulation on speech perception with cochlear prostheses. (45/2052)

Recent research and clinical experience with cochlear implants suggest that subjects' speech recognition with monopolar or broad bipolar stimulation might be equal to or better than that obtained with narrow bipolar stimulation or other spatially restricted electrode configurations. Furthermore, subjects often prefer the monopolar configurations. The mechanisms underlying these effects are not clear. Two hypotheses are (a) that broader configurations excite more neurons resulting in a more detailed and robust neural representation of the signal and (b) that broader configurations achieve a better spatial distribution of the excited neurons. In this study we compared the effects of electrode configuration and the effects of longitudinal placement and spacing of the active electrodes on speech recognition in human subjects. We used experimental processor maps consisting of 11 active electrodes in a 22-electrode scala tympani array. Narrow bipolar (BP), wide bipolar (BP + 6), and monopolar (MP2) configurations were tested with various locations of active electrodes. We tested basal, centered, and apical locations (with adjacent active electrodes) and spatially distributed locations (with every other electrode active) with electrode configuration held constant. Ten postlingually deafened adult human subjects with Nucleus prostheses were tested using the SPEAK processing strategy. The effects of electrode configuration and longitudinal place of stimulation on recognition of CNC phonemes and words in quiet and CUNY sentences in noise (+10 dB S/N) were similar. Both independent variables had large effects on speech recognition and there were interactions between these variables. These results suggest that the effects of electrode configuration on speech recognition might be due, in part, to differences among the various configurations in the spatial location of stimulation. Correlations of subjective judgments of sound quality with speech-recognition ability were moderate, suggesting that the mechanisms contributing to subjective quality and speech-recognition ability do not completely overlap.  (+info)

Audio-visual perception of sinewave speech in an adult cochlear implant user: a case study. (46/2052)

OBJECTIVE: The purpose of this case study was to investigate multimodal perceptual coherence in speech perception in an exceptionally good postlingually deafened cochlear implant user. His ability to perceive sinewave replicas of spoken sentences, and the extent to which he integrated sensory information from multimodal sources was compared with a group of adult normal-hearing listeners to determine the contribution of natural auditory quality in the use of electrocochlear stimulation. DESIGN: The patient, "Mr. S," transcribed sinewave sentences of natural speech under audio-only (AO), visual-only (VO), and audio-visual (A+V) conditions. His performance was compared with the data collected from 25 normal-hearing adults. RESULTS: Although normal-hearing participants performed better than Mr. S for AO sentences (65% versus 53% syllables correct), Mr. S was superior for VO sentences (43% versus 18%). For A+V sentences, Mr. S's performance was comparable with the normal-hearing group (90% versus 86%). An estimate of the amount of visual enhancement, R, obtained from seeing the talker's face showed that Mr. S derived a larger gain from the additional visual information than the normal-hearing controls (78% versus 59%). CONCLUSIONS: The findings from this case study of an exceptionally good cochlear implant user suggest that he is perceiving the sinewave sentences on the basis of coherent variation from multimodal sensory inputs, and not on the basis of lipreading ability alone. Electrocochlear stimulation is evidently useful in multimodal contexts because it preserves dynamic speech-like variation, despite the absence of speech-like auditory qualities.  (+info)

Speech comprehension is correlated with temporal response patterns recorded from auditory cortex. (47/2052)

Speech comprehension depends on the integrity of both the spectral content and temporal envelope of the speech signal. Although neural processing underlying spectral analysis has been intensively studied, less is known about the processing of temporal information. Most of speech information conveyed by the temporal envelope is confined to frequencies below 16 Hz, frequencies that roughly match spontaneous and evoked modulation rates of primary auditory cortex neurons. To test the importance of cortical modulation rates for speech processing, we manipulated the frequency of the temporal envelope of speech sentences and tested the effect on both speech comprehension and cortical activity. Magnetoencephalographic signals from the auditory cortices of human subjects were recorded while they were performing a speech comprehension task. The test sentences used in this task were compressed in time. Speech comprehension was degraded when sentence stimuli were presented in more rapid (more compressed) forms. We found that the average comprehension level, at each compression, correlated with (i) the similarity between the frequencies of the temporal envelopes of the stimulus and the subject's cortical activity ("stimulus-cortex frequency-matching") and (ii) the phase-locking (PL) between the two temporal envelopes ("stimulus-cortex PL"). Of these two correlates, PL was significantly more indicative for single-trial success. Our results suggest that the match between the speech rate and the a priori modulation capacities of the auditory cortex is a prerequisite for comprehension. However, this is not sufficient: stimulus-cortex PL should be achieved during actual sentence presentation.  (+info)

Mice and humans perceive multiharmonic communication sounds in the same way. (48/2052)

Vowels and voiced consonants of human speech and most mammalian vocalizations consist of harmonically structured sounds. The frequency contours of formants in the sounds determine their spectral shape and timbre and carry, in human speech, important phonetic and prosodic information to be communicated. Steady-state partitions of vowels are discriminated and identified mainly on the basis of harmonics or formants having been resolved by the critical-band filters of the auditory system and then grouped together. Speech-analog processing and perception of vowel-like communication sounds in mammalian vocal repertoires has not been demonstrated so far. Here, we synthesize 11 call models and a tape loop with natural wriggling calls of mouse pups and show that house mice perceive this communication call in the same way as we perceive speech vowels: they need the presence of a minimum number of formants (three formants-in this case, at 3.8 + 7.6 + 11.4 kHz), they resolve formants by the critical-band mechanism, group formants together for call identification, perceive the formant structure rather continuously, may detect the missing fundamental of a harmonic complex, and all of these occur in a natural communication situation without any training or behavioral constraints. Thus, wriggling-call perception in mice is comparable with unconditioned vowel discrimination and perception in prelinguistic human infants and points to evolutionary old rules of handling speech sounds in the human auditory system up to the perceptual level.  (+info)