首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
Locating sounds in realistic scenes is challenging because of distracting echoes and coarse spatial acoustic estimates. Fortunately, listeners can improve performance through several compensatory mechanisms. For instance, their brains perceptually suppress short latency (1-10 ms) echoes by constructing a representation of the acoustic environment in a process called the precedence effect. This remarkable ability depends on the spatial and spectral relationship between the first or precedent sound wave and subsequent echoes. In addition to using acoustics alone, the brain also improves sound localization by incorporating spatially precise visual information. Specifically, vision refines auditory spatial receptive fields and can capture auditory perception such that sound is localized toward a coincident visual stimulus. Although visual cues and the precedence effect are each known to improve performance independently, it is not clear whether these mechanisms can cooperate or interfere with each other. Here we demonstrate that echo suppression is enhanced when visual information spatially and temporally coincides with the precedent wave. Conversely, echo suppression is inhibited when vision coincides with the echo. These data show that echo suppression is a fundamentally multisensory process in everyday environments, where vision modulates even this largely automatic auditory mechanism to organize a coherent spatial experience.  相似文献   

2.

Background

The duration of sounds can affect the perceived duration of co-occurring visual stimuli. However, it is unclear whether this is limited to amodal processes of duration perception or affects other non-temporal qualities of visual perception.

Methodology/Principal Findings

Here, we tested the hypothesis that visual sensitivity - rather than only the perceived duration of visual stimuli - can be affected by the duration of co-occurring sounds. We found that visual detection sensitivity (d’) for unimodal stimuli was higher for stimuli of longer duration. Crucially, in a cross-modal condition, we replicated previous unimodal findings, observing that visual sensitivity was shaped by the duration of co-occurring sounds. When short visual stimuli (∼24 ms) were accompanied by sounds of matching duration, visual sensitivity was decreased relative to the unimodal visual condition. However, when the same visual stimuli were accompanied by longer auditory stimuli (∼60–96 ms), visual sensitivity was increased relative to the performance for ∼24 ms auditory stimuli. Across participants, this sensitivity enhancement was observed within a critical time window of ∼60–96 ms. Moreover, the amplitude of this effect correlated with visual sensitivity enhancement found for longer lasting visual stimuli across participants.

Conclusions/Significance

Our findings show that the duration of co-occurring sounds affects visual perception; it changes visual sensitivity in a similar way as altering the (actual) duration of the visual stimuli does.  相似文献   

3.
We simultaneously perturbed visual, vestibular and proprioceptive modalities to understand how sensory feedback is re-weighted so that overall feedback remains suited to stabilizing upright stance. Ten healthy young subjects received an 80 Hz vibratory stimulus to their bilateral Achilles tendons (stimulus turns on-off at 0.28 Hz), a ±1 mA binaural monopolar galvanic vestibular stimulus at 0.36 Hz, and a visual stimulus at 0.2 Hz during standing. The visual stimulus was presented at different amplitudes (0.2, 0.8 deg rotation about ankle axis) to measure: the change in gain (weighting) to vision, an intramodal effect; and a change in gain to vibration and galvanic vestibular stimulation, both intermodal effects. The results showed a clear intramodal visual effect, indicating a de-emphasis on vision when the amplitude of visual stimulus increased. At the same time, an intermodal visual-proprioceptive reweighting effect was observed with the addition of vibration, which is thought to change proprioceptive inputs at the ankles, forcing the nervous system to rely more on vision and vestibular modalities. Similar intermodal effects for visual-vestibular reweighting were observed, suggesting that vestibular information is not a “fixed” reference, but is dynamically adjusted in the sensor fusion process. This is the first time, to our knowledge, that the interplay between the three primary modalities for postural control has been clearly delineated, illustrating a central process that fuses these modalities for accurate estimates of self-motion.  相似文献   

4.
The human brain processes different aspects of the surrounding environment through multiple sensory modalities, and each modality can be subdivided into multiple attribute-specific channels. When the brain rebinds sensory content information (‘what’) across different channels, temporal coincidence (‘when’) along with spatial coincidence (‘where’) provides a critical clue. It however remains unknown whether neural mechanisms for binding synchronous attributes are specific to each attribute combination, or universal and central. In human psychophysical experiments, we examined how combinations of visual, auditory and tactile attributes affect the temporal frequency limit of synchrony-based binding. The results indicated that the upper limits of cross-attribute binding were lower than those of within-attribute binding, and surprisingly similar for any combination of visual, auditory and tactile attributes (2–3 Hz). They are unlikely to be the limits for judging synchrony, since the temporal limit of a cross-attribute synchrony judgement was higher and varied with the modality combination (4–9 Hz). These findings suggest that cross-attribute temporal binding is mediated by a slow central process that combines separately processed ‘what’ and ‘when’ properties of a single event. While the synchrony performance reflects temporal bottlenecks existing in ‘when’ processing, the binding performance reflects the central temporal limit of integrating ‘when’ and ‘what’ properties.  相似文献   

5.
The occipital alpha rhythm (~10 Hz) is the most prominent electrophysiological activity in the awake human brain, yet its functional role and relation to visual perception are little understood. Transient stimuli normally elicit a short series of positive and negative deflections lasting between 300 and 500 ms: the visual-evoked potential (VEP). Alpha oscillations, on the other hand, are generally suppressed by transient visual input; they only augment in response to periodic ("steady-state") inputs around 10 Hz. Here, we applied reverse-correlation techniques to the visual presentation of random, nonperiodic dynamic stimulation sequences and found that the brain response to each stimulus transient was not merely a short-lived VEP but also included a strong ~10 Hz oscillation that lasted for more than 1 s. In other words, the alpha rhythm implements an "echo" or reverberation of the input sequence. These echoes are correlated in magnitude and frequency with the observer's occipital alpha rhythm, are enhanced by visual attention, and can be rendered perceptually apparent in the form of ~10 Hz flicker. These findings suggest a role for the alpha rhythm in the maintenance of sensory representations over time.  相似文献   

6.
Although it is well known that attention to a visual or auditory stimulus can enhance its perception, less is known concerning the effects of attention on the perception of natural tactile stimuli. The present study was conducted to examine the magnitude of the effect of cross-modal manipulations of attention in human subjects on the detection of weak, low-frequency vibrotactile stimuli delivered to the glabrous skin of the finger pad of the right index finger via an Optacon. Three suprathreshold vibrotactile arrays (40 Hz), varying in the number of activated pegs and hence the area of skin stimulated, were used. Subjects were trained to detect the occurrence of vibrotactile or visual stimuli and to respond by pressing a foot pedal as quickly as possible thereafter. Two instructional lights were used to cue the subjects as to which stimulus modality they should attend, in three experimental conditions. In the first cue condition, the forthcoming stimulus modality was indicated by the illumination of its associated light. In the second cue condition, both instructional lights were illuminated, and the subjects were asked to divide their attention equally between the two modalities. In the third cue condition, the stimulus modality was falsely indicated by the illumination of the cue not associated with the stimulus to be presented. Reaction times (RTs) were calculated for each trial. For each modality, tactile and visual, the RTs varied significantly with the cue condition, with the mean RT changing in a graded manner across the experimental conditions (being shortest for the correctly cued condition, intermediate for the neutrally cued condition, and longest for the incorrectly cued condition.(ABSTRACT TRUNCATED AT 250 WORDS)  相似文献   

7.
Unexpected physical increases in the intensity of a frequently occurring “standard” auditory stimulus are experienced as obtrusive. This could either be because of a physical change, the increase in intensity of the “deviant” stimulus, or a psychological change, the violation of the expectancy for the occurrence of the lower intensity standard stimulus. Two experiments were run in which event-related potentials (ERPs) were recorded to determine whether “psychological” increments (violation of an expectancy for a lower intensity) would be processed differently than psychological decrements (violation of an expectancy for a higher intensity). Event-related potentials (ERPs) were recorded while subjects were presented with auditory tones that alternated between low and high intensity. The subjects ignored the auditory stimuli while watching a video. Deviants were created by repeating the same stimulus. In the first experiment, pairs of stimuli alternating in intensity, were presented in separate increment (H-L…H-L…H-H…H-L, in which H = 80 dB SPL and L = 60 dB SPL) and decrement conditions (L-H…L-H…L-L… L-H, in which H = 90 dB SPL and L = 80 dB SPL). The paradigm employed in the second experiment consisted of an alternating intensity pattern (H-L-H-L-H-H-H-L) or (H-L-H-L-L-L-H-L). Importantly, the stimulus prior to the deviant (the standard) and the actual deviants in both increment and decrement conditions in both experiments were physically identical (80 dB SPL tones). The repetition of the lower intensity tone therefore acted as a psychological rather than a physical decrement (a higher intensity tone was expected) while the repetition of the higher intensity tone acted as a psychological increment (a lower intensity tone was expected). The psychological increments in both experiments elicited a larger amplitude mismatch negativity (MMN) than the decrements. Thus, regardless of whether an acoustic change signals a physical increase in intensity or violates an expected decrease in intensity, a large MMN will be elicited.  相似文献   

8.
ERPs to sequences of standard and deviant sinusoidal 100 msec tone pips, high-contrast sinusoidal gratings and to their simultaneously presented combinations were recorded. Mismatch negativity (MMN), an ERP component elicited by deviant stimuli, was estimated for the different stimulus sequences in order to find out whether it reflects modality-specific processes or non-specific attentive phenomena. In addition to the auditory modality, we studied whether the mismatch response could be evoked by a deviant visual stimulus in a visual sequence or by a deviant stimulus in either modality. The results show that only auditory stimuli produced the mismatch response, suggesting that MMN is not a manifestation of a general attentional mechanism but is probably specific to the auditory modality.  相似文献   

9.
Does our perceptual awareness consist of a continuous stream, or a discrete sequence of perceptual cycles, possibly associated with the rhythmic structure of brain activity? This has been a long-standing question in neuroscience. We review recent psychophysical and electrophysiological studies indicating that part of our visual awareness proceeds in approximately 7–13 Hz cycles rather than continuously. On the other hand, experimental attempts at applying similar tools to demonstrate the discreteness of auditory awareness have been largely unsuccessful. We argue and demonstrate experimentally that visual and auditory perception are not equally affected by temporal subsampling of their respective input streams: video sequences remain intelligible at sampling rates of two to three frames per second, whereas audio inputs lose their fine temporal structure, and thus all significance, below 20–30 samples per second. This does not mean, however, that our auditory perception must proceed continuously. Instead, we propose that audition could still involve perceptual cycles, but the periodic sampling should happen only after the stage of auditory feature extraction. In addition, although visual perceptual cycles can follow one another at a spontaneous pace largely independent of the visual input, auditory cycles may need to sample the input stream more flexibly, by adapting to the temporal structure of the auditory inputs.  相似文献   

10.
The auditory system creates a neuronal representation of the acoustic world based on spectral and temporal cues present at the listener''s ears, including cues that potentially signal the locations of sounds. Discrimination of concurrent sounds from multiple sources is especially challenging. The current study is part of an effort to better understand the neuronal mechanisms governing this process, which has been termed “auditory scene analysis”. In particular, we are interested in spatial release from masking by which spatial cues can segregate signals from other competing sounds, thereby overcoming the tendency of overlapping spectra and/or common temporal envelopes to fuse signals with maskers. We studied detection of pulsed tones in free-field conditions in the presence of concurrent multi-tone non-speech maskers. In “energetic” masking conditions, in which the frequencies of maskers fell within the ±1/3-octave band containing the signal, spatial release from masking at low frequencies (∼600 Hz) was found to be about 10 dB. In contrast, negligible spatial release from energetic masking was seen at high frequencies (∼4000 Hz). We observed robust spatial release from masking in broadband “informational” masking conditions, in which listeners could confuse signal with masker even though there was no spectral overlap. Substantial spatial release was observed in conditions in which the onsets of the signal and all masker components were synchronized, and spatial release was even greater under asynchronous conditions. Spatial cues limited to high frequencies (>1500 Hz), which could have included interaural level differences and the better-ear effect, produced only limited improvement in signal detection. Substantially greater improvement was seen for low-frequency sounds, for which interaural time differences are the dominant spatial cue.  相似文献   

11.
The modality of a stimulus and its intermittency affect time estimation. The present experiment explores the effect of a combination of modality and intermittency, and its implications for internal clock explanations. Twenty-four participants were tested on a temporal bisection task with durations of 200-800 ms. Durations were signaled by visual steady stimuli, auditory steady stimuli, visual flickering stimuli, and auditory clicks. Psychophysical functions and bisection points indicated that the durations of visual steady stimuli were classified as shorter and more variable than the durations signaled by the auditory stimuli (steady and clicks), and that the durations of the visual flickering stimuli were classified as longer than the durations signaled by the auditory stimuli (steady and clicks). An interpretation of the results is that there are different speeds for the internal clock, which are mediated by the perceptual features of the stimuli timed, such as differences in time of processing.  相似文献   

12.
To obtain a coherent perception of the world, our senses need to be in alignment. When we encounter misaligned cues from two sensory modalities, the brain must infer which cue is faulty and recalibrate the corresponding sense. We examined whether and how the brain uses cue reliability to identify the miscalibrated sense by measuring the audiovisual ventriloquism aftereffect for stimuli of varying visual reliability. To adjust for modality-specific biases, visual stimulus locations were chosen based on perceived alignment with auditory stimulus locations for each participant. During an audiovisual recalibration phase, participants were presented with bimodal stimuli with a fixed perceptual spatial discrepancy; they localized one modality, cued after stimulus presentation. Unimodal auditory and visual localization was measured before and after the audiovisual recalibration phase. We compared participants’ behavior to the predictions of three models of recalibration: (a) Reliability-based: each modality is recalibrated based on its relative reliability—less reliable cues are recalibrated more; (b) Fixed-ratio: the degree of recalibration for each modality is fixed; (c) Causal-inference: recalibration is directly determined by the discrepancy between a cue and its estimate, which in turn depends on the reliability of both cues, and inference about how likely the two cues derive from a common source. Vision was hardly recalibrated by audition. Auditory recalibration by vision changed idiosyncratically as visual reliability decreased: the extent of auditory recalibration either decreased monotonically, peaked at medium visual reliability, or increased monotonically. The latter two patterns cannot be explained by either the reliability-based or fixed-ratio models. Only the causal-inference model of recalibration captures the idiosyncratic influences of cue reliability on recalibration. We conclude that cue reliability, causal inference, and modality-specific biases guide cross-modal recalibration indirectly by determining the perception of audiovisual stimuli.  相似文献   

13.
Humans may be faster to avoid negative words than to approach negative words, and faster to approach positive words than to avoid positive words. That is an example of affective stimulus–response (S–R) compatibility. The present study identified the reference valence effects of affective stimulus–response (S–R) compatibility when auditory stimulus materials are used. The researchers explored the reference valence effects of affective S–R compatibility using a mixed-design experiment based on visual words, visual pictures and audition. The study computed the average compatibility effect size. A t-test based on visual pictures showed that the compatibility effect size was significantly different from zero, t (22) = 2.43, p<.05 (M = 485 ms). Smaller compatibility effects existed when switching the presentation mode from visual stimuli to auditory stimuli. This study serves as an important reference for the auditory reference valence effects of affective S–R compatibility.  相似文献   

14.
Integrating information across sensory domains to construct a unified representation of multi-sensory signals is a fundamental characteristic of perception in ecological contexts. One provocative hypothesis deriving from neurophysiology suggests that there exists early and direct cross-modal phase modulation. We provide evidence, based on magnetoencephalography (MEG) recordings from participants viewing audiovisual movies, that low-frequency neuronal information lies at the basis of the synergistic coordination of information across auditory and visual streams. In particular, the phase of the 2–7 Hz delta and theta band responses carries robust (in single trials) and usable information (for parsing the temporal structure) about stimulus dynamics in both sensory modalities concurrently. These experiments are the first to show in humans that a particular cortical mechanism, delta-theta phase modulation across early sensory areas, plays an important “active” role in continuously tracking naturalistic audio-visual streams, carrying dynamic multi-sensory information, and reflecting cross-sensory interaction in real time.  相似文献   

15.
This article aims to investigate whether auditory stimuli in the horizontal plane, particularly originating from behind the participant, affect audiovisual integration by using behavioral and event-related potential (ERP) measurements. In this study, visual stimuli were presented directly in front of the participants, auditory stimuli were presented at one location in an equidistant horizontal plane at the front (0°, the fixation point), right (90°), back (180°), or left (270°) of the participants, and audiovisual stimuli that include both visual stimuli and auditory stimuli originating from one of the four locations were simultaneously presented. These stimuli were presented randomly with equal probability; during this time, participants were asked to attend to the visual stimulus and respond promptly only to visual target stimuli (a unimodal visual target stimulus and the visual target of the audiovisual stimulus). A significant facilitation of reaction times and hit rates was obtained following audiovisual stimulation, irrespective of whether the auditory stimuli were presented in the front or back of the participant. However, no significant interactions were found between visual stimuli and auditory stimuli from the right or left. Two main ERP components related to audiovisual integration were found: first, auditory stimuli from the front location produced an ERP reaction over the right temporal area and right occipital area at approximately 160–200 milliseconds; second, auditory stimuli from the back produced a reaction over the parietal and occipital areas at approximately 360–400 milliseconds. Our results confirmed that audiovisual integration was also elicited, even though auditory stimuli were presented behind the participant, but no integration occurred when auditory stimuli were presented in the right or left spaces, suggesting that the human brain might be particularly sensitive to information received from behind than both sides.  相似文献   

16.
A systems theoretical approach was used to compare possible functional roles of theta (4--7 Hz) and alpha (8--15 Hz) response components of brain evoked potentials. These response components were described earlier by Ba\c sar (1980). We recorded EEG and evoked potentials (EPs) from occipital scalp locations in 11 subjects. We used auditory and visual stimuli as inadequate and adequate stimuli, respectively (``cross-modality' measurements). The combined EEG-EP epochs were analysed in frequency domain with fast Fourier transform and adaptive digital filters. Alpha (8--15 Hz) response components turned out to be dependent on whether the stimulus was adequate or not (median amplitude with inadequate vs. adequate stimulation: vs. ). Theta (4--7 Hz) response components were less dependent on stimulus modality (inadequate vs. adequate stimulation: vs. ). In EP recordings the occipital alpha response almost disappeared in the first 250 ms following auditory stimulation. Comparable behaviour was observed in similar experiments with recordings from the cat visual cortex (area 17) and with occipital magnetoencephalographic recordings. Taking into account the above-mentioned previous reports on intracranial recordings in primary sensory areas of the cat brain and preliminary results of magnetoencephalographic measurements, we propose the following hypothesis: alpha responses in a time window of about 250 ms after stimulation might predominantly reflect primary sensory processing whereas the theta responses in the first 250 ms after stimulation might be more involved in supra-modality -- or cross-modality -- associative-cognitive processing. Received: 25 February 1994 / Accepted in revised form: 5 August 1994  相似文献   

17.
Increased force variability constitutes a hallmark of arm disabilities following stroke. Force variability is related to the modulation of force below 1 Hz in healthy young and older adults. However, whether the increased force variability observed post stroke is related to the modulation of force below 1 Hz remains unknown. Thus, the purpose of this study was to compare force modulation below 1 Hz in chronic stroke and age-matched healthy individuals. Both stroke and control individuals (N = 26) performed an isometric grip task to submaximal force levels. Coefficient of variation quantified force variability, and power spectrum density of force quantified force modulation below 1 Hz with a high resolution (0.07 Hz). Analyses indicated that force variability was greater for the stroke group compared with to healthy controls and for the paretic hand compared with the non-paretic hand. Force modulation below 1 Hz differentiated the stroke individuals and healthy controls, as well as the paretic and non-paretic hands. Specifically, stroke individuals (paretic hand) exhibited greater power ∼0.2 Hz (0.07–0.35 Hz) and lesser power ∼0.6 Hz (0.49–0.77 Hz) compared to healthy controls (non-dominant hand). Similarly, the paretic hand exhibited greater power ∼0.2 Hz, and lesser power ∼0.6 Hz than the non-paretic hand. Moreover, variability of force was strongly predicted from the modulation of specific frequencies below 1 Hz (R 2 = 0.80). Together, these findings indicate that the modulation of force below 1 Hz provides significant insight into changes in motor control after stroke.  相似文献   

18.
The mating call of the Atlantic toadfish is generated by bursts of high-frequency twitches of the superfast twitch fibers that surround the swimbladder. At 16°C, a calling period can last several hours, with individual 80–100-Hz calls lasting ∼500 ms interleaved with silent periods (intercall intervals) lasting ∼10 s. To understand the intracellular movements of Ca2+ during the intercall intervals, superfast fibers were microinjected with fluo-4, a high-affinity fluorescent Ca2+ indicator, and stimulated by trains of 40 action potentials at 83 Hz, which mimics fiber activity during calling. The fluo-4 fluorescence signal was measured during and after the stimulus trains; the signal was also simulated with a kinetic model of the underlying myoplasmic Ca2+ movements, including the binding and transport of Ca2+ by the sarcoplasmic reticulum (SR) Ca2+ pumps. The estimated total amount of Ca2+ released from the SR during a first stimulus train is ∼6.5 mM (concentration referred to the myoplasmic water volume). At 40 ms after cessation of stimulation, the myoplasmic free Ca2+ concentration ([Ca2+]) is below the threshold for force generation (∼3 µM), yet the estimated concentration of released Ca2+ remaining in the myoplasm (Δ[CaM]) is large, ∼5 mM, with ∼80% bound to parvalbumin. At 10 s after stimulation, [Ca2+] is ∼90 nM (three times the assumed resting level) and Δ[CaM] is ∼1.3 mM, with 97% bound to parvalbumin. Ca2+ movements during the intercall interval thus appear to be strongly influenced by (a) the accumulation of Ca2+ on parvalbumin and (b) the slow rate of Ca2+ pumping that ensues when parvalbumin lowers [Ca2+] near the resting level. With repetitive stimulus trains initiated at 10-s intervals, Ca2+ release and pumping come quickly into balance as a result of the stability (negative feedback) supplied by the increased rate of Ca2+ pumping at higher [Ca2+].  相似文献   

19.
Auditory feedback from the animal''s own voice is essential during bat echolocation: to optimize signal detection, bats continuously adjust various call parameters in response to changing echo signals. Auditory feedback seems also necessary for controlling many bat communication calls, although it remains unclear how auditory feedback control differs in echolocation and communication. We tackled this question by analyzing echolocation and communication in greater horseshoe bats, whose echolocation pulses are dominated by a constant frequency component that matches the frequency range they hear best. To maintain echoes within this “auditory fovea”, horseshoe bats constantly adjust their echolocation call frequency depending on the frequency of the returning echo signal. This Doppler-shift compensation (DSC) behavior represents one of the most precise forms of sensory-motor feedback known. We examined the variability of echolocation pulses emitted at rest (resting frequencies, RFs) and one type of communication signal which resembles an echolocation pulse but is much shorter (short constant frequency communication calls, SCFs) and produced only during social interactions. We found that while RFs varied from day to day, corroborating earlier studies in other constant frequency bats, SCF-frequencies remained unchanged. In addition, RFs overlapped for some bats whereas SCF-frequencies were always distinctly different. This indicates that auditory feedback during echolocation changed with varying RFs but remained constant or may have been absent during emission of SCF calls for communication. This fundamentally different feedback mechanism for echolocation and communication may have enabled these bats to use SCF calls for individual recognition whereas they adjusted RF calls to accommodate the daily shifts of their auditory fovea.  相似文献   

20.
General anesthesia is not a uniform state of the brain. Ongoing activity differs between light and deep anesthesia and cortical response properties are modulated in dependence of anesthetic dosage. We investigated how anesthesia level affects cross-modal interactions in primary sensory cortex. To examine this, we continuously measured the effects of visual and auditory stimulation during increasing and decreasing isoflurane level in the mouse visual cortex and the subiculum (from baseline at 0.7 to 2.5 vol % and reverse). Auditory evoked burst activity occurred in visual cortex after a transition during increase of anesthesia level. At the same time, auditory and visual evoked bursts occurred in the subiculum, even though the subiculum was unresponsive to both stimuli previous to the transition. This altered sensory excitability was linked to the presence of burst suppression activity in cortex, and to a regular slow burst suppression rhythm (∼0.2 Hz) in the subiculum. The effect disappeared during return to light anesthesia. The results show that pseudo-heteromodal sensory burst responses can appear in brain structures as an effect of an anesthesia induced state change.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号