-
The Journal of the Acoustical Society... Oct 2017Meaningful sounds represent the majority of sounds that humans hear and process in everyday life. Yet studies of human sound localization mainly use artificial stimuli... (Comparative Study)
Comparative Study
Meaningful sounds represent the majority of sounds that humans hear and process in everyday life. Yet studies of human sound localization mainly use artificial stimuli such as clicks, pure tones, and noise bursts. The present study investigated the influence of behavioral relevance, sound category, and acoustic properties on the localization of complex, meaningful sounds in the horizontal plane. Participants localized vocalizations and traffic sounds with two levels of behavioral relevance (low and high) within each category, as well as amplitude-modulated tones. Results showed a small but significant effect of behavioral relevance: localization acuity was higher for complex sounds with a high level of behavioral relevance at several target locations. The data also showed category-specific effects: localization biases were lower, and localization precision higher, for vocalizations than for traffic sounds in central space. Several acoustic parameters influenced sound localization performance as well. Correcting localization responses for front-back reversals reduced the overall variability across sounds, but behavioral relevance and sound category still had a modulatory effect on sound localization performance in central auditory space. The results thus demonstrate that spatial hearing performance for complex sounds is influenced not only by acoustic characteristics, but also by sound category and behavioral relevance.
Topics: Acoustic Stimulation; Adult; Cues; Female; Humans; Male; Noise, Transportation; Psychoacoustics; Sound Localization; Voice; Young Adult
PubMed: 29092572
DOI: 10.1121/1.5003779 -
The Journal of the Acoustical Society... Apr 2018The psychoacoustical literature contains multiple reports about small differences in performance depending upon the sex and phase of the menstrual cycle of the subjects....
The psychoacoustical literature contains multiple reports about small differences in performance depending upon the sex and phase of the menstrual cycle of the subjects. In an attempt to verify these past reports, a large-scale study was implemented. After extensive training, the performance of about 75 listeners was measured on seven common psychoacoustical tasks. For most tasks, the signal was a 3.0-kHz tone. The initial data analyses failed to confirm some past outcomes. Additional analyses, incorporating the limited information available about the racial background of the listeners, did confirm some of the past reports, with the direction and magnitude of the differences often diverging for the White and Non-White listeners. Sex differences and race differences interacted for six of the seven tasks studied. These interactions suggest that racial background needs to be considered when making generalizations about human auditory performance, and when considering failures of reproducibility across studies. Menstrual differences were small, but generally larger for Whites than Non-Whites. Hormonal effects may be responsible for the sex and cycle differences that do exist, and differences in intra-cochlear melanocytes may account for the race differences.
Topics: Acoustic Stimulation; Adolescent; Adult; Auditory Pathways; Differential Threshold; Female; Humans; Male; Menstrual Cycle; Psychoacoustics; Racial Groups; Sex Factors; Young Adult
PubMed: 29716303
DOI: 10.1121/1.5030998 -
PloS One 2020Bayesian staircases are widely used in psychophysics to estimate detection thresholds. Simulations have revealed the importance of the parameters selected for the...
Bayesian staircases are widely used in psychophysics to estimate detection thresholds. Simulations have revealed the importance of the parameters selected for the assumed subject's psychometric function in enabling thresholds to be estimated with small bias and high precision. One important parameter is the slope of the psychometric function, or equivalently its spread. This is often held fixed, rather than estimated for individual subjects, because much larger numbers of trials are required to estimate the spread as well as the threshold. However, if this fixed value is wrong, the threshold estimate can be biased. Here we determine the optimal slope to minimize bias and maximize precision when measuring stereoacuity with Bayesian staircases. We performed 2- and 4AFC disparity detection stereo experiments in order to measure the spread of the disparity psychometric function in human observers assuming a Logistic function. We found a wide range, between 0.03 and 3.5 log10 arcsec, with little change with age. We then ran simulations to examine the optimal spread using the empirical data. From our simulations and for three different experiments, we recommend selecting assumed spread values between the percentiles 60-80% of the population distribution of spreads (these percentiles can be extended to other type of thresholds). For stereo thresholds, we recommend a spread around the value σ = 1.7 log10 arcsec for 2AFC (slope β = 4.3 /log10 arcsec), and around σ = 1.5 log10 arcsec for 4AFC (β = 4.9 /log10 arcsec). Finally, we compared a Bayesian procedure (ZEST using the optimal σ) with five Bayesian procedures that are versions of ZEST-2D, Psi, and Psi-marginal. In general, for the conditions tested, ZEST optimal σ showed the lowest threshold bias and highest precision.
Topics: Adolescent; Adult; Bayes Theorem; Child; Child, Preschool; Empirical Research; Female; Humans; Male; Middle Aged; Psychoacoustics; Psychometrics; Sensory Thresholds; Visual Perception; Young Adult
PubMed: 31895925
DOI: 10.1371/journal.pone.0226822 -
NeuroImage Feb 2018Binaural hearing, the ability to detect small differences in the timing and level of sounds at the two ears, underpins the ability to localize sound sources along the...
Binaural hearing, the ability to detect small differences in the timing and level of sounds at the two ears, underpins the ability to localize sound sources along the horizontal plane, and is important for decoding complex spatial listening environments into separate objects - a critical factor in 'cocktail-party listening'. For human listeners, the most important spatial cue is the interaural time difference (ITD). Despite many decades of neurophysiological investigations of ITD sensitivity in small mammals, and computational models aimed at accounting for human perception, a lack of concordance between these studies has hampered our understanding of how the human brain represents and processes ITDs. Further, neural coding of spatial cues might depend on factors such as head-size or hearing range, which differ considerably between humans and commonly used experimental animals. Here, using magnetoencephalography (MEG) in human listeners, and electro-corticography (ECoG) recordings in guinea pig-a small mammal representative of a range of animals in which ITD coding has been assessed at the level of single-neuron recordings-we tested whether processing of ITDs in human auditory cortex accords with a frequency-dependent periodic code of ITD reported in small mammals, or whether alternative or additional processing stages implemented in psychoacoustic models of human binaural hearing must be assumed. Our data were well accounted for by a model consisting of periodically tuned ITD-detectors, and were highly consistent across the two species. The results suggest that the representation of ITD in human auditory cortex is similar to that found in other mammalian species, a representation in which neural responses to ITD are determined by phase differences relative to sound frequency rather than, for instance, the range of ITDs permitted by head size or the absolute magnitude or direction of ITD.
Topics: Adult; Animals; Auditory Cortex; Electrocorticography; Female; Guinea Pigs; Humans; Magnetoencephalography; Male; Psychoacoustics; Sound Localization; Time Factors; Young Adult
PubMed: 29122721
DOI: 10.1016/j.neuroimage.2017.11.012 -
Animal Cognition May 2016Metrical phonology is the perceptual "strength" in language of some syllables relative to others. The ability to perceive lexical stress is important, as it can help a... (Comparative Study)
Comparative Study
Metrical phonology is the perceptual "strength" in language of some syllables relative to others. The ability to perceive lexical stress is important, as it can help a listener segment speech and distinguish the meaning of words and sentences. Despite this importance, there has been little comparative work on the perception of lexical stress across species. We used a go/no-go operant paradigm to train human participants and budgerigars (Melopsittacus undulatus) to distinguish trochaic (stress-initial) from iambic (stress-final) two-syllable nonsense words. Once participants learned the task, we presented both novel nonsense words, and familiar nonsense words that had certain cues removed (e.g., pitch, duration, loudness, or vowel quality) to determine which cues were most important in stress perception. Members of both species learned the task and were then able to generalize to novel exemplars, showing categorical learning rather than rote memorization. Tests using reduced stimuli showed that humans could identify stress patterns with amplitude and pitch alone, but not with only duration or vowel quality. Budgerigars required more than one cue to be present and had trouble if vowel quality or amplitude were missing as cues. The results suggest that stress patterns in human speech can be decoded by other species. Further comparative stress-perception research with more species could help to determine what species characteristics predict this ability. In addition, tests with a variety of stimuli could help to determine how much this ability depends on general pattern learning processes versus vocalization-specific cues.
Topics: Adult; Animals; Conditioning, Operant; Cues; Female; Humans; Loudness Perception; Male; Melopsittacus; Pitch Perception; Psychoacoustics; Speech Acoustics; Speech Perception
PubMed: 26914456
DOI: 10.1007/s10071-016-0968-3 -
Journal of Experimental Psychology.... Dec 2015A target sound can become more audible and may "pop out" from a simultaneously presented masker if the masker is presented first by itself, as a precursor. This...
A target sound can become more audible and may "pop out" from a simultaneously presented masker if the masker is presented first by itself, as a precursor. This phenomenon, known as auditory enhancement, may reflect the general perceptual principle of contrast enhancement, which facilitates adaptation to ongoing acoustic conditions and the detection of new events. Little is known about the mechanisms underlying enhancement, and potential confounding factors have made the size of the effect and its time course a point of contention. Here we measured enhancement as a function of precursor duration and delay between precursor offset and target onset, using 2 single-interval pitch comparison tasks, which involve either same-different or up-down judgments, to avoid the potential confounds of earlier studies. Although these 2 tasks elicit different levels of performance and may reflect different underlying mechanisms, they produced similar amounts of enhancement. The effect decreased with decreasing precursor duration, but remained present for precursors as short as 62.5 ms, and decreased with increasing gap between the precursor and target, but remained measurable 1 s after the precursor. Additional conditions, examining the effect of precursor/masker similarity and the possible role of grouping and cueing, suggest multiple sources of auditory enhancement.
Topics: Acoustic Stimulation; Adolescent; Adult; Auditory Threshold; Female; Humans; Male; Perceptual Masking; Psychoacoustics; Young Adult
PubMed: 26280269
DOI: 10.1037/xhp0000115 -
The Journal of the Acoustical Society... Sep 2002The dynamics of sound localization were studied using a free-field direct localization task (pointing to sound sources) and an observer-weighting analysis that assessed...
The dynamics of sound localization were studied using a free-field direct localization task (pointing to sound sources) and an observer-weighting analysis that assessed the relative influence of each click in a click-train stimulus. In agreement with previous studies of the precedence effect and binaural adaptation, weighting functions showed increased influence of the onset click when the interclick interval (ICI) was short (<5 ms). For longer ICIs, all clicks in a train contributed roughly the same amount to listeners' localization responses. Finally, when a short gap was introduced in the middle of a train, the influence of the click immediately following the gap increased, in agreement with the "restarting" results obtained by Hafter and Buell [J. Acoust. Soc. Am. 88, 806-812 (1990)].
Topics: Acoustic Stimulation; Attention; Functional Laterality; Humans; Loudness Perception; Pitch Perception; Psychoacoustics; Sound Localization; Sound Spectrography; Time Perception
PubMed: 12243153
DOI: 10.1121/1.1497366 -
The Journal of the Acoustical Society... Dec 2003The term informational masking has traditionally been used to refer to elevations in signal threshold resulting from masker uncertainty. In the present study, the method...
The term informational masking has traditionally been used to refer to elevations in signal threshold resulting from masker uncertainty. In the present study, the method of constant stimuli was used to obtain complete psychometric functions (PFs) from 44 normal-hearing listeners in conditions known to produce varying amounts of informational masking. The listener's task was to detect a pure-tone signal in the presence of a broadband noise masker (low masker uncertainty) and in the presence of multitone maskers with frequencies and amplitudes that varied at random from one presentation to the next (high masker uncertainty). Relative to the broadband noise condition, significant reductions were observed in both the slope and the upper asymptote of the PF for multitone maskers producing large amounts of informational masking. Slope was affected more for some listeners and conditions while asymptote was affected more for others; consequently, neither parameter alone was highly predictive of individual thresholds or the amount of informational masking. Mean slopes and asymptotes varied nonmonotonically with the number of masker components in a manner similar to mean thresholds, particularly when the estimated effect of energetic masking on thresholds was subtracted out. As in past studies, the threshold data were well described by a model in which trial-by-trial judgments are based on a weighted sum of levels in dB at the output of independent auditory filters. The psychometric data, however, complicated the model's interpretation in two ways: First, they suggested that, depending on the listener and condition, the weights can either reflect a fixed influence of masker components on each trial or the effect of occasionally mistaking a masker component for the signal from trial to trial. Second, they indicated that in either case the variance of the underlying decision variable as estimated from PF slope is not by itself great enough to account for the observed changes in informational masking.
Topics: Acoustic Stimulation; Adolescent; Adult; Attention; Auditory Threshold; Child; Female; Humans; Male; Perceptual Masking; Pitch Perception; Psychoacoustics; Psychometrics; Reference Values; Sound Spectrography
PubMed: 14714808
DOI: 10.1121/1.1629303 -
Journal of Speech, Language, and... Feb 2018The purpose of this study was to evaluate whether listeners with normal hearing perceiving noise-vocoded speech-in-speech demonstrate better intelligibility of target...
PURPOSE
The purpose of this study was to evaluate whether listeners with normal hearing perceiving noise-vocoded speech-in-speech demonstrate better intelligibility of target speech when the background speech was mismatched in language (linguistic release from masking [LRM]) and/or location (spatial release from masking [SRM]) relative to the target. We also assessed whether the spectral resolution of the noise-vocoded stimuli affected the presence of LRM and SRM under these conditions.
METHOD
In Experiment 1, a mixed factorial design was used to simultaneously manipulate the masker language (within-subject, English vs. Dutch), the simulated masker location (within-subject, right, center, left), and the spectral resolution (between-subjects, 6 vs. 12 channels) of noise-vocoded target-masker combinations presented at +25 dB signal-to-noise ratio (SNR). In Experiment 2, the study was repeated using a spectral resolution of 12 channels at +15 dB SNR.
RESULTS
In both experiments, listeners' intelligibility of noise-vocoded targets was better when the background masker was Dutch, demonstrating reliable LRM in all conditions. The pattern of results in Experiment 1 was not reliably different across the 6- and 12-channel noise-vocoded speech. Finally, a reliable spatial benefit (SRM) was detected only in the more challenging SNR condition (Experiment 2).
CONCLUSION
The current study is the first to report a clear LRM benefit in noise-vocoded speech-in-speech recognition. Our results indicate that this benefit is available even under spectrally degraded conditions and that it may augment the benefit due to spatial separation of target speech and competing backgrounds.
Topics: Acoustic Stimulation; Adolescent; Humans; Linguistics; Noise; Perceptual Masking; Psychoacoustics; Speech Perception; Young Adult
PubMed: 29396580
DOI: 10.1044/2017_JSLHR-H-17-0215 -
The Journal of the Acoustical Society... Aug 2014Monaural spectral features are important for human sound-source localization in sagittal planes, including front-back discrimination and elevation perception. These...
Monaural spectral features are important for human sound-source localization in sagittal planes, including front-back discrimination and elevation perception. These directional features result from the acoustic filtering of incoming sounds by the listener's morphology and are described by listener-specific head-related transfer functions (HRTFs). This article proposes a probabilistic, functional model of sagittal-plane localization that is based on human listeners' HRTFs. The model approximates spectral auditory processing, accounts for acoustic and non-acoustic listener specificity, allows for predictions beyond the median plane, and directly predicts psychoacoustic measures of localization performance. The predictive power of the listener-specific modeling approach was verified under various experimental conditions: The model predicted effects on localization performance of band limitation, spectral warping, non-individualized HRTFs, spectral resolution, spectral ripples, and high-frequency attenuation in speech. The functionalities of vital model components were evaluated and discussed in detail. Positive spectral gradient extraction, sensorimotor mapping, and binaural weighting of monaural spatial information were addressed in particular. Potential applications of the model include predictions of psychophysical effects, for instance, in the context of virtual acoustics or hearing assistive devices.
Topics: Acoustic Stimulation; Acoustics; Cues; Discrimination, Psychological; Humans; Models, Theoretical; Motion; Pattern Recognition, Physiological; Psychoacoustics; Sound; Sound Localization; Speech Perception; Time Factors
PubMed: 25096113
DOI: 10.1121/1.4887447