Learn More
All 10 forms of the test of Speech Perception in Noise (SPIN) were presented to 128 listeners who had some degree of sensorineural hearing loss. Presentation of the speech track was at 50 dB above the estimated threshold for the babble track. Signal-to-babble ratio was 8 dB. Half of the subjects listened through headphones and half via loudspeaker. Half(More)
Consonant confusion matrices were obtained from 22 outpatient listeners with sensorineural hearing loss for four sets of CV and VC nonsense syllables, presented monaurally at SRT + 40 DB. Testing was typically conducted for six hours on each of two separate days. Overall performance and patterns of confusions were stable over time. Analysis of the matrices(More)
Spontaneous otoacoustic emissions (SOAEs), which occur in about 40% of normal-hearing humans, do not have a firm explanation in auditory theory nor are their distributional properties well understood. To enhance our understanding of SOAEs, we have pooled data from three reports (Hammel, 1983; Strickland, Burns, & Tubis, 1985; Zurek, 1981) to assemble a(More)
This paper presents a two-microphone technique for localization of multiple sound sources. Its fundamental structure is adopted from a binaural signal-processing scheme employed in biological systems for the localization of sources using interaural time differences (ITD). The two input signals are transformed to the frequency domain and analyzed for(More)
Two psychophysical methods, a method of adjustment (MOA) and a forced-choice double-staircase adaptive procedure (FCDS) (Jesteadt, 1980), were used to measure the predominant pitch and loudness of tinnitus for 11 subjects during one test session. The FCDS within-session variability of matches to tinnitus pitch was smaller than for the MOA and comparable to(More)
It has been found that listeners with sensorineural hearing loss who show similar patterns of consonant confusions also tend to have similar audiometric profiles. The present study determined whether normal listeners, presented with filtered speech, would produce consonant confusions similar to those previously reported for the hearing-impaired listener.(More)
Extraction of a target sound source amidst multiple interfering sound sources is difficult when there are fewer sensors than sources, as is the case for human listeners in the classic cocktail-party situation. This study compares the signal extraction performance of five algorithms using recordings of speech sources made with three different two-microphone(More)
This paper describes algorithms for signal extraction for use as a front-end of telecommunication devices, speech recognition systems, as well as hearing aids that operate in noisy environments. The development was based on some independent, hypothesized theories of the computational mechanics of biological systems in which directional hearing is enabled(More)