The effects of intense pure tones on human auditory temporal acuity were examined. Five college age adults (four males) were exposed to (SPL) pure tones of sufficient duration to induce an 8 to 12dB temporary threshold shift (TTS) a half octave above the test frequency 2 minutes after exposure. The subjects were then exposed to octave band noise having frequencies of 1.7, 2.4, and 3.4kHz (when TTSs were induced by the 1.7kHz tone) or 0.4, 0.56, or 0.8kHz (when TTSs were induced by the 0.5kHz tone) at levels of 35, 55, and 75dB SPL. Gap detection thresholds (GDTs) were measured before and 2 to 16 minutes after exposure to the three octave band noise. Absolute thresholds (ATs) were also measured. Significant increases in GDT occurred only after exposure to 2.4, 3.4, 0.56, and 0.8kHz broad band noise at intensities of 35 and 55dB. The shifts in GDT were correlated with the TTS; however, the coefficients of correlation were smaller for the low frequency noise. Following exposures to the 2.4 and 3.4kHz centered noise, the shifts in GDT and AT required similar amounts of time to recover. Shifts in GDT following exposure to 0.56 and 0.8kHz centered noise required a longer time to recover than the shifts in AT. The authors conclude that assessing temporary acuity may be a useful technique for quantifying the effects of intense low frequency sounds.