On the use of a real-time convolution system to study perception of and response to self-generated speech and music in variable acoustical environments

ORAL

Abstract

A real-time convolution system has been developed to quickly manipulate the auditory room-acoustical experiences of human subjects. This system is used to study the perception of self-generated speech and music, and the responses of talkers and musicians to varying conditions. Simulated and measured oral-binaural room impulse responses are used within the convolution system. Subjects in an anechoic environment experience room responses excited by their own voices or instruments via the convolution system. Direct sound travels directly to the ear, but the convolved room response is heard through specialized headphones spaced away from the head. The development of the convolution system and future research regarding its use are discussed.

Authors

  • Jennifer Whiting

    Brigham Young University

  • Timothy Leishman

    Brigham Young University

  • Eric Hunter

    Michigan State University