Individualized HRTFs using computer vision and computational acoustics

For a given sound source location and frequency content, the sound received at the two ears of a person is first scattered by his/her torso, head, and pinnae, leading to interaural differences in the intensity and spectral features of the sound received. These features are used as cues to perceive the spatial location of the source, and are encoded in an extremely individual ‘‘head‐related transfer function’’ (HRTF) that depends on the person’s external anatomy (structure of the torso, head, and pinnae). This individuality has made it difficult to use the HRTF to render perceptually convincing spatially distributed audio for applications such as entertainment or virtual reality. We have commenced a long‐term project that will use numerical methods to compute individualized HRTFs from accurate 3D surface models of a person’s torso, head, and pinnae (obtained using computer vision techniques). These surface models will be used with boundary‐element methods to solve the wave equation and compute the HRTFs. A...