Super-realistic environmental sound synthesizer for location-based sound search system

This paper presents a proposal of a super realistic environmental sound synthesizer for a location-based sound search system. The location-based sound search system enables a user to experience any environmental sound in any desired remote place. This synthesizer is achieved by spatial audio mixing considering real-world conditions such as geographical features and townscapes, as well as dynamic situations such as those of town events and weather. Furthermore, a new velocity vector-based clustering method is proposed to reduce the cost of audio processing on the server side, which achieves a lower number of composing and decomposing clusters that have a direct and strong impact on high-cost audio processing such as spatial audio mixing1.

[1]  Richard Kronland-Martinet,et al.  A 3-D Immersive Synthesizer for Environmental Sounds , 2010, IEEE Transactions on Audio, Speech, and Language Processing.

[2]  C.-C. Jay Kuo,et al.  Environmental sound recognition using MP-based features , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.

[3]  George Drettakis,et al.  Progressive perceptual audio rendering of complex scenes , 2007, SI3D.

[4]  Vincenzo Lombardo,et al.  A Framework for Soundscape Analysis And Re-synthesis , 2009 .

[5]  He Yuzhu,et al.  Sound rendering and its application in virtual cockpit , 1998, ICSP '98. 1998 Fourth International Conference on Signal Processing (Cat. No.98TH8344).

[6]  David B. Shmoys,et al.  A Best Possible Heuristic for the k-Center Problem , 1985, Math. Oper. Res..

[7]  George Drettakis,et al.  Perceptual audio rendering of complex virtual environments , 2004, ACM Trans. Graph..

[8]  Andrea Valle,et al.  A graph-based system for the dynamic generation of soundscapes , 2009 .

[9]  Takao Onoye,et al.  Embedded implementation of acoustic field enhancement for stereo sound sources , 2003, IEEE Trans. Consumer Electron..

[10]  Gregory H. Wakefield,et al.  Introduction to Head-Related Transfer Functions (HRTFs): Representations of HRTFs in Time, Frequency, and Space , 2001 .

[11]  J. Walker,et al.  A parametric model for spectral sound synthesis of musical sounds , 2008, 2008 International Conference on Audio, Language and Image Processing.

[12]  Luca Turchet,et al.  Sound Synthesis and Evaluation of Interactive Footsteps and Environmental Sounds Rendering for Virtual Reality Applications , 2011, IEEE Transactions on Visualization and Computer Graphics.

[13]  Jordi Janer,et al.  Soundscape Generation for Virtual Environments using Community-Provided Audio Databases , 2010 .

[14]  Brian Gygi,et al.  Similarity and categorization of environmental sounds , 2007, Perception & psychophysics.

[15]  X. Anguera XBIC : Real-Time Cross Probabilities measure for speaker segmentation , 2005 .

[16]  Takao Onoye,et al.  Embedded 3D sound movement system based on feature extraction of head-related transfer function , 2005, 2005 Digest of Technical Papers. International Conference on Consumer Electronics, 2005. ICCE..