Blind compensation of interchannel sampling frequency mismatch for ad hoc microphone array based on maximum likelihood estimation

In this paper, we propose a novel method for the blind compensation of drift for the asynchronous recording of an ad hoc microphone array. Digital signals simultaneously observed by different recording devices have drift of the time differences between the observation channels because of the sampling frequency mismatch among the devices. On the basis of a model in which the time difference is constant within each short time frame but varies in proportion to the central time of the frame, the effect of the sampling frequency mismatch can be compensated in the short-time Fourier transform (STFT) domain by a linear phase shift. By assuming that the sources are motionless and have stationary amplitudes, the observation is regarded as being stationary when drift does not occur. Thus, we formulate a likelihood to evaluate the stationarity in the STFT domain to evaluate the compensation of drift. The maximum likelihood estimation is obtained effectively by a golden section search. Using the estimated parameters, we compensate the drift by STFT analysis with a noninteger frame shift. The effectiveness of the proposed blind drift compensation method is evaluated in an experiment in which artificial drift is generated. HighlightsModeling of drift as frame shift.STFT-domain compensation as linear phase shift.Probabilistic model of drift.Maximum likelihood estimation of sampling frequency mismatch.Efficient resampling with modified STFT analysis with noninteger frame shift.

[1]  Tomohiro Nakatani,et al.  Distributed microphone array processing for speech source separation with classifier fusion , 2012, 2012 IEEE International Workshop on Machine Learning for Signal Processing.

[2]  Shoji Makino,et al.  Blind compensation of inter-channel sampling frequency mismatch with maximum likelihood estimation in STFT domain , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[3]  Michael S. Brandstein,et al.  Microphone Arrays - Signal Processing Techniques and Applications , 2001, Microphone Arrays.

[4]  Shoji Makino,et al.  Optimizing frame analysis with non-integrer shift for sampling mismatch compensation of long recording , 2013, 2013 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics.

[5]  Rémi Gribonval,et al.  Performance measurement in blind audio source separation , 2006, IEEE Transactions on Audio, Speech, and Language Processing.

[6]  Martin Vetterli,et al.  Rate-Constrained Collaborative Noise Reduction for Wireless Hearing Aids , 2009, IEEE Transactions on Signal Processing.

[7]  Danilo P. Mandic,et al.  Kurtosis-Based Blind Source Extraction of Complex Non-Circular Signals with Application in EEG Artifact Removal in Real-Time , 2011, Front. Neurosci..

[8]  Nobutaka Ono,et al.  Stable and fast update rules for independent vector analysis based on auxiliary function technique , 2011, 2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA).

[9]  Shigeki Sagayama,et al.  Blind Estimation of Locations and Time Offsets for Distributed Recording Devices , 2010, LVA/ICA.

[10]  Zicheng Liu,et al.  Energy-Based Sound Source Localization and Gain Normalization for Ad Hoc Microphone Arrays , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.

[11]  Alexander Bertrand,et al.  Applications and trends in wireless acoustic sensor networks: A signal processing perspective , 2011, 2011 18th IEEE Symposium on Communications and Vehicular Technology in the Benelux (SCVT).

[12]  Ted S. Wada,et al.  On Dealing with Sampling Rate Mismatches in Blind Source Separation and Acoustic Echo Cancellation , 2007, 2007 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics.

[13]  Minerva M. Yeung,et al.  On the importance of exact synchronization for distributed audio signal processing , 2003, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)..

[14]  Zicheng Liu SOUND SOURCE SEPARATION WITH DISTRIBUTED MICROPHONE ARRAYS IN THE PRESENCE OF CLOCK SYNCHRONIZATION ERRORS , 2008 .

[15]  Israel Cohen,et al.  Blind Sampling Rate Offset Estimation and Compensation in Wireless Acoustic Sensor Networks with Application to Beamforming , 2012, IWAENC.

[16]  Shigeru Katagiri,et al.  ATR Japanese speech database as a tool of speech recognition and synthesis , 1990, Speech Commun..

[17]  Nobutaka Ito,et al.  Blind alignment of asynchronously recorded signals for distributed microphone array , 2009, 2009 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics.

[18]  Te-Won Lee,et al.  Blind Speech Separation , 2007, Blind Speech Separation.

[19]  Rainer Lienhart,et al.  Position calibration of microphones and loudspeakers in distributed computing platforms , 2005, IEEE Transactions on Speech and Audio Processing.

[20]  M. Eckert Slowing Down: Age-Related Neurobiological Predictors of Processing Speed , 2011, Front. Neurosci..

[21]  Reinhold Häb-Umbach,et al.  Sampling rate synchronisation in acoustic sensor networks with a pre-trained clock skew error model , 2013, 21st European Signal Processing Conference (EUSIPCO 2013).

[22]  Christian Schindelhauer,et al.  Acoustic Self-calibrating System for Indoor Smartphone Tracking (ASSIST) , 2012, 2012 International Conference on Indoor Positioning and Indoor Navigation (IPIN).

[23]  Christian Schindelhauer,et al.  Self-localization application for iPhone using only ambient sound signals , 2010, 2010 International Conference on Indoor Positioning and Indoor Navigation.