Cochleotopic/AMtopic (CAM) and Cochleotopic/Spectrotopic (CSM) map based sound sourcce separation using relaxatio oscillatory neurons

We use a two-layered unsupervised bio-inspired neural network to segregate sound sources, e.g. double-vowels or vowels intruded by nonstationary noise sources. The network consists of spiking neurons. The spiking neurons in both layers are modeled by relaxation oscillators. The first layer of the network is locally connected, while the second layer is a fully connected network. We show that in order to correctly segregate sound sources, we should either use Cochleotopic/AMtopic map (CAM) or Cochleotopic/Spectrotopic map (CSM) depending on the nature of the intruding sound source.

[1]  J. Guinan,et al.  The ipsilaterally evoked olivocochlear reflex causes rapid adaptation of the 2f1-f2 distortion product otoacoustic emission. , 1996, The Journal of the Acoustical Society of America.

[2]  Te-Won Lee,et al.  A Maximum Likelihood Approach to Single-channel Source Separation , 2003, J. Mach. Learn. Res..

[3]  Jean Rouat,et al.  A pitch determination and voiced/unvoiced decision algorithm for noisy speech , 1995, Speech Commun..

[4]  Robert D. Frisina,et al.  Effects of Age on Contralateral Suppression of Distortion Product Otoacoustic Emissions in Human Listeners with Normal Hearing , 2002, Audiology and Neurotology.

[5]  Phil D. Green,et al.  Robust automatic speech recognition with missing and unreliable acoustic data , 2001, Speech Commun..

[6]  G.-J. Jang,et al.  Single-channel signal separation using time-domain basis functions , 2003, IEEE Signal Processing Letters.

[7]  Sam T. Roweis,et al.  One Microphone Source Separation , 2000, NIPS.

[8]  P. Woodland,et al.  A computational model of the auditory periphery for speech and hearing research. II. Descending paths. , 1994, The Journal of the Acoustical Society of America.

[9]  Ch. von der Malsburg,et al.  A neural cocktail-party processor , 1986, Biological Cybernetics.

[10]  Daniel P. W. Ellis,et al.  The auditory organization of speech and other sources in listeners and computational models , 2001, Speech Commun..

[11]  Fabrice Plante,et al.  Improvement of speech spectrogram accuracy by the method of reassignment , 1998, IEEE Trans. Speech Audio Process..

[12]  Guy J. Brown,et al.  Separation of speech from interfering sounds based on oscillatory correlation , 1999, IEEE Trans. Neural Networks.

[13]  Jean Rouat,et al.  Nonlinear speech processing with oscillatory neural networks for speaker segregation , 2002, 2002 11th European Signal Processing Conference.

[14]  Jean Rouat,et al.  Oscillatory Dynamic Link Matching for Pattern Recognition , 2003 .