A model for adaptive reduced-dimensionality equalisation

We present a method for mapping between the input space of a parametric equaliser and a lower-dimensional representation, whilst preserving the effect’s dependency on the incoming audio signal. The model consists of a parameter weighting stage in which the parameters are scaled to spectral features of the audio signal, followed by a mapping process, in which the equaliser’s 13 inputs are converted to (x, y) coordinates. The model is trained with parameter space data representing two timbral adjectives (warm and bright), measured across a range of musical instrument samples, allowing users to impose a semantically-meaningful timbral modification using the lower-dimensional interface. We test 10 mapping techniques, comprising of dimensionality reduction and reconstruction methods, and show that a stacked autoencoder algorithm exhibits the lowest parameter reconstruction variance, thus providing an accurate map between the input and output space. We demonstrate that the model provides an intuitive method for controlling the audio effect’s parameter space, whilst accurately reconstructing the trajectories of each parameter and adapting to the incoming audio spectrum.

[1]  H. Hotelling Analysis of a complex of statistical variables into principal components. , 1933 .

[2]  Bryan Pardo,et al.  Social-EQ: Crowdsourcing an Equalization Descriptor Map , 2013, ISMIR.

[3]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[4]  R. Fisher THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMS , 1936 .

[5]  Joe Wolfe,et al.  Does timbral brightness scale with frequency and spectral centroid , 2006 .

[6]  Jörn Loviscach,et al.  subjEQt: controlling an equalizer through subjective terms , 2006, CHI EA '06.

[7]  Joshua D. Reiss,et al.  Analysis of Musical Timbre Semantics through Metric and Non-Metric Data Reduction Techniques , 2012 .

[8]  Tim Brookes,et al.  Perceptually-Motivated Audio Morphing: Brightness , 2007 .

[9]  James W. Beauchamp,et al.  Synthesis by Spectral Amplitude and "Brightness" Matching of Analyzed Musical Instrument Tones , 1981 .

[10]  J. Grey Multidimensional perceptual scaling of musical timbres. , 1977, The Journal of the Acoustical Society of America.

[11]  György Fazekas,et al.  SAFE: A System for the Extraction and Retrieval of Semantic Audio Descriptors , 2014 .

[12]  Bryan Pardo,et al.  2DEQ: an intuitive audio equalizer , 2009, C&C '09.

[13]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[14]  Joshua D. Reiss,et al.  An Additive Synthesis Technique for Independent Modification of the Auditory Perceptions of Brightness and Warmth , 2011 .

[15]  Tim Brookes,et al.  Perceptually-Motivated Audio Morphing: Warmth , 2010 .

[16]  Peter J. Bickel,et al.  Maximum Likelihood Estimation of Intrinsic Dimension , 2004, NIPS.

[17]  A. de Cheveigné,et al.  The effect of fundamental frequency on the brightness dimension of timbre. , 2007, Journal of the Acoustical Society of America.

[18]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.