Deep Learning Approaches to Chemical Property Prediction from Brewing Recipes

Despite the explosion of craft beer brewing over the last decade, there is virtually no work in the public domain exploring machine learning approaches to understand and optimize the brewing process. Learning to map between representations of an object across different domains is one of the fundamental challenges in machine learning. There are at least three distinct representations of beer that one may wish to learn to map between: 1) the brewing recipe, 2) the chemical composition of the resulting beer and 3) the written reviews of the beer. The mapping between any pair of these three domains is highly non-linear. Brewing beer involves complicated biological and chemical processes, while the human qualitative perception of a beer may be even more complex. In the work described in this paper, we focus on the former: mapping between the recipe and chemical attribute domains. We use two deep learning architectures to model the non-linear relationship between beer in these two domains, classifying coarse- and fine-grained beer type and predicting ranges for original gravity, final gravity, alcohol by volume, international bitterness units and color. Such models could be used to optimize recipes to produce desired chemical properties of the beer, allowing brewers to design better tasting beer, faster and with less waste. Using a set of approximately 223K brewing recipes from homebrewing site http://brewtoad.com, we find that deep and recurrent neural network models significantly outperform several baselines in predicting these attributes, offering relative reductions in classification error by 20%+ and reducing the root mean squared error for the attribute ranges by 44% relative to the best baseline.

[1]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[2]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[3]  Sharad Vikram,et al.  Capturing Meaning in Product Reviews with Character-Level Generative Text Models , 2015, ArXiv.

[4]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[5]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[6]  P. McCullagh,et al.  Generalized Linear Models , 1984 .

[7]  Jure Leskovec,et al.  From amateurs to connoisseurs: modeling the evolution of user expertise through online reviews , 2013, WWW.

[8]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[9]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[10]  Jeffrey L. Elman,et al.  Finding Structure in Time , 1990, Cogn. Sci..

[11]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[12]  Phil Blunsom,et al.  Recurrent Continuous Translation Models , 2013, EMNLP.

[13]  Yejin Choi,et al.  Mise en Place: Unsupervised Interpretation of Instructional Recipes , 2015, EMNLP.

[14]  Carson James Carroll THE BOTTLEFLY IOS APPLICATION FOR WINE RECOMMENDATIONS , 2016 .

[15]  Alexander J. Smola,et al.  Deep Sets , 2017, 1703.06114.

[16]  Martín Abadi,et al.  TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.

[17]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[18]  Chloé Kiddon,et al.  Learning to Interpret and Generate Instructional Recipes , 2016 .

[19]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .