Ensemble learning reveals dissimilarity between rare-earth transition-metal binary alloys with respect to the Curie temperature

We propose a data-driven method to extract dissimilarity between materials, with respect to a given target physical property. The technique is based on an ensemble method with Kernel ridge regression as the predicting model; multiple random subset sampling of the materials is done to generate prediction models and the corresponding contributions of the reference training materials in detail. The distribution of the predicted values for each material can be approximated by a Gaussian mixture models. The reference training materials contributed to the prediction model that accurately predicts the physical property value of a specific material, are considered to be similar to that material, or vice versa. Evaluations using synthesized data demonstrate that the proposed method can effectively measure the dissimilarity between data instances. An application of the analysis method on the data of Curie temperature () of binary 3d transition metal- 4f rare-earth binary alloys also reveals meaningful results on the relations between the materials. The proposed method can be considered as a potential tool for obtaining a deeper understanding of the structure of data, with respect to a target property, in particular.

[1]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[2]  Satoshi Morinaga,et al.  Fully-Automatic Bayesian Piecewise Sparse Linear Models , 2014, AISTATS.

[3]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[4]  M. Stone Cross‐Validatory Choice and Assessment of Statistical Predictions , 1976 .

[5]  Thomas G. Dietterich Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.

[6]  Sanguthevar Rajasekaran,et al.  Accelerating materials property predictions using machine learning , 2013, Scientific Reports.

[7]  Yingjie Tian,et al.  A Comprehensive Survey of Clustering Algorithms , 2015, Annals of Data Science.

[8]  Christian P. Robert,et al.  Machine Learning, a Probabilistic Perspective , 2014 .

[9]  Shuichi Iwata,et al.  The Pauling File, Binaries Edition , 2004 .

[10]  Rampi Ramprasad,et al.  Adaptive machine learning framework to accelerate ab initio molecular dynamics , 2015 .

[11]  R. Dennis Cook,et al.  Cross-Validation of Regression Models , 1984 .

[12]  Zhi-Hua Zhou,et al.  Ensemble Methods: Foundations and Algorithms , 2012 .

[13]  Hieu Chi Dam,et al.  Important Descriptors and Descriptor Groups of Curie Temperatures of Rare-earth Transition-metal Binary Alloys , 2018, Journal of the Physical Society of Japan.

[14]  Matthias Rupp,et al.  Machine learning for quantum mechanics in a nutshell , 2015 .

[15]  Pierre Baldi,et al.  The dropout learning algorithm , 2014, Artif. Intell..

[16]  Joseph N. Wilson,et al.  Twenty Years of Mixture of Experts , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[17]  Jon Atli Benediktsson,et al.  Multiple Classifier Systems , 2015, Lecture Notes in Computer Science.

[18]  Josef Kittler,et al.  Multiple Classifier Systems , 2004, Lecture Notes in Computer Science.

[19]  Kevin P. Murphy,et al.  Machine learning - a probabilistic perspective , 2012, Adaptive computation and machine learning series.

[20]  Truyen Tran,et al.  Committee machine that votes for similarity between materials , 2018, IUCrJ.

[21]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[22]  Kohei Hayashi,et al.  Factorized Asymptotic Bayesian Inference for Latent Feature Models , 2013, NIPS.

[23]  Geoffrey E. Hinton,et al.  Adaptive Mixtures of Local Experts , 1991, Neural Computation.