A Directional-Linear Bayesian Network and Its Application for Clustering and Simulation of Neural Somas

Neural somas perform most of the metabolic activities in the neuron and support the chemical process that generates the basic elements of the synapses, and consequently the brain activity. The morphology of the somas is one of the fundamental features for classifying neurons and their functionality. In this paper, we characterize the morphology of the 39 three-dimensional reconstructed human pyramidal somas in terms of their multiresolutional Reeb graph representation, from which we extract a set of directional and linear variables to perform model-based clustering. To deal with this dataset, we introduce the novel Extended Mardia-Sutton mixture model whose mixture components are distributed according to a newly proposed multivariate probability density function that is able to capture the directional-linear correlations. We exploit the capabilities of Bayesian networks in combination with the Structural Expectation-Maximization algorithm to learn the finite mixture model that clusters the neural somas by their morphology and the conditional independence constraints between variables. We also derive the Kullback–Leibler divergence of the Extended Mardia-Sutton distribution to be used as a measure of similarity between soma clusters. The proposed finite mixture model discovered three subtypes of human pyramidal somas. We performed Weltch t-tests and Watson-Williams tests, as well as rule-based identification of clusters to characterize each group by its most prominent features. Furthermore, the resulting model allows us to simulate the 3D virtual representations of somas from each cluster, which can be a useful tool for neuroscientists to reason and suggest new hypotheses.

[1]  Francesco Lagona,et al.  A Latent-Class Model for Clustering Incomplete Linear and Circular Data in Marine Studies , 2021 .

[2]  Concha Bielza,et al.  3D morphology-based clustering and simulation of human pyramidal cell dendritic spines , 2018, PLoS Comput. Biol..

[3]  I. Vitoriano Directional-linear Bayesian networks and applications in neuroscience , 2018 .

[4]  Anabik Pal,et al.  JCLMM: A finite mixture model for clustering of circular-linear data and its application to psoriatic plaque segmentation , 2017, Pattern Recognit..

[5]  김경민,et al.  Finite mixture models and model-based clustering , 2017 .

[6]  María Concepción Bielza Lozoya,et al.  Hybrid Gaussian and von Mises model-based clustering , 2016, ECAI 2016.

[7]  Concha Bielza,et al.  A univocal definition of the neuronal soma morphology using Gaussian mixture models , 2015, Front. Neuroanat..

[8]  Concha Bielza,et al.  Article in Press G Model Artificial Intelligence in Medicine Classifying Gabaergic Interneurons with Semi-supervised Projected Model-based Clustering , 2022 .

[9]  Christophe Ley,et al.  A tractable, parsimonious and flexible model for cylindrical data, with applications , 2015, 1505.08113.

[10]  Utpal Roy,et al.  SWGMM: a semi-wrapped Gaussian mixture model for clustering of circular–linear data , 2014, Pattern Analysis and Applications.

[11]  A. Maruotti,et al.  Bayesian hidden Markov modelling using circular‐linear general projected normal distribution , 2014, 1408.4834.

[12]  Rafael Yuste,et al.  Age-based comparison of human dendritic spine structure using complete three-dimensional reconstructions. , 2013, Cerebral cortex.

[13]  Charles C. Taylor,et al.  Non‐parametric Regression for Circular Responses , 2013 .

[14]  Kanti V. Mardia,et al.  Mixtures of concentrated multivariate sine distributions with applications to bioinformatics , 2012 .

[15]  A. Maruotti,et al.  A Multivariate Hidden Markov Model for the Identification of Sea Regimes from Incomplete Skewed and Circular Time Series , 2012 .

[16]  Shengrui Wang,et al.  Measuring the component overlapping in the Gaussian mixture model , 2011, Data Mining and Knowledge Discovery.

[17]  A. Maruotti,et al.  A non-homogeneous hidden Markov model for the analysis of multi-pollutant exceedances data , 2011 .

[18]  Karel Svoboda,et al.  The Past, Present, and Future of Single Neuron Reconstruction , 2011, Neuroinformatics.

[19]  Jose Miguel Puerta,et al.  Learning Bayesian networks by hill climbing: efficient methods based on progressive restriction of the neighborhood , 2010, Data Mining and Knowledge Discovery.

[20]  Shi-Qing Xin,et al.  Improving Chen and Han's algorithm on the discrete geodesic problem , 2009, TOGS.

[21]  Nir Friedman,et al.  Probabilistic Graphical Models - Principles and Techniques , 2009 .

[22]  Demian Battaglia,et al.  Classification of NPY-Expressing Neocortical Interneurons , 2009, The Journal of Neuroscience.

[23]  K. Shimizu,et al.  Dependent models for observations which include angular ones , 2008 .

[24]  J. A. Carta,et al.  A joint probability density function of wind speed and direction for wind energy analysis , 2008 .

[25]  Kanti V. Mardia,et al.  A multivariate von mises distribution with applications to bioinformatics , 2008 .

[26]  Daniel Cohen-Or,et al.  Consistent mesh partitioning and skeletonisation using the shape diameter function , 2008, The Visual Computer.

[27]  Constantin F. Aliferis,et al.  The max-min hill-climbing Bayesian network structure learning algorithm , 2006, Machine Learning.

[28]  Rafael Yuste,et al.  Dendritic size of pyramidal neurons differs among mouse cortical regions. , 2006, Cerebral cortex.

[29]  Michael M. Kazhdan,et al.  Poisson surface reconstruction , 2006, SGP '06.

[30]  Inderjit S. Dhillon,et al.  Clustering on the Unit Hypersphere using von Mises-Fisher Distributions , 2005, J. Mach. Learn. Res..

[31]  Remco C. Veltkamp,et al.  A survey of content based 3D shape retrieval methods , 2004, Proceedings Shape Modeling Applications, 2004..

[32]  H. Sung Gaussian Mixture Regression and Classification , 2004 .

[33]  G. Cooper,et al.  A Bayesian Method for the Induction of Probabilistic Networks from Data , 2004, Machine-mediated learning.

[34]  Adrian E. Raftery,et al.  Model-Based Clustering, Discriminant Analysis, and Density Estimation , 2002 .

[35]  Taku Komura,et al.  Topology matching for fully automatic similarity estimation of 3D shapes , 2001, SIGGRAPH.

[36]  S. R. Jammalamadaka,et al.  Topics in Circular Statistics , 2001 .

[37]  Nir Friedman,et al.  Learning Belief Networks in the Presence of Missing Values and Hidden Variables , 1997, ICML.

[38]  G. McLachlan,et al.  The EM algorithm and extensions , 1996 .

[39]  William W. Cohen Fast Effective Rule Induction , 1995, ICML.

[40]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[41]  Judea Pearl,et al.  Probabilistic reasoning in intelligent systems - networks of plausible inference , 1991, Morgan Kaufmann series in representation and reasoning.

[42]  D. N. Geary Mixture Models: Inference and Applications to Clustering , 1989 .

[43]  A. F. Smith,et al.  Statistical analysis of finite mixture distributions , 1986 .

[44]  N. Fisher,et al.  The BIAS of the maximum likelihood estimators of the von mises-fisher concentration parameters , 1981 .

[45]  Richard A. Johnson,et al.  Some Angular-Linear Distributions and Related Regression Models , 1978 .

[46]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[47]  Kanti V. Mardia,et al.  A Model for Cylindrical Variables with Applications , 1978 .

[48]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[49]  E. Batschelet,et al.  Angular-linear correlation coefficient for rhythmometry and circannually changing human birth rates at different geographic latitudes. , 1973, International journal of chronobiology.

[50]  K. Mardia Statistics of Directional Data , 1972 .

[51]  G. S. Watson,et al.  ON THE CONSTRUCTION OF SIGNIFICANCE TESTS ON THE CIRCLE AND THE SPHERE , 1956 .

[52]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[53]  Welch Bl THE GENERALIZATION OF ‘STUDENT'S’ PROBLEM WHEN SEVERAL DIFFERENT POPULATION VARLANCES ARE INVOLVED , 1947 .