A tiered, Bayesian approach to estimating of population variability for regulatory decision-making.

Characterizing human variability in susceptibility to chemical toxicity is a critical issue in regulatory decision-making, but is usually addressed by a default 10-fold safety/uncertainty factor. Feasibility of population-based in vitro experimental approaches to more accurately estimate human variability was demonstrated recently using a large (~1000) panel of lymphoblastoid cell lines. However, routine use of such a large population-based model poses cost and logistical challenges. We hypothesize that a Bayesian approach embedded in a tiered workflow provides efficient estimation of variability and enables a tailored and sensible approach to selection of appropriate sample size. We used the previously collected lymphoblastoid cell line in vitro toxicity data to develop a data-derived prior distribution for the uncertainty in the degree of population variability. The resulting prior for the toxicodynamic variability factor (the ratio between the median and 1% most sensitive individuals) has a median (90% CI) of 2.5 (1.4-9.6). We then performed computational experiments using a hierarchical Bayesian population model with lognormal population variability with samples sizes of n = 5 to 100 to determine the change in precision and accuracy with increasing sample size. We propose a tiered Bayesian strategy for fit-for-purpose population variability estimates: (1) a default using the data-derived prior distribution; (2) a pilot experiment using samples sizes of ~20 individuals that reduces prior uncertainty by > 50% with > 80% balanced accuracy for classification; and (3) a high confidence experiment using sample sizes of ~50-100. This approach efficiently uses in vitro data on population variability to inform decision-making.

[1]  Robert Goble,et al.  A STRAW MAN PROPOSAL FOR A QUANTITATIVE DEFINITION OF THE RfD , 2002, Drug and chemical toxicology.

[2]  D. Rubin,et al.  Inference from Iterative Simulation Using Multiple Sequences , 1992 .

[3]  Alexander Golbraikh,et al.  A Novel Two-Step Hierarchical Quantitative Structure–Activity Relationship Modeling Work Flow for Predicting Acute Toxicity of Chemicals in Rodents , 2009, Environmental health perspectives.

[4]  Jonathan I Levy,et al.  Science and Decisions: Advancing Risk Assessment , 2010, Risk analysis : an official publication of the Society for Risk Analysis.

[5]  P. Royston A Remark on Algorithm as 181: The W‐Test for Normality , 1995 .

[6]  Alexander Tropsha,et al.  Chembench: a cheminformatics workbench , 2010, Bioinform..

[7]  J. Bailar,et al.  Toxicity testing in the 21st century—a vision and a strategy , 2012 .

[8]  David M. Reif,et al.  Aggregating Data for Computational Toxicology Applications: The U.S. Environmental Protection Agency (EPA) Aggregated Computational Toxicology Resource (ACToR) System , 2012, International journal of molecular sciences.

[9]  Igor Linkov,et al.  Environment models and decisions , 2014, Environment Systems and Decisions.

[10]  Harvey J. Clewell,et al.  Physiologically Based Pharmacokinetic (PBPK) Modeling of Interstrain Variability in Trichloroethylene Metabolism in the Mouse , 2014, Environmental health perspectives.

[11]  Karsten M. Borgwardt,et al.  Prediction of human population responses to toxic compounds by a collaborative competition , 2015, Nature Biotechnology.

[12]  Ruili Huang,et al.  Population-Based in Vitro Hazard and Concentration–Response Assessment of Chemicals: The 1000 Genomes High-Throughput Screening Study , 2015, Environmental health perspectives.

[13]  N. Mantel The detection of disease clustering and a generalized regression approach. , 1967, Cancer research.

[14]  Weihsueh A. Chiu,et al.  A Unified Probabilistic Framework for Dose–Response Assessment of Human Health Effects , 2015, Environmental health perspectives.

[15]  Wout Slob,et al.  Benchmark dose and the three Rs. Part II. Consequences for study design and animal use , 2014, Critical reviews in toxicology.

[16]  Ivan Rusyn,et al.  Toxicogenetics: population-based testing of drug and chemical safety in mouse models. , 2010, Pharmacogenomics.

[17]  J. Hartigan,et al.  The Dip Test of Unimodality , 1985 .

[18]  W. Russell,et al.  Ethical and Scientific Considerations Regarding Animal Testing and Research , 2011, PloS one.

[19]  David M. Reif,et al.  Update on EPA's ToxCast program: providing high throughput decision support tools for chemical risk management. , 2012, Chemical research in toxicology.

[20]  A. Gelman,et al.  Stan , 2015 .

[21]  Ruili Huang,et al.  Quantitative high-throughput screening for chemical toxicity in a population-based in vitro model. , 2012, Toxicological sciences : an official journal of the Society of Toxicology.

[22]  I. Y. Chen,et al.  Induced pluripotent stem cells: at the heart of cardiovascular precision medicine , 2016, Nature Reviews Cardiology.

[23]  William R. Bell,et al.  Using the t-distribution to Deal with Outliers in Small Area Estimation , 2006 .

[24]  Wout Slob,et al.  Benchmark dose and the three Rs. Part I. Getting more information from the same number of animals , 2014, Critical reviews in toxicology.

[25]  Ruili Huang,et al.  CERAPP: Collaborative Estrogen Receptor Activity Prediction Project , 2016, Environmental health perspectives.

[26]  C. Austin,et al.  Improving the Human Hazard Characterization of Chemicals: A Tox21 Update , 2013, Environmental health perspectives.