Quantifying risk associated with clinical trial termination: A text mining approach

Abstract Clinical trials that terminate prematurely without reaching conclusions raise financial, ethical, and scientific concerns. Scientific studies in all disciplines are initiated with extensive planning and deliberation, often by a team of highly trained scientists. To assure that the quality, integrity, and feasibility of funded research projects meet the required standards, research-funding agencies such as the National Institute of Health and the National Science Foundation, pass proposed research plans through a rigorous peer review process before making funding decisions. Yet, some study proposals successfully pass through all the rigorous scrutiny of the scientific peer review process, but the proposed investigations end up being terminated before yielding results. This study demonstrates an algorithm that quantifies the risk associated with a study being terminated based on the analysis of patterns in the language used to describe the study prior to its implementation. To quantify the risk of termination, we use data from the clinicialTrials.gov repository, from which we extracted structured data that flagged study characteristics, and unstructured text data that described the study goals, objectives and methods in a standard narrative form. We propose an algorithm to extract distinctive words from this unstructured text data that are most frequently used to describe trials that were completed successfully vs. those that were terminated. Binary variables indicating the presence of these distinctive words in trial proposals are used as input in a random forest, along with standard structured data fields. In this paper, we demonstrate that this combined modeling approach yields robust predictive probabilities in terms of both sensitivity (0.56) and specificity (0.71), relative to a model that utilizes the structured data alone (sensitivity = 0.03, specificity = 0.97). These predictive probabilities can be applied to make judgements about a trial's feasibility using information that is available before any funding is granted.

[1]  L. Hooft,et al.  Premature trial discontinuation often not accurately reflected in registries: comparison of registry records with publications. , 2017, Journal of clinical epidemiology.

[2]  Tony Tse,et al.  Terminated Trials in the ClinicalTrials.gov Results Database: Evaluation of Availability of Primary Outcome Data and Reasons for Termination , 2015, PloS one.

[3]  R. Califf,et al.  Prevalence, characteristics, and predictors of early termination of cardiovascular clinical trials due to low recruitment: insights from the ClinicalTrials.gov registry. , 2014, American heart journal.

[4]  Keith Marsolo,et al.  Building Gold Standard Corpora for Medical Natural Language Processing Tasks , 2012, AMIA.

[5]  David Robinson,et al.  Text Mining with R: A Tidy Approach , 2017 .

[6]  Chung-Ho Hsieh,et al.  Novel solutions for an old disease: diagnosis of acute appendicitis with random forest, support vector machines, and artificial neural networks. , 2011, Surgery.

[7]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[8]  Ignacio Ferreira-González,et al.  Prevalence, characteristics, and publication of discontinued randomized trials. , 2014, JAMA.

[9]  Richard Smith,et al.  Peer Review: A Flawed Process at the Heart of Science and Journals , 2006, Journal of the Royal Society of Medicine.

[10]  Walter Daelemans,et al.  Counting trees in Random Forests: Predicting symptom severity in psychiatric intake reports. , 2017, Journal of biomedical informatics.

[11]  Brian B. Avants,et al.  The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS) , 2015, IEEE Transactions on Medical Imaging.

[12]  I. Baldi,et al.  Early termination of cardiovascular trials as a consequence of poor accrual: analysis of ClinicalTrials.gov 2006–2015 , 2017, BMJ Open.

[13]  Eric Fosler-Lussier,et al.  Textual inference for eligibility criteria resolution in clinical trials , 2015, J. Biomed. Informatics.

[14]  Arie Ben-David,et al.  Comparison of classification accuracy using Cohen's Weighted Kappa , 2008, Expert Syst. Appl..

[15]  D Demner-Fushman,et al.  Aspiring to Unintended Consequences of Natural Language Processing: A Review of Recent Developments in Clinical and Consumer-Generated Text Processing , 2016, Yearbook of Medical Informatics.

[16]  Rey-Long Liu,et al.  Medical query generation by term-category correlation , 2011, Inf. Process. Manag..

[17]  A. Doorenbos,et al.  Indications of Recruitment Challenges in Research with U.S. Military Service Members: A ClinicalTrials.gov Review. , 2017, Military medicine.

[18]  Fytton Rowland,et al.  The peer‐review process , 2002, Learn. Publ..

[19]  Manolis Tsiknakis,et al.  Semantic biomedical resource discovery: a Natural Language Processing framework , 2015, BMC Medical Informatics and Decision Making.

[20]  Sushma Jain,et al.  A survey towards an integration of big data analytics to big insights for value-creation , 2018, Inf. Process. Manag..

[21]  A. Jamjoom,et al.  Randomized controlled trials in neurosurgery: an observational analysis of trial discontinuation and publication outcome. , 2017, Journal of neurosurgery.