Original Article
Models developed by three techniques did not achieve acceptable prediction of binary trauma outcomes

https://doi.org/10.1016/j.jclinepi.2005.05.007Get rights and content

Abstract

Background and Objectives

To develop prediction models for outcomes following trauma that met prespecified performance criteria. To compare three methods of developing prediction models: logistic regression, classification trees, and artificial neural networks.

Methods

Models were developed using a 1996–2001 dataset from a major trauma center in Victoria, Australia. Developed models were subjected to external validation using the first year of data collection, 2001–2002, from a state-wide trauma registry for Victoria. Different authors developed models for each method. All authors were blinded to the validation dataset when developing models.

Results

Prediction models were developed for an intensive care unit stay following trauma (prevalence 23%) using information collected at the scene of the injury. None of the three methods gave a model that satisfied the performance criteria of sensitivity >80%, positive predictive value >50% in the validation dataset. Prediction models were also developed for death (prevalence 2.9%) using hospital-collected information. The performance criteria of sensitivity >95%, specificity >20% in the validation dataset were not satisfied by any model.

Conclusion

No statistical method of model development was optimal. Prespecified performance criteria provide useful guides to interpreting the performance of developed models.

Introduction

Interest in developing prognostic models for binary outcomes is widespread, and guidelines exist for their creation [1] and validation [2] as applied to health outcomes. A key element of model creation is the use of a large database containing variables that will be available if the new model is applied in routine practice. This creation step involves trade off between (1) complexity of model; the more complex the model, the more attuned it becomes to observed features of the database from which it is developed, and (2) transportability of the model, that is, how well it performs with different databases and when it is used in practice. Validation involves two critical elements. First, “external validation”: the performance of the newly created model must be tested on a second dataset that was not a part of model creation but that is representative of the same population [3], [4]. Second, the model must be validated in the context of prespecified utility scores for correct predictions and/or indicators that represent acceptable model performance when used in routine practice [2]. These indicators represent minimum performance thresholds for the model to be of practical use when applied.

Numerous statistical methods have been applied to create prognostic models for binary outcomes. Many comparisons between different methods exist [4], [5], [6], [7], [8], [9], [10], [11]; however, there is no consensus as to an optimal method. It is therefore prudent to explore different methods.

We present a case study of prognostic model development for binary outcomes that involves separate model-creation and validation datasets and prespecified performance indicators. Three statistical methods were employed for the creation of a prognostic model; logistic regression, classification trees, and neural networks. Logistic regression has become a mainstay of medical research; the latter two techniques are commonly used with data involving interactions and nonlinear relationships [12] and are long established [13], [14]. Each method was applied by a different investigator, blinded to the validation dataset when creating their models.

Section snippets

Case study

Data from the adult trauma database of the Royal Melbourne Hospital (RMH; a major metropolitan trauma service of Melbourne) for the period of January 1, 1996 to April 30, 2001, was used for developing models. A total of 4,014 blunt trauma cases were in the dataset; penetrating trauma cases were excluded because of their differing clinical presentation, management, and likely outcomes. To evaluate the performance of newly developed prognostic models, the Victorian State Trauma Registry (VSTR)

Logistic regression

Logistic regression is common in the analysis of medical data [15]. A statistical model is specified for the probability of an outcome event, Y, based on a function of predictor variables, X, including covariates such as age: probability of Y = exp()/[1 + exp()]. These models give a predicted probability of Y for any future individual based on their covariate values and the parameter vector, β. By assessing whether an individual's predicted probability is greater than or less than some

Prehospital prediction of ICU stay

Of the 4,014 blunt trauma cases, 1,324 (33%) had complete data for the creation of prognostic models for an ICU stay. The prevalence of ICU stay was 301/1,324 (23%). The final logistic regression model was

  • Xβ = −2.643

  • + 0.972∗I(GCS eye opening response = abnormal)

  • + 1.998∗I(GCS motor response = abnormal)

  • + 0.320∗I(Cause of injury = pedestrian)

  • + 0.941∗I(Cause of injury = motorcycle)

  • + 0.448∗I(Cause of injury = vehicle)

  • − 0.00826∗(SBP-126)

  • + 0.000198∗(SBP-126)2

  • + 0.0488∗(Respiratory rate—19)

  • +

Discussion

We pursued more in-depth analyses using logistic regression [44], [45] partly for convenience but also because this methodology has been accepted into mainstream use within trauma research, and our results show that its performance in our database was not inferior to the other two methods.

For ICU stay, differing prevalence in the RMH and VSTR datasets led to calibration problems for our developed models and was one reason for the models not satisfying the performance criteria. The cutoff on the

Conclusion

We did not find an optimal statistical method for the development of prognostic models for binary outcomes. All of our developed models failed to meet prespecified performance criteria. One possible reason for this is because we did not consider important predictor variables. Our databases provided access to most previously reported predictors of mortality and ICU stay. Variables not considered, but potentially able to be collected in the prehospital setting, for example, the presence of

Acknowledgment

The work presented in this article was performed with the assistance of funding from the Victorian Trauma Foundation.

References (48)

  • R.L. Lammers et al.

    Prediction of traumatic wound infection with a neural network-derived decision model

    Am J Emerg Med

    (2003)
  • C. Senkowski et al.

    Trauma scoring systems: a review

    J Am Coll Surg

    (1999)
  • F. Harrell et al.

    Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors

    Stat Med

    (1996)
  • D.G. Altman et al.

    What do we mean by validating a prognostic model?

    Stat Med

    (2000)
  • B. Ripley et al.

    Neural networks as statistical methods in survival analysis

  • H. Allore et al.

    A case study found that a regression tree outperformed multiple linear regression in predicting the relationship between impairments and social and preventive activities scores

    J Clin Epidemiol

    (2005)
  • B. Eftekhar et al.

    Comparison of artificial neural network and logistic regression models for prediction of mortality in head trauma based on initial clinical data

    BMC Med Inform Decision Making

    (2005)
  • D.C. Becalick et al.

    Comparison of artificial intelligence techniques with UKTRISS for estimating probability of survival after trauma: UK Trauma and Injury Severity Score

    J Trauma

    (2001)
  • M.C. Costanza et al.

    Binary classification of dyslipidemia from the waist-to-hip ratio and body mass index: a comparison of linear, logistic, and CART models

    BMC Med Res Methodol

    (2004)
  • T. Hastie et al.

    Elements of statistical learning: data mining, inference and prediction

    (2001)
  • W.A. Belson

    Matching and prediction on the principle of biological classification

    Appl Stat

    (1959)
  • F. Rosenblatt

    The perceptron: a probabilistic model of information storage and organization in the brain

    Psychol Rev

    (1958)
  • D.G. Altman

    Practical statistics for medical research

    (2000)
  • StataCorp

    Stata statistical software version 7

    (2001)
  • Cited by (0)

    View full text