Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Classification performance of administrative coding data for detection of invasive fungal infection in paediatric cancer patients

  • Jake C. Valentine ,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    Jake.Valentine@petermac.org

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, Paediatric Integrated Cancer Service, Royal Children’s Hospital, Parkville, Victoria, Australia, Sir Peter MacCallum Department of Oncology, University of Melbourne, Parkville, Victoria, Australia

  • Leon J. Worth,

    Roles Conceptualization, Investigation, Methodology, Supervision, Writing – review & editing

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, Sir Peter MacCallum Department of Oncology, University of Melbourne, Parkville, Victoria, Australia, Department of Infectious Diseases, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia

  • Karin M. Verspoor,

    Roles Methodology, Supervision, Visualization, Writing – review & editing

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, School of Computing and Information Systems, University of Melbourne, Parkville, Victoria, Australia

  • Lisa Hall,

    Roles Supervision, Writing – review & editing

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, School of Public Health, University of Queensland, Brisbane, Queensland, Australia

  • Daniel K. Yeoh,

    Roles Writing – review & editing

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, Department of Infectious Diseases, Perth Children’s Hospital, Perth, Western Australia, Australia

  • Karin A. Thursky,

    Roles Writing – review & editing

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, Sir Peter MacCallum Department of Oncology, University of Melbourne, Parkville, Victoria, Australia, Department of Infectious Diseases, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia

  • Julia E. Clark,

    Roles Data curation, Writing – review & editing

    Affiliation Infection Management Service, Queensland Children’s Hospital, Brisbane, Queensland, Australia

  • Gabrielle M. Haeusler

    Roles Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Visualization, Writing – review & editing

    Affiliations National Centre for Infections in Cancer, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, Paediatric Integrated Cancer Service, Royal Children’s Hospital, Parkville, Victoria, Australia, Sir Peter MacCallum Department of Oncology, University of Melbourne, Parkville, Victoria, Australia, Department of Infectious Diseases, Peter MacCallum Cancer Centre, Melbourne, Victoria, Australia, Infectious Diseases Unit, Department of General Medicine, Royal Children’s Hospital, Parkville, Victoria, Australia

Abstract

Background

Invasive fungal infection (IFI) detection requires application of complex case definitions by trained staff. Administrative coding data (ICD-10-AM) may provide a simplified method for IFI surveillance, but accuracy of case ascertainment in children with cancer is unknown.

Objective

To determine the classification performance of ICD-10-AM codes for detecting IFI using a gold-standard dataset (r-TERIFIC) of confirmed IFIs in paediatric cancer patients at a quaternary referral centre (Royal Children’s Hospital) in Victoria, Australia from 1st April 2004 to 31st December 2013.

Methods

ICD-10-AM codes denoting IFI in paediatric patients (<18-years) with haematologic or solid tumour malignancies were extracted from the Victorian Admitted Episodes Dataset and linked to the r-TERIFIC dataset. Sensitivity, positive predictive value (PPV) and the F1 scores of the ICD-10-AM codes were calculated.

Results

Of 1,671 evaluable patients, 113 (6.76%) had confirmed IFI diagnoses according to gold-standard criteria, while 114 (6.82%) cases were identified using the codes. Of the clinical IFI cases, 68 were in receipt of ≥1 ICD-10-AM code(s) for IFI, corresponding to an overall sensitivity, PPV and F1 score of 60%, respectively. Sensitivity was highest for proven IFI (77% [95% CI: 58–90]; F1 = 47%) and invasive candidiasis (83% [95% CI: 61–95]; F1 = 76%) and lowest for other/unspecified IFI (20% [95% CI: 5.05–72%]; F1 = 5.00%). The most frequent misclassification was coding of invasive aspergillosis as invasive candidiasis.

Conclusion

ICD-10-AM codes demonstrate moderate sensitivity and PPV to detect IFI in children with cancer. However, specific subsets of proven IFI and invasive candidiasis (codes B37.x) are more accurately coded.

Introduction

Invasive fungal infections (IFIs) represent significant challenges in the management of paediatric cancer patients with impaired immunity [13] and are an important cause of morbidity and mortality [1, 4]. Current methods for detecting IFI are manual, time consuming and often labour intensive [1, 2, 5, 6], and are reliant on a suite of clinical, laboratory and radiological data. There is therefore limited capacity to routinely capture IFIs to assess the epidemiology, detect potential outbreaks and inform optimal antifungal use in children with cancer [7].

Uniform case definitions for IFI are widely accepted as measurable outcomes in clinical trials (i.e. European Organization for Research and Treatment of Cancer/Invasive Fungal Infections Cooperative Group and the National Institute of Allergy and Infectious Diseases Mycoses Study Group [EORTC/MSG]) [8]. However, these are complex and require detailed case review. Administrative coding data possess potentially favourable attributes for simplified surveillance [9], including standardised classification and availability of specific codes for yeast and mould infections [4, 10]. In Australia, the International Statistical Classification of Diseases and Related Health Problems, Tenth Revision, Australian Modification (ICD-10-AM) are a monohierarchical, codified, medical lexicon used for coding inpatient diagnoses and is a commonly used ontology to inform activity-based funding models [11].

Earlier data have suggested the sensitivity of administrative coding data for classifying invasive aspergillosis to be moderate (63%) [12], but findings were restricted to filamentous fungi in adult allogeneic and autologous haematopoietic stem cell transplantation recipients and excluded invasive candidiasis, one of the most prevalent IFIs in the paediatric haematology-oncology setting [1]. Despite the high incidence and poor survival prognoses of IFI in cancer patients [13], there is a paucity of available evidence examining the utility of administrative coding data for reliable and reproducible surveillance of IFI in vulnerable paediatric cancer populations.

The objectives of this study were to: (i) determine the sensitivity, positive predictive value (PPV) and F1 score of administrative coding data for case ascertainment of IFI; and (ii) describe the misclassification rate of ICD-10-AM in paediatric haematology-oncology patients.

Materials and methods

Study design and population

This was a retrospective, single-site, cohort study of paediatric patients (<18-years) diagnosed with a haematological malignancy or solid tumour neoplasm between the 1st April 2004 and 31st December 2013 at the Royal Children’s Hospital (RCH) in Melbourne, Victoria, Australia. Study design was consistent with criteria endorsed in the STrengthening the Reporting of OBservational studies in Epidemiology (STROBE; S1 Table) [14] and the REporting of studies Conducted using Observational Routinely-collected health Data (RECORD; S2 Table) statements [15].

Gold-standard invasive fungal infection dataset

Data collected as part of the multisite The Epidemiology and Risk Factors for Invasive Fungal Infections in Immunocompromised Children (TERIFIC) study and restricted to episodes collected at the RCH (denoted as r-TERIFIC), were used for the current study [1, 2]. Detailed study methodology is available elsewhere [1, 2]. Briefly, this 10-year retrospective study identified all episodes of IFI in children with cancer or haematological malignancy from hospital microbiology, pharmacy-dispensing, radiology, oncology diagnostic and clinical management records as well as diagnostic coding data. Invasive fungal infection episodes were classified as proven, probable, possible or modified possible in accordance with EORTC/MSG criteria [8] and modifications described elsewhere [1, 2].

Administrative coding dataset

Episode-level, administrative coding data were sourced from the Victorian Admitted Episodes Dataset (VAED) and mapped to each patient record captured in the r-TERIFIC dataset. The VAED is Australia’s largest hospital morbidity database, and consists of diagnostic ICD-10-AM and procedural Australian Classification of Health Interventions (ACHI) codes for paediatric cancer patients admitted to private and public hospitals in Victoria [16].

Patients with haematological malignancy or a solid tumour were defined using the principal diagnosis codes denoting a primary malignant neoplasm (ICD-10-AM codes: C00.x - C76.x, C80.x, C81.x [0/1]—C88.x [0/1], C90.x [0/1]—C96.x [0/1], and D46), where “x” denotes any number (S3 Table). Invasive fungal infection was defined when an additional diagnosis code (Australian Coding Standards 0002 Additional diagnoses [17]) denoting IFI was reported in the VAED (ICD-10-AM codes: B37.x, B42.x - B50.x) (S3 Table). Hospitalisations for autologous or allogeneic haematopoietic stem cell transplantation were defined by corresponding ACHI codes 13706–00, -06, -07, -08, -09, -10 [802] (S4 Table). Updates to the ICD-10-AM and ACHI codes from the Third to Eighth Edition were elucidated. Duplicate IFI codes denoting the same IFI in the same hospitalisation, as well as those reported at the time of admission in subsequent hospitalisations, were considered the same IFI and were counted only once per patient. Multiple discrete IFI codes appearing in the same hospitalisation per patient were counted as separate IFI episodes. Accordingly, patients with ≥2 mutually exclusive gold-standard IFI diagnoses in the r-TERIFIC dataset, diagnosed in the same or in discrete hospitalisations, were counted as individual gold-standard cases for each IFI diagnosis (for example, one patient with both invasive aspergillosis and invasive candidiasis was counted as one case of invasive aspergillosis and one case of invasive candidiasis). Index hospitalisation was defined as the first admission date at the RCH.

Exclusion criteria

Cancer patients with superficial fungal infections (codes B36.x), including dermatophytes (codes B35.x), and patients with no underlying malignancy were excluded.

Statistical analyses

For patient and clinical characteristic data, normality was tested on histogram analysis and the skewness and kurtosis test [18]. The mean (±standard deviation) and median (interquartile range) were reported for parametric and non-parametrically distributed data, respectively.

Classification accuracy.

To determine the accuracy of ICD-10-AM codes for IFI case detection, sensitivity, PPV and F1 scores were calculated, stratified by IFI type, EORTC/MSG classification and underlying cancer diagnosis [19, 20]. Sensitivity and PPV of the coding data were calculated as the number of clinically-confirmed IFI patients in receipt of at least one IFI code (i.e. true positives; cases where the ICD-10-AM code agrees with the clinical label) divided by the total number of clinically-confirmed IFI cases in the r-TERIFIC dataset and the total number of patients assigned an ICD-10-AM code for IFI (code positives), respectively. Exact binomial 95% confidence intervals (CI) were calculated for all sensitivity and PPV calculations. The F1 score was used to measure the harmonic mean of the sensitivity and PPV of the coding data according to the formula [21]: (1) where F1 ranges in [0,1] = {F1:0≤F1≤1} and an F1 = 1 indicates perfect sensitivity and PPV. To identify which coding abstraction yields the highest sensitivity, PPV and F1 score within each combination of IFI codes, the union of different ICD-10-AM code sets for IFI (represented as Ak) was evaluated. The union of code sets A1 and A2, denoted A1A2, is equivalent to the set of patients in the r-TERIFIC dataset that are correctly assigned either code A1 (Pr(A1)) or code A2 (Pr(A2)) or codes A1 and A2 (Pr(A1A2)). Classification performance was determined according to increasing numbers of assigned code sets (Pr(A1A2 ∪…∪ Ak)). Sensitivity, PPV and F1 estimates of 0% indicate IFI code sets that were not assigned to true positive cases in the r-TERIFIC dataset, denoted A1′ ∩ A2′. The number of different combinations (C) of codes (n) in increasing set sizes (r) was determined according to the following formula: (2) Classification statistics are reported in accordance with the Standards for Reporting Diagnostic (STARD) accuracy studies statement [22] (S5 Table).

Misclassification rate.

Misclassification rate was calculated as a proportion of discordant-coded IFIs (e.g. the proportion of invasive candidiasis cases coded as invasive aspergillosis).

All statistical analyses were undertaken using Stata/SE v15.1 software (StataCorp® LLC, College Station, Texas, U.S.A.) A two-sided p value <0.05 was considered statistically significant.

Ethics

Ethics approval was granted by the Royal Children’s Hospital Human Research Ethics Committee (project number: 59636) and the need for informed consent was waived in accordance with the National Statement on Ethical Conduct in Human Research 2007 (Updated May 2015) [23].

Results

Study population

From 1st April 2004 to 31st December 2013, there were 1,671 paediatric cancer patients admitted to RCH according to the coding dataset (Fig 1). Of the 1,671 cancer patients, 114 (6.82%) were in receipt ≥1 ICD-10-AM code denoting IFI in the coding dataset and 113 (6.76%) fulfilled gold-standard definitions for IFI in the r-TERIFIC dataset. Sixty-eight of the 113 patients (60%) in the r-TERIFIC dataset were coded with ≥1 IFI (Fig 1; Table 1). Of the 45 patients in the r-TERIFIC dataset that did not receive an ICD-10-AM code for IFI, nine had invasive aspergillosis (20%), 11 invasive candidiasis (24%) and 4 other/unspecified IFI (8.89%). There were 46 false positive cases in the coding dataset that were not captured in the r-TERIFIC dataset (Table 1), of which 38 (83%) were coded as ‘candidiasis of other sites’ (ICD-10-AM code: B37.88).

thumbnail
Fig 1. Consort diagram of the study methodology and the number (N) of linked patient records across the r-TERIFIC and administrative coding datasets.

ICD-10-AM, International Statistical Classification of Diseases and Related Health Problems, Tenth Revision, Australian Modification; IFI, invasive fungal infection; r-TERIFIC, The Epidemiology and Risk Factors for Invasive Fungal Infections in Immunocompromised Children (Royal Children’s Hospital); VAED, Victorian Admitted Episodes Dataset.

https://doi.org/10.1371/journal.pone.0238889.g001

thumbnail
Table 1. Clinical agreement between the gold-standard and administratively-coded cases of invasive fungal infection in the study cohort.

https://doi.org/10.1371/journal.pone.0238889.t001

Baseline characteristics of the study cohort are presented in Table 2. Acute lymphoblastic leukaemia was the predominant underlying malignancy (n = 516; 31%), followed by neuroblastoma (n = 397; 24%) and acute myeloid leukaemia (n = 110; 6.58%) (Table 2). Of the 113 patients defined according to EORTC/MSG criteria, 46 (41%) were classified as possible, 30 (27%) proven, 26 probable (23%) and 11 (9.73%) modified possible or treated for IFI.

thumbnail
Table 2. Baseline characteristics of the study cohort, N = 1,671.

https://doi.org/10.1371/journal.pone.0238889.t002

Sensitivity, positive predictive value and F1 scores of the ICD-10-AM codes

Sensitivity, PPV and F1 scores of the ICD-10-AM codes are provided in Table 3. Sixty-eight of the 113 IFI patients in the r-TERIFIC dataset were in receipt of ≥1 ICD-10-AM code(s) for IFI, resulting in an overall sensitivity of 60% (95% CI: 51–69; 68/113 cases) (Table 3). Sixty-eight of the 114 IFI-coded patients in the ICD-10-AM coding dataset were identified in the r-TERIFIC dataset (PPV: 60% [95% CI: 50–69]), resulting in an F1 score of 60% (Table 3). After stratifying by type of IFI, invasive candidiasis codes resulted in the highest sensitivity, PPV and F1 score of 83% (95% CI: 61–95), 70% (95% CI: 58–81) and 76%, respectively, followed by invasive aspergillosis (sensitivity: 42% [95% CI: 20–67]; PPV: 32% [95% CI: 15–54]; F1 = 36%) (Table 3). After stratifying by underlying malignancy, sensitivity for all coded IFI was highest in patients with neuroblastoma (88% [95% CI: 47–99]; F1 = 52%) and acute lymphoblastic leukaemia (69% [95% CI: 55–80]; F1 = 68%) (Table 3), and PPV was highest in patients with acute myeloid leukaemia (76% [95% CI: 55–91]; F1 = 66%) (Table 3). Overall classification performance of the IFI codes was highest for proven EORTC/MSG (sensitivity, 77% [95% CI: 58–90]; PPV, 34% [95% CI: 23–46]; F1 = 47%) and possible EORTC/MSG IFI diagnoses (sensitivity, 61% [95% CI: 45–75]; PPV, 41% [95% CI: 29–54]; F1 = 49%) in the studied population (N = 1,671).

thumbnail
Table 3. Performance (in percent) of administrative coding data for case detection of proven, probable and possible invasive fungal infection.

https://doi.org/10.1371/journal.pone.0238889.t003

ICD-10-AM codes B44.2, B44.8 and B44.9 denoting invasive aspergillosis, codes B37.5 and B37.6 denoting invasive candidiasis, and codes B48.0-B48.7 and B49 denoting other/unspecified IFI returned F1 scores of 0%, yielding no improvement to the sensitivity and PPV (Table 4). The overall classification performance of the ICD-10-AM codes improved with larger sets of combined codes. Assignment of codes B44.0, B44.1 or B44.7, and B37.7 or B37.88, as a combined union set of codes, yielded the highest sensitivity, PPV and F1 scores for invasive aspergillosis and invasive candidiasis, respectively (Tables 3 and 4). Only code B48.8 returned a sensitivity, PPV and F1 estimate for the other/unspecific IFI category (Table 4).

thumbnail
Table 4. Performance (in percent) of different coding abstractions stratified by invasive fungal infection in the study cohort, N = 1,671.

https://doi.org/10.1371/journal.pone.0238889.t004

The sensitivity of B44.x codes denoting invasive aspergillosis decreased for r-TERIFIC patients in receipt of 2 codes (Pr(B44.0 ∩ B44.1) = 5.26% [95% CI: 0.13–26]; Pr(B44.0 ∩ B44.7) = 0%; Pr(B44.1 ∩ B44.7) = 0%) compared to assignment of ≥1 code (42% [95% CI: 20–67], Table 3). Likewise, the sensitivity of B37.x codes denoting invasive candidiasis decreased for r-TERIFIC patients in receipt of 2 codes (Pr(B37.7 ∩ B37.88) = 17% [95% CI: 4.95–39]) compared to assignment of ≥1 code (83% [95% CI: 61–95], Table 3).

Misclassification rate

Misclassification was greatest in patients with invasive aspergillosis coded as invasive candidiasis (n = 2; 13%). Of the 21 patients with invasive candidiasis, 2 (9.52%) were coded as invasive aspergillosis.

Discussion

This study is the first to describe the performance of administrative coding data to detect IFI in immunocompromised children with cancer. Overall sensitivity and PPV of ICD-10-AM codes for detection of clinically-confirmed IFI were moderate. However, sensitivity was improved for ascertainment of proven and possible IFI cases, in particular for invasive candidiasis, suggesting there is potential merit in using administrative coding data to signal medical record review for these discrete IFIs.

We found that ICD-10-AM codes alone were not sufficient to accurately classify IFI cases. In keeping with earlier estimates reported in Chang et al. [12], we observed an overall sensitivity and PPV of 60%, indicating that administrative coding data alone are not sufficient to reliably detect true cases of IFI in paediatric patients. The performance of coding data for IFI case detection was enhanced when subsets of proven IFI were examined, suggesting that where confirmatory laboratory results are available, then the quality of coding may be improved. Accuracy and completeness of medical record documentation likely contributes to this variation with one study showing that 97% of fungaemia cases were assigned an IFI code when fungaemia was explicitly documented in the medical record, as opposed to only 42% of cases when only microbiology results were used [24]. While underlying malignancy is important for evaluating IFI risk [4], our findings suggest that cancer diagnosis is less relevant to understanding the classification performance of administrative coding data for IFI.

In addition, we observed a difference in the performance of coding data for accurate detection of specific subsets of fungal infection. Cases of invasive candidiasis were more accurately coded than invasive aspergillosis cases. We propose that this may be related to readily available and simple diagnostic tests for yeast infections, in comparison to heterogenous diagnostic testing and the requirement for interpretation of imaging and laboratory results in order to identify mould infections. These factors could impact upon coding practices, particularly where microbiology, histology and radiological findings require integration by clinicians, with documentation in medical files, to facilitate accurate coding by clinical coders.

Cases of invasive aspergillosis were most frequently misclassified as invasive candidiasis in the coded data. Although the number of invasive aspergillosis cases in the gold-standard data were small (N = 15), our findings are likely indicative of the uncertainty in discriminating between yeast and mould infections at the clinical coding level. A recent qualitative study [25] identified clinical coders’ experience and awareness of IFI as a factor associated with discordant coding. Although it is a reasonable assumption that clinical coder experience is associated with our misclassification estimates, in the setting of IFI where clinical case definitions are complex, it is conceivable that other factors are at play. This includes the complexity of translating clinical data indicating invasive aspergillosis into ICD-10-AM [24, 26], the absence of clear definitions [27, 28], subjective interpretation of existing guidelines [24, 25, 27], delays in diagnosis [29], and the review of multiple data sources to make a confirmatory diagnosis of mould infection [1, 2, 30]. This setting underscores the importance of clear, complete, legible and standardised documentation of IFI to mitigate misclassification in current coding workflows.

We noted variation in classification performance according to specific code sets. Our results indicate that algorithms including the largest combination of specific IFI code sets yield the highest probability for case ascertainment in hospitalised paediatraic cancer patients. Notwithstanding, the fact that the F1 score for specific invasive aspergillosis code abstractions (B44.0 ∪ B44.1 ∪ B44.7) is still low-to-moderate (F1 = 36%, Tables 3 and 4) underscores that although these specific codes are the most sensitive starting point to signal medical chart review, existing coding rules are an unreliable indicator for invasive mould infections when used in isolation. Importantly, the sensitivity of ICD-10-AM decreases from 42% to 5.26% and 83% to 17% when comparing patients assigned one versus two codes denoting invasive aspergillosis and invasive candidiasis, respectively. Mathematically, the subset of true positive cases (numerator) diminishes as the number of assigned IFI codes increases (and the case definition therefore becomes more specific), whilst the number of gold-standard IFI cases (denominator) remains fixed. For example, true positives with 3 IFI codes is a subset of true positives with ≥2 IFI codes, which is a subset of true positives with ≥1 IFI code. Alternatively, {patients with 3 codes} ⊆ {≥2 codes} ⊆ {≥1 code}.

Methodological differences between the Australian Coding Standards and EORTC/MSG definitions likely contribute to our moderate overall F1 score of 60%. Clinical coders must adhere to rigid coding rules in accordance with Australian Coding Standards in the same way that clinicians adhere to complex and comprehensive criteria for IFI (i.e. EORTC/MSG), but these two sets of criteria may not directly match. This disconnect in clinical case definitions is a fundamental drawback in using ICD-10 codes as a reproducible proxy for IFI given cases detected according to clinical criteria may not reflect coded cases using ICD-10-AM. For example, clinical coders’ reliance on microbiology and histology records to identify cases of IFI in line with current coding rules can be subject to ascertainment bias in the coded data, given many IFIs are diagnosed according to a combination of metrics, namely clinical acumen, radiological findings and serum antigen testing [12, 26]. Notwithstanding, strategic imperatives to mitigate erroneous coding of IFI are likely two-fold. First, harmonisation of clinical EORTC/MSG definitions with existing Australian Coding Standards may help safeguard accurate detection of IFI in the coded data by reducing ascertainment of false positive cases (for example, our high number of false positive cases [N = 38] coded as ‘candidiasis of other sites’ [code B37.88]). In fact, recent qualitative research proposes the use of Systematized Nomenclature of Medicine–Clinical Terms (SNOMED-CT) codes in electronic health records as a more granular tool to standardise terminology and facilitate clinical coding of complex diseases [25, 31, 32]. Second, ensuring that chart documentation is complete, legible and streamlined will ensure clinical coders have sufficient access to the data required to assign the appropriate IFI code(s) [25, 33, 34].

Our high classification estimates for invasive candidiasis suggest that administrative coding data may be a feasible proxy to facilitate existing surveillance methods of yeast infection. Owing to the availability and easier interpretation of confirmatory diagnostic data indicating invasive candidiasis compared to invasive aspergillosis [35], the sensitivity and PPV of the administrative coding data are high (F1 = 76%). These findings substantiate potential merit in its use as a signal to trigger medical record review. Current surveillance of IFI is manual, onerous, time-consuming and resource-intensive [1, 2, 4, 12, 24]. However, use of ICD-10 codes as a feasibly available surrogate measure for invasive candidiasis may help restrict medical chart reviews to patients most likely presenting with yeast infection, therefore mitigating unnecessary record review. Our promising classification results also suggest there may be value in using ICD-10-AM codes for population-based monitoring of invasive candidiasis (codes B37.x) in paediatric populations.

Limitations of the current study include the fact that single-centre experience was evaluated, and findings may not reflect clinical coding performance and differences in other paediatric haematology-oncology units [36, 37]. Second, ICD-10 is an amalgamation of diagnostic information into a codified, monohierarchical, medical lexicon which does not discriminate between EORTC/MSG classifications, therefore rendering the data insufficient for fungal surveillance based on classification of proven/probable/possible IFI. Third, the wide 95% confidence intervals for our classification estimates (Table 3) are attributed to a small sample size of true positive cases stratified by type of IFI and EORTC/MSG criteria. Further, although the one-month average time lag [38] for hospital diagnoses to be coded make ICD-10-AM unsatisfactory for real-time IFI surveillance, our data indicate potential merit in using invasive candidiasis codes (B37.x) to signal retrospective detection of potentially missed cases.

Conclusions

In conclusion, we demonstrate moderate performance of ICD-10-AM codes for detection of IFI in children with cancer. Coding of invasive fungal infections having greater diagnostic certainty according to EORTC/MSG criteria (i.e. proven IFI), as well as yeast infections, resulted in higher sensitivity for case ascertainment. Findings suggest that while administrative coding data are not an accurate reflection of overall IFI disease burden, these data may provide an acceptable reflection of relative disease burden and signal a medical chart review for specific IFI categories (namely, proven/possible IFI and yeast infections) in paediatric patients with cancer. Future studies are required to assess the utility of ICD-10-AM data for these specific infections to detect changes in disease burden and longitudinally monitor quality improvement activities.

Supporting information

S1 Table. STROBE statement—checklist of items that should be included in reports of cohort studies.

https://doi.org/10.1371/journal.pone.0238889.s001

(PDF)

S2 Table. The RECORD statement—checklist of items, extended from the STROBE statement, that should be reported in observational studies using routinely collected health data.

https://doi.org/10.1371/journal.pone.0238889.s002

(PDF)

S3 Table. ICD-10-AM coding conventions for invasive fungal infection, haematological malignancy and solid tumour neoplasms.

https://doi.org/10.1371/journal.pone.0238889.s003

(PDF)

S4 Table. Australian Classification for Health Intervention codes denoting allogeneic and autologous haematopoietic stem cell transplantations.

https://doi.org/10.1371/journal.pone.0238889.s004

(PDF)

S5 Table. The STARD 2015 statement—checklist of essential items for reporting diagnostic accuracy studies.

https://doi.org/10.1371/journal.pone.0238889.s005

(PDF)

S6 Table. Performance (in percent) of administrative coding data for case detection of ‘modified possible’ and ‘treated for invasive fungal infection’ classifications.

https://doi.org/10.1371/journal.pone.0238889.s006

(PDF)

Acknowledgments

The authors thank Ms. Belinda Zambello of the Victorian Paediatric Integrated Cancer Service for preparing the administrative coding data from the Victorian Admitted Episodes Dataset for collection.

References

  1. 1. Bartlett AW, Cann MP, Yeoh DK, Bernard A, Ryan AL, Blyth CC, et al. Epidemiology of invasive fungal infections in immunocompromised children; an Australian national 10-year review. Pediatr Blood Cancer. 2019;66(4):e27564. Epub 2018/12/05. pmid:30511385.
  2. 2. Wang SS, Kotecha RS, Bernard A, Blyth CC, McMullan BJ, Cann MP, et al. Invasive fungal infections in children with acute lymphoblastic leukaemia: Results from four Australian centres, 2003–2013. Pediatr Blood Cancer. 2019;66(10):e27915. Epub 2019/07/17. pmid:31309711.
  3. 3. Sahbudak Bal Z, Yilmaz Karapinar D, Karadas N, Sen S, Onder Sivis Z, Akinci AB, et al. Proven and probable invasive fungal infections in children with acute lymphoblastic leukaemia: results from an university hospital, 2005–2013. Mycoses. 2015;58(4):225–32. pmid:25728069.
  4. 4. Valentine JC, Morrissey CO, Tacey MA, Liew D, Patil S, Peleg AY, et al. A population-based analysis of invasive fungal disease in haematology-oncology patients using data linkage of state-wide registries and administrative databases: 2005–2016. BMC Infect Dis. 2019;19(1):274. Epub 2019/03/23. pmid:30898090; PubMed Central PMCID: PMC6429824.
  5. 5. Kontoyiannis DP, Marr KA, Park BJ, Alexander BD, Anaissie EJ, Walsh TJ, et al. Prospective surveillance for invasive fungal infections in hematopoietic stem cell transplant recipients, 2001–2006: Overview of the Transplant-Associated Infection Surveillance Network (TRANSNET) Database. Clin Infect Dis. 2010;50(8):1091–100. pmid:20218877.
  6. 6. Valentine JC, Hall L, Verspoor KM, Worth LJ. The current scope of healthcare-associated infection surveillance activities in hospitalized immunocompromised patients: a systematic review. Int J Epidemiol. 2019;48(6):1768–82. Epub 2019/08/01. pmid:31363780.
  7. 7. Blyth CC, Haeusler GM, McMullan BJ, Kotecha RS, Slavin MA, Clark JE. Critical review of current clinical practice guidelines for antifungal therapy in paediatric haematology and oncology. Support Care Cancer. 2017;25(11):3289–90. Epub 2017/03/30. pmid:28353037.
  8. 8. De Pauw B, Walsh TJ, Donnelly JP, Stevens DA, Edwards JE, Calandra T, et al. Revised definitions of invasive fungal disease from the European Organization for Research and Treatment of Cancer/Invasive Fungal Infections Cooperative Group and the National Institute of Allergy and Infectious Diseases Mycoses Study Group (EORTC/MSG) Consensus Group. Clin Infect Dis. 2008;46(12):1813–21. Epub 2008/05/09. pmid:18462102; PubMed Central PMCID: PMC2671227.
  9. 9. Jhung MA, Banerjee SN. Administrative coding data and health care-associated infections. Clin Infect Dis. 2009;49(6):949–55. Epub 2009/08/12. pmid:19663692.
  10. 10. Valentine JC, Morrissey CO, Tacey MA, Liew D, Patil S, Ananda‐Rajah M. A population‐based analysis of attributable hospitalisation costs of invasive fungal diseases in haematological malignancy patients using data linkage of state‐wide registry and costing databases: 2009 ‐ 2015. Mycoses. 2020;63(2):162–71.
  11. 11. Heslop L. Activity-based funding for safety and quality: A policy discussion of issues and directions for nursing-focused health services outcomes research. Int J Nurs Pract. 2019;25(5):e12775. Epub 2019/08/16. pmid:31414554.
  12. 12. Chang DC, Burwell LA, Lyon GM, Pappas PG, Chiller TM, Wannemuehler KA, et al. Comparison of the use of administrative data and an active system for surveillance of invasive aspergillosis. Infect Control Hosp Epidemiol. 2008;29(1):25–30. pmid:18171183.
  13. 13. Valentine JC, Hall L, Spelman T, Verspoor KM, Seymour JF, Rischin D, et al. Burden and clinical outcomes of hospital-coded infections in patients with cancer: an 11-year longitudinal cohort study at an Australian cancer centre. Support Care Cancer. 2020:
  14. 14. Connell L, MacDonald R, McBride T, Peiperl L, Ross A, Simpson P, et al. Observational studies: getting clear about transparency. PLoS Med. 2014;11(8):e1001711. Epub 2014/08/27. pmid:25158064; PubMed Central PMCID: PMC4144975.
  15. 15. Benchimol EI, Smeeth L, Guttmann A, Harron K, Moher D, Petersen I, et al. The REporting of studies Conducted using Observational Routinely-collected Data (RECORD) Statement. PLoS Med. 2015;12(10):e1001885.
  16. 16. Victorian paediatric oncology care pathways: Providing optimal care for children and adolscents—Acute leukaemia, central nervous system tumours and solid tumours. Melbourne, Australia: Paediatric Integrated Cancer Service; 2019. p. 1–116.
  17. 17. Victorian Additions to the Australian Coding Standards. Melbourne, Victoria, Australia: Victorian Agency for Health Information, State Government of Victoria; 2019. p. 1–9.
  18. 18. Gould WW, Rogers WH. sg3.4: Summary of tests of normality. Stata Technical Bulletin. 1991;1(3):106–10.
  19. 19. Düntsch I, Gediga G. Confusion marices and rough set data analysis. J Phys Conf. 2019;1229(1):012055.
  20. 20. Becker SO, Caliendo M. Sensitivity analysis for average treatment effects. Stata J. 2007;7(1):71–83.
  21. 21. El Kafrawy P, Mausad A, Esmail H. Experimental comparison of methods for multi-label classification in different application domains. Int J Comput Appl. 2015;114(19):1–9.
  22. 22. Bossuyt PM, Reitsma JB, Bruns DE, Gatsonis CA, Glasziou PP, Irwig L, et al. STARD 2015: an updated list of essential items for reporting diagnostic accuracy studies. BMJ. 2015;351:h5527. Epub 2015/10/30. pmid:26511519; PubMed Central PMCID: PMC4623764.
  23. 23. National Statement on Ethical Conduct in Human Research 2007 (Updated May 2015). Canberra, Australian Capital Territory, Australia: National Health and Medical Research Council, Australian Research Council and the Australian Vice-Chancellor’s Committee. Commonwealth of Australia; 2007.
  24. 24. Nguyen Q, Reid B. A comparison of two sources of data on fungaemia in two hospitals. Health Inf Manag. 2005;34(3):88–93.
  25. 25. Doktorchik C, Lu M, Quan H, Ringham C, Eastwood C. A qualitative evaluation of clinically coded data quality from health information manager perspectives. Health Inf Manag. 2020;49(1):19–27. Epub 2019/07/10. pmid:31284769.
  26. 26. Donnelly JP, Chen SC, Kauffman CA, Steinbach WJ, Baddley JW, Verweij PE, et al. Revision and Update of the Consensus Definitions of Invasive Fungal Disease From the European Organization for Research and Treatment of Cancer and the Mycoses Study Group Education and Research Consortium. Clin Infect Dis. 2019: Epub 2019/12/06. pmid:31802125.
  27. 27. Tsitsikas DA, Morin A, Araf S, Murtagh B, Johnson G, Vinnicombe S, et al. Impact of the revised (2008) EORTC/MSG definitions for invasive fungal disease on the rates of diagnosis of invasive aspergillosis. Med Mycol. 2012;50(5):538–42. Epub 2011/11/15. pmid:22074309.
  28. 28. Valentine JC, Thursky KA, Worth LJ. Sepsis in cancer: a question of definition. Aust N Z J Public Health. 2020;44(3):245.
  29. 29. Ruhnke M, Böhme A, Buchheidt D, Cornely O, Donhuijsen K, Einsele H, et al. Diagnosis of invasive fungal infections in hematology and oncology-guidelines from the Infectious Diseases Working Party in Haematology and Oncology of the German Society for Haematology and Oncology (AGIHO). Ann Oncol. 2012;23(4):823–33. Epub 2011/09/29. pmid:21948809.
  30. 30. Hage CA, Carmona EM, Epelbaum O, Evans SE, Gabe LM, Haydour Q, et al. Erratum: Microbiological laboratory testing in the diagnosis of fungal infections in pulmonary and critical care practice. An official American Thoracic Society Clinical Practice Guideline. Am J Respir Crit Care Med. 2019;200(10):535–50. Epub 2019/11/16. pmid:31729907; PubMed Central PMCID: PMC6945799.
  31. 31. Alonso V, Santos JV, Pinto M, Ferreira J, Lema I, Lopes F, et al. Health records as the basis of clinical coding: Is the quality adequate? A qualitative study of medical coders' perceptions. Health Inf Manag. 2020;49(1):28–37. Epub 2019/02/13. pmid:30744403.
  32. 32. Thun S, Dewenter H. ICD-11, ICHI und SNOMED CT—was bedeuten die Systematiken für E-Health-Anwendungen? Bundesgesundheitsblatt Gesundheitsforschung Gesundheitsschutz. 2018;61(7):812–20. Epub 2018/05/31. pmid:29846742.
  33. 33. Shepheard J. What do we really want from clinical documentation improvement programs? Health Inf Manag. 2018;47(1):3–5. Epub 2017/09/28. pmid:28952382.
  34. 34. Hay P, Wilton K, Barker J, Mortley J, Cumerlato M. The importance of clinical documentation improvement for Australian hospitals. Health Inf Manag. 2020;49(1):69–73. Epub 2019/07/16. pmid:31303061.
  35. 35. Arvanitis M, Anagnostou T, Fuchs BB, Caliendo AM, Mylonakis E. Molecular and nonmolecular diagnostic methods for invasive fungal infections. Clin Microbiol Rev. 2014;27(3):490–526. Epub 2014/07/02. pmid:24982319; PubMed Central PMCID: PMC4135902.
  36. 36. Assareh H, Achat HM, Stubbs JM, Guevarra VM, Hill K. Incidence and variation of discrepancies in recording chronic conditions in Australian hospital administrative data. PLoS One. 2016;11(1):e0147087. Epub 2016/01/26. pmid:26808428; PubMed Central PMCID: PMC4726608.
  37. 37. Ridder HG, Doege V, Martini S. Differences in the implementation of diagnosis-related groups across clinical departments: a German hospital case study. Health Serv Res. 2007;42(6):2120–39. Epub 2007/11/13. pmid:17995556; PubMed Central PMCID: PMC2151394.
  38. 38. Sundararajan V, Romano PS, Quan H, Burnand B, Drösler SE, Brien S, et al. Capturing diagnosis-timing in ICD-coded hospital data: recommendations from the WHO ICD-11 topic advisory group on quality and safety. Int J Qual Health Care. 2015;27(4):328–33. Epub 2015/06/06. pmid:26045514; PubMed Central PMCID: PMC4521616.