<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Predicting unplanned hospital readmissions</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Martin H RENNHACKKAMP</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Assoc Prof Graeme K HART</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chief Data Scientist</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>PBT Australia</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Director Dept Intensive Care</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Austin Health</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>The number of unplanned readmissions is an important quality improvement indicator for Australian hospitals, as it has a direct impact on costs and patient outcomes. It is a useful indicator for patient multi-morbidity to identify patients that should potentially have different treatment or discharge plans. The objective of the Health Insights Challenge project was to develop a predictive model to determine the likelihood of an unplanned hospital readmission. The outcome was a validated analytical model that predicts the likelihood of an unplanned hospital readmission occurring within 30 days after a given in-patient episode. An 'all-cause' model does not pay attention to any specific disease or diagnosis groups. The predicted likelihood of an unplanned readmission occurring can assist care workers and clinicians in planning treatment and discharge processes.</p>
      </abstract>
      <kwd-group>
        <kwd />
        <kwd>Patient readmission</kwd>
        <kwd>patient admission</kwd>
        <kwd>patient discharge</kwd>
        <kwd>multimobidity</kwd>
        <kwd>inpatient</kwd>
        <kwd>episode of care</kwd>
        <kwd>predictive analytics</kwd>
        <kwd>hospital analytics</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>1. Background</title>
      <p>
        Austin Health is a public provider of tertiary health services, education and research in Melbourne.
It operates 980 beds across acute, sub-acute and mental health facilities [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. During the 2014
financial year (1 July – 30 June), Austin Health had 95,142 inpatient admissions, 177,027
outpatient attendances and 75,366 emergency presentations respectively.
      </p>
      <sec id="sec-2-1">
        <title>1.1. Problem definition</title>
        <p>Austin Health’s objectives for predicting the likelihood of 30-day unplanned readmissions were to:




identify the most likely patient groups for 30-day unplanned readmissions;
identify the drivers of 30-day unplanned hospital readmission;
redesign the admission, treatment and discharge planning processes; and
predict the impact of avoidable bed days on costs, throughput and waiting lists.</p>
        <p>Austin Health uses the concept of a spell, which is a continuous series of inpatient episodes,
where episodes may be separated by statistical separations to account for different cost centres.</p>
        <p>
          Unplanned readmissions are measured in Victoria as a hospital outcome indicator by the
Australian Commission on Safety and Quality in Health Care [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. Austin Health defines an
unplanned readmission based on the Dr Foster business rules [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. It occurs when a patient with an
emergency admission at a hospital on or before the 30th day after discharge is admitted to acute
care where either the principal diagnosis for the readmission is related to the previous admission,
or the previous admission has a surgical diagnosis code for any episode. Regular attenders (e.g.
haemodialysis, chemotherapy, radiotherapy and mental health) are not considered unplanned
readmissions.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>1.2. Literature review</title>
        <p>
          Preliminary research identified 294 relevant papers on 30-day all-cause readmissions. Five
relevant papers, clearly identifying the variables, the type of model and the model validation
approach, were used as reference case studies, as summarised in table 1 [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ],[6],[7],[8],[9]. They
used mostly administrative data, however, they were performed on different datasets of varying
numbers of patients and episodes, and over different time periods.
        </p>
        <p>The variables used in these models were compared to the data from Austin Health and a large
correspondence gave confidence that Austin Health’s dataset was suitable for this project.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>2. Methodology</title>
      <p>The project followed the SEMMA (Sample, Explore, Modify, Model and Assess) framework
developed by SAS Institute [11].</p>
      <sec id="sec-3-1">
        <title>2.1. Data preparation</title>
        <p>Austin Health provided de-identified data for a 5-year period (2009 to 2014) depicted in Table
Table 2: Data Extracts (1 July 2009 to 31 October 2014):</p>
        <p>No.</p>
        <p>1
2
3
4</p>
        <p>Activity Data Tables
Emergency Department
Inpatient Spells
Inpatient Episodes
Inpatient Comorbidities</p>
        <sec id="sec-3-1-1">
          <title>The following data preparation steps were performed:</title>
          <p>Episodes and spells were joined to allocate the spell attributes across the episodes.
Episodes resulting from unplanned readmissions were linked to the indexed episodes, to
enable the calculation of impact in terms of bed days and direct hospitalisation costs.
Lookup tables were joined in to enable analysis and reporting on descriptive attributes.
The number of Emergency Department presentations since the previous inpatient episode
were determined and added as a new field on each inpatient episode record.</p>
          <p>
            A binary target variable for model training was added to indicate the episodes resulting in
30-day unplanned readmissions according to the Dr Foster’s business rules [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ].
          </p>
          <p>This resulted in a dataset comprising 141 variables and 280,078 episodes, with the proportion
of unplanned readmissions (4.27%) to non-readmitting episodes (95.73%) highly skewed.







</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Data exploration</title>
        <p>An extensive data exploration exercise was performed, the results of which will be published
elsewhere. The most important findings were:</p>
        <p>A majority of the patients having unplanned readmissions were between 50 – 89 years old,
were diagnosed with chronic obstructive pulmonary disease (COPD), heart failure,
ascites and other diagnoses affecting the heart, lungs and liver, and they had many and
frequent inpatient episodes as well as many presentations at the emergency department.
Many of the patients having unplanned readmissions had high Charlson scores, which is a
measure of comorbidity [10].</p>
        <p>Many anecdotal indicators for unplanned readmissions were investigated, such as care at
home, hospital in the home days, method of transport, interpreter required and so on, but
the data exploration did not surface any positive evidence that these were relevant.</p>
      </sec>
      <sec id="sec-3-3">
        <title>2.3. Predictive modelling</title>
        <p>SAS Enterprise Miner was used to develop the models [11]. In order to provide opportunities for
intervention, the models were trained and evaluated on the data available at admission. Stepwise,
Forward and Main Effects regression, and one Decision Tree models were built. Using lift and the
Receiver Operator Characteristic (ROC) index of the validation dataset as selection criteria, the
Forward Regression model was the best performing model. It was named the Predictive model for
All-Cause unplanned Readmissions (PACR).</p>
        <p>Twenty variables were significant in the model, as illustrated in Table 3, calculated using a
decision tree algorithm [12]. The top three predictors were counts of emergency presentations,
inhospital spells per lifetime and the accumulated length of stay over 5 years. Other important
predictors include the admission ward, the Charlson Score and the admission unit.</p>
        <p>Age
Renal Disease Flag at Admission
Mild Liver Disease Flag at Admission
Malignancy Disease Flag at Admission
COPD Disease Flag at Admission
Moderate – Severe Liver Disease Flag at Admission
No of Spells
Admission Source
Care Type
Cerebrovascular Disease Flag at Admission
Non Chronic Disease Flag at Admission</p>
        <p>Dementia Disease Flag at Admission</p>
      </sec>
      <sec id="sec-3-4">
        <title>2.4. Model assessment</title>
        <p>Episode
Episode
Episode
Episode
Episode
Episode
5 years
Episode
Episode
Episode
Episode
Episode
The gains chart shows the percentiles of the patient episodes, ranked by their likelihood to have an
unplanned readmission. As depicted in Figure 2, the Cumulative Captured Response Rate depicts
the number of unplanned readmissions that were correctly identified at each percentile. At the 35th
percentile, almost 75% of predicted readmissions were identified. Comparatively, a model with no
predictive power (i.e. a random model) would be expected to predict approximately 5% of events
in each percentile. The cumulative lift at the 5th percentile shows the model is four times more
likely to correctly predict unplanned readmissions, as compared to random chance.</p>
        <p>
          For the models identified in the literature review studies, only the ROC statistic was published
to give an indication of model performance [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ],[6],[7],[8] and [9]. The PACR model came in
second when benchmarked against this metric (Table 4). Note that this is only an indicative
comparison; as the models were not evaluated against the same dataset, nor for the same period.
        </p>
        <sec id="sec-3-4-1">
          <title>Predicting 30-day all-cause hospital readmission [5]</title>
          <p>PACR model
A Public-Private Partnership Develops and Externally Validates a 30-Day
Hospital Readmission Risk Predictive Model [8]
Potentially Avoidable 30-Day Hospital Readmissions in Medical Patients [7]
Risk Factors for All-Cause Hospital Readmission Within 30 Days of Hospital
Discharge [9]
Development of a predictive model to identify inpatients at risk of
readmission within 30 days of discharge (PARR-30) [6]</p>
        </sec>
      </sec>
      <sec id="sec-3-5">
        <title>2.5. External validation</title>
        <p>ROC
0.80
0.77
0.76
0.71
0.70</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Predictive ability</title>
      <p>High
Moderate
Moderate</p>
      <sec id="sec-4-1">
        <title>Moderate Moderate 0.67 Low</title>
        <p>To avoid bias and to ascertain the model’s discriminative ability in an actual operational setting,
PACR was tested on an out-of-sample dataset, which consisted of data for the first eight months of
the 2015 financial year. It consisted of 41,666 inpatient episode records, with their corresponding
emergency department presentations. As depicted in Table 5, the model had a marginally higher
specificity (74%), as compared to its sensitivity (70%) for this dataset.</p>
        <sec id="sec-4-1-1">
          <title>2.6. Model deployment</title>
          <p>Description
Probability of identifying patients at risk of unplanned readmissions within
30 days of discharge (event)
Probability of identifying patients not at risk of unplanned readmissions
within 30 days of discharge (non-event)
Accurate classification of ‘events’ and ‘non-events’ as a % of total episodes</p>
          <p>Misclassification rate
Patients predicted to have unplanned readmissions were categorised using a risk stratification
indicator. The top 1/6 were scored as highly likely, the next 1/3 as medium likely and the
remaining 1/2 with a low likelihood, based on their predicted likelihood, as illustrated in Figure 3.</p>
          <p>Using this stratification, clinicians can prioritise their attention on the predicted episodes.
Error! Reference source not found. illustrates the impact of paying attention to each episode
predicted to be succeeded by an unplanned readmission in the 8-month validation dataset, using
the likelihood stratification introduced in Figure 3. Between 955 patients highly likely to have
unplanned readmissions, they had a total of 1747 episodes over the 8 month period. By focusing
on an average of 7 cases per day, the highly likely to readmit patients can all be considered for
specialised treatment or planning. Should these interventions be successful, 5,509 bed-days of
resulting readmissions could potentially be saved (shown as the “readmission LOS days” in figure
4). This represents a theoretical potential saving as not all such interventions would necessarily be
successful.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>3. Discussion</title>
      <p>A sensitivity score of 70% indicates that the model has a relatively good discriminative ability for
predicting unplanned readmissions – even when only using administrative data.</p>
      <p>The inclusion of Charlson comorbidity score at admission is an interesting point. As an
indicator of disease complexity, it significantly contributes to the model.</p>
      <sec id="sec-5-1">
        <title>3.1. Alignment with objectives</title>
        <p>The most likely patients for 30-day unplanned readmissions were identified through data
exploration.</p>
        <p>The most likely to readmit episodes were identified through the significant variables of the
model. The model outputs indicate which values or value bands are most significant [12].</p>
        <p>Discussions are underway with Austin Health regarding field trials to utilise the model outputs
to redesign admission, treatment and discharge planning processes.</p>
        <p>The number of potentially avoidable bed days and their direct hospital costs were shown.
These form the basis of the business case to justify the redesign of treatment and discharge
processes. A quick simulation study also showed the impact on throughput and waiting lists.</p>
      </sec>
      <sec id="sec-5-2">
        <title>3.2. Model’s potential impact</title>
        <p>Unplanned readmission is a key quality improvement indicator. Managing highly likely unplanned
readmissions can reduce their related costs; increase the capacity for new patients needing care;
streamline processes to improve operations; provide higher and more targeted quality of care for
patients requiring care; and indirectly improve the hospital’s reputation in the industry.</p>
      </sec>
      <sec id="sec-5-3">
        <title>3.3. Recommended future work</title>
        <p>Validating the model on an out-of-sample dataset has given good insight on the model’s predictive
ability and its potential benefits in an operational setting. The next step would be to validate the
effectiveness of the model through an on-site field trial at the Austin Health hospitals.</p>
        <p>Prediction models that use the Charlson comorbidity index need to be compared to models
that do not, as applied to the same datasets. Many Australian hospitals do not use Charlson scores.</p>
        <p>Although 30 days is an accepted metric in Victoria, some hospitals and health insurers
elsewhere use 28 days. The impact of the length of the interval before an unplanned readmission
on the model’s outcomes should also be investigated. Readmissions that occurred just outside the
30 day cut-off are also “medically and financially speaking” just as worthy of intervention.</p>
        <p>A thorough simulation study will better show the impact that reducing unplanned
readmissions have on patient throughput and waiting times.</p>
        <p>Additional improvements could further boost the model’s discriminative ability, for example,
by incorporating additional clinical and pharmacy variables.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>4. Acknowledgements</title>
      <p>Thank you to executive sponsors Mark Petty, Fiona Webster, and analysts Ray Robbins and
Ronald Ma of Austin Health. Also thanking PBT Australia, Deakin University (Ali Tammadoni),
the Australian Centre for Health innovation and EntityThree for collaborating on this project.</p>
      <p>Special mention is due to the review board: Dr. Chris Bain, currently Chief Health
Information Services Manager at Mercy Public Hospitals; Shmuley Goldberg, former Director of
Corporate Development for Victorian Comprehensive Cancer Centre (VCCC), Robert Heyes,
Workforce Analyst at the Victorian Department of Health and Professor Damminda Alahakoon,
currently Professor in Business Analytics at La Trobe University.
[6] J. Billings, I. Blunt, A. Steventon, T. Georghiou, G. Lewis and M. Bardsley, Development of a predictive model to
identify inpatients at risk of re-admission within 30 days of discharge (PARR-30), BMJ Open, 2,4 (2012).
[7] J. Donzé, D. Aujesky, D, Williams and J.L. Schnipper, Potentially avoidable 30-day hospital readmissions in medical
patients: derivation and validation of a prediction model, JAMA Internal Medicine, 173,8 (2013) 632-638.
[8] S.A. Choudhry, J. Li, D. Davi, C. Erdmann, R. Sikka and B. Sutariya, A public-private partnership develops and
externally validates a 30-day hospital readmission risk prediction model, Online Journal Of Public Health Informatics,
5,2 (2013) 219.
[9] L.K. Robin, D.H. Harlen, W.M. Richard, F.E. Matthew, S.W. Douglas and R.M. David, Risk Factors for All-Cause
Hospital Readmission Within 30 Days of Hospital Discharge, Journal of Clinical Outcomes Management, 20,5
(2013) 203-214.
[10] M.E. Charlson, P. Pompei, K.L. Ales and C.R. MacKenzie, A new method of classifying prognostic comorbidity in
longitudinal studies: Development and validation, Journal of Chronic Diseases, 40,5 (1987) 373–383.
[11] SAS Institute Inc., Data Mining Using SAS® Enterprise MinerTM: A Case Study Approach, Third Edition, SAS</p>
      <p>Institute Inc. Cary NC, 2013.
[12] M. Maldonado, Tip: How to Interpret SAS Rapid Predictive Modeler Results, [Online] Available at:
https://communities.sas.com/docs/DOC-6813, 2014.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Austin</given-names>
            <surname>Health</surname>
          </string-name>
          <string-name>
            <surname>Victoria</surname>
          </string-name>
          , About Us, [Online] Available at: http://www.austin.org.au/about-us/,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Health</given-names>
            <surname>Information Management Association Australia</surname>
          </string-name>
          , FAQ - What is ICD-10
          <string-name>
            <surname>-AM</surname>
          </string-name>
          , ACHI and ACS?, [Online] Available at: www.himaa2.org.au/education/?q=node/81,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>State</given-names>
            <surname>Government</surname>
          </string-name>
          of Victoria,
          <source>Victorian Health Service Performance Framework</source>
          <year>2014</year>
          -
          <volume>15</volume>
          , [Online] Available at: http://docs.health.vic.gov.au/docs/doc/184F7ACEBB1167FECA257D3800093A93/$FILE/1408007_VPMF%20web2 .pdf.,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Dr</given-names>
            <surname>Foster</surname>
          </string-name>
          <string-name>
            <surname>Intelligence</surname>
          </string-name>
          ,
          <source>Quality Investigator Victoria Business Rules, March</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Shulan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gao</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Moore</surname>
          </string-name>
          ,
          <article-title>Predicting 30-day all-cause hospital readmissions</article-title>
          ,
          <source>Health Care Management Science</source>
          ,
          <volume>16</volume>
          ,
          <issue>2</issue>
          (
          <year>2013</year>
          ),
          <fpage>167</fpage>
          -
          <lpage>175</lpage>
          , Springer Science + Business Media, New York,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>