Abstract
A hospital readmission is when a patient who was discharged from the hospital is admitted again for the same or related care within a certain period. Hospital readmissions are a significant problem in the healthcare domain, as they lead to increased hospitalization costs, decreased patient satisfaction, and increased risk of adverse outcomes such as infections, medication errors, and even death. The problem of hospital readmissions is particularly acute in intensive care units (ICUs), due to the severity of the patients’ conditions, and the substantial risk of complications. Predicting Unplanned Readmissions in ICUs is a challenging task, as it involves analyzing different data modalities, such as static data, unstructured free text, sequences of diagnoses and procedures, and multivariate time-series. Here, we investigate the effectiveness of each data modality separately, then alongside with others, using state-of-the-art machine learning approaches in time-series analysis and natural language processing. Using our evaluation process, we are able to determine the contribution of each data modality, and for the first time in the context of readmission, establish a hierarchy of their predictive value. Additionally, we demonstrate the impact of Temporal Abstractions in enhancing the performance of time-series approaches to readmission prediction. Due to conflicting definitions in the literature, we also provide a clear definition of the term Unplanned Readmission to enhance reproducibility and consistency of future research and to prevent any potential misunderstandings that could result from diverse interpretations of the term. Our experimental results on a large benchmark clinical data set show that Discharge Notes written by physicians, have better capabilities for readmission prediction than all other modalities.
Similar content being viewed by others
Introduction
The process of discharging patients from hospitals is complicated and presents many difficulties, with more than 35 million hospital discharges taking place each year in the United States alone1. One such difficulty is hospital readmission. Hospital readmission refers to the occurrence of a patient being admitted to the hospital again for the same or related care within a certain period, often within 30 days, after being discharged. Hospital readmissions cause a significant financial burden, an avoidable waste of medical resources, and can also put a strain on the healthcare system2. The cost of unplanned readmissions in the United States is estimated at 15 to 20 billion dollars annually3.
The creation of the Hospital Readmissions Reduction Program (HRRP) in 2010 by the Affordable Care Act (ACA) aimed to tackle the problem of unplanned hospital readmissions. The program imposes penalties on general acute-care hospitals whose rates of readmissions within 30 days exceed a maximal allowed readmission rate per clinical category. According to data released by the Centers for Medicare and Medicaid Services (CMS), since the beginning of this program, 93% of the 3,139 general acute hospitals subject to HRRP evaluation have been penalized at least once, with a total of 2.5 billion dollars of penalties for readmissions4.
When focusing on intensive care unit (ICU) readmissions, severe health consequences are added to the economic implications. On average, about 16% of unplanned ICU readmissions occur within 30 days after the initial hospital discharge5. Patients who are readmitted to the ICU have a significantly higher mortality rate (ranging from 21% to 40%) compared to those who are not readmitted (with a mortality rate ranging from 3.6 to 8.4%)6.
Reducing ICU readmission rates is an important goal for healthcare providers and policymakers, as it can improve patient outcomes, reduce length of stay that ultimately effects the availability of ICU beds, and reduce healthcare costs. To achieve this goal, hospitals need to identify patients with a high risk of ICU readmission. An accurate decision support system can assist physicians in identifying such patients, therefore reducing their risk for readmissions by prolonging their ICU stay, improving their follow-up care, or providing a comprehensive discharge plan.
The usual and expected candidates for readmission are patients that were hospitalized within the past year7, who may also suffer from chronic conditions, and thus are potentially easier to identify by physicians. As the aim of predicting unplanned readmissions is to detect instances of unexpected and unforeseen readmissions, it is important to identify and exclude such patients from the analysis, to avoid mixing expected and unplanned readmissions. Throughout the rest of this paper we use readmission synonymously with this excluding definition of unplanned readmissions.
Every ICU stay involves a vast amount of heterogeneous data: (1) static, such as patients demographics, (2) sequential, such as diagnoses and prescriptions, (3) unstructured, high-dimensional and sparse information such as clinical discharge notes, and (4) dynamic multivariate time-series data, such as lab tests and near bedside measurements, with problems common in clinical medicine, of missing data, different sampling frequencies, and random noise. The multimodal nature of the data makes predicting readmission a difficult task.
In this paper, we investigate each data modality, and its influence on unplanned ICU readmission prediction. Figure 1 illustrates our proposed approach. It is comprised of four building blocks that represent every data modality of input that is being processed using three communicative models: (A) raw, time-stamped multivariate temporal data collected by various sensors, which are abstracted into several interval-based abstract concepts, each belonging to one of two abstraction types, as we explain in “Time-series and temporal abstraction” section; (B) diagnoses and procedures performed on patients, represented as sequences of relevant ICD-9 (International Classification of Diseases, Ninth Revision) codes (B1), or ICD-9 textual descriptions (B2); (C) discharge notes, which are unstructured free text, written by physicians when releasing patients from the ICU; and (D) Static data on patients that includes their demographics, insurance type, etc. Besides (D), every data modality is examined separately, and in combination with others, as illustrated at the bottom of Fig. 1.
Our contributions in this study are as follows: (1) Demonstrating the effectiveness of each data modality separately, and every combination of them, on the prediction results. To the best of our knowledge, this is the first time multimodality analysis is being conducted in the task of ICU readmission prediction. (2) Combining different temporal-abstraction types in addition to domain knowledge, to better capture the temporal behavior of the data, thus improving the performance of temporal modalities. (3) Redefining the Unplanned Readmission problem, and creating a specific subset of patients to use in future works, enhancing reproducibility and cross study comparisons.
The rest of the paper is structured as follows. The next section provides background on time-series, temporal abstraction and on natural language processing. “Related work” section reviews related work. “Data collection” section describes the process of creating the dataset used in the evaluation. “Evaluation methods” section presents an experimental evaluation and all methods involved, in the ICU domain. “Results” section provides the evaluation’s results. “Discussion and conclusions” section discusses our insights from the results and concludes the paper.
Background
Time-series and temporal abstraction
Time series are sequences of discrete or continuous real-valued elements collected over time, such as network monitoring data, stock market data, and patient data stored in medical records. The entity (e.g., a patient) is sampled at fixed or varying time periods. The data may later be represented as time points for some variables (e.g., a body temperature of 38 °C at 19:30:00 on 01/03/23), or as time intervals (e.g., a fever lasting from 01/03/23 to 02/03/23). Often, the entity’s collected data are multivariate, originating from multiple sensors that sample the data in frequencies that vary between sensors and within each sensor. Calculating similarities between different sets of such multivariate time series is challenging, as we need to deal with different sampling frequencies, lack of synchronization between sensors, missing data, and random changes in entities’ behavior over time.
Basing the analysis on interval-based temporal abstract concepts and domain knowledge (e.g., a fever over eight hours), rather than the original time-stamped raw data, might be helpful in reducing random noise, avoiding problems resulting from sampling the data at different frequencies, summarizing large numbers of data in a succinct fashion8, and helping algorithms to cope with missing data9.
Temporal abstraction (TA) is the task of representing time-stamped raw data as a set of time intervals, typically at a higher level of abstraction10,11. Given a set of time-stamped data, external events, and sometimes also abstraction goals (which provide additional contexts for the TA process), TA produces a set of interval-based abstractions of the data that represent past and present states and trends that are relevant for the given set of goals. The interval-based abstractions are sensitive to the context in which each datum appears12, such as a blood-glucose measurement taken within the temporal span of the effect of a short-acting Insulin administration, or weight measurements of infants versus those of adults.
TA, which in most cases includes some form of interpolation13, is often helpful when dealing with data that were sampled at different frequencies or that are characterized by multiple missing values due to its smoothing effect on the generated abstractions. Temporal abstraction techniques usually include a discretization module as a pre-processing step, which discretizes raw data into a small number of possible ordinal categories (e.g., ’low’, ’normal’, ’high’), by determining cut-offs values. The discretization can be performed either by automatic unsupervised techniques14,15, supervised methods16, or by using knowledge acquired from a domain expert17, allowing us to embed human knowledge in the machine learning process.
Natural language processing
Natural Language Processing (NLP) is a sub-field of machine learning that focuses on allowing machines to understand and generate human language, as well as perform other downstream language-based tasks.
Classic NLP methods have typically involved building rule-based systems that attempted to mimic human language processing. These systems relied on hand-crafted rules and linguistic analysis to identify and extract features from text. While these methods were effective to some extent, they were limited in their ability to handle the complexity and variability of natural language. Further advancements included explicit probabilistic modeling of language (LMs) using tools such as Markov chains and bag-of-words approaches. With the rise in popularity of neural networks, variations of recurrent neural networks (RNNs) have also been used with increasing success18.
In recent years, the transformer architecture19 has become the state-of-the-art approach in NLP. Transformer models are deep neural networks, usually trained on massive amounts of text data, which use a technique called self-attention to learn contextual relationships between words. This allows them to generate more natural and fluent language, as well as produce state-of-the-art results for most NLP tasks, including language translation, sentiment analysis, and question answering.
Following the introduction of transformer based large language models (LLMs), and BERT20 in particular, tasks involving deep understanding of natural language have shown dramatic improvement. Furthermore, pre-training BERT architectures on medical text has shown that LLMs can achieve even better performance on downstream clinical tasks21. More specifically, Gu et al. have shown that their model, BioBert achieves state-of-the-art performance on multiple clinical tasks.
Related Work
The combination of machine learning in general, and deep learning in particular, with data from electronic health records (EHR) has been proven to be useful for various clinical applications, including readmission22. While useful, EHR data is often very complex in structure, spanning multiple modalities, namely, tabular data, time series, and free text. As such, many works have focused on the EHR records from the perspective of either time-series methods or NLP methods. Multiple works have demonstrated that deep learning yields better results than classic machine learning algorithms23,24,25. Nonetheless, various studies in the past26,27 have utilized linear and ensemble models, combined with feature engineering, to achieve varying degrees of success in classification of readmission.
Proper utilization of the temporal data available in EHRs is a non-trivial task. EHR data is characterized by irregular sampling frequencies, both across measurement types as well as internally. This leads to many missing values when treating the data as a multivariate time series. Various works have bench-marked a variety of time series architectures23, 24,28, predominantly favoring variations of RNNs, including combinations of RNNs with convolutional neural networks (CNNs). While multiple works, have demonstrated success using temporal methods, to the best of our knowledge only23 have conducted ablation studies to directly assess the predictive value of EHR temporal data. All other works have incorporated additional data modalities, most importantly tabular data describing doctors’ diagnoses, making it difficult to isolate the contribution of each component.
In contrast to the temporal data, utilizing the free text available has been much more straightforward. Huang et al.29 have shown that pre-training a BERT based model, ClinicalBERT, on EHR doctor notes, specifically discharge summaries, followed by fine-tuning, can yield better results than other NLP models. To overcome limits on input size to the model, a sentence aggregation method was employed. Further improvements haven been shown to be possible30 using note embeddings to create a similarity based patient graph, and then using graph neural networks. Currently, these type of models are the state-of-the-art in readmission prediction.
A major issue that arises when comparing the various works in the field is that prior to the introduction of MIMIC-III31, and even afterwards, many works focus on proprietary datasets, making their results unreproducible. More so, every selection of a subset of patients yields different results, even within a single dataset (for example, results reported on ClinicalBERT in29 vs.30). This makes direct comparison of results reported in different studies highly problematic. In this study we aim to solve this last problem, by presenting a standardized subset of patient admissions for future evaluation of Unplanned Readmission methods.
Data collection
Our evaluation uses the MIMIC-III dataset31, which is comprised of de-identified health-related data associated with over 40,000 patients who stayed in intensive care units (ICUs) at Beth Israel Deaconess Medical Center between 2001 and 2012. The database includes information such as demographics, vital sign measurements made at the bedside, laboratory test results, doctor notes, and more. It consists of 26 tables that record every piece of data for the patient’s admission to the hospital.
Data modalities
We view the data in MIMIC-III as being composed of three main data modalities, temporal, tabular and textual.
Following28,32, we identified 17 clinical concepts from the temporal data in MIMIC-III that are related to readmission. These can be divided into two groups: (1) lab tests (i.e., WBC, Hemoglobin, etc.), sampled with low frequency (daily, not hourly); (2) chart items (body temperature, blood pressure, heart rate, etc.), sampled with high frequency (hourly/minutely). Then we established knowledge-based state and gradient abstraction definitions, using guidelines which are described in Table 1. The definitions were used for the abstraction of the raw data into three to six states, and three gradients (increasing, decreasing or stable values). We will refer to this data as Charts.
Throughout a patient’s stay, hospital staff also record all doctor diagnoses, including pre-existing medical conditions, as well as the procedures a patient went through. This data is recorded using standardized ICD-9 codes. While the data is provided in a table using the codes, a textual description is also available for each code. This data modality is pseudo-tabular, since it can be viewed as a low frequency time series or it can be flattened into table form. We will refer to this data as ICD-9.
Finally, a large body of notes is included for each stay. Following29, we chose to focus on the discharge notes. These contain a brief summary of the patients stay as well as any other relevant information the doctor chooses to include, mostly in unstructured free-form text. We will refer to this data as Notes.
Dataset creation
We examined the data of adult patients (18 years old or older), that were admitted to the ICU and remained there for at least 1 day and no more than 30 days (to remove outliers). We labeled our data as positive if a patient has returned to the ICU within the 30 days following their discharge from the hospital, and negative otherwise. In order to properly evaluate the contribution of each data modality we removed admissions that did not have at least one of each lab tests measurement, and at least five of each chart items measurements. We also removed patients that died during the stay or throughout the 30 following days, since they clearly cannot be readmitted. Finally, we removed all but the first ICU stay of each patient in every year. This is important for two reasons: many hospital admissions may include multiple ICU stays before being discharged, and the usefulness of predicting readmission for patients that have had many recent readmissions is limited, as explained in 1. Those conditions resulted in a total of 14,837 patients across 15,424 ICU stays: 1752 readmissions (11.3%) and 13,672 that were not readmitted. These represent the entirety of patients relevant to Unplanned Readmissions in MIMIC-III, per our definition. The demographic distribution of the patients can be seen in Table 2, where numbers in bold indicate the number of positive labels and the remainder indicate the number of negative labels.
We then proceeded to divide the patients into five folds randomly, leaving (20%) as a test set for each fold. The split was done by patient and not by ICU stay, to avoid data leakage across a patient’s various stays. The splits were done in a stratified way, to ensure a similar distribution of labels in all five test sets. These splits were used for all of our experiments, and the reported results are aggregations across all folds.
Evaluation methods
Due to the multiple modalities, our experiments were split so that an ablation study regarding the predictive value of each modality could be conducted properly. All models and modalities were trained using the same training and evaluation loop as described in “Appendix A”. Our code was all written in Python, with deep learning components using PyTorch33, and Random Forest and evaluation metrics using Scikit-learn34.
Key architectures
The two main architectures used throughout our experiments are BERT20 and a Gated Recurrent Network (GRU)35 based bidirectional RNN network (BIRNN).
For the BERT models we used the standard architecture, available from Hugging Face’s Transformers library36, followed by a fully connected layer for classification.
The BIRNN’s architecture is composed of four main steps: an Embedding layer, two GRUs, an attention layer, and a fully connected layer. The embedding layer takes a vector from \({\mathbb {R}}^L\) and returns a \({\mathbb {R}}^{L \times \lfloor \root 4 \of {D} \rfloor }\) matrix, where L is the maximal sequence length and D is the number of unique values in the input. The embeddings are then passed to two GRUs, with the second one receiving a reversal (in the time dimension) of the input. The output of both GRUs is then concatenated together, and passed to the attention layer, providing a weighted version of its input. Finally, the previous output is concatenated to any additional modality if necessary, and a fully connected layer reduces each sample to a single value representing the score. A visual representation of the architecture can be seen in Fig. 2.
Charts
Being the most complicated data modality, we experimented with multiple data preparation methods. We compared each method from 5.2 both on its own, as well as combined with the method from 5.2. We also performed a one hot encoding of the patient’s demographics (demographics 1-hot) and concatenated it to the input of each model trained on Charts.
Discretization
As our baseline, we used the method described by28. We discretized the raw charts data into severity bins based on32,37, followed by flattening all measurements into a temporally ordered vector, i.e. a univariate time series. An illustration of this can be seen in Figure 1.A, Temporal Abstraction, e.g. Body Temperature below 36.2 °C is categorized as “Hypothermia”, between 36.2 and 37.2 °C as “Normal” and above 37.2 °C as “Fever”, and so on for each measurement, at each time step, using guidelines which are described in Table 1. This vector was then one hot encoded and padded to the maximal length (Charts 1-hot).
Our next method was to use a simple temporal-linear interpolation before the discretization, as described in Eq. 1:
where \(x_i^{(\tau _{t-1})}, \; x_i^{(\tau _{t+1})}\) are two consecutive known data points of measurement i, and there is at least one other measurement that is known at time \(\tau _t\). We then performed the same discretization as described above, except that instead of flattening, we left the data as a multivariate time series (Charts interpolated). Note that the number of observations in both methods is the same, but now \(x^{(\tau _{t})}\) is a 17-dimensional vector, rather than a single data point.
Temporal abstraction
Similar to what is described in “Time-series and temporal abstraction” section, we created TAs, describing the direction of change in the data. We will refer to these as gradients. We used a simple method to characterize subsequent gaps, given as follows:
Time series models
We trained three different model architectures and selected the best architecture for the rest of the evaluations. The three we selected to compare were a BIRNN model based on the architecture described in “Key architectures” section, an MLSTM-FCN model based on the architecture proposed in38, and a TimeTransformer model based on the architecture proposed in39. To make the comparison fair, we evaluated all three architectures on Charts 1-hot interpolated.
ICD-9
We prepared the data in two different ways.
In the first method, we took all the ICD-9 codes and flattened them into a single vector, essentially a one hot encoding. As each admission may have a varying number of codes, we padded all vectors to the maximal size (ICD-9 1-hot).
In the second method, we simply concatenated the textual descriptions, separated by commas, instead of the codes (ICD-9 text). We added a textual description of the patient’s demographics as well as the text “Procedures patient went through and doctor’s diagnoses:” at the beginning.
As our baseline we used a random forest (RF) model, due to the pseudo-tabular nature of the data. Its input was ICD-9 1-hot and demographics 1-hot concatenated together. We then trained a BIRNN model, as described in “Key architectures” section. We elected to use a recurrent model over a linear model due to the sequential nature of the data, i.e. diagnoses are provided throughout the ICU stay, and ordered by time. Finally, we trained a BioBERT model21 using the pretrained weights Gu et al. released to fine tune a model with ICD-9 text as its input.
Notes
We used the methods described in29 to pre-process the discharge notes. This includes text cleaning, and tokenization. We also followed their suggested method for dealing with the note size, by dividing each note into n chunks and recombining them by performing their suggested aggregation (Eq. 3) on the model’s output.
We used the pre-trained model weights that they have published and perform fine-tuning for readmission classification. It is important to note that while the model may have been pre-trained using notes from some fold’s test set, we do not see this as a data leakage issue. Our rationale is that a model can always be pre-trained with new discharge notes that do not yet have a label, similar to any other unannotated data used for unsupervised tasks.
Multi-modalities
For completeness we also trained two multi-modal models. The first model was a combination of the best Charts model with the BIRNN ICD-9 model. The second model was a combination of the previous model with the ClinicalBERT Notes model. We combined the individual models into a single multi-modality model using the method described in “Key architectures” section.
Results
Throughout our evaluation we used five metrics to determine the quality of a model’s predictions: area under the receiver operating characteristic curve (AUROC), F1 score, area under the precision recall curve (AUPRC), precision, and recall. While AUROC and AUPRC take the raw outputs of the model as input, the others were calculated by using an optimal threshold. Since some of these metrics are not directly correlated, i.e., a model may outperform another in AUROC but not in F1, we say a model is conclusively better than another if it receives a higher score in at least three metrics.
As can be seen in Table 3, we found no conclusive winner between the BIRNN and the TimeTransformer architectures, and opted for the BIRNN architecture only because of its similarity to previous works. Both clearly outperformed the MLSTM-FCN.
In the comparison of four different Charts data preparation methods, as shown in Table 4, we found that Charts interpolated yielded significantly better results than Charts 1-hot. This result seems to highlight the power of simple data imputation methods when dealing with missing values. Surprisingly, while adding gradients conclusively improved both methods, Charts 1-hot combined with gradients emerged as the best Charts model, outperforming Charts interpolated with gradients. The additional information gained when including gradients seems to yield an improved effect when compared to interpolating. We believe this may be due to the improved ability to capture the relationship between states without providing “artificial” data.
Deep learning methods have proven to be conclusively better than classic ML methods for ICD-9, as shown in Table 5. Results in italic signifies that they are the best when using only ICD-9. It is interesting to observe that using the full textual description ICD-9 text was conclusively better than its one-hot encoded counterpart. We think that this stems from the pretraining knowledge incorporated within BioBERT, allowing the model a deeper semantic understanding. Following the same line of reasoning, we were not surprised that the ClinicalBERT model, trained on Notes, achieved the best scores across the board, and by a significant margin. Furthermore, Notes is written in formats that are intended to be informative, allowing for an easier understanding of an admission, which we believe may assist the model.
As expected, the multi-modality model that included both ICD-9 and Charts (Table 5 method 6), was a conclusive improvement. It is easy to observe that each of these two modalities contain distinctly different information, thus the combination allowed for a richer representation of an admission. Nonetheless, the overall contribution of adding Charts was relatively small. On the other hand, the second multi-modality model (Table 5 method 8) demonstrated that combining all three together yields worse results that simply using Notes on its own. A possible explanation for this is that unlike the other two modalities, Notes may already include all the important information from other modalities, pre-filtered by doctors.
Discussion and conclusions
One of the main problems when dealing with clinical data is the vast amount of heterogeneous raw data originating from various sources. It includes static demographic data on patients, sequences of diagnoses and procedures, doctor notes as free unstructured text, and multivariate time-series data that in most cases is sampled irregularly (in time), with various sampling frequencies and missing values. In this paper, we demonstrated the effectiveness of every type of data on the challenging task of predicting patients readmissions. We defined the unplanned readmission problem in a strict way: we only took into account unplanned readmissions, i.e., ICU hospitalizations of patients who had not been hospitalized in an ICU during the previous year. We did so to avoid cases of patients who have chronic conditions and a high likelihood of being readmitted for the same issue in a short period, patients who are the usual candidates for readmission. It is important to identify and avoid including such patients in our evaluation, as they are easily recognizable by physicians, and our goal was to identify unplanned and unexpected cases of readmission.
While previous research on Readmission prediction usually focused on one or two data modalities, in this paper we investigated the prediction performance using every modality separately, and using different combinations of them, as each data modality may capture different aspects of a patient’s hospitalization.
When focusing on the multivariate time-series modality, using knowledge-based abstracted data, rather than the raw-time stamped data, improves the results. The creation of interval-based temporal abstractions of the raw, time-stamped data solves many of the missing value and multiple granularity sampling issues through a form of smoothing. Moreover, combining different levels of abstraction by adding to the pre-processing step Gradient abstraction, which focus on different resolution-level changes in the variables’ values, beyond the State abstraction (which focuses on discretization of the values), added considerable prediction power.
Yet another important aspect is the fact that ICD-9 sequences (both as codes or with their textual descriptions) achieve better prediction results than a model that is based solely on the multivariate time-series data (or their abstracted form). A combination between those two only slightly improves the results. Our assumption is that those ICD-9 sequences manage to summarize patients’ hospitalization with greater accuracy, focusing on aspects that are more relevant to readmission chances, then the information captured in long sequences of chart items and lab tests.
Finally, our experiments clearly show that a model that is based on discharge notes made be physicians has the best prediction results. Discharge notes are a rich source of unstructured data that contains valuable information about a patient’s health status, medical history, and treatment plan. NLP techniques that are geared toward clinical language, such as ClinicalBert, are able to extract key information from these notes and capture complex relationships between clinical variables that might be missed when using structured data alone. Additionally, it is indeed possible that this kind of clinical text can provide context that is not available through structured data, such as patient preferences or social determinants of health, which can further improve the accuracy of readmission predictions. When combining such data with the other data modalities (time-series, ICD-9 sequences) we did not observe any improvement in prediction capabilities. Overall, the discharge notes were found to be the most valuable resource for predicting patients readmission.
Our next objective is to investigate the influence of different sections in the discharge notes on the predication capabilities, explore additional approaches for integrating all data modalities, with the aim of enhancing the accuracy of prediction outcomes, and find ways to summarize the multivariate time-series data and ICD-9 sequences in a way that could potentially replace the discharge notes in their role as predictors for readmission, in situations where such data does not exist.
Training details
All models (except the RF model) were trained using a single NVIDIA A-100 GPU. To avoid over-fitting, all models used a stop loss of 7 instead of an epoch limitation. We consider a model to have improved if the AUPRC score for the validation set has risen between two successive evaluations, conducted every 200 steps. If the answer is positive, we save the model, otherwise we reduce the learning rates. If the stop loss has been exceeded, we stop the training and return the last saved model. We then use the saved model to evaluate it on the test set. We used a batch size of 64. The NLP models use a learning rate of 2·10−5, and a reduction rate of 0.9, the rest use 10−3 and 0.97 respectively. For each model we perform this training loop for each of the five folds.
Data availability
The dataset generated and/or analysed during the current study are available in the MIMIC-III repository through: mimic.mit.edu. Once access to MIMIC-III has been gained, reproduction of the list of ICU stays can be done using the code found at: https://github.com/meniData/Unplanned-Readmission.
References
Centers for Disease Control and Prevention (CDC). Hospital Utilization (in non-Federal short-stay hospitals). https://www.cdc.gov/nchs/fastats/hospital.htm (2018).
Lin, Y. W., Zhou, Y., Faghri, F., Shaw, M. J. & Campbell, R. H. Analysis and prediction of unplanned intensive care unit readmission using recurrent neural networks with long short-term memory. PloS One 14(7), e0218942 (2019).
Jencks, S. F., Williams, M. V. & Coleman, E. A. Rehospitalizations among patients in the Medicare fee-for-service program. New Engl. J. Med. 360(14), 1418–1428 (2009).
Rau, J. 10 Years of hospital readmissions penalties. https://www.kff.org/health-reform/slide/10-years-of-hospital-readmissions-penalties/ (2021).
Pishgar, M., Theis, J., Del Rios, M., Ardati, A., Anahideh, H. & Darabi, H. Prediction of unplanned 30-day readmission for ICU patients with heart failure. BMC Med. Inform. Decis. Mak. 22(1), 1–12 (2022).
Mcneill, H. et al. Impact of Intensive Care Unit readmissions on patient outcomes and the evaluation of the national early warning score to prevent readmissions: Literature review. JMIR Perioper. Med. 3(1), e13782 (2020).
Mudge, A. M. et al. Recurrent readmissions in medical patients: A prospective study. J. Hosp. Med. 6(2), 61–7 (2011).
Goldstein, A. & Shahar, Y. An automated knowledge-based textual summarization system for longitudinal, multivariate clinical data. J. Biomed. Inform. 61(2016), 159–175 (2016).
Moskovitch, R. & Shahar, Y. Fast time intervals mining using the transitivity of temporal relations. Knowl. Inf. Syst. 42(1), 21–48 (2015).
Combi, C. et al. Temporal reasoning and temporal data maintenance in medicine: Issues and challenges. Comput. Biol. Med. 27(5), 353–368 (1997).
Shahar, Y. & Combi, C. Timing is everything. Time-oriented clinical information systems. West. J. Med. 168(2), 105 (1998).
Shahar, Y. Dynamic temporal interpretation contexts for temporal abstraction. Ann. Math. Artif. Intell. 22(1–2), 159 (1998).
Shahar, Y. Knowledge-based temporal interpolation. J. Exp. Theor. Artif. Intell. 11(1), 123–144 (1999).
Dougherty, J., Kohavi, R. & Sahami, M. Supervised and unsupervised discretization of continuous features. In Machine Learning Proceedings 1995, 194–202 (Elsevier, 1995).
Mörchen, F. & Ultsch, A. Optimizing time series discretization for knowledge discovery. In Proceedings of the Eleventh ACM SIGKDD International Conference on Knowledge Discovery in Data Mining, 660–665 (ACM, 2005).
Moskovitch, R. & Shahar, Y. Classification-driven temporal discretization of multivariate time series. Data Min. Knowl. Discov. 29(4), 871–913 (2015).
Shahar, Y. A framework for knowledge-based temporal abstraction. Artif. Intell. 90(1–2), 79–133 (1997).
Jurafsky, D. & Martin, J. H. Speech and language processing—an introduction to natural language processing, computational linguistics, and speech recognition. In Prentice Hall series in artificial intelligence (2000).
Vaswani, A., et al. Attention is all you need. arXiv:1706.03762 (2017).
Devlin, J., Chang, M.-W., Lee, K. & Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv:1810.04805 (2019).
Yu, Gu. et al. Domain-specific language model pretraining for biomedical natural language processing. ACM Trans. Comput. Healthc. (HEALTH) 3(2020), 1–23 (2020).
Morgan, D. J., et al. Assessment of machine learning vs standard prediction rules for predicting hospital readmissions. JAMA Netw. Open 2 (2019).
Lin, Y.-W., Zhou, Y., Faghri, F., Shaw, M. J. & Campbell, R. H. Analysis and prediction of unplanned intensive care unit readmission using recurrent neural networks with long short-term memory. PLoS ONE 14 (2018).
Pham, T., Tran, T., Phung, D. Q. & Venkatesh, S. DeepCare: A deep dynamic memory model for predictive medicine. arXiv:1602.00357 (2016).
Wang, H. et al. predicting hospital readmission via cost-sensitive deep learning. IEEE/ACM Trans. Comput. Biol. Bioinform. 15(2018), 1968–1978 (2018).
Caruana, R., Lou, Y., Gehrke, J., Koch, P., Sturm, M. & Elhadad, N. Intelligible models for HealthCare: Predicting pneumonia risk and hospital 30-day readmission. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (2015).
Sushmita, S. et al.Predicting 30-day risk and cost of “all-cause” hospital readmissions. In Expanding the Boundaries of Health Informatics Using AI, in AAAI Workshop (2016).
Barbieri, S., et al. Benchmarking deep learning architectures for predicting readmission to the ICU and describing patients-at-risk. Sci. Rep. 10 (2019).
Huang, K., Altosaar, J. & Ranganath, R. ClinicalBERT: Modeling clinical notes and predicting hospital readmission. arXiv:1904.05342 (2019).
Golmaei, S. N. & Luo, X. DeepNote-GNN: Predicting hospital readmission using clinical notes and patient network. In Proceedings of the 12th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (2021).
Johnson, A. E. W. et al. MIMIC-III, a freely accessible critical care database. Sci. Data 3(1), 1–9 (2016).
Sheetrit, E., Nissim, N., Klimov, D. & Shahar, Y. Temporal probabilistic profiles for sepsis prediction in the ICU. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining (2019).
Paszke, A., et al. PyTorch: An imperative style high-performance deep learning library. In Neural Information Processing Systems (2019).
Pedregosa, F. et al. Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 12(2011), 2825–2830 (2011).
Chung, J., Gülçehre, Ç., Cho, K. & Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv:1412.3555 (2014).
Wolf, T., et al. Transformers: State-of-the-art natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, Online, 38–45. https://www.aclweb.org/anthology/2020.emnlp-demos.6.
Johnson, A. E. W., Kramer, A. A. & Clifford, G. D. A new severity of illness scale using a subset of acute physiology and chronic health evaluation data elements shows comparable predictive accuracy*. Crit. Care Med. 41(2013), 1711–1718 (2013).
Karim, F., Majumdar, S., Darabi, H. & Harford, S. Multivariate LSTM-FCNs for time series classification. Neural Netw. 116(2018), 237–245 (2018).
Wu, N., Green, B., Ben, X. & O-Banion, S. Deep transformer models for time series forecasting: The influenza prevalence case. arXiv:2001.08317 (2020).
Author information
Authors and Affiliations
Contributions
E.S. and M.B. designed the study and experimental plan, analyzed the data, conducted the experiments, and wrote the manuscript. O.E. provided guidance throughout the study, designed the experimental plan, and provided substantial revision to the manuscript.All authors have read and approved the final version of the manuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Sheetrit, E., Brief, M. & Elisha, O. Predicting unplanned readmissions in the intensive care unit: a multimodality evaluation. Sci Rep 13, 15426 (2023). https://doi.org/10.1038/s41598-023-42372-y
Received:
Accepted:
Published:
Version of record:
DOI: https://doi.org/10.1038/s41598-023-42372-y
This article is cited by
-
Deep learning models for ICU readmission prediction: a systematic review and meta-analysis
Critical Care (2025)
-
The Rothman Index predicts unplanned readmissions to intensive care associated with increased mortality and hospital length of stay: a propensity-matched cohort study
Patient Safety in Surgery (2024)




