key: cord-0850980-acn5tfk5 authors: Han, Xiaoyang; Yu, Ziqi; Zhuo, Yaoyao; Zhao, Botao; Ren, Yan; Lamm, Lorenz; Xue, Xiangyang; Feng, Jianfeng; Marr, Carsten; Shan, Fei; Peng, Tingying; Zhang, Xiao-Yong title: The value of longitudinal clinical data and paired CT scans in predicting the deterioration of COVID-19 revealed by an artificial intelligence system date: 2022-04-08 journal: iScience DOI: 10.1016/j.isci.2022.104227 sha: 95a0f241429ac2ada7994e09b84065b333a614f5 doc_id: 850980 cord_uid: acn5tfk5 The respective value of clinical data and CT examinations in predicting COVID-19 progression is unclear, because the CT scans and clinical data previously used are not synchronized in time. To address this issue, we collected 119 COVID-19 patients with 341 longitudinal CT scans and paired clinical data, and developed an AI system for the prediction of COVID-19 deterioration. By combining features extracted from CT and clinical data with our system, we can predict whether a patient will develop severe symptoms during hospitalization. Complementary to clinical data, CT examinations show significant add-on values for the prediction of COVID-19 progression in the early stage of COVID-19, especially in the 6th to 8th day after the symptom onset, indicating that this is the ideal time window for the introduction of CT examinations. We release our AI system to provide clinicians with additional assistance to optimize CT usage in the clinical workflow. During COVID-19 pandemic, fast and accurate prediction of disease progression in the early stage is crucial for clinical decision-making and optimal medical resource allocation. Chest computed tomography (CT) plays an important role for evaluating COVID-19 patients by showing specific image features such as ground-glass opacification and consolidation. (Wong et al., 2020 ) Based on CT images or clinical data, or both, artificial intelligence (AI) based methods have shown advantages in disease diagnosis and progression prediction. (Shi et al., 2020; Wong et al., 2020; Roberts et al., 2021; Zhu et al., 2021) However, most of these AI studies focused on CT or clinical data acquired at a single time point during COVID-19 progression, which limits the prediction accuracy. To address this issue, the evaluation of both longitudinal CT images and clinical data is desirable because it can help to better understand dynamical changes of pneumonia lesions during COVID-19 deterioration as well as the progression process following therapy and thus resulting in more accurate prognostication of patient outcomes. (Feng, Ding and Zhang, 2020; Huang et al., 2021) Longitudinal measurements can provide temporal information that comprehensively reflects dynamic changes of COVID-19 deterioration. So far, only a few studies explored temporal information from longitudinal CT or clinical measurements using deep learning (DL) based methods. (Fang et al., 2021; Kim et al., 2021; Pu et al., 2021; R. Wang et al., 2021; Shamout et al., 2021) Zhou, et al., (Zhou et al., 2021) used clinical features only to predict COVID-19 severity. Pu, et al., (Pu et al., 2021) and Kim, et al.,(Kim et al., 2021) registered longitudinal CT scans and visualized the difference in diseased vs. healthy areas between scans, and found that the CT based biomarkers may be used to monitor the development of the infection process in COVID-19. Another work (Fang et al., 2021) extracted features directly from sequential CT scans without an extra registration step and is therefore more practical using a recurrent neural network to predict the malignant progression of COVID-19. Shamout, at al., (Shamout et al., 2021) provided visually intuitive saliency maps to help clinicians interpret the model predictions of risk (Heagerty and Zheng, 2005) of deterioration over different time horizons, ranging from 24 hours to 96 hours with combination of Chest X-rays images and routinely collected clinical variables at only one time point. Wang Robin, et al.,(R. Wang et al., 2021) developed an AI system to predict future deterioration to clinical illness in COVID-19 patients using chest CT and clinical data and proves that the combination of CT and clinical data improves the performance of disease progression prediction. Although the above studies provided important information for the prediction of COVID-19 progression, the respective values of clinical data and parallel CT scans remain unclear because longitudinal time points of clinical data and CT scans previously used were limited and not paired. To address this issue, based on well-designed longitudinal paired CT scans and clinical data, we developed a new AI model to predict whether the patient will progress to a severe status, and thus to reveal the respective value of clinical data and parallel CT scans for comprehensive prediction of COVID-19 progression reflecting its dynamic deterioration changes ( Figure 1 ). Since accurate pneumonia lesion segmentation is the key step for disease progression prediction, we evaluated lesion segmentation performance of BCL-Net along with four state-of-the-art methods on four datasets which are summarized in Table 3 . Annotations of lung and COVID-19 lesions of the Coronacases dataset and parts of the Radiopedia dataset are publicly available, (Ma et al., 2021) which we used to train our model and to evaluate it with a five-fold cross-validation. We annotated the remaining data by two experienced radiologists and used it for testing only. In all four datasets, BCL-Net achieves better DC of 0.840 (95% CI 0.813~0.867) than other methods for lesion segmentation ( Figure 4A , B, Table 3 ), suggesting that BCL-Net is robust for variations in lesion size, CT image intensities, and slice thickness. By contrast, the performance of a full 3D convolution deteriorates with increasing slice thickness ( Figure S2 ) and could completely fail when the information of slice thickness is missing and the interpolation along the z-axis is not properly done ( Figure 4B, nnUNet3D) . Moreover, compared to other methods, BCL-Net is three times faster than the second fastest method, Sensor3D, and hundreds of times faster than nnUNet ( Figure 4C , Table 2 ). This is an important advantage, as local hospitals generally face enormous time pressure when a COVID-19 wave hits a region. The CT features quantified from lung and pneumonia lesion segmentation are summarized in Table S2 . The z-position of progression patients is significantly higher than non-progression patients (53.70 vs 47.05, p < 0.001), consistent with previous studies. (Yu et al., 2020) J o u r n a l P r e -p r o o f 5 In order to accurately predict the progress of COVID-19, based on a combination of features extracted from CT scans and paired clinical data, we implemented an ensemble learning method to assess whether the patient will develop a severe state ( Figure 5 ) and achieved a prediction with an AUC of 0.900 (95% CI 0.891~0.909) ( Figure 5A ) in a five-fold cross validation. If we use the CT data or clinical data only, lower AUCs are achieved of only 0.857 (95% CI 0.846~0.868) and 0.844 (95% CI 0.833~0.856) for CT scan only model and clinical data only model respectively. Practically, we may use the prediction of disease progression as guidance for hospital triage to distribute the limited hospital beds to patients who are predicted to have a severe disease progression. For interpretation, the 10 most features that contribute to the prediction were listed ( Figure 5B ). In addition to CT-related features and clinical data, age was found to be a key factor leading to different disease progression patterns, consistent with previous studies. (Huang et al., 2020; Wang et al., 2020) The detailed comparison of performance on disease progression prediction are summarized in Table 4 . We visualized the progression of pneumonia in individual patients by plotting the PLV and the average trajectory of three clinical data (including CK, NLR, and LDH) over time ( Figure 5C , E) with normal range in orange dotted lines (Zhou et al., 2021) . The progression of pneumonia lesions of a progression patient and a non-progression patient are visualized in Figure 5D . Although individual differences are large, patients who progress to a severe state usually show a sharp increase in the PLV, CK and LDH in the first week of hospitalization, peaking at around 7-10 days, and then slowly recover in the following days. In contrast, patients who showed only mild symptoms during hospitalization had consistently lower PLV, CK, NLR, and LDH. Furthermore, in order to reveal the respective values of clinical data and paired CT scans in predicting the deterioration of COVID-19, we divided the data into six groups longitudinally every three days from the onset of COVID-19 symptoms, and analyzed them using our AI model. As shown in Figure 6A , the progression prediction model achieves time-dependent AUCs of 0.688, 0.833, 0.901, 0.942 and 0.946 with combination of CT and clinical data over three-day time points in 14 days from symptom onset. The time-dependent AUC of the time group Day 15reaching 1.000 trained and evaluated on only ten paired CT scans and clinical data (see Figure S3 ) was excluded from the discussion for the limitation of data scale. As a comparison, AUCs of the disease progression prediction model are 0.656, 0.800, 0.842, 0.915, and 0.944 using CT only, respectively. Using only clinical data, the AUCs of the model are 0.672, 0.748, 0.846, 0.869, and 0.890, respectively. In order to more appropriately describe the performance of COVID-19 progression prediction models on the dataset with imbalanced positive and negative samples (273 vs. 68) over three-day time groups, we compared the changes in F1 scores ( Figure 6B ). The progression prediction model achieved time-dependent F1 scores of 0.286, 0.563, 0.783, 0.857, and 0.889 with combination of CT and clinical data over 3-day time groups. With only features extracted from CT the prediction model achieved F1 scores of 0.286, 0.424, 0.694, 0.815, and 0.889, and with clinical data only the model achieved 0.250, 0.400, 0.638, 0.815, and 0.750. Time-dependent AUCs and F1 scores with different time group combinations were detailed in Figure S3 and Table S3 . Note that in the very early stage (Day 0 -Day 2), the performance of disease progression prediction models is limited with either CT or clinical data alone, even with both. From the 3rd to 14th day after the onset of symptoms, compared to using clinical data alone, the introduction of CT examinations significantly increases the prediction accuracy (p <0.05). The add-on value of CT over clinical data is maximized during the 6th to 8th day (AUC improves from 0.833 to 0.901, F1-score from 0.563 to 0.783) These results suggest that the ideal time window for a CT scan is the 6th to 8th day after the COVID-19 symptom onset. In the current work, based on well-designed longitudinal clinical data and paired CT scans, we developed a new AI system to reveal the respective value of clinical data and parallel CT scans for comprehensive prediction of COVID-19 progression. CT scans and clinical data capture different characteristics of COVID-19 patients, but the complementarities between them have not been fully leveraged. Several machine learning studies have addressed this issue and are attempting to assess the risk of critical illness for COVID-19 patients at the hospital admission. (Liang et al., 2020; Vaid et al., 2020; Wu et al., 2020) Although these studies are very promising to identify patients at high risk, they usually rely on CT scans (Kim et al., 2021; X. Wang et al., 2021) or clinical measurements (Yan et al., 2020; Zhou et al., 2021) alone, which can lead to incomplete clinical observation. By contrast, as shown in Figure 7 , our correlation analysis between CT features and clinical measurements revealed there is a strong correlation between LDH and PLV (r=0.68, p<0.01) and PCV (r=0.65, p<0.01). Additionally, the univariate and bivariate distributions and the correlation of the features are shown in Figure S4 . Based on the combination of chest CT scans and clinical data, our AI system reaches an AUC score of 0.900, which has been proven outperforming CT scan or clinical data alone. Therefore, our tool is complementary to existing AI tools for COVID-19 triage and prognosis. Early prediction of disease deterioration is crucial because it allows for timely and effective treatment, which can potentially improve outcomes and reduce mortality for critically ill COVID-19 patients. (Li et al., 2020) DL methods have the potential to help identify patients at risk for progression to critical illness in an early stage of disease. In contrast to published work, with longitudinal measurements, we found that the performance of the predictive model improves as the disease progresses and reaches considerable accuracy after five days. In the very early stage (the first three days from the date of symptom onset), the performance of disease progression prediction is limited with either CT scans or clinical data. The performance of the model using combined features is greatly improved in three to eight days compared to using clinical data only. Given the prediction accuracy of the combined approach reaches a satisfying F1-score of around 0.80, the most recommended time point for CT scan is the 5th to 8th day from the date of symptom onset. DL-based methods often encounter performance degradation in the multicenter study, mainly due to the large data distribution discrepancy between different cohorts. As shown in Table 3 , BCL-Net achieves better Dice scores than other methods for both lung and lesion segmentation in all four cohorts from multiple centers, suggesting that BCL-Net is robust for variations in lesion size, CT image intensities, and slice thickness. We also found that our new volumetric segmentation algorithm, BCL-Net, increases the prediction accuracy by 6~8% over the current state-of-the-art method, DeepLabV3+. Yet DeepLabV3 is a pure 2D segmentation algorithm that neglects the rich inter-slice contexts in CT scans. By contrast, BCL-Net is a hybrid 2D-3D segmentation network that combines classical 2D intra-slice feature extraction and a bidirectional ConvLSTM network for inter-slice feature learning, yielding an improved performance. Sensor3D fuses 2D segmentation with ConvLSTM, whilst BCL-Net uses a dual path attention mechanism to reduce the computation cost. On the other hand, compared to full 3D volumetric segmentation such as 3D U-Net, BCL-Net is several orders of magnitude faster, lower in memory consumption, and robust with respect to thick-slice CT as well as thinslice CT ( Figure S2 ). Hence, it is more generalizable in clinical settings where different CT scanners and computing resources are used. One limitation of this study is that the sample size is not big. Unlike the acquisition of routine clinical data, due to the radiation in CT examinations, unnecessary multiple CT scans are not only harmful to health, but also consume excessive medical resources. Therefore, for the same patient, it is often impractical to obtain longitudinal multi-time CT data that exactly matches the clinical data. Fortunately, in this study we have collected the paired CT and clinical data. We have taken advantage of these data to predict disease progression after the onset of COVID-19 symptoms. The results indicate that the addition of CT examinations on 5th -8th day after the onset of symptoms may be ideal for predicting COVID-19 progress. In future work, we will collect more CT data at multiple time points that are paired with clinical data, and accurately analyze the predictive value of CT examinations for the prediction COVID-19 progression, and evaluate its necessity in tracking clinical treatment outcome. CT images and paired clinical data were collected at the same time point. Based on these data, we have developed an AI system to derive a pneumonia progression score by volumetric segmentation of lung (transparent) and COVID-19 lesions (red), and to combine with clinical features to predict whether the patient will progress to a severe status, and thus to reveal the respective value of clinical data and parallel CT scans for comprehensive prediction of COVID-19 progression. . The overview of our proposed AI system. (a) CT features were quantified using our BCL-Net, which combines a 2D U-net to process intra-slice spatial information with an LSTM to leverage inter-slice context. BCL-Net uses the share-weighted 2D convolution in both encoding and decoding paths, avoiding computational expensive 3D convolution. Instead, it uses a BC-LSTM Module (right) to integrate inter-slice context from multiple CT slices. In addition, a dual path attention mechanism including a global average pooling and a depthwise 2D convolution followed by 1 × 1 convolution is applied to reduce the computation cost. (b) We implemented an ensemble learning method to predict the disease progression using the paired CT scans and clinical data. 21 clinical measurements were combined with 12 lung and lesion volume CT features to predict the disease progression. See also Table S1. With an average inference time of 16 ms/slice, it takes less than 5 seconds to segment a conventional CT scan with about 300 slices in clinics. In comparison, nnUnet3D, due to its computationally expensive interpolation preprocessing step, needs almost 20 minutes to process the same CT image. See also Figure S2 . Table 3 . Comparison of different network architectures and segmentation performances. In addition to BCL-Net, we tested four state-of-the-art segmentation methods, including DeepLabV3+, Sensor3D, nnUNet2D and nnUNet3D. Among all these methods, DeepLabV3+ and nnUNet2D are purely 2D segmentation methods; nnUNet3D is a full 3D volumetric segmentation method with 3D convolution kernels; both Sensor3D and BCL-Net are hybrid 2D-3D methods, whilst Sensor3D fuses 2D segmentation with convolutional LSTM (ConvLSTM). Compared to other methods, BCL-Net achieves the highest dice score for both lung and lesion segmentation as well as the fastest inference speed. Segmentation performances are presented as dice scores along with 95% CI. See also Figure S2 . Further information and requests for resources should be directed to and will be fulfilled by the lead contact, Xiao-Yong Zhang (xiaoyong_zhang@fudan.edu.cn). This study did not generate any new unique materials. • Partial data reported in this paper will be shared by the lead contact upon request. The longitudinal data reported in this study cannot be deposited in a public repository because patients do not want their data to be made public. • All original code has been deposited at https://robin970822.github.io/DABC-Net-for-COVID-19/ and is publicly available as of the date of publication. • Any additional information required to reanalyze the data reported in this paper is available from the lead contact upon request. This longitudinal cohort study was approved by the Ethics Committee of Shanghai Public Health Clinical Center (YJ-2020-S035-01). Figure 2 illustrates the flowchart of the patient selection. As shown in Table1, a total number of 218 COVID-19 patients with 735 CT scans were collected from January 21, 2020 to April 29, 2020. From public datasets available online, we obtained 27 CT scans with lesion labels from Wuhan, 19 CT scans with lesion and lung labels from Radiopaedia.org and 10 CT scans with lesion and lung labels from Coroncase.org. From Shanghai Public Health Clinical Center, we obtained 679 CT scans from 162 patients including 43 CT scans with lesion and lung labels and 341 longitudinally measured CT scans and corresponding clinical data from 119 patients. The CT scans from Shanghai Public Health Clinical Center were obtained on dedicated CT scanners (Hitachi, Philips and UIH) with the following parameters: slice thickness 1-6 mm, slice gap 0 mm. All CT scans with lesion and lung labels were included to segment the lesion and lung volume on a deep learning method. 119 patients with 341 longitudinally measured CT scans and paired clinical data Shanghai Public Health Clinical Center were enrolled to build an AI system to predict the disease progression and investigate the respective value of clinical data and CT scans. The longitudinal clinical data including patient demographics, medical history, presenting signs and symptoms and laboratory examinations were all collected from the hospital information system (HIS). The laboratory examinations including absolute neutrophil count (ANC), absolute lymphocyte count (ALC), neutrophil lymphocyte ratio (NLR), lactate dehydrogenase (LDH), direct bilirubin (D-BIL), and creatine kinase (CK) were measured and matched with CT scans. The time points of symptoms onset and the beginning of the severe/critical stage were recorded for the time alignment. According to the guidelines of national diagnosis and treatment protocols for COVID-19 (the seventh version),((Released by National Health Commission & National Administration of Traditional Chinese Medicine on March 3, 2020), 2020) patients with COVID-19 can be divided into four subtypes: mild, common, severe, and critically ill. The progress of the severe illness with COVID-19 is usually rapid and there is no clear separation between the severe illness and the critical illness. (Li et al., 2020) Therefore, patients deteriorating into these two subtypes were combined to be the progression patients. The 119 patients who were common subtypes at admission were divided into two groups, that is, 90 non-progression patients with 273 paired CT scans and clinical data who were discharged from hospital in the common subtype, and 29 progression patients with 68 paired CT scans and clinical data who deteriorated into severe or critically ill subtype or died. Overview of our proposed AI system An AI system (Figure 3 ) was used to predict the disease progression with paired clinical data and CT scans at each time point for each patient. CT features are obtained by BCL-Net. Both clinical features (patient demographics, medical history, signs and symptoms, and laboratory tests) and CT features are combined and fed into the AI system for the prediction. As a hybrid 2D-3D network, BCL-Net combines a U-shaped network (UNet) (Çiçek et al., 2016; Falk et al., 2019) with shared-weighted encoder and decoder to process in-plane context and a BC-LSTM Module that uses bidirectional convolutional LSTM to integrate cross-plane context ( Figure 3B ). BCL-Net uses a share-weighted convolution operator in the encoding and decoding paths of 2D U-Net to extract intra-slice features. More specifically, each convolution and transposed convolution block consists of two 3 × 3 shared convolution filters followed by a 2 × 2 max pooling layer and ReLU function. These parameters are shared between all input slices. In order to process the cross-plane context along the z-axis and retain the original high-resolution xy inplane information, a convolutional long-short term memory (C-LSTM) (Shi et al., 2015) is used to integrate in-plane features extracted by 2D U-Net. Unlike temporal sequential data (e.g., video clips), where information flows in only one forward direction, structural CT scans have two orientations that need to be considered. Hence, a bidirectional C-LSTM module (BC-LSTM Module) is used to model both forward and backward information. To reduce computational cost, we propose a dual path attention mechanism including a global average pooling (GAP) to generate channel-wise maps and a channel-dimension squeeze procedure with a depthwise convolution followed by 1 × 1 convolution to generate spatial-wise maps. BC-LSTM only fuses the information distilled by the channel-wise maps and the spatialwise maps and reduces the computation cost. We trained two BCL-Nets, respectively for lung and pneumonia lesion segmentation. We further multiplied the output of lesion BCL-Net with the corresponding output of lung BCL-Net to remove possible false positive lesions outside the lung organ. Consolidations were found more common in patients >50 years old (Poyiadji et al., 2020) and could be a warning sign of severe progression. Therefore, we further outlined the consolidation region thresholding at 0.5 on normalized intensity within the lesion region. (Liu et al., 2020) Additionally, we calculated the weighted volume from the inner product of the lesion and the intensity, determining the center of the lesion in the z-axis. (Yu et al., 2020) In all, 12 AI-derived CT features including lung volume (LuV), lesion volume (LeV), consolidation volume (CV), percentage of lesion (PLV) and consolidation volume (PCV), and the related center of the lesion in the z-axis (z-position) in left and right lung were obtained. We compared BCL-Net with four state-of-the-art medical segmentation methods, namely DeepLabV3+, (Chen et al., 2018) Sensor3D, (Novikov et al., 2019) nnUNet2D & nnUNet3D. (Isensee et al., 2021) Table 3 summarizes the individual features of all five methods including BCL-Net in terms of model architecture, parameter size and inference speed. As illustrated in Figure 3A , we implemented an ensemble learning method to predict the disease progression using a total of 33 features including 12 AI-derived features extracted from CT scans and paired 21 clinical measurements collected from HIS. The ensemble learning method was implemented in Python scikit-learn library with base learners consisting of support vector machine (SVM), k-nearest neighbors (KNN), naive bayes (NB), multilayer perceptron (MLP), random forest (RF), gradient boost (GB), logistic regression (LR), adaptive boosting (Adaboost), and extreme gradient boosting (XGBoost). (Chen and Guestrin, 2016) The averaged output of base learners was calculated as the final output for prediction of disease prediction. Table S1 summarizes the implementation details of base learners. We compared our ensemble learning method with four independently trained and evaluated machine learning methods including SVM, MLP, GB, and XGBoost in a five-fold cross validation. In addition to prediction, we also highlighted important features that mostly contribute to our prediction using the average feature importances from base learners including RF, GB, Adaboost, and XGBoost. Given the fact that patients from Shanghai Public Health Clinical Center were under daily medical observation and scanned for CT about every 3 days ( Figure S1 ), we divided 341 paired CT and clinical data in six three-day time groups, that is 27, 74, 79, 73, 78, and 10 paired CT and clinical data in group Day0-Day2 (within two days from symptom onset), group Day3-Day5 (3-5 days), group Day6-Day8 (6-8 days), group Day9-Day11 (9-11 days), group Day12-Day14 (12-14 days), and group Day15-(more than 15 days from symptom onset), separately. We compare the performance changes over the different three-day time groups of the model trained with a combination of CT and clinical data at the same time-point, the model with only CT and the model with only clinical data. Using our proposed AI system (Figure 3) , the respective value of longitudinal clinical data and paired CT scans in predicting the deterioration of COVID-19 was evaluated with AUC (area under the receiver operating characteristic curve, ROC) and F1 score in a five-fold cross validation in each time group except group Day0-2 and Day15-in a three-fold cross validation owing to the insufficiency of progression patients. Statistical analysis of features extracted from CT scans and clinical characteristics is performed in the Python statsmodel library. Continuous variables are represented by median and range, and a comparison between non-progression vs progression patients is performed using the Student's t-test for normally distributed variables and the Mann-Whitney U statistics for non-normally distributed variables. Categorical variables are represented as numbers with percentages, and a comparison between patient groups is performed by chi-square. The performances of lung and pneumonia lesion segmentation are evaluated in dice coefficient (DC) and compared with four state-of-the-art in the Friedmann test. The performances of disease progression prediction are evaluated in AUC and F1 score and compared in the permutation test. All the metrics of the AI system are presented with 95% confidence interval (CI). Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation XGBoost: A Scalable Tree Boosting System 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation', Medical Image Computing and Computer-Assisted Intervention -MICCAI 2016 U-Net: deep learning for cell counting, detection, and morphometry Deep learning for predicting COVID-19 malignant progression', Medical image analysis Dynamic evolution of lung abnormalities evaluated by quantitative CT techniques in patients with COVID-19 infection Survival model predictive accuracy and ROC curves Clinical features of patients infected with 2019 novel coronavirus in Wuhan Dynamic changes in chest CT findings of patients with coronavirus disease 2019 (COVID-19) in different disease stages: a multicenter study nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation Longitudinal Quantitative Assessment of COVID-19 Infection Progression from Chest CTs Early triage of critically ill COVID-19 patients using deep learning Therapeutic strategies for critically ill patients with COVID-19', Annals of Intensive Care CT quantification of pneumonia lesions in early days predicts progression to severe illness in a cohort of COVID-19 patients Toward data-efficient learning: A benchmark for COVID-19 CT lung and infection segmentation Deep Sequential Segmentation of Organs in Volumetric Medical Scans Acute Pulmonary Embolism and COVID-19 Automated quantification of COVID-19 severity and progression using chest CT images Released by National Health Commission & National Administration of Traditional Chinese Medicine on Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans', Nature Machine Intelligence An artificial intelligence system for predicting the deterioration of COVID-19 patients in the emergency department Review of Artificial Intelligence Techniques in Imaging Data Acquisition, Segmentation and Diagnosis for COVID-19 Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting Machine Learning to Predict Mortality and Critical Events in a Cohort of Patients With COVID-19 in New York City: Model Development and Validation Clinical Characteristics of 138 Hospitalized Patients With 2019 Novel Coronavirus-Infected Pneumonia in Wuhan Artificial intelligence for prediction of COVID-19 progression using CT imaging and clinical data', European radiology Joint Learning of 3D Lesion Segmentation and Classification for Explainable COVID-19 Diagnosis Frequency and Distribution of Chest Radiographic Findings in Patients Positive for COVID-19 Development of a clinical decision support system for severity risk prediction and triage of COVID-19 patients at hospital admission: an international multicentre study', The European respiratory journal: official journal of the European Society for An interpretable mortality prediction model for COVID-19 patients Multicenter cohort study demonstrates more consolidation in upper lungs on initial CT increases the risk of adverse clinical outcome in COVID-19 patients Clinically Applicable AI System for Accurate Diagnosis, Quantitative Measurements, and Prognosis of COVID-19 Pneumonia Using Computed Tomography Eleven routine clinical features predict COVID-19 severity uncovered by machine learning of longitudinal measurements Joint prediction and time estimation of COVID-19 developing severe symptoms using chest CT scan', Medical image analysis This work was supported by the National Natural Science Foundation of China (92043301)