key: cord-0222078-flhs37w2 authors: Xie, Weiyi; Jacobs, Colin; Charbonnier, Jean-Paul; Ginneken, Bram van title: Dense Regression Activation Maps For Lesion Segmentation in CT scans of COVID-19 patients date: 2021-05-25 journal: nan DOI: nan sha: b0255ad2c81689cfd14ec8573780400a5fbf7b99 doc_id: 222078 cord_uid: flhs37w2 Automatic lesion segmentation on thoracic CT enables rapid quantitative analysis of lung involvement in COVID-19 infections. However, obtaining a large amount of voxel-level annotations for training segmentation networks is prohibitively expensive. Therefore, we propose a weakly-supervised segmentation method based on dense regression activation maps (dRAMs). Most weakly-supervised segmentation approaches exploit class activation maps (CAMs) to localize objects. However, because CAMs were trained for classification, they do not align precisely with the object segmentations. Instead, we produce high-resolution activation maps using dense features from a segmentation network that was trained to estimate a per-lobe lesion percentage. In this way, the network can exploit knowledge regarding the required lesion volume. In addition, we propose an attention neural network module to refine dRAMs, optimized together with the main regression task. We evaluated our algorithm on 90 subjects. Results show our method achieved 70.2% Dice coefficient, substantially outperforming the CAM-based baseline at 48.6%. T HE coronavirus disease 2019 (COVID-19) has been declared a global pandemic since March, 2020. Infected cases have reached over 252 million worldwide, with more than five million deaths. Unfortunately, both numbers are still increasing. To reduce the fatality rate, effective diagnosis and treatment planning are essential. As COVID-19 mainly damages the lungs of infected subjects, Computed Tomography (CT) of the chest plays a critical role in rapid diagnosis and progression monitoring of COVID-19 infection. Based on chest CT analysis, standardized CT scoring systems, such as the COVID-19 Reporting and Data System (CO-RADS) [18] , were defined to quantify the degree of suspicion of COVID-19 according to CT findings into 1-5 scores with an increasing level of suspicion. Similarly, a CT severity scoring system [12] was designed to assess the extent of parenchymal involvement of the disease. These scoring systems may be applied more accurately, objectively, and rapidly when automatic segmentation of infected ( areas (lesions) is available. Therefore, this work aims at developing an algorithm that can automatically segment lesions related to COVID-19 on chest CT scans. One of the major obstacles of semantic segmentation is the difficulty of acquiring a large amount of voxel-wise annotations for training as manual outlining in high-resolution 3D scans is extremely laborious. Therefore, we present a novel weakly-supervised segmentation method that only requires lobe-wise severity scores, such as routinely reported using the CT severity scoring system, to supervise training. Using only these lobe scores, we aim to produce high-resolution lesion segmentation maps. Weakly-supervised segmentation (WSS) has been extensively studied in recent years, where reference standards can be provided using scribbles [9] , or surface points [20] . Both these approaches seek a trade-off between annotation efforts and location information in need. However, because COVID-19 CT abnormalities often have bilateral lung involvement with a peripheral and diffuse distribution [21] , annotating scribbles or extreme points could still be demanding. A less demanding approach is to label the entire image volume or only regions within a volume. Early WSS methods using image-level labels were based on multi-instance learning frameworks [17] and the expectation-maximization algorithm [15] . Compared to these early works, approaches based on class activation maps (CAMs) [1] , [7] , [23] , [25] significantly improved the segmentation performance on major benchmarks [23] . CAMs were originally generated at low resolution, usually from the features before global pooling or fully connected layers in training a classification network. Low-resolution CAMs do not provide local details. To obtain high-resolution activation maps, MS-CAM [13] exploited both low-and high-resolution convolution features in a multi-scale framework. High-resolution CAMs may produce segmentation maps with local details. However, they still do not necessarily align with object segmentation, because naturally CAMs only reflect discriminative regions responsible for classification. This may causes certain objects or parts being ignored in the CAM, as long as the classification was made correctly. To mitigate this issue, research efforts were made to expand or refine CAMs, often using the original CAMs as the initial object cues or seed regions. Wei et al. [25] proposed to progressively erase already-found object maps and force the network to discover new and complement CAM regions at later runs. A similar iterative approach was applied on lesion localisation in color fundus images [5] . A seeded region growing module was proposed in [7] to expand CAMs to cover the complete object boundaries. AffinityNet [1] exploited inter-pixel affinities as the transition probability matrix and applied random walks to expand and refine CAMs. Instead of relying on CAMs, we obtain object cues based on dense regression activation maps (dRAMs) by training a segmentation network for regressing the per-lobe lesion percentage. dRAMs can provide high-resolution segmentation maps. Moreover, regression training allows the network to be aware of the object size, which was not done in previous work that relied on training with categorical labels. As described in [11] , lobe-wise severity scores provided by radiologists can be translated into intervals of lesion percentage per lobe. We propose a novel interval regression loss to enforce the predicted lesion percentage to fall in a particular range to use these intervals as reference standards. Furthermore, we introduce an attention module for revising dRAMs, trained together with the regression task. This refinement module updates each location in dRAMs using information from its local neighbors, according to voxelwise affinities embedded in convolution features. This process mimics the random-walker in [1] and seeded-region growth in [7] , but contrary to those approaches, it was trained end-to-end in our framework. Our key contributions are as follows: 1) we propose a lesion segmentation framework that produces high-resolution segmentation maps using only lobe-wise labels for training; 2) we convert the lesion segmentation problem to regression of the per-lobe lesion percentage. The regression problem is solved using a proposed interval regression loss. These ideas are generic and can be extended to other WSS problems; 3) we refine the dense regression activation maps using a proposed attention neural network module trained together with the main regression target and show this obviates the need for ad hoc post-processing steps. Several recent works on COVID-19 lesion segmentation attempted to reduce the demand for voxel-level supervision in training. Fan et al. [3] proposed a semi-supervised training strategy that requires a few labeled images to train the initial segmentation model and leverages primarily unlabeled data to fine-tune the model progressively. Laradji et al. [10] proposed to use point-level labels in an active learning schema to generate lesion segmentation maps. Yao et al. [28] superimposed synthesized lesions on healthy CT scans to train networks to separate lesions from other structures. Xu et al. [27] relied on scan-level labels and fractional voxel-level labels to train a generative adversarial learning framework for segmenting COVID-19 lesions. Wang et al. [22] proposed to train a binary classifier based on the presence of COVID-19 on CT scans and used the classifier to generate CAMs for detecting lesions. Our approach is closely related to CAMs-based WSS approaches in three building blocks: 1) the generation of CAMs by training a convolution neural network, often for a classification task. 2) regularization, to stabilize training. 3) CAM refinement, which is often needed because CAMs do not necessarily align with object boundaries. In these three aspects, our approach generates dense class regression maps (dRAMs) by training a segmentation network towards a regression target. Meanwhile, we introduce selfsupervised learning by equivariant regularization [23] for improving the consistency of extracted dRAMs over various affine transformed input images. In terms of the refinement, our method is motivated by [1] , which revised CAMs by a random-walker in post-processing steps where the transitional probabilities were derived from learned local affinities. A similar but end-to-end solution can be found in [23] where an attention-based neural network module was proposed to capture global affinities for refining CAMs. In our method, we capture affinities via an attention module inspired by [23] but computed within a local neighborhood, similar to that in [1] . We selected CT scans from patients who presented at the emergency wards of the Radboud University Medical Center, the Netherlands, from March to September 2020. Patients were referred for CT imaging because of suspicion of moderate to severe COVID-19 pneumonia. The ethical review board approved the retrospective and anonymous collection of this data (Radboudumc CMO2016-3045, Project 20027). All CT scans were obtained with a low-dose thin slice protocol without administration of contrast (details in [11] ). Following the guidelines of the Dutch Radiological Society [18] , the radiology report for each scan contained CO-RADS and lobe-wise severity scores. CO-RADS 1 is defined as a scan that is normal or has non-infectious etiologies, and thus a very low level of suspicion for COVID-19. CO-RADS 2 indicates the CT-scan has features typical for infections other than COVID-19. CO-RADS 3 indicates equivocal findings: features compatible with COVID-19 but also with other diseases. CO-RADS 4 and 5 indicate a high and very high level of COVID-19 suspicion, respectively. CO-RADS 6 was given to scans from patients already known to be positive for COVID-19 with reverse transcription-polymerase chain reaction (RT-PCR) tests at the time of reporting. Lobe-wise severity scores indicate the extent of lobar involvement of the COVID-19 infection. A score from 0 to 5 is assigned to each lobe according to the visually assessed lesion percentage of that lobe. The mapping between lobe-wise severity score and lesion percentage per lobe can be found in Table I (b). We used lobe severity scores to generate the weak labels in training our algorithm. 1) Data Selection and Partitioning: We randomly selected 390 subjects (shuffled into 300 for training and 90 for testing). This selection included all subjects that were available when this project started. A single scan was used for each subject. Thirty subjects in the training set were used as the validation set during model development to prevent over-fitting. The distribution of CO-RADS and lobe severity scores are provided in Table I (a and b) . 2) Reference Standard: For evaluating our method, lesion segmentation references on 90 test scans were obtained from Thirona (Nijmegen, the Netherlands), a medical image analysis service company specializing in chest CT analysis. First, lung parenchyma regions with a higher attenuation were identified by thresholding and morphologic operations. Next, automatic methods were used to suppress vessels and airways. Following the radiology report, a certified image analyst with at least one year of experience corrected the segmentations obtained by the automatic methods. The analysts also labeled segmented lesions into the ground glass, consolidation, and mixed type, to allow us to evaluate segmentation performance for different lesion subtypes. The analyst could consult a radiologist in cases of doubt during the annotation process. The overview of the proposed weakly-supervised lesion segmentation framework is shown in Fig. 1 . We train a regression network to predict the lesion percentage per lobe, and in the process, we generate the dense regression activation maps (dRAMs). The network is trained using an interval regression loss because the exact lesion percentage per lobe is not available for training. Instead, we only know the range of lesion percentage from lobe-wise severity scores, as given in Table I (b) . For example, radiologists would label a lobe with a severity score of 1 when the estimated lesion percentage in that lobe is between 1% and 5%. All lobe-wise severity scores were translated into lesion percentage intervals. Before using these intervals as the regression target, we calibrate the lesion percentage intervals using a candidate proposal technique to reduce the annotation errors. Moreover, equivariant regularization is used to improve the consistency of dRAMs among various affine-transformed input images. Finally, dRAMs are refined in an auxiliary training task using a proposed attention mechanism in addition to the regression and regularization. The following subsections elaborate on each of these steps. 1) Low-level features for lesion candidate proposal: Lowlevel features were commonly used in weakly-supervised semantic segmentation. For example, Wei et al. [24] used lowlevel features-based saliency maps as the reference for training the initial segmentation network. We propose to isolate high attenuation areas in the lung by applying Otsu's threshold [14] separately for each lobe. Meanwhile, we coarsely suppress vessels via vessel enhancement filtering [4] . The high attenuation lung regions after vessel suppression are marked as candidate lesions. 2) Regression training for generating dRAMs: The dense regression activation map is generated by training a regression network for predicting the lesion percentage per lobe. We adopt the 3D U-Net [2] as the regression network (shown in Fig. 1 (a) ). The 3D U-Net has three down-sampling layers in the encoding path, and each layer consists of two convolutions and a max-pooling operation. Following the down-sampling path, two more convolutions are used to double the number of convolution filters. In the up-sampling path, three layers are used to reconstruct the resolution, and each contains one trilinear interpolation, followed by two convolutions to reduce the interpolation artifacts. Convolution kernels have 3 × 3 × 3 kernel size, a stride of 1 voxel, and zero padding. At the final upsampling layer, feature maps are dense (having the same resolution as the network input). Then a regression head with a single 1 × 1 × 1 convolution and a softmax layer is used to output a C-channel dense prediction map. The regression network takes a lobe chunk image as the input, which is cropped around each segmented lobe and resized to have a fixed spatial size D × W × H, where D, W , and H are the number of slices, width, and height in the input lobe chunk image. The pulmonary lobe segmentation was performed using a publicly available algorithm [26] . For each lobe chunk input, areas outside the lobe of interest are masked out. Denoting the lobe chunk input image as I and the regression network as F (·), forwarding I through F (·) produces a C×D×W ×H dimensional dense prediction map F (I), referred to as the dense regression activation map (dRAM). In this paper, C channels are channels corresponding to the background, vessels, and abnormal regions in ascending order. For any channel, we can compute the per-lobe percentage of the object corresponding to the channel by simply average-pooling F (I) over all voxels within the lobe (lobe-wise mean pooling), assuming dRAMs are segmentation maps. Denoting the lobe-wise mean pooling as P (·), the parameters of F (·) can be trained given a target percentage range (r l , r u ) via the interval regression loss L IN T : Note that P (F (I)), r l , and r u are float values between 0 and 1 representing a percentage. L IN T can also be interpreted as the quadratic version of the piecewise linear loss function that The initial lesion percentage range (r * l ,r * l ) translated from lobe-wise severity scores (See Table I (b) for the definition) may contain measurement errors due to inaccurate estimation of lesion percentage when annotating severity scores. To reduce these measurement inaccuracies, we first calculate an estimation of per-lobe lesion percentage p * using candidate lesions (III-A1) as the volume of the candidate lesions divided by the volume of the corresponding lobe. Then we use p * to calibrate the initial range via: After the calibration, r l and r l are aware of the percentage of candidate lesions. The calibration can avoid overestimating the percentage of lesions during labeling lobe severity scores when both r * l , r * l or one of them are beyond the percentage of candidate lesions. We added 5% tolerance for the possible errors in lesion candidate proposals (p * ± 0.05). 3) Equivariant regularization: Training on weak labels may lead to a trivial segmentation solution because no location information is given. Therefore, regularization techniques are commonly used to stabilize training. Wang et al. [23] introduced an implicit equivariant constraint that enforces CAMs produced by an affine-transformed input to be similar to the affine-transformed CAMs produced by the original input. Denote a predefined spatial affine transformation as T (·), the input lobe chunk image as I and F (I) as the dRAM. Equivariant regularization loss L ER can be formulated as min: ||F (A(I)) − T (F (I))|| 1 . Equivariant regularization introduces the self-supervising correspondences among affine-transformed images. The affine transformations used in this work combine random resizing ([80% ∼ 120%] of size on each axis) and rotation of 90, 180, or 270 degree at randomly selected axes. Without having the exact lesion percentage per lobe, dRAMs trained by enforcing the predicted percentage to fall into a range may not suffice to delineate lesions accurately. To further improve the segmentation performance, we introduce a refinement step. First, we generate voxel-wise pseudo labels t * , where t * is one-hot encoded in a C × D × W × H dimensional matrix. Lesion labels take the overlapping regions between the dRAMs and the candidate lesions (III-A1). Vessel labels are detected vessels from the candidate proposal step. The remaining voxels are considered the background. Because these pseudo labels are generated using automatic methods and may contain errors, we adopt a bootstrapping loss [19] as the refinement loss, which minimizes where C is the number of classes (3 in our case, including background, vessel, and lesion), t c is the pseudo reference for the class label c and z c is the bootstrapping target produced by the network output. q c is the softmax probability of assigning a voxel into the class c. β is set to 0.9. The idea of this loss is to leverage the knowledge learned during training to provide hints of the true labels. The refinement loss and regression loss are trained simultaneously along with the equivariant regularization loss. The total loss is the weighted sum, where the regression loss is weighted 2.0 and others are weighted 1.0. 5) Attention-based dRAM refinement: To further improve the quality of dRAM, we apply an attention module for refining dRAMs. First, features from the first and second layers of the regression network are squeezed to have l channels via a 1 × 1 × 1 convolution. Before squeezing, features are detached from the back-propagation, such that refinement training does not update these features. Second, squeezed features are resized to the shape of the input image and concatenated with the input image. Denote this concatenated feature map as x, x ∈ R (1+2l)×D×W ×H . We use x to compute the local voxel-wise affinities in the dRAM y, y ∈ R C×D×W ×H (C classes are 0 as background, 1 as vessels, 2 as lesions). The affinity a(x i , x j ) between two locations i and j in y can be computed via a gated embedded Gaussian function a(x i , x j ) = e max((W θ xi) T (W φ xj ),0) , where W θ and W φ are linear transformations to project x into a l dimensional subspace. We eliminate weak affinities using the gate operation by max(·, 0). For each location i, only local voxel-wise affinities are measured between i and surrounding locations j in a 3 × 3 × 3 spatial window taking i as the center within a connectivity of 2, resulting in a total of 18 neighbors. Computing local affinities for all locations in x produces an attention map A, A ∈ R D×W ×H×18 . For each location, local affinities are normalized within its neighbors Ωj. The normalizing factor ζ(x) can be computed by a summation over the neighboring locations j as ζ(x) = j∈Ωj a(x i , x j ). In matrix form, this normalization is equivalent to applying a softmax function over the last dimension on A. To revise the dRAM y, we linearly project y into an l dimensional subspace. Then we apply matrix multiplication between the projected dRAM and the attention map A to generate the revised dRAM. This matrix multiplication can be seen as selectively collecting information from its local neighbors for each location in y. The impact from local neighbors is determined by the voxel-wise affinities reflecting similarities in terms of low-level image semantics. Finally, we project the revised y back to the input space using a linear transformation r(·). The process of refining the dRAM y at the location i can be formulated as follows: This attention module, in graph terms, defines a message passing graph where a node (voxel location) connects to its neighbors within a 3 × 3 × 3 grid using connectivity of 2. And to propagate information among the nodes, we simply matrix multiply the node embeddings (the dRAM y) with the attention map (A) as 5. Considering A as the probability transitional matrix, this attention module acts similarly to the use of random-walks based on local affinity matrix for refining CAMS as used in [1] . Our method could also have an effect similar to the deep seeded-region growth method proposed in [7] , where the authors adopted deep features to define the pairwise similarities between neighboring locations and seeds were defined based on CAMs. The goal of these approaches, in graph terms, is to propagate information among the nodes. Our method enables end-to-end optimization process, whereas random-walks in [1] and seeded-region growth in [7] were used as a separate step and not included in the neural network training. Training, validation, and testing of each experiment were carried out on a machine with an NVIDIA A100 with 40 GB GPU memory. Train, validation, and test data split can be found in Table I . Training each weakly-supervised method took 30 hours, stopped at 200 epochs. Weakly-supervised methods were implemented using Python 3.8, and the Pytorch 1.7.1 library [16] . Model parameters were initialized according to [6] and were optimized using stochastic gradient descent with a momentum of 0.9. The initial learning rate was set to 10 −5 . For both training and testing of WSS methods, the size of the lobe chunk input images was 80 × 80 × 80 cropped from the input scan resampled to 1.4 millimeters in isotropic spacing. Intensities in input scans were clipped into the range [−1000 ∼ 400] HU before re-scaling into [0 ∼ 1]. Dense activation maps for separate lobes were tiled together to form a scan-level activation map in the test phase. Then segmentation prediction was obtained by assigning labels with the maximum activation. For the attention module in the proposed method, we set the dimension of the projected subspace to 8 (l=8, refer to III-A5). Vessel enhancement filter used for detecting vessels in finding candidate lesions (III-A1) was implemented using the Gaussian scales of 0.8, 1.0, 1.5, 2.0, and 4.0 millimeters, with α equals to 0.5, β as 0.5, and γ of 15 as correction constants. The fully-supervised experiment was based on the official nn-UNet [8] implementation 1 . We limited the nn-UNet to only operate on 3D full-resolution mode, omitting 2D, cascading, and model ensemble options. We randomly selected 85 scans from the training dataset of 300 scans (Refer to II-1) and let analysts manually labeled lesions as the foreground and the background following the same protocols as annotating the test set. We splitted 85 scans into 65 for training and 20 for validation for training the nn-UNet method. The nn-UNet required training for two full weeks before reaching the required 1000 epochs (Detailed self-configured hyperparameters listed in the appendix). The Dice coefficient (DSC), absolute percentage difference (APD), the absolute difference in surface to volume ratio (SVRD), false discovery rate (FDR), and true positive ratio (TPR) were used for measuring the segmentation performance. Given the predicted lesion segmentation map X and the segmentation reference Y , DSC measures the degree of overlap between the two as: where ∩ indicates the intersection between X and Y . V ol(·) computes the volume size of the underlying region. Using the same notations, APD measures the absolute difference of lesion percentage per lung between the prediction X and the reference Y as: given the volume of the lung as V ol(Lung). APD serves as a volume-based measure. The surface-to-volume ratio (SVR) measures the irregularity or compactness of a shape. With the same object size, the ball shape has the smallest SVR, and SVR grows larger when object size increases. SVRD measures the absolute difference in two shapes in terms of the shape irregularity as: where X denotes the predicted lesion mask, Y is the reference mask, and Sur(·) computes the surface area of the underlying region. SVRD serves as a shape-based measurement. FDR is a precision-based measure that computes the number of falsely-labeled voxels divided by the total number of voxels in prediction as: between the prediction X and the reference Y . TPR, or recall, is a measure that indicates the amount of lesions that are detection in the segmentation, computed as: between the prediction X and the reference Y . We compute TPR for each lesion subcategory, namely consolidation, ground-glass, and mixed types. In addition to segmentation measurements, we also evaluate the classification performance in predicting lobe severity scores. We translate the predicted lesion percentage per lobe computed as a ratio between segmented lesions and lobes to the lobe severity score using the definition in Table I(b) . We measure classification accuracy (ACC) to count the number of correctly predicted lobes divided by the total number of lobes in terms of lobe-wise severity. Also, a linearly weighted kappa is computed for measuring the degree of reader agreement using R software package (version 3.6.2; R Foundation for Statistical Computing, Vienna, Austria). Besides the quality measurements, the computational complexity was measured by counting the number of Multi-Adds operations (MAC) and the number of network parameters. We denote the regression training with only interval regression loss to obtain segmentation maps as the dRAM method. The proposed method is the extension of dRAM where we trained the regression network with interval regression loss and added equivariant regularization plus refinement loss. We compare the proposed method and dRAM method with the CAM-based methods using the same network. The regression network can be turned into a classification network by replacing the regression head with a standard classification head. Given the dense feature maps before the regression head, the classification head computes a mean pooling of dense features per lobe and reshapes the pooled features according to the number of classes using a linear layer. With the classification head, the network can be trained using cross-entropy loss for classifying the presence of lesions. The categorical labels for training the classification network are binary and determined as positive if lobe severity score is non-zero, negative otherwise. CAMs are then generated by multiplying dense feature maps with class-specific weights from the linear layer at the end of the classification. We denote this approach as dCAM because of the use of dense features. The dCAM method is analogous to the MS-CAM method proposed in [13] because U-Net dense features are intrinsically multi-scale, due to the skip-connections between down-sampling and upsampling layer hierarchies. Both dCAM and MS-CAM were designed to generate high-resolution CAMs for capturing local details. In order to compare with CAM-based methods where CAMs were generated at a low resolution, we skipped layers in the upsampling path of the 3D-UNet, transforming the classification network to a network structure typical for generating CAMs where features are convolved in progressively down-sampled resolution. This slimmed classification network is trained for binary classification as that in dCAM, and CAMs are resized by trilinear interpolation to provide lesion segmentation maps at the input resolution. We denote this method as CAM. CAMs in the test phase are rescaled to be in the [0,1] range separately for each lobe by eliminating negative values using a Rectified Linear Unit (Relu), and subtracting by the minimal and dividing by the maximum. To generate segmentation maps, rescaled CAMs are binarized using Otsu's threshold. Because raw CAMs produce poor segmentation performance, we designed a simple yet efficient post-processing technique. In post-processing, we apply the lesion proposal technique (described in III-A1) on CAMs by excluding regions outside candidate lesions. The best performing method based on CAMs is the dCAM with post-processing, which is referred to as the baseline method in this paper. Quantitative results of CAM, dCAM, dRAM, and the proposed method are shown in Table II . Without equivariant regularization and refinement loss, training with only interval regression loss (dRAM) reaches 59.90% in DSC, outperforming all CAM-based methods by a large margin. Using equivariant regularization and refinement loss, the proposed regression training method achieved 70.24% in DSC, the best weakly-supervised method in comparison. Not surprisingly, a fully-supervised method (nn-UNet) still outperforms weaklysupervised methods, but for several metrics differences with the proposed method are small. Relatively large standard deviations in DSC of all methods are due to the high variation in lesion volume in different cases. The post-processing improves the performance of the CAMs-based method, from 40.42% DSC in dCAM to 48.63% after post-processing. This is because post-processing reduces a large amount of falsely detected regions in CAMs. In Fig. 2 , CAMs encloses mostly healthy lung parenchyma when there II: Segmentation and classification results using CAM, dCAM, dRAM, and the proposed methods on the test set. nn-UNet indicates the results from the fully-supervised training. Post-processed results are suffixed by '-p'. We use dCAM with post-processing as the baseline method (dCAM-p). Boldface denotes the best result among all WSS methods. The computational complexities in the right column are reported for the worst-case scenario. DSC scores are shown in mean ± standard deviation. is only a focal lesion (2nd row, 4th column) due to its lowresolution nature. dCAMs, on the other hand, include a large number of false positives in the perivascular area (3rd row). dRAM does not benefit from post-processing, possibly because post-processing may create small disconnected structures by intensity thresholding, which are rarely seen in manual segmentation. This phenomenon is quantitatively viewed as the dramatic increases in SVRD between raw and post-processed methods (visual examples of post-processing shown in the appendix). In terms of shape and volume-based measurements, the proposed method reaches 3.52 APD and 3.58 in SVRD, representing a low error rate in predicting lesion volumes and shape compactness compared to manual segmentation. We also notice that the proposed method has a higher voxel-level precision resulting in a lower FDR, but a substantially lower recall (TPRs are lower in all lesion sub-types) than the fullysupervised method. In terms of predicting lobe-wise severity scores using the segmentation produced by the listed methods, the proposed method achieved the best classification accuracy at 47.33% (confusion matrix attached in the appendix), approaching the fully-supervised method at 50.67%. We also computed the linearly weighted kappa for the baseline, the proposed, and the fully-supervised methods to measure the agreement against a radiologist. Visual inspection of the generated activation maps (Fig. 2 ) reveals that blob-like CAMs (2nd row) do not recognize object boundaries as they were was created at low-resolution features. dCAMs (3rd row) produces high-resolution activation maps using dense features, but still suffer from a large amount of over-segmentation and under-segmentation. This can be seen, for example, in the right upper lobe of the first subject (3rd row, 1st column), The activation map in dCAM misses a small pleural lesion, whereas many vessels and opacities near vessels in other subjects are falsely detected. These errors are greatly reduced in the proposed method (4th row). The main reason is that the proposed method was trained to quantify lesions in the image through an interval regression loss. Supplying information about the object size allows the model to learn and 16.32M GPU memory (the memory consumption for the proposed attention module is negligible). In terms of the runtime speed, the proposed method processes each scan at testing time only 2 seconds on average, while the nn-UNet takes roughly 480 seconds to process a scan due to the heavy build-in pre-processing. 1) Ablation study on dRAM-based methods: We conducted an ablation study to evaluate the effectiveness and computational complexity of each component in the proposed WSS framework. These components are the equivariant regularizer (ER), refinement loss (REF), and the proposed attention module (AT). Quantitative results are shown in Table III . dRAM trained only with interval regression loss achieves 59.91% in DSC due to the lack of voxel-level cues. By introducing the dRAM refinement process, we greatly improve DSC to 64.62% without imposing additional computational burden. The reason is that suppression of vessels and falsedetected regions beyond the intensity threshold encourages the network to discover new areas to satisfy the regression target during training. Self-supervised training using the equivariant regularizer contributes to an improvement from 59.91% DSC to 61.74% DSC, with a doubled computation cost (925.84 GMacs in the worst-case scenario) during training. The high computation cost is due to the siamese networks taking both original and affine-transformed input images in equivariant regularization. The proposed attention module alone boosts the segmentation performance from 59.91% DSC to 62.17% DSC. The module updates activation maps at each location selectively using its neighboring information. This is similar to popular CAM-refinement techniques, i.e., conditional random field [7] , random walkers [13] , and seeded-region growth [7] . Fig. 3 shows the activation maps before and after the attention module in two input images during training. We notice that neighboring information is gathered and weighted more towards high attenuation areas. This makes low attenuation regions (probably healthy lung parenchyma) less involved in further computation, suppressing false alarms and forcing the network to look for other regions for finding lesions. With the equivariant regularizer, refinement loss, and the attention module, the proposed method ultimately reaches a DSC of Fig. 3 : Visualization of dense feature maps before (2nd col) and after (3rd col) the attention module in two input volumes in coronal views (1st col) during training. We use color map jet for this plot. 70.24%. Even though the computation cost of the proposed method is high, we argue that the large computational burden only occurs during training. For testing, all listed methods in the ablation study are the same. All listed approaches do not differ much in terms of classification performance, where the proposed method achieves the best accuracy at 47.33%. This study aims to examine if the segmentation method performs worse or better for a specific lesion subtype. We computed the correlation between the volume ratio of each subtype and the segmentation performance (in DSC,%) on every test image. Each point in Fig. 4 tallies the correlation on each test image, where the top-left image shows the distribution of DSC scores for the baseline, the proposed, and the fully-supervised methods. The other images rank the percentage of each lesion subtype (the volume of each lesion subtype divided by the overall lesion volume) at x-axis, and y-axis shows the DSC scores. The top-right subplot shows that all methods perform well on mixed lesions because points are more concentrated at the top right corner. However, the performance is relatively weak for the ground-glass and consolidated lesions because points are densely distributed at the top-left corner on the subplots corresponding to the ground-glass and consolidation subtypes. As shown in Fig. 5 , the result from the baseline method (2nd row) exhibit substantial under-segmentation in pleural lesions (2nd, 3rd, 5th, 6th columns). Also, the baseline method often over-segments perivascular regions (3rd, 4th columns). The proposed method (3rd row) generally performed well on lesion segmentation. However, compared with the nn-UNet results, the proposed method under-segments small pleural groundglass lesions (3rd and 5th column). One reason is that the lobe-wise severity scores only represent an interval of the perlobe lesion percentage and not the exact percentage, which potentially allows the network to tolerate certain mistakes. On the other hand, ground-glass opacities may create challenges in visual recognition. This challenge may also cause measurement errors for radiologists in labeling severity scores, further contributing to confusion regarding ground-glass opacities in our methods. V. DISCUSSION AND CONCLUSION We proposed a novel weakly-supervised segmentation method. This method can train a segmentation network using only severity scores provided for individual lobes, where these scores correspond to a range of percentages of affected regions in these lobes. From only these lobe scores, the network can generate dense regression activation maps (dRAMs). These dRAMs were refined by the proposed attention module using low-level semantics. The proposed attention module enriches the semantic representation at each voxel based on its local neighbors (affinities). The proposed method achieved significant improvements in segmentation performance compared with the baseline approach. In terms of the prediction of lobe severity scores, the proposed method reached a moderate agreement with the scores assigned by the radiologist, while the baseline method only reached a fair agreement. Our results showed that the proposed model sometimes under-segments small groundglass lesions in the subpleural region of the lung periphery. However, as weak labels are cheap to collect, more advanced approaches can be built upon our model using our methods as the initial seed for interactive (e.g., adaptive learning scenarios) or iterative refinement (e.g., knowledge distillation). The proposed method is generic and can be easily adapted to other weakly-supervised segmentation problems if specific object statistics are given and can be used as the regression target. Visually assessing the amount of affected regions is common in radiology, and therefore we believe that the proposed weakly-supervised segmentation framework can be used for many segmentation problems in medical imaging, where automatic segmentation is often used for quantification analysis. In these scenarios, our framework can directly use visually assessed quantification results from radiological scoring systems as the regression target. B. Effect of post-processing Fig. 6 : Post-processed segmentation results of CAM (1st row), dCAM (2nd row), and dRAM (3rd row) methods in three different test subjects listed column-wise. Segmented contours from raw activation maps were drawn in , post-processed activation maps in , and reference masks in . Contours were drawn in orders of raw, post-processed, and the reference. Best viewed in color. C. Confusion Matrix of the proposed method in predicting lobe-wise severity scores Fig. 7 : Confusion matrix of the proposed method in predicting lobe-wise severity scores on the test set, rows represent predicted score and columns show the target scores. Learning pixel-level semantic affinity with image-level supervision for weakly supervised semantic segmentation 3D U-Net: Learning dense volumetric segmentation from sparse annotation Inf-Net: Automatic COVID-19 lung infection segmentation from CT images Multiscale vessel enhancement filtering Iterative augmentation of visual evidence for weaklysupervised lesion localization in deep interpretability frameworks: application to color fundus images Delving deep into rectifiers: Surpassing human-level performance on imagenet classification Weakly-supervised semantic segmentation network with deep seeded region growing nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation Scribblebased hierarchical weakly supervised learning for brain tumor segmentation A weakly supervised region-based active learning method for COVID-19 segmentation in CT images Automated assessment of COVID-19 reporting and data system and chest CT severity scores in patients suspected of having COVID-19 using artificial intelligence The clinical and chest CT features associated with severe and critical COVID-19 pneumonia MS-CAM: Multi-scale class activation maps for weaklysupervised segmentation of geographic atrophy lesions in SD-OCT images A threshold selection method from gray level histograms Weakly-and semi-supervised learning of a deep convolutional network for semantic image segmentation Pytorch: An imperative style, high-performance deep learning library From image-level to pixel-level labeling with convolutional networks CO-RADS -a categorical CT assessment scheme for patients with suspected COVID-19: definition and evaluation Training deep neural networks on noisy labels with bootstrapping Weakly supervised segmentation from extreme points Radiological findings from 81 patients with COVID-19 pneumonia in Wuhan, China: a descriptive study A weakly-supervised framework for COVID-19 classification and lesion localization from chest CT Self-supervised equivariant attention mechanism for weakly supervised semantic segmentation STC: A simple to complex framework for weaklysupervised semantic segmentation Object region mining with adversarial erasing: A simple classification to semantic segmentation approach Jean-Paul Charbonnier, and Bram van Ginneken. Relational modeling for robust and efficient pulmonary lobe segmentation in CT scans Weakly-supervised framework for COVID-19 lesion segmentation Label-free segmentation of COVID-19 lesions in lung CT