key: cord-0786376-jzt8j4wp authors: Liu, Caixia; Zhao, Ruibin; Xie, Wangli; Pang, Mingyong title: Pathological lung segmentation based on random forest combined with deep model and multi-scale superpixels date: 2020-08-07 journal: Neural Process Lett DOI: 10.1007/s11063-020-10330-8 sha: c78c04b491986bb79c46cc9cf38f561d2a9cf7a9 doc_id: 786376 cord_uid: jzt8j4wp Accurate segmentation of lungs in pathological thoracic computed tomography (CT) scans plays an important role in pulmonary disease diagnosis. However, it is still a challenging task due to the variability of pathological lung appearances and shapes. In this paper, we proposed a novel segmentation algorithm based on random forest (RF), deep convolutional network, and multi-scale superpixels for segmenting pathological lungs from thoracic CT images accurately. A pathological thoracic CT image is first segmented based on multi-scale superpixels, and deep features, texture, and intensity features extracted from superpixels are taken as inputs of a group of RF classifiers. With the fusion of classification results of RFs by a fractional-order gray correlation approach, we capture an initial segmentation of pathological lungs. We finally utilize a divide-and-conquer strategy to deal with segmentation refinement combining contour correction of left lungs and region repairing of right lungs. Our algorithm is tested on a group of thoracic CT images affected with interstitial lung diseases. Experiments show that our algorithm can achieve a high segmentation accuracy with an average DSC of 96.45% and PPV of 95.07%. Compared with several existing lung segmentation methods, our algorithm exhibits a robust performance on pathological lung segmentation. Our algorithm can be employed reliably for lung field segmentation of pathologic thoracic CT images with a high accuracy, which is helpful to assist radiologists to detect the presence of pulmonary diseases and quantify its shape and size in regular clinical practices. Pulmonary disease is one of the major causes of morbidity and mortality around the world [1, 2] . For example, the recent global outbreak of COVID-19 has killed tens of thousands of people in just a few months. Early diagnosis of pulmonary disease with computed tomography (CT) technique is crucial for making treatment decisions. In non-invasive detection B Mingyong Pang panion@netease.com 1 Institute of EduInfo Science and Engineering, Nanjing Normal University, Nanjing, China and diagnosis of pulmonary disease, accurate lung segmentation is often a prerequisite for assessing the disease severity, it ensures that disease detection is not confounded by regions outside lungs [2] . However, inner structures of thoracic CT images are usually various with different textures and pixel densities. Additionally, intensities of pathological images are inhomogeneous and it is difficult to provide a reliable generic solution for a wide spectrum of lung abnormalities, which cause difficulties in lung segmentation. Lung segmentation refers to the computer-based process of identifying the boundaries of lungs from surrounding thoracic tissue on CT images [3] , see Fig. 1 for a reference. In this paper, the key issue we want to tackle is to segment pathological lungs from thoracic CT images accurately overcoming external distractions of lung diseases and abnormalities. Thus, we proposed a novel pathological segmentation algorithm based on random forest, deep convolutional network, and multi-scale superpixels. Our contributions in the paper are: • We propose a novel pathological lung segmentation algorithm combining three principal processes: feature extraction, classification fusion, and contour correction, which can generate more complete segmentations. • We put forward an effective classification fusion method based on a fractional-order gray correlation, which can produce more accurate fusion results in multi-scale classifications. • We present a new lung segmentation refinement approach based on a divide-and-conquer strategy of contour correction of left lungs and region repair of right lungs, which contributes to generating more accurate lung segmentations. The remainder of this paper is organized as follows: in Sect. 2, we introduce the related research in lung segmentation, and in Sect. 3, we make a detailed description of our lung segmentation algorithm. In Sect. 4, we provide a set of experimental results and in Sect. 5 and Sect. 6, we make a discussion and summarize our algorithm. The basic pipeline of our algorithm is illustrated in Fig. 2. Many methods have been proposed to segment lungs from thoracic CT images in the past decades. In [4] , three thresholding based approaches, connected threshold, neighborhood connected, and threshold level set were performed on lung segmentation. Prabin et al. [5] Fig. 2 The pipeline of our algorithm. A DCNN model is trained with samples captured from CT images and their ground truths. An input image is segmented into three groups of superpixels: P 1 , P 2 , and P 3 , with three different scales, respectively. The deep features, texture, and intensity features are extracted from each group of superpixels, which are classified with random forests (RFs). We fused the classification results (R 1 , R 2 , and R 3 ) and refined the fused segmentation result with a divide-and-conquer strategy and information propagation mechanism segmented thoracic CT images using a region growing algorithm along with a combination of supervised contextual clustering technique. Mansoor et al. [6] segmented pathological lungs from CT scans by combining region-based segmentation with a local descriptor based classification. Chen et al. [7] segmented pathological lungs from 3D low-dose CT images using an eigenspace sparse shape composition by integrating a sparse shape composition with an eigenvector space shape prior model. Revathi et al. [8] introduced a pathological lung identification system, where, FC was used for segmenting lungs with a diverse range of lung abnormalities, RF was then applied to refine the segmentation by identifying pathology and non-pathology tissues according to the features extracted from the gray-level co-occurrence matrix (GLCM), gray level run length matrices, histograms and so on. Soliman et al. [9] segmented pathological lungs from CT images based on a 3D joint Markov-Gibbs random field model which integrated the first-order visual appearance model, the second-order spatial interaction model, and a shape prior model. Hua et al. [10] used a graph search driven by a cost function combining the intensity, gradient, boundary smoothness, and rib information for pathological thoracic CT image segmentation. Hosseini-Asl et al. [11] proposed a nonnegative matrix factorization (NMF)-based pathological lung segmentation approach, which included three stages: removing image background from CT images by a conventional 3D region growing method, modeling visual appearance of the remaining chest-lung image with an NMF technique, extracting 3D lung voxels by a two-step data clustering and a region map cleaning approach. An automated lung segmentation algorithm was developed combining unsupervised and supervised techniques in [12] . The method combined an unsupervised MRF technique to provide an initial estimate of lung borders, a supervised texture analysis scheme based on an SVM classifier was applied on searching border regions and distinguishing lung tissues from their surrounding tissues. Liu et al. [13] segmented lungs from CT images with a random forest (RF) classifier, where texture and intensity features extracted from superpixels were taken as the input of RF Classifier. Meng et al. [14] presented a lung segmentation algorithm based on anatomical knowledge and a snake model. By setting the snake model's initial curve on human's rib edge in the thoracic CT image, their model could capture the concavities locating on lung boundaries well. Abdollahi et al. [15] segmented multi-scale initial lungs from CT images using a linear combination of discrete Gaussian approach and a Markov-Gibbs random field (GMRF) model. The initial segmentations were fused together using a Bayesian approach to obtain the final segmentation of lung regions. The potential benefits of deep learning techniques in image analysis have also generated remarkable results [16, 17] . Harrison et al. [1] presented a bottom-up deep-learning-based approach unaffected by any variations in lung shape. In the method, a deep model, progressive holistically-nested network, was used to produce finer detailed lung masks. Park et al. [18] employed a twodimensional U-Net for lung parenchyma segmentation. Anthimopoulos et al. [19] used a deep purely convolutional neural network for the semantic segmentation of interstitial lung disease (ILD) patterns. Lung CT images of arbitrary sizes were taken as inputs and produced corresponding label maps. In lung segmentation, manually crafted features such as shape, color, and/or texture play an important part in some algorithms. Nevertheless, the features are complementary in medical images and cannot make representations of high-level problem domain concepts [20] . Deep learning is often used to improve the descriptive ability in feature representation [21] [22] [23] . For example, Hong et al. [21] used deep neural networks to compute features for face pose estimation. Zhang et al. [22] extracted deep features with contractive autoencoders for unsupervised dimension reduction. Convolutional neural network (CNN) has an ability of representation learning, where input information is extracted and filtered layer by layer. The convolutional layers of a CNN can be seen as feature extractors, which generate local features of image patches in each layer, and the features are combined to produce a global deep feature vector in the last. Recent work demonstrated that 2D CT slices are expressive enough for segmenting complex organs [24] . Consequently, we fuse the deep features and low-level traditional features to character different regions of thoracic CT images and extract lung regions from CT images with RF according to the fused features. Furthermore, the lung segmentation is refined by contour correcting and region repairing. The superpixel algorithm [25] can group pixels into perceptually meaningful atomic regions with similar features, such as intensity, texture, and so on. Approximately equally-sized superpixels with boundaries aligning to local image edges are more suitable for being taken as classification units than isolated image patches. Consequently, we take superpixels as classification units of deep convolutional neural network (DCNN) and RF classifiers for preserving lung contours well. Our lung segmentation algorithm mainly includes four steps: superpixel segmenting, feature extracting, RF classification, and segmentation refining. After preprocessed with morphological filters [26] , thoracic CT images are segmented into a group of superpixels with the simple linear iterative clustering (SLIC) approach [25] . SLIC employs k-means clustering approach to efficiently generate superpixels with approximately equal sizes. The scales of superpixels in an image depend on a user-specified parameter, P. Assume the size of a thoracic CT image is Z , the scale of a superpixel is roughly √ Z /P. Hence, smaller P produces larger-scaled superpixels, which is helpful for segmenting inhomogeneous intensities images involving pathologic tissues with strong anti-noise interference ability. Conversely, bigger P generates smaller-scaled superpixels, which can capture detailed regions. Here, a thoracic CT image is segmented by SLIC with three Ps (P 1 , P 2 , P 3 ), respectively. As a consequence, we obtain three groups of superpixels with three different scales. We map the superpixels into the thoracic CT image and extract deep, texture, and intensity features from superpixels. CNN is one of deep neural networks with multilayer perceptrons and usually consists of an input and an output layers, as well as multiple hidden layers typically including a series of convolutional layers, pooling layers, fully connected layers, and normalization layers [27] . Here, we construct an 8-layer DCNN model for deep feature extraction: The outputs of the convolutional layers are all activated by a Sigmoid function, Three kinds of regions, lungs, pleural tissues, and backgrounds, are captured from thoracic CT images according to their ground truths. They are resized and used for training and testing of our DCNN model. Finally, the output of the last mean-pooling layer, a feature vector of 40 elements, is extracted and taken as our deep features. GLCM is a widely used texture statistical analysis and measurement tool. Its statistical characteristics are superior to those of fractal dimension, Markov model, Gabor filter and so on [28, 29] . We extract 46 features from GLCM, such as energy, contrast, homogeneity, correlation and so on. Some of them are defined as Energy = , where μ and σ are the mean and standard deviations of GLCM. Besides GLCM, moment invariants, ηs, have been extensively used to characterize the patterns of images in a variety of applications [30] . Hu [31] introduced seven moment invariants φ 1 ∼ φ 7 . Moment invariants are useful properties, which are robust to image scaling, translation, and rotation. Collectively, we obtain a total of 54 texture features after adding image entropy [32] into texture features. Since color information is mainly distributed in low-order moments, color moments have been proved an effective tool in representing the color distribution in images [33] . Stricker et al. [34] introduced three color moments to represent image color distribution including first-order moments (μ), second-order moments (δ), and third-order moments (s) defined as 3 , where p i, j stands for the ith color value of pixel i, N and μ respectively represent the total number of pixels in the image and the mean value of ith color channel. Here, we convert a CT image into a grayscale image and extract intensity features from it, and in total, we concatenate 97 features listed in Table 1 . We capture the maximum inscribed square patch from each superpixel as a region of interest (ROI), resize, and take them as inputs of our DCNN model to extract deep features. Combining with the deep, texture, and intensity features extracted from superpixels, Random forest (RF) [35] classifies the superpixels into three classes: lungs, pleural tissues, and image backgrounds. The lung segmentation results in the initial classification with an RF are usually incomplete due to the existence of pathologic lung tissues. Because superpixels with different scales can cluster regions with different scales and details, we adopt the RF to classify multiscale superpixels according to their features and fuse different segmentation results to get a relatively complete lung segmentation result. However, some image patches located at the same positions are assigned different classes in multiple classifications. In order to deal with the issue, we adopt a fusion technology based on a gray correlation algorithm instead of a simple and crude addition fusion. The similarity of two data series can be determined by their slopes in corresponding periods according to the theory of gray correlation. If the slopes of two series are equal or similar in each period, the two series have a large correlation coefficient and vice versa. Thus, image patch similarity in terms of grayscale and texture is calculated by the gray correlation of two vectors. Assume two data series are x (0) 0 (k), k = 1, 2, . . . n and x (0) 1 (k), k = 1, 2, . . . n , the similarity between them is calculated as follows: 1. Fractional accumulation [36] We reduce the noise affect by fractional accumulation which is expressed as where p q 0 < p q < 1 is fractional accumulation operator. Let C 0 p q −1 = 1, C k+1 k = 0, k = 0, 1, . . . , n − 1, and The purpose of initialization is to make each sequence comparable. Inverse accumulation is to find slops of the curve at each time point. Here, we use the absolute values of difference of adjacent data in data series to reflect the overall distribution trend of data series. The correlation coefficient is calculated by Eq. 7. 5. Fractional-order gray correlation degree calculation Fractional-order gray correlation degree of two data series is defined as Eq. 8. With the fractional-order gray correlation degree, we can distinguish any two image patches and assign them an appropriate class in classification. The segmentation fusion is described detailedly in Algorithm 1. Algorithm 1 Segmentation fusion based on fractional-order gray correlation. 1 : Input grayscale lung image G and corresponding multi-scale classification results M k (k = 1, 2, . . . , K ). 2: Capture regions R i assigned the same class, C i , (i = 1, 2, 3), in M k , M 2 , . . ., and M K , respectively: 1, 2, 3) in G into a 1D vector, Rv i . 4: Find regions, S j ( j = 1, 2, . . . m), belonging to different classes in M k , m is the number of regions: S j = M 1 ⊕ M 2 ⊕ · · · , ⊕M K . 5: for each sub-region S j do 6: Convert pixels in S j in G into a 1D vector, Sv j . 7: Intercept sub-vectors Rvs i with same length as Sv j from Rv i . 8: Calculate correlation degrees e i according to Eq. 8 between Sv j and Rvs i . 9: Find max e i and assign i to region S j . 10: Update M o according to S j . 11: Output final fused mask M o . In Fig. 4 , we set the numbers of superpixels, Ps, in SLIC method to 200, 300, and 500, respectively, which generates three scales of superpixels. Because lung tissues and image backgrounds have low intensities, which are very different from pleural tissues with high intensities, we group lung tissues and image backgrounds into one class and pleural tissues into another class in the output of our algorithm. By fusing the different classification results with Algorithm 1, we obtain a relatively complete segmentation.Additionally, because we use three small-sized Ps, the time cost is much less than that with a big-sized P. For example, in Fig. 4 , the time cost of the fusion algorithm with P 1 = 200, P 2 = 300, and P 3 = 500 is 0.68 times of that with P = 2000. At the same time, the fused segmentation is better than one-time segmentation. We obtain the initial lung segmentation results by extracting lung regions from the fusion results of RFs with morphological operations. Pathologic lungs usually have similar densities with surrounding tissues, which often leads to incomplete segmentation of lung regions. Additionally, the left and right lungs closed to each other usually connected in some segmentation results. In order to tackle these problems, we utilize a lung contour correction approach and lung region repair method to refine segmentations after lung separation, see Fig. 5 for a reference. As shown in Fig. 6 , in some pathological lung CT images, the right lung regions usually have larger defects and the left lung regions are relatively more complete. Accordingly, we adopt a divide-and-conquer strategy to refine segmentation results by combing contour correction of left lung contours and region repair of the right lungs. We first presented a contour correction approach based on SUSAN operator for left lung contour correction. Because the left lung and the right lung are approximately symmetrical and they usually have similar contours in a thoracic CT image, the right lung region is repaired supervised by left lung contours. It should be noted that the segmentation refinement methods for the left and right lungs can be switched in the other general images. Anyway, the first step of our segmentation refinement is to separate two connected lungs. In a thoracic CT slice sequence (I 1 · · · , I k , · · · I L ), left and right lungs are closed to each other and may connected from I k in a segmentation sequence (I k · · · , I k+1 , · · · I k+n ). Here, we utilize a lung separation line propagation approach to separate the connected lungs by using the separation line of its former one. The Smallest Univalue Segment Assimilating Nucleus (SUSAN) algorithm is a famous corner detection technique. In order to repair lung contour concaves caused by pathological abnormalities, we used a contour correction approach based on SUSAN operator [37] for lung contour correcting, which covers three stages: candidate corner detecting, validate corner filtering and corner connecting. We create a copy, M 1 , of lung mask M, remove the right lung region from M 1 and retain left lung. The main steps of contour correction are described in Algorithm 2. The left lung and right lung in a thoracic CT image usually have similar contours. Accordingly, instead of using the whole mask of the left lung to repair the right lung, we take a part of corrected left lung contour as the contour mask of the right lung. We divide the repair process into three stages: extraction of reference contour, repair of the right part, and repair of the left part of the lung. 3: Count the number of pixels having similar values with the central point p 0 in w: n( p0) = p∈w C( p, p 0 ) 4: Extract corner point according to Eq. 9. Calculate corner response function, R( p 0 ), of point p 0 . Point p 0 is marked as a candidate corner if n( p 0 ) less than a threshold g. where g=(r * r )/2. 5: Perform non-maxima suppression for every candidate corner to remove false corners. 6: Merge corners locating within a distance d. Here, we set d=4. 7: Remove the corner located at lung junction region. 8: Connect corner pairs with circle dilating method [13] . 9: Fill holes. 10: Output: lung mask M 1 with lung contour corrected. (4) Extract the middle segment (reference contour) of E l . b. Repair of right part (1) Flip E l horizontally, move it to the right until E l reaches the right boundary of right lung M 2 . (2) Stretch E l from its two endpoint according to its slope, respectively. Interstitial lung disease (ILD) is the leading cause of mortality and characterized by widespread fibrotic and inflammatory abnormalities of lungs [38] . The images used for training and testing our algorithm come from ILDs database [39] . It contains high-resolution computed tomography (HRCT) image series with pathologically proven diagnoses of ILDs. The images which are difficult to segment by conventional methods are used to test the performance of our segmentation approach. In the following, we first introduce metrics in our algorithm evaluation and then evaluate the influence of preset superpixel numbers. Finally, we perform our algorithm on a set of pathological lung CT images and analyze the experiment results. TNR, TPR, PPV, ACC, Error (Er) are used to evaluate our lung segmentation performance [40] . Four classical metrics [41] , over-segmentation rate (OR), under-segmentation rate (UR), Dice similarity coefficient (DSC) and Jaccard's similarity index (JSI) are also utilized for evaluating the performance of our algorithm. Larger values of TNR, TPR, PPV, ACC, DSC and JSI and smaller values of Er, UR and OR indicate more accurate segmentation of lung images. The preset superpixel number, P, affects the efficiency and accuracy of our segmentation algorithm greatly. In order to detect a suitable group of Ps, we randomly select 80 CT images affected with sarcoidosis to test our algorithm efficiency and accuracy with different Ps. We test four groups of Ps: G 1 (P 1 = 100, P 2 = 200, P 3 = 300), G 2 (P 1 = 300, P 2 = 400, P 3 = 500), G 3 (P 1 = 500, P 2 = 600, P 3 = 700), G 4 (P 1 = 700, P 2 = 800, P 3 = 900) and the corresponding average time costs are depicted in Fig. 8 . For example, the average time coats on P 1 =100, P 2 =200 and P 3 =300 are 14.2s, 21.8s, and 27.8s, respectively. The total time cost of G 1 is 64.93s. The segmentation is implemented in MATLAB environment on a computer with Intel i7-6500U CPU and 16GB of RAM. It can be seen, the time cost increases with the increase of P. In Fig. 9 , we show the performance of our algorithm with four groups of Ps. We can see that the segmentation accuracies in terms of DSC and JSI increase with an increase of Ps. Nevertheless, the DSCs are all above 97%. In order to reduce the time cost while keeping a high segmentation accuracy, P i is ranged from 100 to 700. We chose a challenging data set of thoracic CT images containing density pathologies in varying degrees of severity from ILDs database to test the performance of our lung segmentation algorithm: images affected with ground glass (GG), fibrosis (F), nodules (N), reticulation (R) and PCP (P). The number of the tested images in each type thoracic CT images ranged from 40 to 80, and we list the average performances of our algorithm in terms of TNR, TPR, PPV, ACC and Er in Table 2 , and DSC, JSI, OR and UR in Table 3 . It can be seen that the segmentation accuracy of R images has the highest accuracy such as TPR because of relatively homogeneous intensities in the images. Conversely, P images have the lowest accuracy among our segmentations due to the severe uneven distribution of gray scales. The average metric values (AVE) in terms of TNR, TPR, PPV, ACC and Er is shown in Fig. 10 , and average DSC, JSI, OR and UR in Fig. 11 . It can be seen that ACC and DSC are Fig. 12 . We randomly select 270 images from the above mentioned five types of images and the error distance defined as (F N + F P)/(F N + T P) is shown in Fig. 13 . Error distances are mainly concentrated in a range of (0, 0.2). The P images usually be destructed by pneumocystis pneumonia (PCP) and the intensities are inhomogeneous. Fig. 14 depicts a comparison of the segmentation accuracy of initial segmentation and refined segmentation of lungs in P images. It can be seen that the accuracies of the refined segmentation results in terms of TPR, ACC, DSC, and JSI are all higher than those of the initial segmentation results. In Fig. 15 , we show a group of segmentation results with some state-of-art methods [13, 18] . It can be seen that the other methods cannot deal with some intensity inhomogeneous images well. Conversely, our algorithm can segment more complete lungs from pathological lung CT images. In Fig. 16 , we compare our algorithm with the other methods on a set of CT images affected with fibrosis and PCP. The results show that our algorithm achieves higher accuracies in terms of DSC, JSI, TPR, and ACC, which implies the segmentation ability of deep learning networks [18] on pathological lung CT images is hindered by a limited size of a training set (1000). The segmentation accuracy of a single-scale RF method [13] is also smaller than ours on the whole. Our algorithm is relatively more robust on lung segmentation than other methods because of the multi-scale classification fusion and effective post-processing operations. The fusion algorithm can generate high segmentation accuracy. However, it may generate a little bit higher OR shown in Fig. 11 due to two main factors. One factor arises from superpixel segmentation. We take superpixels as the input of the RF classifier. In order to accelerate our algorithm, we use three Ps with smaller values, which generates a group of multi-scaled superpixels with the SLIC approach. However, a bigger superpixel is difficult to capture narrow regions and sometimes brings about high ORs, see Fig. 17b for a reference. Accordingly, OR of fusion result is high by simply adding the multi-scale segmentation results together. In our algorithm, the correlation approach can solve this tissue and reduce OR. Another factor lies in region repair of right lung supervised by the left lung contour, and the over-segmentation of the left lung may propagate to the right lung. In order to overcome the issue, we capture a middle segment of the left lung contour and stretch it according to its local slope to match the right lung contour, which reduces the probability of wrong contour propagation, as well as OR. Pathologic thoracic CT image segmentation is a challenging issue for the presence of inhomogeneous intensities and various abnormalities. In this paper, we introduce a novel pathologic lung segmentation algorithm based on the DCNN model and random forest. In our algorithm, two fusion operations enhance the performance of our segmentation algorithm. First, the fusion of features of deep, texture and intensity enriches the classification information, which contributes to the classification of random forest. Second, the fusion of classification results of random forest based on multi-scale superpixels can produce more accurate results than one-time segmentation with relatively low time cost due to small-sized Ps. In order to improve segmentation accuracy, we introduce a divide-and-conquer strategy for refining segmentation results by correcting and repairing the left and right lungs. In our future work, we will improve the fusion techniques, augment training samples, and further enhance the performance of our lung algorithm on pathologic lung segmentation. Progressive and multi-path holistically nested neural networks for pathological lung segmentation from CT images A generic approach to pathological lung segmentation Segmentation and image analysis of abnormal lungs at CT: current approaches, challenges, and future trends Comparison of image segmentation of lungs using methods: connected threshold, neighborhood connected, and threshold level set segmentation Automatic segmentation of lung CT images by CC based region growing Near-optimal keypoint sampling for fast pathological lung segmentation Automatic pathological lung segmentation in low-dose CT image using eigenspace sparse shape composition Lung segmentation and classification for pathological lung identification Segmentationof pathological lungs from CT chest images Segmentation of pathological and diseased lung tissue in CT images using a graph-search algorithm. From Nano to Macro Automatic segmentation of pathological lung using incremental nonnegative matrix factorization Exploiting unsupervised and supervised classification for segmentation of the pathological lung in CT Lung segmentation based on random forest and multiscale edge detection A new lung segmentation algorithm for pathological CT images A novel Gaussian scale space-based joint MGRF framework for precise lung segmentation Spatial pyramid-enhanced NetVLAD with weighted triplet loss for place recognition Multimodal deep autoencoder for human pose recovery Lung segmentation on HRCT and volumetric CT for diffuse interstitial lung disease using deep convolutional neural networks Semantic segmentation of pathological lung tissue with dilated fully convolutional networks Medical image classification based on deep features extracted by deep model and statistic feature fusion with multilayer perceptron Multimodal face-pose estimation with multitask manifold deep learning Local deep-feature alignment for unsupervised dimension reduction Image-based three-dimensional human pose recovery by multiview locality-sensitive sparse retrieval U-net: convolutional networks for biomedical image segmentation SLIC superpixels compared to state-of-the-art superpixel methods Morphological grayscale reconstruction in image analysis: applications and efficient algorithms Gradient-based learning applied to document recognition Performance evaluation for four classes of textural features Color texture segmentation for clothing in a computer-aided fashion design system Analysis of Hu's moment invariants on image scaling and rotation Visual pattern recognition by moment invariants A novel method of determining parameters of CLAHE based on image entropy Color moments. School of Informatics Similarity of color images Do we need hundreds of classifiers to solve real world classification problems? Discrete grey model based on fractional order accumulate SUSAN-a new approach to low level image processing Connective tissue disease-associated interstitial lung diseases (CTD-ILD) report from OMERACT CTD-ILD Working Group Building a reference multimedia database for interstitial lung diseases Classifiers and their metrics quantified A new hybrid approach using fuzzy clustering and morphological operations for lung segmentation in thoracic CT images Publisher's Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations Conflict of interest No potential conflict of interest was reported by the authors.