key: cord-0075514-e0bmj2yq authors: AlZubi, Ahmad Ali; Tiwari, Shailendra; Walia, Kuldeep; Alanazi, Jazem Mutared; AlZobi, Firas Ibrahim; Verma, Rohit title: An Efficient Stacked Deep Transfer Learning Model for Automated Diagnosis of Lyme Disease date: 2022-02-28 journal: Comput Intell Neurosci DOI: 10.1155/2022/2933015 sha: 68fffc6e1ce07c1f82c0826ff81d2f45fc514fa8 doc_id: 75514 cord_uid: e0bmj2yq Lyme disease is one of the most common vector-borne infections. It typically causes cardiac illnesses, neurologic illnesses, musculoskeletal disorders, and dermatologic conditions. However, most of the time, it is poorly diagnosed due to many similarities with other diseases such as drug rash. Given the potentially serious consequences of unnecessary antimicrobial treatments, it is essential to understand frequent and uncommon diagnoses that explain symptoms in this population. Recently, deep learning models have been used for the diagnosis of various rash-related diseases. However, these models suffer from overfitting and color variation problems. To overcome these problems, an efficient stacked deep transfer learning model is proposed that can efficiently distinguish between patients infected with Lyme (+) or infected with other infections. 2(nd) order edge-based color constancy is used as a preprocessing approach to reduce the impact of multisource light from images acquired under different setups. The AlexNet pretrained learning model is used for building the Lyme disease diagnosis model. To prevent overfitting, data augmentation techniques are also used to augment the dataset. In addition, 5-fold cross-validation is also used. Comparative analysis indicates that the proposed model outperforms the existing models in terms of accuracy, f-measure, sensitivity, specificity, and area under the curve. Lyme disease is one of the most common vector-borne infections, generally due to one of the three pathogenic genospecies of the spirochete Borrelia [1, 2] . It typically causes cardiac illnesses, neurologic illnesses, musculoskeletal disorders, dermatologic conditions, etc. [3, 4] . However, most of the time, it is poorly diagnosed due to many similarities with other diseases such as drug rash [5] , pityriasis rosea rash [6] , and ringworm [7] . Figure 1 shows the example of Lyme disease along with other similar diseases. It is clearly found that the drug rash, pityriasis rosea rash, and ringworm visually seem to be similar and so many times, Lyme disease is either underdiagnosed or overdiagnosed. Overdiagnosis or underdiagnosis of Lyme disease leads to unnecessary antibiotic treatments. Numerous problems and adverse consequences because of medicines being given to patients longer than recommended, needless antibiotics, or unusual therapies for Lyme disease were reported, like cholecystitis, catheter-associated bloodstream infection, clots from venous catheters, Clostridioides difficile infections, and death. Given the potentially serious consequences of unnecessary antimicrobial treatments, it is essential to understand frequent and uncommon diagnoses that explain symptoms in this population [2] [3] [4] . erefore, it is required to build such a framework or model which can clearly distinguish among patients infected with Lyme (+) or infected with other infections. Recently, the imaging dataset of Lyme patients has been published on the Kaggle website [8] . Many researchers have utilized it for distinguishing between Lyme infections and infections with other rashes using deep learning models. However, these models suffer from the overfitting problem. Also, suitable preprocessing techniques are required to improve the quality of images under consideration to achieve efficient results. To overcome these problems, in this paper, an efficient stacked deep transfer learning model is proposed to classify Lyme patients. e main contributions of this paper are as follows: (a) An efficient stacked deep transfer learning model is proposed to classify Lyme patients (b) 2 nd order edge-based color constancy is used as a preprocessing approach to reduce the impact of multisource light from images acquired under different setups (c) e AlexNet pretrained learning model is used (d) Data augmentation techniques are also used to augment the dataset e remaining paper is organized as follows: Section 2 presents the literature work, Section 3 presents the proposed model, Section 4 discusses various experimental results, and Section 5 concludes the paper. In [9] , an ensemble deep learning pipeline (EDLP) was designed by using the 34-layer ResNet model. ResNet was used to extract the features from the limited skin disease dataset. Eleven skin conditions were classified. It has achieved 91.7% precision and 92.55% recall, respectively. In [10] , skin cancer was evaluated from the rashes. A convolutional neural network (CNN) was utilized to predict images of rashes or skin cancer. It has achieved an accuracy of 80.2% for 20 epochs. In [11] , a mobile-enabled expert system named i-Rash was designed for the diagnosis of inflammatory skin lesions. It can predict the given image as psoriasis, eczema, acne, and healthy. i-Rash was trained using pretrained SqueezeNet. In [12] , a lightweight attention-based deep learning model (LWADL) was designed to predict eleven skin diseases. LWADL achieved better accuracy as compared to VGG19, VGG16, ResNet50, and InceptionV3. In [13] , a UNet-based dense CNN (UNet-dCNN) model was designed. MobileNetV2 was also used to achieve better results. It was utilized for histopathological image-based skin cancer diagnosis. It has shown an average accuracy of 87.7%. In [14] , a multitask deep learning model was designed. It was utilized for automatic analysis and classification of skin lesions. A focal loss and a Jaccard distance-based loss function were designed. A three-phase joint training approach was used to assure significant feature learning. In [15] , a pretrained ResNet-50 model was utilized for the classification of melanoma. Ensemble learning was also considered to obtain better results. In [16] , a fully automated deep ensemble model (FADEM) was designed for lesion boundary segmentation. DeeplabV3+ and Mask R-CNN were ensembled and tested on the ISIC-2017 dataset. FADEM has achieved specificity and sensitivity of 97.94% and 89.93%, respectively. In [17] , ensembling of CNN models was achieved by integrating with a test-time regularly spaced shifting. e proposed ensemble model was used for the classification of skin lesions. In [18] , an adaptive dual attention module-based CNN model was designed for the segmentation of skin lesions. e proposed model has shown better performance during the computation of potential features in classifying the skin lesions' boundary. From the existing literature, it is found that the existing models suffer from overfitting and hyperparameter tuning problems. In addition, preprocessing techniques are required to reduce the impact of multiple light sources on the images. Also, none of the researchers have focused on the diagnosis of Lyme disease. erefore, in this paper, an efficient model is designed to achieve better results. is section discusses the proposed methodology. Initially, the obtained images are improved by using the 2 nd order edge color constancy. ereafter, a ResNet-based model is trained to achieve better results [18] . Figure 2 shows the proposed automated Lyme disease diagnosis model. It clearly shows that the proposed model is decomposed into three phases, i.e., data augmentation, 2 nd order edge-based color constancy, and an AlexNet-based pretrained model for extracting the features which are used for the classification using fully connected layers. To achieve regularization and to prevent overfitting, dropouts are also used. Binary crossentropy is used as a loss function. Color constancy has the ability to restore the impact infected multiple light sources. us, the obtained images are independent of colors of the light source. In this paper, a 2 nd order-edge based color constancy approach is used. It states that the distribution of color derivatives exhibits the principal dissimilarity in the direction of a light source [18] . Minkowski's norm is then applied to the computed derivatives to predict the direction of a light source [19, 20] . A step-by-step algorithm for the 2 nd edge-based color constancy is presented in Algorithm 1. Residual network (ResNet) is a well-known pretrained model used as a backbone for classifying many imaging datasets. It allowed us to successfully build an enormously deep model with more than 150 layers. Before ResNet, the existing models suffered from vanishing gradients whenever we tried to train them deeply. It has achieved better results with the help of skip connections. It prevents the vanishing gradient problem by using a substitute shortcut route for the gradient to flow through. It allows the model to build an identity function that assures the topmost layer will achieve better performance, the same as the lower layer. Figure 3 shows the ResNet model in the paper. Initially, the images obtained from the color constancy model are utilized for building the trained model. It utilizes various convolution layers, followed by ReLU, normalization, and pooling operations. After using 5 convolution layers, a fully connected layer is utilized along with ReLU and dropouts. Finally, after using three fully connected layers, the softmax function is used to obtain the results. e experiments for the proposed model are performed on the online MATLAB 2021a using a benchmark Kaggle dataset. Comparisons are also performed by considering the competitive models. In addition, we have also validated the proposed model with and without considering the 2nd order edge-based color constancy. erefore, data augmentation is used to augment the dataset. Since the obtained images were captured using different machines under different light sources, using these images directly for diagnosis may result in poor performance of the model. erefore, in this paper, to prevent the effect of multiple light sources, color constancy is used. It can restore the impact of color light sources from the images to achieve better performance of the models. Figure 4 shows the training and validation analysis of the proposed model without the use of 2 nd order edge-based color constancy. It clearly shows that the proposed model has achieved 95.71% validation accuracy. But also, it is found that the proposed model suffers from the overfitting issue since the training accuracy is 100%. erefore, still, there is room for improvement in it. Figure 5 shows the training and validation accuracy of the proposed model with 2 nd order edge-based color constancy. It is found that the proposed model with color constancy achieves 98.69% validation accuracy. erefore, the proposed model is least affected by the overfitting problem. Besides, the validation accuracy of the proposed model has shown better convergence speed than in the results shown in Figure 4 . Step 1: Divide C I into R I , G I , and B I . Step 2: Remove saturated color points. It represents those pixels that are greatly influenced by the light direction [21, 22] . Step 3Computed aggregated color values as [23] : Here, R T , G T , and B T define the aggregated color values of R I , G I , and B I , respectively. Step 4: Compute the average of all color channels as [24] : Step 5: Computer the impact of color saturation as [25, 26] : Step 6: Remove the color saturation as [18, 27] : Step 7: Evaluate the effect illuminance (C p ) using 2 nd order edge-based approach as: Step 8: Evaluate the impact of light on each color channel as [18, 28] : N R � (C P (: , : , 1) . * m R ) 1/m n N G � (C P (: , : , 2) . * m G ) 1/m n N B � (C P (: , : , 3) . * m B ) 1/m n Here, m R , m G , and m B show the mask containing the saturated pixels. Step 9: Evaluate the aggregated impact of normalized whiteness in the color channels as [18, 29] : Step 10: Compute the impact of light source as: S R � S R /I L S G � S G /I L S B � S B /I L Step 11: Compute the restored color channels as Step 12: Return concatenated (R o , G O , and B o ) ALGORITHM 1: Second-order edge-based color constancy algorithm. 4 Computational Intelligence and Neuroscience Computational Intelligence and Neuroscience are widely accepted to compute the performance of the various classification models. It utilizes the concepts of true positive (TP), true negative (TN), false positive (FP), and false negative (FN) that can be computed using the actual and predicted classes. Figure 6 shows the confusion matrix of the proposed model without the use of 2 nd order edge-based color constancy. It is found that without the use of color constancy, the proposed model achieves an average accuracy of 95.7%. It is found that this model has achieved better results, but with a 4.3% error rate. Figure 7 demonstrates the confusion matrix analysis of the proposed model with 2 nd order edge-based color constancy. It is found that the proposed model has achieved 98.7% accuracy, which is 2.0% better than the proposed model without the use of 2 nd order edge-based color constancy (EBCC). ere are only 7 cases that are poorly classified by the proposed model, compared to 23 in the model with the use of 2 nd order edge-based color constancy. is section presents the discussion of the proposed and competitive models when they are applied to the Lyme disease dataset. e hyperparameters of the competitive models are obtained from their published papers. ese competitive models are CNN [10] , EDLP [9] , SqueezeNet [11] , LWADL [12] , Unet-dCNN [13] , ResNet-50 [15] , FADEM [16] , and Ensemble CNN [17] . Table 2 demonstrates the testing analysis of the proposed and competitive models. It is found that the proposed model achieves remarkably better performance than the competitive models. Bold values represent the better-performing model. e proposed model has achieved an average improvement in terms of accuracy, f-measure, sensitivity, specificity, and area under the curve (AUC) as 2.9787%, 2.7891%, 3.0875%, 2.1578%, and 2.1579%, respectively. It is found that the proposed model achieves remarkable performance for Lyme disease. erefore, the proposed model can be used for real-time diagnosis of Lyme disease and can help doctors treat patients with the correct medicines. Computational Intelligence and Neuroscience is paper proposes an efficient stacked deep transfer learning model that can efficiently distinguish between patients infected with Lyme (+) or infected with other infections. 2 nd order edge-based color constancy was used as a preprocessing approach to reduce the impact of multisource light from images acquired under different setups. e AlexNet pretrained learning model was utilized for building the Lyme disease diagnosis model. Data augmentation techniques were also used to augment the dataset. Extensive comparative analyses have shown that the proposed model outperforms the competitive models in terms of accuracy, f-measure, sensitivity, specificity, and AUC of 2.9787%, 2.7891%, 3.0875%, 2.1578%, and 2.1579%, respectively. e used dataset is freely available at https://www.kaggle. com/sshikamaru/lyme-disease-rashes. Mistaken identity: many diagnoses are frequently misattributed to Lyme disease Death from inappropriate therapy for Lyme disease Serious bacterial infections acquired during treatment of patients given a diagnosis of chronic Lyme disease -United States e consequences of overdiagnosis and overtreatment of Lyme disease: an observational study Drug rash with eosinophilia and systemic symptoms and severe renal injury induced by proton pump inhibitor therapy: a case report SARS CoV 2 spike protein positivity in pityriasis rosea like and urticaria like rashes of COVID 19 Medical social workers as mediators between patients, physicians, and the court: the case of former ringworm patients Lyme Disease Rashes DermaDetect: a novel computer vision model for an accurate diagnosis of skin conditions and rashes Detection and differentiation of skin cancer from rashes An intelligent inflammatory skin lesions classification scheme for mobile devices A visually interpretable deep learning framework for histopathological image-based skin cancer diagnosis Deep learning classification of systemic sclerosis skin using the MobileNetV2 model An end-to-end multi-task deep learning framework for skin lesion analysis Deep learning and handcrafted method fusion: higher diagnostic accuracy for melanoma dermoscopy images Skin lesion segmentation in dermoscopic images with ensemble deep learning methods Skin lesion classification by ensembles of deep convolutional networks and regularly spaced shifting Edge-based color constancy Multiobjective evolutionary optimization techniques based hyperchaotic map and their applications in image encryption Graphology based handwritten character analysis for human behaviour identification Novel deep transfer learning model for COVID-19 patient detection using X-ray chest images Deep neural networkbased speaker-aware information logging for augmentative and alternative communication Visibility improvement and mass segmentation of mammogram images using quantile separated histogram equalisation with local contrast enhancement Metaheuristic-based deep COVID-19 screening model from chest X-ray images A fully convolutional neural network-based regression approach for effective chemical composition analysis using near-infrared spectroscopy in cloud Deep learning approach for microarray cancer data classification Multi-modality medical image fusion technique using multi-objective differential evolution based deep neural networks Human activity recognition and embedded application based on convolutional neural network Providing security and privacy to huge and vulnerable songs repository using visual cryptography