key: cord-0440772-cas5nwx2 authors: Kumar, Ahlad; Singh, Harsh Vardhan title: Tchebichef Transform Domain-based Deep Learning Architecture for Image Super-resolution date: 2021-02-21 journal: nan DOI: nan sha: 09a393d76891a0be0d824f70c445ede22808fdfb doc_id: 440772 cord_uid: cas5nwx2 The recent outbreak of COVID-19 has motivated researchers to contribute in the area of medical imaging using artificial intelligence and deep learning. Super-resolution (SR), in the past few years, has produced remarkable results using deep learning methods. The ability of deep learning methods to learn the non-linear mapping from low-resolution (LR) images to their corresponding high-resolution (HR) images leads to compelling results for SR in diverse areas of research. In this paper, we propose a deep learning based image super-resolution architecture in Tchebichef transform domain. This is achieved by integrating a transform layer into the proposed architecture through a customized Tchebichef convolutional layer ($TCL$). The role of TCL is to convert the LR image from the spatial domain to the orthogonal transform domain using Tchebichef basis functions. The inversion of the aforementioned transformation is achieved using another layer known as the Inverse Tchebichef convolutional Layer (ITCL), which converts back the LR images from the transform domain to the spatial domain. It has been observed that using the Tchebichef transform domain for the task of SR takes the advantage of high and low-frequency representation of images that makes the task of super-resolution simplified. We, further, introduce transfer learning approach to enhance the quality of Covid based medical images. It is shown that our architecture enhances the quality of X-ray and CT images of COVID-19, providing a better image quality that helps in clinical diagnosis. Experimental results obtained using the proposed Tchebichef transform domain super-resolution (TTDSR) architecture provides competitive results when compared with most of the deep learning methods employed using a fewer number of trainable parameters. Tchebichef Transform Domain-based Deep Learning Architecture for Image Super-resolution Ahlad Kumar Abstract-The recent outbreak of COVID-19 has motivated researchers to contribute in the area of medical imaging using artificial intelligence and deep learning. Super-resolution (SR), in the past few years, has produced remarkable results using deep learning methods. The ability of deep learning methods to learn the non-linear mapping from low-resolution (LR) images to their corresponding high-resolution (HR) images leads to compelling results for SR in diverse areas of research. In this paper, we propose a deep learning based image superresolution architecture in Tchebichef transform domain. This is achieved by integrating a transform layer into the proposed architecture through a customized Tchebichef convolutional layer (T CL). The role of T CL is to convert the LR image from the spatial domain to the orthogonal transform domain using Tchebichef basis functions. The inversion of the aforementioned transformation is achieved using another layer known as the Inverse Tchebichef convolutional Layer (IT CL), which converts back the LR images from the transform domain to the spatial domain. It has been observed that using the Tchebichef transform domain for the task of SR takes the advantage of high and lowfrequency representation of images that makes the task of superresolution simplified. We, further, introduce transfer learning approach to enhance the quality of Covid based medical images. It is shown that our architecture enhances the quality of Xray and CT images of COVID-19, providing a better image quality that helps in clinical diagnosis. Experimental results obtained using the proposed Tchebichef transform domain superresolution (TTDSR) architecture provides competitive results when compared with most of the deep learning methods employed using a fewer number of trainable parameters. Index Terms-Image Super-Resolution (SR), Tchebichef moments, Deep Learning, Convolutional neural network, Covid-19. T HE coronavirus disease (COVID-19) is a newly emerging viral disease that caused a worldwide pandemic. The World Health Organization has listed it as the sixth international public health emergency. It has impacted around 170 countries and has almost taken the lives of 2 million people as of January 2021. COVID-19 diagnosis by X-ray and CT images is very popular due to the quick results and that with great accuracy. This paper proposes a deep learning based SR architecture to enhance the quality of COVID-19 medical images for clinical diagnosis. possible solutions existing for one single image. It refers to the process of obtaining high-resolution (HR) images from the corresponding low-resolution (LR) images. It's used in wide variety of applications such as in security, mobile cameras, medical imaging, etc. [1] have attracted researchers over the past few years. SR problems are categorized into Single Image Super-resolution (SISR), and Multi-Image Super-resolution (MISR) [2] - [5] . SISR aims to recover the HR image from a single LR image and the SISR based methods have an edge over MISR based methods, as they produce better perceptual quality of the images. The conventional SR based methods uses dictionary-based approaches, that consist of two dictionaries of HR and LR images or patches [6] - [9] . These dictionaries are often learned with sparse-coding methods to get high-quality SR images. Some methods also use the dictionary features that need to be handcrafted [10] . Recent advancements in deep learning methods have shown the state of the art results in SR over the different dataset of images [11] . The earliest deep learning method was SRCNN [12] and later its improved version was reported in [13] . Recursive networks [14] and residual learning [15] is also incorporated in the deep learning architectures to boost the training process. We propose a new orthogonal domain based deep learning architecture for image SR. It uses Tchebichef moment to transform spatial domain to orthogonal domain and then finds the difference between the Tchebichef coefficients of HR and LR image pairs. It has been observed that HR-LR image pairs have huge differences in the coefficient values at higher frequencies and small to negligible differences in the lower frequencies. This key observation plays a central role in developing the proposed architecture. The key aspects of this paper are summarized as follows • A deep neural network to solve the SR problem in an orthogonal transform domain is introduced. The architecture includes both the forward and inverse mapping, hence provides a complete pipeline for image SR. • The proposed architecture exploits the Tchebichef kernels to generate the representation of images in the transform domain. Two custom convolutional layers are designed; one for converting the image to transform domain (T CL) and the other, for doing inverse transformation (IT CL). T CL layer is kept fixed and non-trainable, while the IT CL as trainable to get the optimized reconstruction kernels, which are used in IT CL layer for converting from transform domain to spatial domain. • The proposed architecture consist of high and lowfrequency paths. The high frequency path employs Inception-Resnet based structure with local residual connection to boost the training process. The low-frequency path employs a simple convolutional neural network (CNN) architecture. • To handle the artifacts that occur during the reconstruction phase after ITCL, additional convolutional layers are employed to process the spatial domain images that results in the enhanced SR images. II. RELATED WORK A. Single Image Super Resolution (SISR) The SISR poses a highly challenging problem that is not so easy to solve due to the ill-posed nature, and hence, SISR involves multiple solutions for one LR image or patch. SISR can mainly be divided into three categories: learning, interpolation and reconstruction-based methods. Learning-based SISR methods [7] , [16] are computationally fast and have good performance, specifically, sparse-coding has shown compelling results. The method involves two dictionaries of images, separately for LR and HR patches. LR image or patch are represented in terms of their corresponding sparse code from the LR dictionary. Similarly, an HR image or patch is generated by the same sparse code, but it is applied to the HR dictionary. Interpolation SISR methods involve techniques, like bicubic interpolation [17] and Lanczos resampling [18] , although they are speedy and straightforward, but lack satisfactory accuracy. Reconstruction based methods [19] - [22] are time-consuming, SISR relies on advantages from prior knowledge but these methods highly degrade the SR image quality when the scaling factor is increased. Earlier due to lack of computational capabilities and not so advanced architectures, the conventional SR methods have produced average results. Recent advancements both in computational architectures and effective research using deep neural networks, have produced state-of-the-art results in the SR domain. Dong et al. [12] introduced a CNN network for the SR problem, known as SRCNN. It involved three layers of architecture that outperformed the previous methods based on sparse-coding. SRCNN introduced the non-linear mapping function between the LR and HR image. Image patches are fed to the CNN to generate the feature representation followed by more convolutional layers to produce the higher representation of images. Although SRCNN produced great results, but continuous advancements in the SR domain led to newer methods and architectures. Since then fully connected neural networks have also shown considerable improvements in SR by combining several ideas from different architectures; such as sparse-coding, residual learning, etc. Prior information based deep neural networks has shown promising results. FSRNet [23] used to generate the human face SR images. Wang et al. [24] focused on using the structural feature priors for effective recovery of detailed texture features in an image. A huge improvement in SR image quality was observed using Generative Adversarial Networks (GANs) [25] . Johnson et al. [26] focused on creating the photo-realistic image with high perceptual quality and focused less on maintaining the pixel-wise difference between the images. GANs generate visually better images but training a GAN is a complex and time-consuming process making it unsuitable for many practical applications. Li et al. [27] introduced the combined architecture using image transform domain as well as CNN, and converted the input image/patch into the Fourier transform domain. The discrete Fourier transform (DFT) coefficients, thus obtained are passed to a CNN architecture. As the convolution of the kernels and image in the spatial domain is equivalent to the multiplication of kernels and images in Fourier domain, the claim is that the CNN architecture performs the element-wise multiplication to speed up the training process. Experimentally, it was observed that the performance was similar to previous works but not at par with state-of-the-art methods. Similar work is reported in [28] which uses a combination of transform domain and CNN, where the images are transformed using discrete cosine transform (DCT) and incorporated the pre and post-processing of DCT and inverse discrete cosine transform (IDCT) coefficients within the CNN, using custom convolutional layer. This paper exploits the image transform domain using Tchebichef basis functions that are integrated as kernels with the CNN network. A custom convolutional layers (T CL) for transforming the image into orthogonal domain and the reverse using inverse transform (IT CL) is proposed. The transform domain representations of an image is utilized to learn the high-frequency details that are lost during the degradation using bicubic interpolation. The inverse transformation layer (IT CL) is kept trainable and optimizable, hence the kernels used in IT CL optimize with training process. The optimized kernels obtained after training process provides the best possible reconstruction of the images from its corresponding transform domain. Moreover, the reconstruction artifacts like softness and haziness are removed by the help of additional layers, incorporated in the proposed architecture. The main objective of this paper is to performs SR effectively by preserving the visual attributes of an image. In this section, a brief review about the definition of Tchebichef moment [29] is discussed. They have been used recently in many pattern recognition [30] - [32] and denoising applications [33] , [34] . Its robust feature representation capabilities allows to reconstruct images with promising results. The Tchebichef moment of order (m + n) for an image with intensity function g(x, y) is given as [29] T n,m (u) = and t m (y; N ) are the normalized Tchebichef polynomials, given by and To simplify the notation, t n (x) is used to represent t n (x; N ). Here, t n (x) is the ortho-normal version of Tchebichef polynomials and it can be calculated using recurrence relation as [29] t where The initial conditions for the above recurrence relationship is given ast Image can be reconstructed back from Tchebichef moments using the inverse Tchebichef transformation as given by Tchebichef moment in (1) can also be implemented in matrix format. The set of Tchebichef moments upto order (m + n) in matrix form is given as where G is a square image matrix. P and Q are Tchebichef polynomials in matrix form upto orders of p and q, respectively given as Similarly, the inverse transformation given in (7) can be represented in matrix form as Tchebichef moments of an image can be interpreted as the projection of the image on the basis (kernel) functions, w pq which is given as The complete set of w pq basis functions is shown in Fig. 1 . Tchebichef moments can also be viewed as correlation between the basis functions and image G. A high value is recorded, if there is a strong similarity between the content of the image and basis function, and vice versa. In the proposed architecture, Tchebichef basis functions are used as filters and are re-arranged in a zig-zag order as shown in Fig. 2 This zig zag reordering is inspired from the JPEG compression procedure [35] . The average value of the coefficients generated by the convolution of the Tchebichef kernels with HR and LR images respectively, is shown in Fig. 3. Here, Figs. 3(a)-(b) show the coefficients of a LR and HR version of the medical image, while Fig. 3(c) shows the differences between the HR and LR image coefficients. Please note that the values obtained in Fig. 3 (c) is scaled for proper visualization. It can be observed from this figure that with the increase in kernels complexity, there is a substantial loss of the coefficients in the high frequency when compared to lower frequency region. In the Tchebichef domain, the problem of SR becomes recovering the high frequency Tchebichef coefficients of the HR image from their corresponding LR images. This observation is incorporated in the proposed architecture discussed next. Similar analysis is carried out on natural image and the results are shown in Figs. 3(d)-(f). In this section, a detail description of the proposed architecture, shown in Fig. 4 for SR is discussed. The architecture consist of the following blocks: (1) Tchebichef convolutional layer (T CL) (2) Frequency cube (3) Non-linear mapping for low frequency. (4) Inception-residual connection for high frequency (5) Inverse Tchbeichef transformation layer (IT CL). A. Network Structure 1) Tchebichef convolutional layer (T CL): This block transforms images from spatial domain to Tchebichef moment domain, and has basis function w i as the kernels. There are 64 such kernels of size 8×8 arranged in a zig-zag manner for increasing complexity with an increase in index i of the kernels. The detail about this is discussed later in Sec. III-D. The transformation from spatial to Tchebichef moment domain works as follows: T CL layer creates a 64 feature maps f i for the entire image by performing convolution using w i with image G as given in (14) . Here, represent the convolution operation and is performed using the stride S = 1 and same padding in order to preserve the dimension of the image. Kernels of the T CL layer are kept fixed and non-trainable during the training phase as the primary role of this layer is to convert images into the transform domain. 2) Frequency Cube: Frequency domain feature maps f i=0,...,63 , obtained from (14) , are used to form a cube (see label 2 marked in Fig. 4 ). This cube is re-organized version of Tchebichef coefficients, calculated for the whole image and is ordered in increasing frequency content (complexity). Based on the detail discussion carried out for Fig. (3) in Sect. III-D, it has been observed that there is a substantial loss of the coefficients in the high frequency region compared to lower frequency region. Due to this reason, partition of the frequency cube is done into two parts with a particular split point T . The low and high frequency maps are defined as f low = f i=1,...,T and f high = f i=T +1,...,63 , respectively. Fig. 5 shows the details about this partitioning process. The calculation of the split point T is experimental and its optimal value is obtained as 5. The discussion about its optimal value is carried out in experimental section. The proposed architecture process the partitioned cubes f low and f high separately. It can be observed from Figs. 3(c) and (f), that there is a higher amount of coefficient loss in highfrequency region and thereby the high frequency block f high requires more robust and complex mapping to recover the HR image from LR image. On the other hand, the coefficient loss in low-frequency region is not so significant, but does play an important role in image quality. Hence, processing the low frequency block f low is done, using simpler non-linear convolutional mapping to recover the image details. Next, we discuss the simple and complex deep learning architectures for f low an f high . 3) Architecture for f low : The mapping of low-frequency coefficients of LR image to corresponding low frequency coefficients of HR image is accomplished via CNN network consisting of two convolutional layers (see green arrow in Fig. 4 ). The first layer is a 5×5 followed by 1×1 convolutional layer. Leaky rectified linear unit (ReLU) is used as the activation function in both the layers. The non-linear mapping is given as where k represents the index to the two convolutional layers, z [k] low is the output of k th layer, W 1 are the weights and biases of the k th layer, α is the leaky ReLU parameter having value of 0.1. The non-linear mapping of (15) recovers the information loss in the lower frequency spectra of the image 4) Architecture for f high : To recover the information loss in high frequency spectra of the image, a non-linear mapping is implemented using the deep learning architecture inspired from the inception network [36] (see green arrow in Fig. 4) . The high frequency feature maps f high of LR image is split into three convolutional paths; each of which are using different kernel sizes, namely, 3×3, 5×5, and 7×7. Larger kernel sizes are used in gathering the global information, while the smaller kernel size gather information that is distributed more locally in the feature map. This allows the model to take advantage of the multi-level feature extraction. Finally, concatenating the features obtained from all the levels is done followed by a 1×1 convolution, which serves two purposes. First it creates a linear projection of a stack of feature maps, and secondly, it reduces the depth of the network. The nonlinear mapping for the process described above is given as Here, z T high is the combination of all the feature maps obtained via three parallel paths denoted by k. This layer is required to transform image from Tchebichef moment domain to spatial domain. It takes input which is obtained by combining both, the low and high frequency cubes, z [2] low and z T high , respectively. The output of this layer reconstruct the image y in spatial domain and is given aŝ Here, the weights of the Tchebichef kernel w i are trainable, so that during the training process the kernels adapt to the data and provides efficient inverse transformation. 6) Fine-Tuning Network: The reconstructed image y obtained using (18) is further processed through a small fine tuning network shown in Fig. 4 , which consist of three convolutional layers. The main purpose of introducing this additional network is to get rid of minor artifacts from the image y. This section discuss the training details of the proposed TTDSR architecture. In order to learn the end to end mapping function F for SR task, optimized values of the network parameters θ ∈ (W 2 ) are required. These parameters can be obtained by minimizing the loss between the network generated reconstructed SR image F(Y i , θ), and the high-resolution ground truth image X. Given the batch of highresolution images X i and the corresponding low-resolution images Y i , the loss function is given as where M is the total number of training images, λ is the regularization parameter and l is the total number of kernels used in the architectures discussed in Sections IV-A3 and IV-A4. The loss is minimized using the Adaptive moment estimation i.e Adam [37] optimizer with standard back-propagation [38] , that computes adaptive learning rates of each parameter. The training phase of the network involves Adam as optimizer with default parameters as: β 1 = 0.9 (the first-order moments), β 2 = 0.999 (the second-order moments), = 1e−7 (a small constant for numerical stability). Learning rate η is initialised as 1e−3. The filter weights for each layer in the network are initialized with Glorot-uniform that draws the samples from the uniform distribution. It is observed that without using regularization, the network becomes highly un-stable, hence, L2 regularization of λ = 0.01 is applied on the network weights to penalize the weights. The T CL is kept as nontrainable, while the IT CL is kept as trainable to get the best optimized Tchebichef kernels. There are 14 convolutional layers in the TTDSR architecture leading to total number of parameters as 94k, out of which 90k are the trainable parameters and the remaining are fixed parameters which are used in the T CL layer. The network is trained for 100 epochs with batch size of 64. Training and testing phase are conducted on NVIDIA GeForce RTX 2080 Ti GPU, with the TensorFlow as support package. There are several widely used dataset for image SR. The combination of images from T91 [39] and DIV2K [11] dataset have been used to create a combined training dataset. From a robust model, data augmentation technique is used, where the images are augmented using three methods, i.e, a. The images are rotated using 45°, 90°, 135°, 180°, 225°. b. Horizontal and vertical flips of the images are done. c. Scaling of images is performed by factor of 0.6, 0.7, 0.8 and 0.9 respectively. These augmented images are the variations in the HR image, which are then down-sampled by a factor of η. The downsampled images are scaled up using the bi-cubic interpolation of same factor η to form the degraded LR images for training. Training images are first converted from RGB to Y CbCr format. Inspired from [9] , [15] , [40] , luminance (Y ) channel is used as the input to the architecture while Cb and Cr channels are directly up-scaled, using the bi-cubic interpolation of LR images. Finally, combining the up-scaled Cb and Cr channels, with the predicted luminance (Y ), channel from the architecture is used to generate the SR image which is then converted back into RGB format. As the proposed architecture is trained on single channel, i.e., Y in Y CbCr domain, this allow for the flexibly of performing transfer learning on Covid-19 medical images, which are gray-scale images. SR results for the same can be analyzed in the experimental work. During the testing phase, several standard datasets like Set5 [41] , Set14 [42] , BSDS100 [43] and Urban100 [44] have been utilized to evaluate the performance of the proposed architecture. The metrics used for image quality assessment are PSNR and SSIM [45] . Few published methods work with larger datasets like DIV2K [11] , ImageNet [46] and MS-COCO [47] . However, our choice of datasets for comparison has been used to keep it consistent with the majority of the methods that make use of it. The above datasets used for testing mainly consist of natural images. As pointed out earlier, transfer learning is carried out on the proposed architecture, which is now tested on medical images. For this, COVID-19 image database which contains a set of images collected by Cohen et al. [48] are used. The dataset contains of chest Xray and computed tomography (CT) images. The images are mainly in gray-scale format and is a collection of anteriorposterior view of chest X-rays. The dataset is continuously updated and it is worth mentioning that the resolution of images varies from image to image. A sample of these images can be found in Fig. 6 . In this section, the performance of the proposed architecture with the other existing methods is discussed. For this, several standard datasets (Sect. V-B) are being used for its evaluation. Here, following methods are used for comparison with our architecture a. ScSR [9] : Sparse coding based SR method, constructs LR-HR image patch dictionary. b. A+ [49] : Adjusted anchored neighbourhood regression for fast super resolution is the updated and modified version of [50] . c. SelfEx [44] : Self similarity based method that measures the similarity within the images. d. SCN [51] : Sparse prior method implemented with the help of CNN. e. SRCNN [12] : Earliest deep learning method for image SR based on CNN architecture. f. FSRCNN [13] : An advanced and modified version of SRCNN with deeper architecture and transpose convolution approach. g. VBPS [52] : Recent method for image SR that exploits the inherent self-similarities found in images. Tables I and II report the PSNR and SSIM results of TTDSR and other methods, respectively. Out of all the methods, FSRCNN and VBPS gives competitive score when compared with TTDSR. However, TTDSR on an average performs well on all kinds of datasets. We now make a subjective comparison of SR results using various methods. Figs. 8 and 7 show the SR results for the TTDSR and other methods in enhancing the quality of the image degraded due to bi-cubic interpolation. Fig. 7 shows the SR results on monarch image. The enlarged version shows the thin black edge at head of monarch image. It can be observed that bi-cubic interpolated image shows heavy loss of thin edge details with discontinuity. Also, other methods fails to generate the edges gracefully. The proposed TTDSR architecture generates clear edge and overcomes the discontinuity artifact observed in other methods and gives better results in terms of PSNR and SSIM. Fig. 8 shows the SR results on bi-cubic interpolated zebra image. It can be observed that the black and white strips present on zebra lacks details and fails to capture the orientation of the edges. Further, the FSRCNN method shows slightly better results compared to our method in terms of PSNR and SSIM, but the diagonal edges are overlapping leading to poor image visualization. On the other hand, though TTDSR gives second best results but it exploits the frequency domain details to overcome this degradation and thereby, generate correctly oriented black and white strips. Next, SR results on COVID-19 medical images is carried out. Here, the aim is not to detect the infection of COVID-19 through images, but to enhance the quality of these image to provide better diagnosis. The dataset used for this purpose is presented in Cohen et al. [48] . As our model is trained on single channel, i.e. Y(luminance) and medical images are gray-scale image that contains only luminance information of the pixel and no color information. This provides a flexibly to use TTDSR architecture on medical images using transfer learning. The SR results on the image can be seen in Fig.9 . The average PSNR and SSIM comparison on COVID-19 dataset can be seen in Table III . It can be observed that the proposed method gives better results compared to other methods. D. Network Parameters and its impact 1) Split Point for Tchebichef Frequency Cube: Tchebichef polynomials are treated as filters and create a frequency cube in the image transform domain as shown in Fig. 5 . In the architecture (Fig. 4) , it can be observed that there are two subnetworks, one for recovering the loss in high-frequency content and the other works for recovering loss in low-frequency content. The frequency cube is split into two halves at a split point T. The splitting of the frequency cube is experimental and the performance of the network varies with the different values of T. In this paper, the value of T is taken as 5 based on the experiment conducted in Fig. 10 , where for different datasets, the average PSNR reported by the network is highest when T is taken as 5. 2) Impact of Residual Connection: Initially, the network was structured without the residual connection and only relied on the inception based module. In this case the network performed with limited capability due to vanishing gradient problem. To overcome this problem, two major residual connections are added in the network, local residual connection for high frequency components (f high ) and the overall global residual connection for both high and low frequency component. The local residual connection [53] in the inception module was introduced to boost the gradients in the training phase for the recovery of high-frequency components of the image. The use of local residual connectivity overcomes the vanishing gradient problem and helps the optimizer to reach the minima faster. Experimental analysis for same is shown Fig. 11 , where it can be observed that the architecture with residual connections converges to smaller loss L, when used without residual connections. 3) Optimized Learned Tchebichef filters: As discussed, the network architecture consists of two custom layers namely T CL and IT CL. The kernel functions used in T CL are fixed; whereas in that of IT CL are kept trainable to adapt to the training phase of the network. Fig. 12 shows the optimized kernels obtained after training process. These optimized kernels are used to reconstruct the image from Tchebichef moment domain and hence gives part of its contribution in providing better image quality when compared to other methods. A deep learning architecture for super resolution of natural and COVID-19 medical images is presented. It makes use of Tchebichef transform domain that helps in exploiting the low and high frequency details present in the images to enhance its quality. The detail analysis of various parameters that affects the performance of the architecture is discussed. The objective and visual comparison of the SR results with the existing methods shows that the proposed architecture provides superior results when evaluated in terms of average PSNR and SSIM metrics. The visual comparison of the result shows that our work restore the details present in an image effectively. This work opens up the possibility to explore SR architectures using different transform domains. Super-resolution image reconstruction: A technical overview Fast and robust multiframe super resolution Advances and challenges in super-resolution Multiframe super-resolution employing a spatially weighted total variation model A multi-frame image super-resolution method Super-resolution with sparse mixing estimators Super-resolution through neighbor embedding Super-resolution from a single image Image super-resolution via sparse representation Image restoration: From sparse and low-rank priors to deep priors Ntire 2017 challenge on single image super-resolution: Methods and results Image super-resolution using deep convolutional networks Accelerating the super-resolution convolutional neural network Deeply-recursive convolutional network for image super-resolution Accurate image super-resolution using very deep convolutional networks Example-based superresolution Cubic convolution interpolation for digital image processing Lanczos filtering in one and two dimensions Softcuts: A soft edge smoothness prior for color image super-resolution Image super-resolution using gradient profile prior Single image superresolution based on gradient profile sharpness Image super-resolution by tvregularization and bregman iteration Fsrnet: End-to-end learning face super-resolution with facial priors Recovering realistic texture in image super-resolution by deep spatial feature transform Generative adversarial nets Perceptual losses for real-time style transfer and super-resolution A frequency domain neural network for fast image super-resolution Adaptive transform domain image super-resolution via orthogonally regularized deep networks Image analysis by tchebichef moments Translation and scale invariants of tchebichef moments Radial tchebichef moment invariants for image recognition Computing invariants of tchebichef moments for shape based image retrieval Tchebichef and adaptive steerable-based total variation model for image denoising Image denoising via overlapping group sparsity using orthogonal moments as similarity measure The jpeg still picture compression standard Going deeper with convolutions Adam: A method for stochastic optimization Gradient-based learning applied to document recognition Fast and accurate image upscaling with super-resolution forests Image deblurring and superresolution by adaptive sparse domain selection and adaptive regularization Christine Guillemot, and Marie line Alberi Morel. Low-complexity single-image super-resolution based on nonnegative neighbor embedding On single image scaleup using sparse-representations A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics Single image super-resolution from transformed self-exemplars Image quality assessment: from error visibility to structural similarity Imagenet: A large-scale hierarchical image database Microsoft coco: Common objects in context Covid-19 image data collection A+: Adjusted anchored neighborhood regression for fast super-resolution Anchored neighborhood regression for fast example-based super-resolution Deep networks for image super-resolution with sparse prior Heavy-tailed selfsimilarity modeling for single image super resolution Identity mappings in deep residual networks