key: cord-0283905-wf9mwifo authors: Chen, Meixu; Goodall, Todd; Patney, Anjul; Bovik, Alan C. title: Learning to Compress Videos without Computing Motion date: 2020-09-29 journal: nan DOI: 10.1016/j.image.2022.116633 sha: be43c0ff99da77ec701e47e6fc569b9717b50ebd doc_id: 283905 cord_uid: wf9mwifo With the development of higher resolution contents and displays, its significant volume poses significant challenges to the goals of acquiring, transmitting, compressing, and displaying high-quality video content. In this paper, we propose a new deep learning video compression architecture that does not require motion estimation, which is the most expensive element of modern hybrid video compression codecs like H.264 and HEVC. Our framework exploits the regularities inherent to video motion, which we capture by using displaced frame differences as video representations to train the neural network. In addition, we propose a new space-time reconstruction network based on both an LSTM model and a UNet model, which we call LSTM-UNet. The new video compression framework has three components: a Displacement Calculation Unit (DCU), a Displacement Compression Network (DCN), and a Frame Reconstruction Network (FRN). The DCU removes the need for motion estimation found in hybrid codecs and is less expensive. In the DCN, an RNN-based network is utilized to compress displaced frame differences as well as retain temporal information between frames. The LSTM-UNet is used in the FRN to learn space-time differential representations of videos. Our experimental results show that our compression model, which we call the MOtionless VIdeo Codec (MOVI-Codec), learns how to efficiently compress videos without computing motion. Our experiments show that MOVI-Codec outperforms the Low-Delay P veryfast setting of the video coding standard H.264 and exceeds the performance of the modern global standard HEVC codec, using the same setting, as measured by MS-SSIM, especially on higher resolution videos. In addition, our network outperforms the latest H.266 (VVC) codec at higher bitrates, when assessed using MS-SSIM, on high-resolution videos. V IDEO traffic is predicted to reach 82 percent of all consumer Internet traffic by 2021 [1] , and to continue this rapid growth even further. The increasing share of video in Internet traffic is being driven by several factors, including the great diversity and extraordinary popularity of streaming and social media services, the rise of video teleconferencing and online video education (accelerated by the Coronavirus M. Chen and A. C. Bovik are with the Department of Electrical and Computer Engineering, University of Texas at Austin, Austin, USA (e-mail: chenmx@utexas.edu; bovik@ece.utexas.edu). T. Goodall was with Facebook Reality Labs when this work was conducted, and now he is with Apple, Inc (email: beyondmetis@gmail.com). A. Patney was with Facebook Reality Labs when this work was conducted, and now he is with Nvidia. (email: rwebb@fb.com). Crisis), and significant increases in video resolution. Indeed, it is estimated that by 2023, two-thirds of installed flat-panel television sets will be UHD, up from 33 percent in 2018 [2] . Given significant strains on available bandwidth, it is crucial to continue and greatly accelerate the evolution of video compression systems. Traditional video compression codecs, like H.264, HEVC and the latest VVC/H.266 process videos through a sequence of hand-designed algorithms and modules, including block motion estimation, and local decorrelating decompositions like the Discrete Cosine Transform (DCT). Although the component modules of modern hybrid codecs have been carefully designed over several generations, the overall codecs have not been globally optimized other than by visual examination or post-facto objective measurement of results, typically by the highly fallible PSNR [3] . Naturally, one could expect the performances of video codecs to be improved by collective, end-to-end optimization. Because of their tremendous ability to learn efficient visual representations, deep learning models are viewed as highly promising vehicles of developing alternative, globally optimal video codecs, and a variety of deep learning based image compression architectures have been proposed [4] - [19] . These new models have deployed Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), autoencoders, and Generative Adverserial Networks (GAN) yielding rate-distortion efficiencies that are reportedly comparable to those of traditional image compression codecs like JPEG, JPEG 2000, and BPG. Encouraged by these advances, several authors have devised deep video compression models that suggest the considerable promise of this general approach. Wu et al. [20] proposed the first end-to-end trained deep video codec, using a hierarchical frame interpolation scheme. A block-based deep video compression codec was proposed by Chen et al. [19] . Lu et al. proposed an end-toend video compression model (DVC) [21] which replaces each component of the traditional hybrid video codec with a deep learning model, which are jointly trained as a global hybrid architecture against a single loss function. Another hierarchical video compression architecture called HLVC (Hierarchical Learned Video Compression) was proposed by Yang et al. [22] . In both traditional video codecs and recent deep learningbased ones, motion estimation and compensation has occupied a significant portion of the system resources. Motion estimation requires an expensive search process that we avoid, by instead training the network to efficiently represent the residuals between each current frame and a set of spatially-displaced neighboring frames. Computing a set of frame differences, even over many displacement directions is much cheaper than effective search processes. Moreover, while the statistics of motion are generally not regular, the intrinsic statistics of frame differences exhibit strong regularities [23] , including those of differences between spatially displaced frames [24] . The strong internal structure of these frame differences makes them easier to efficiently represent in a deep architecture. Our idea is inspired by the way the human vision system processes natural time-varying images. Many studies have produced strong evidence suggesting that the early stages of the vision system are primarily implicated in reducing redundancies in the sensed input visual signals [25] , [26] . Indeed, much of early visual processing along the retinocortical pathway appears to be devoted to process of spatial and temporal decorrelation [27] - [32] . We have found that sets of spatially displaced frame differences, which are space-time processes, supply a rich and general way to exploit space-time redundancies [23] , [24] . Importantly, our idea is also related to recent theories of the role of microsaccades in human visual information processing [28] - [31] . Microsaccades create small spatial displacements of visual field from moment to moment. While microsaccades have been theorized to play roles in avoiding retinal saturation, maintaining accurate fixation in the presence of drifts, and preserving the perception of fine spatial details [31] , they are more recently thought to play an important role in efficiently representing locally changing and shifting space-time visual information [28] - [31] . We believe that micro-saccadic eye movements deployed by the human eye have adapted to the local regularities induced by small spatial displacements over time, in order to achieve more efficient visual (neural) representations. This has inspired us to, in like manner, trained a deep coder-decoder network to compress videos using regular displaced residual representations as inputs. By capturing displaced frame differences from a large database of videos, and feeding them into a deep spacetime coding-decoding network, we have formulated a new breed of deep video compression algorithms that are motion computation free, statistically motivated, and have perceptual relevance. The contributions of this work can be summarized as follows: • We innovate the use of displaced frame differences to capture efficient representations of structures induced by motion. • Our method avoids the computational overhead of motion estimation and motion compensation. • A combined LTSM-UNet efficiently captures both spatial and temporal information which it uses to recreate video frames from the abstracted video code. • The entire video compression system is collectively jointly optimized using a single loss function. Our results show that video compression can be efficiently accomplished without explicitly computing motion predictions. We trained the new MOVI-Codec architecture end-to-end on the Kinetics-600 dataset and the Vimeo-90K dataset, using a single perceptual loss function (MS-SSIM), and tested it on the UVG dataset, the VTL dataset, and the HEVC Standard The rest of the paper is organized as follows. Section II briefly introduces current progress on learning-based methods for image/video compression and motion estimation. Section III describes details of the architecture and training protocol of the new MOVI-Codec model. Section IV discusses the experiments we conducted and their outcomes, along with a data analysis along several dimensions. Section V concludes the paper with a discussion of future research directions. A variety of standardized image compression engines have been proposed over the years to meet the needs of increasingly picture-centric technologies. JPEG algorithm [33] , and later challengers JPEG 2000 [34] , BPG [35] , and VP9 [36] . These methods have proven to be quite practical, and in the case of JPEG, ubiquitous. Yet they are all handcrafted, highly modularized without the benefit of collective optimization of all their elements. Each of these standards maps pixels to a less correlated representation, regardless of the attributes of the input image. These transformed values are then nonuniformly quantized, typically with reference to a human visual sensitivity model. A variety of authors have recognized the potential of deep learning to advance progress on the image compression problem (a still timely goal given the senectitude of the prevailing JPEG standard), and many learning-based architectures have been devised [4] - [18] . Given that Convolutional Neural Networks (CNN) [37] were the first deep learning models to obtain standout performance on image analysis problems, it was natural that it be the first deep architecture to be applied to learning-based image compression. Ballé et al. [8] proposed a CNN-based image compression framework that was optimized end-to-end, which was shown to outperform JPEG2000 with respect to both MS-SSIM and PSNR image quality measures. Their framework was later extended by incorporating a hyperprior to capture spatial dependencies in the latent representation for entropy estimation [6] . In [14] , Minnen et al. further enhanced the entropy model, by combining autoregressive and hierarchical priors to exploit the probabilistic structure in the latents. The resulting model was reported to outperform BPG with respect to both PSNR and MS-SSIM. Another architecture favored for learning-based image compression are Recurrent Neural Networks (RNN), because of their ability to exploit representative memories. Long Short-Term Memory (LSTM) models were proposed [38] to address the vanishing gradient problem of RNNs. Toderici et al. [4] , [5] was the first to deploy a deep RNNbased architecture for image compression by utilizing a scaleadditive framework. This architecture allows for variable bit rates and only needs to be trained once. The authors also presented results using different types of RNNs, including LSTM, associative LSTM and a hybrid of a Gated Recurrent Unit (GRU) [39] and a ResNet, reporting that the performance of the model was better than JPEG. Generative Adversarial Networks (GAN) have been applied in several learning-based image compression models. Early on, Rippel et al. [11] proposed a GAN-based image compression framework that they claim outperformed all existing codecs with respect to MS-SSIM, while being lightweight and deployable. In [12] , a GAN framework is presented to build an extreme image compression system which the authors report as achieving state-of-the-art performance, especially at very low bit rates, based on a user study. It is natural to also consider learning-based methods for video compression [19] - [22] , [40] - [43] . Wu et al. [20] proposed a video compression architecture based on the idea that video compression is repeated image compression. They define two types of frames: key frames and other frames. Key frames are compressed using an RNN-based image compression network [5] , while the other frames are interpolated in a hierarchical manner. Another hierarchical video compression architecture, called Hierarchical Learned Video Compression (HLVC), was proposed by Yang et al. [22] . In this method, there are three quality layers: an image compression layer, a Bi-Directional Deep Compression (BDDC) layer, and a Single Motion Deep Compression (SMDC) layer. In an attempt to match the pipeline structure of hybrid codecs, Lu et al. proposed an end-to-end video compression model (DVC) [21] that replaces each traditional hybrid component, with deep learning models, then jointly optimized all the components against a single loss function. This work was further extended to two models, a lightweight version called DVC_Lite, and an advanced version called DVC_Pro, by adjusting various components of the architecture. Later, Habibian et al. [40] proposed a deep generative model for video compression using an autoregressive prior to conduct entropy coding. Generally, all learning-based video compression models implement traditional block-based motion estimation or optical flow, both of which have a high computational overhead. The most related work to ours is [43] , whereby an interpolation loop is used as an alternative to motion estimation/compensation. However, the frame interpolation network still requires training, which adds to the complexity of the overall method. Motion estimation (ME) and motion compensation (MC) are crucial components in modern hybrid video codecs. These are used to exploit the temporal redundancy of video frames via inter-frame prediction. In traditional hybrid video codecs like H.264 and H.265, video frames are first partitioned into blocks, then motion vectors (MV) associated with each block are estimated with respect to predictions of neighboring reference frames via expensive block search methods, which is the most intensive aspect of video compression. A few deep learning methods have been proposed to solve the ME problem. For example, Choi et al. [44] trained a CNN to measure the similarity of pairs of image patches and used this to estimate MVs. However, this method still requires a search process to find the best match. In [45] , the authors developed a CNN that was trained to conduct both uni-and bi-directional ME, using separate networks so that motion information need not be transferred from the encoder to the decoder. The CNN does require two frames from the decoded picture buffer and their temporal indices as inputs, which it uses to produce filter coefficients that synthesize patches of a new frame, which is then used to predict the current frame. A drawback of this approach is that it requires the CNN to be resident at both the encoder and the decoder, which reduces decoding efficiency. Another popular alternative to block matching algorithms are optical flow routines, which seek to obtain a dense vector field mapping the movements of pixel. A variety of deep learning based optical flow estimation methods have been proposed to reduce the computational overhead of dense optical flow vectors [46] . FlowNet [47] showed that it was possible to train a network from two input images to predict optical flow while matching or exceeding the accuracies of traditional methods. Later improvements introduced a stacked architecture that included warping of the second image via intermediate optical flow estimates, and a sub-network specialized to predict small motions [48] . Other approaches have tried to combine networks with traditional methods. Ranjan et al. [46] proposed such a network called SpyNet, which adopted a traditional coarse-to-fine computational hierarchy using a spatial pyramid. Later, another network competitive with FlowNet2 was proposed, called LiteFlowNet [49] , but with a significantly decreased model size. Our approach avoids even these methods of deep flow computation, by instead feeding the network a set of directional inter-frame residuals containing adequate information for the network to seek the most efficient perceptual representation. A. Framework Figure 1 exemplifies the flow of our deep video compression network. A current frame is input to the network, along with multiple displaced frame differences from adjoining, previously coded and then decoded frames (lower part of figure) . This is similar to the classic hybrid coding loop, which also includes the decoder as part of the encoder loop, to reduce reconstruction errors. The key components in our network is: Displacement Calculation Unit (DCU), Displacement Compression Network (DCN), and Frame Reconstruction Network (FRN). The details of each key component in our network will be discussed in the following sections. The flow of our network is: Given an input video with frames x 1 , x 2 , ..., x T , for every frame x t , displaced frame differences between the current frame x t and previous reconstructed framex t−1 are calculated via the DCU, after which the displaced frame differences d t are input into the DCN. The DCN compresses the incoming displaced frame differences which are used to capture statsitical redundancies. An illustration of displaced frame differences, i.e. differences between spatially displaced frames, is shown in Figure 2 . Given a compressed outputd t from the DCN, FRN uses the reconstructed displaced frame differencesd t and the reconstructed previous framex t−1 to reconstruct a current framê x t . Every frame is processed following this except for the first frame. The first frame x 1 is processed differently as it does not have previous reconstructed frame. As a result, an all-zero image is chosen as its previous reconstructed frame and it is otherwise processed the same as other frames. Pseudo code of the flow is shown in Algorithm 1. By using this architecture, we are able to reconstruct the videos without the use of motion. Algorithm 1 Flow of MOVI-Codec for an Input Video x 1 to x T : video frames. x 0 : previous reconstructed frame for x 1 . d t ,d t : displaced frame differences and corresponding reconstructed ones, respectively. d 1 ,d 1 : displaced frame differences between x 1 andx 0 , and corresponding reconstructed ones, respectively. 1: procedure MOVI-CODEC 2: for t in 1 to T do 3: if t is 1 then 4:x 0 = all zero frame 5: The DCU removes the need for any kind of motion vector search. Instead, it allows the DCU network to learn to optimally represent time-varying images as sets of spatially displaced frame differences. Given a video with T frames x 1 , x 2 , ..., x T of width w and height h, two directional (spatially displaced) temporal differences are computed between each pair of adjacent frames, as shown in Figure 2 . In the DCU, the inputs are a current frame x t and the reconstructed previous framex t−1 . Then, at each spatial coordinate (i, j), a set of spatially displaced differences is calculated as: where s = 0, ±3, ±5, ±7 in our experiment. The set of 13 displaced frame differences (residuals) is then fed into the Displacement Compression Network, which delivers as output the reconstructed set of displaced residualsd t . As mentioned in Section II, the statistics of non-displaced frame differences have been observed to be nicely regular. As shown in [24] , the statistics of displaced frame differences are also highly regular, and more so in the direction of local motion. This makes them good video representations to learn to exploit spacetime redundancies, while avoiding the computational burden of motion estimation and compensation. Although the range of motion between frames can be larger than our largest choice of displacement, larger motions can be captured by various combinations of our set of displacements. C. Displacement Compression Network (DCN) After a set of 13 displaced frames are generated from the Displacement Calculation Unit, they are fed into the Displacement Compression Network, where each displacement occupies three channels (RGB), hence the overall input to the DCN comprises 39 channels. The compression network comprises four parts, displacement encoder, displacement decoder, hyper encoder, and hyper decoder. Displacement encoder takes the displaced frame differences calculated from DCU and generates the latent representation y t using several convolutional layers and convolutional LSTM layers similar to other deep learning-based compression architectures [5] , [20] . LSTM [50] as a special RNN structure has proven stable and powerful for modeling long-range dependencies in sequence modeling. The major innovation of LSTM is its memory cell which keeps accumulating the state information. As a result, it helps hold the spatio-temporal information provided by displaced frame differences generated by DCU. The hyper autoencoder uses y t as input to generate side information, which is then used to better compress quantized latent representationŷ t . Finally, the reconstructedd t is generated usingŷ t . The detailed processing flow of the hyper autoencoder is explained in later sections. Traditional quantization inevitably produces zero gradients during backpropagation (BP) which halts network training. Our network deploys BP via stochastic gradient descent, which requires differentiability of all network elements. Hence, we implemented a modified quantizer as in [6] , as follows, wherê y is the binarization of the latent representation of displaced frame differences, which lie between -1 and 1, and represents quantization noise:ŷ = y + ∈ −1, 1 Following quantization, the size ofŷ is H 16 × W 16 × C, where H and W are the height and width of the frame, and C is the number of channels of the last convolution layer in the displacement encoder. In our architecture, C = 128, as shown in Figure 3 . To estimate the entropy of the compressed codes H(ŷ), whereŷ is the quantized latent representation of y, we adopted the hyper-prior scheme proposed by Ballé et al. [6] , where they use an additional set of random variablesẑ to capture the spatial dependencies and model the latent representationŝ y as Gaussian distribution as follows: where pẑ(ẑ) is modeled using the factorized entropy model [8] . The hyperprior autoencoder architecture is indicated by Hyper Encoder and Hyper Decoder in Figure 3 , which is responsible for estimating the parameters of the Gaussian model used for entropy coding. After the displacement encoder encoded the input set of displaced frame differences d t , the resulting latent representation y t with spatially varying standard deviations is fed into the hyper encoder, which summarizes the distribution of standard deviations in the latent representation z t . After quantization and arithmetic coding, the quantizedẑ t is transmitted as side information. The hyper decoder uses the quantizedẑ t as input to obtain Gaussian model parameter σ (µ=0 in our implementation). During modeling training, the Gaussian model parameters can be used to calculate pŷ t and then estimate H(ŷ t ) to guide model optimization. While during model validation and/or testing, the Gaussian model can be used to calculate the cumulative distribution function (CDF) ofŷ t and then guide the arithmetic encoding and decoding of y t , which could further losslessly compressŷ t to bitstream. Figure 4 shows the structure of the Frame Reconstruction Network (FRN). The FRN uses the reconstructed displaced frame differencesd t and the reconstructed previous framê x t−1 as the model input to reconstruct the current frame. The architecture of FRN incorporates Convolutional LSTM (C-LSTM) blocks into a UNet architecture. The UNet architecture, which is an encoder-decoder style network with skip connections, makes it possible to extract and represent meaningful descriptors over multiple image scales. However, without modification, the UNet architecture cannot account for temporal relationships between frames of video data, which are deeply relevant to the efficiency of video compression. The C-LSTM is a convolutional version of the original LSTM, which replaces the matrix multiplication operation of the traditional LSTM with convolutions. It is quite useful for analyzing temporal image sequences, where the C-LSTM layers act as a temporal buffer and capture the long-short dependency of previously processed displaced frame differences. By introducing C-LSTM blocks into the UNet architecture, the FRN is able to process evolving frame properties over multiple scales, by relating compact representations of them in the C-LSTM memory units, leading to better reconstructed frame quality and higher compression rates. We modeled the loss function considering the rate-distortion trade-off as follows: where D and R represent the distortion and rate, respectively. λ controls the trade-off between the number of bits and distortion. D 1 denotes the distortion between the input frame x t and reconstructed framex t measured by MS-SSIM or MSE, and D 2 denotes the distortion between displaced frame differences d t and the reconstructed displaced frame differencesd t measured by MSE. β controls the trade-off between the perceptual distortion D 1 and the pixel-to-pixel distortion D 2 . H(·) represents the bitrates for encoding the latent representationsŷ andẑ estimated by the hyperprior autoencoder. To leverage multi-frame information using our RNN-based codec structure, we update the network parameters every set of N frames during model training, using the loss function in Equation 6 but modified as a sum of losses over the kth set of the N frames indexed x t k +1 , ..., x t k +N : IV. EXPERIMENTS The MOVI-Codec networks were trained end-to-end on the Kinetics-600 dataset [51] , [52] and the Vimeo-90K dataset [53] . The Kinetics-600 videos are downloaded from YouTube, each video having duration of about 10s and various resolutions and frame rates. We used part of the testing set from Kinetics-600, which consists of around 10,000 videos, to conduct our experiments. From each video, a random 128 × 128 patch with 49 frames was randomly selected for training, and the values of each input video were normalized to [-1,1] . We randomly downsampled the original frame and extracted a 128 × 128 patch to reduce any previously introduced compression artifacts. The Vimeo-90K dataset consists of 4,278 videos of fixed resolution 448 × 256. Since the Vimeo-90K dataset has 7 frames per video, we randomly selected a patch of the same size as mentioned before with 7 frames for training. In the Vimeo-90K dataset, the consecutive frames are selected so that the average motion magnitude is between 1-8 pixels, whereas there is no limitation to the motion magnitude between frames in the Kinetics-600 dataset. The mini-batch size is set as 8 for training, and the step length N in our recurrent network is set as 7. By training on both the Vimeo-90K and the Kinetics-600 dataset, we are able to generalize our model to a wide range of natural motions. We tested the MOVI-Codec on the VTL dataset [54] , the JCT-VC [55] (Class B, C, D and E) datasets, and the UVG datasets [56] . These datasets cover a variety of resolutions as shown in Table I . For fair comparison with [21] , [57] and [22] , we tested our framework on the JCT-VC datasets using the first 100 frames, and tested on VTL and UVG using all frames. To evaluate the quality of the reconstructed videos, we used two quality models: the perception-based MS-SSIM [58] and the non-perceptual PSNR. Multiscale SSIM (MS-SSIM) is a widely used image quality assessment model which captures local luminance, contrast, and structural information. For each quality metric, we trained 5 models with different values of the weighting parameter λ to cover different bitrate ranges. For the MS-SSIM model, λ was set to 0.01, 0.05, 0.1, 0.5 and 1.0, respectively. For PSNR based models, λ was set to 0.0005, 0.0025, 0.005, 0.025 and 0.05. We fixed β = 1, since we didn't observe any significant differences in model performance as it was varied over the range 0.1 to 10.0. We compared our method with both traditional and recent deep learning models. H.264 [59] , H.265 [60] and the most recent H.266 [61] were included as representatives of traditional hybrid compression codecs. We follow [21] [22] , and used the x264 and x265 "LDP very fast" mode. For H.266, we followed [62] to implement the "faster" mode. So that we could compare against another motion-free method, we also included the H.265 zero motion setting, using x265 with merange set to zero, which allows exploiting temporal redundancy using an IB prediction structure but without performing motion estimation. In this regard, this setting is most similar to our architecture [63] . Among recent deep learning models, DVC [21] and Wu et al. [20] are optimized for PSNR, Habibian et al. [40] and Cheng et al. are optimized for MS-SSIM, and HLVC [22] has both MS-SSIM optimized and PSNR optimized results. In this section, we compare our video compression engine against the standards H.264, HEVC, and H.266/VVC, and with other deep learning-based video compression architectures (Wu [20] , DVC [21] , [57] , and Cheng [43] ) on the UVG dataset, the VTL dataset, and the HEVC Standard Test Sequences (Class B, Class C, Class D, and Class E). When compressing videos using the H.264 and HEVC codecs, we followed the settings in [21] and used FFmpeg with the very fast mode 1 . When implementing H.266, we followed [62] 1 H.264: ffmpeg -pix_fmt yuv420p -s WxH -r FR -i Video.yuv -vframes N -c:v libx264 -preset veryfast -tune zerolatency -crf Q -g GOP -bf 2 -b_strategy 0 -sc_threshold 0 output.mkv H.265: ffmpeg -pix_fmt yuv420p -s WxH -r FR -i Video.yuvvframes N -c:v libx265 -preset veryfast -tune zerolatency -x265-params "crf=Q:keyint=GOP" output.mkv FR, N, Q, GOP represents the frame rate, the number of encoded frames, quality, GOP size, respectively. N is set to 100 for HEVC datasets. using the faster mode. We also provide visual examples of our approach against other approaches in Figure 5 . More exemplar reconstructed videos are included on our project page with link given in the Abstract. Figures 6, 7, 8, and 9 show the experimental results on the VTL dataset, the UVG dataset, and the HEVC Standard Test Sequences (Class B, Class C, Class D, and Class E). These results show that our network outperformed both H.264 and the HEVC standard against MS-SSIM. On datasets with higher resolution videos (UVG dataset, HEVC Class B dataset, and HEVC Class E dataset), our network was able to outperform the latest H.266 codec at higher bitrates as assessed using the perceptually relevant MS-SSIM algorithm. We also compared our model against several deep learning-based compression models, including a frame interpolation-based model by Wu et al. [20] , DVC [57] , HLVC [22] and the video compression framework proposed by Cheng et al., which uses an added spatial energy compaction penalty in the loss function [43] . Among these, DVC and HLVC were trained on both PSNR and MS-SSIM, to obtain better results against each metric. In our comparison, we include the best performance for these two methods for each metric. It is worth noting that our model only uses one previous frame as input, whereas in Wu's framework, both neighboring frames are utilized when reconstructing the middle frame. Additionally, our framework replaces the classical motion estimation and compensation module by instead training the network to optimally interpolate displaced frame differences. For completeness, we also evaluated all models against the PSNR, where MOVI-Codec did not always perform as well. However, this is a problem with the PSNR, which is not perceptually relevant, and which produces significantly inferior quality predictions than perception-based quality predictors like MS-SSIM [3] . Indeed, the high quality of the reconstructions that we make available on the model page (see link in Abstract) further attests to this. As has been observed by others [8] , [11] , [12] , [41] , perceptual measures are better arbiters of deep compressed video quality than absolute fidelity models like the PSNR. It is worth noting that when comparing our model against the H.265 zero motion setting, while both methods do not utilize motion estimation, MOVI-Codec was able to perform better with respect to MS-SSIM than the H.265 zero motion setting, while delivering similar performance against PSNR. We conducted ablation studies to assess the choices we made in our approach, specifically with respect to the choice of displaced frame differences, and the effectiveness of the proposed LSTM-UNet. The results are shown in Figure 10 and Figure 11 . Figure 10 shows the experimental results on different combinations of displaced frame differences, where s = 0 refers to frame differences with no displacements, which gives the worst performance of all combinations evaluated. This shows the value of "displaced" frame differences as a way of training the network on more diverse motion induced displacements. Including displacements as large as s = 7 greatly increases the overall performance, by allowing interpolation of larger motions in videos. We also tried adding s = 9 to our choice of displacement combinations, but this new combination did not improve the overall performance, meaning that our combination of displacements was adequate to capture motions of various sizes. It is worth noting that as compared with the H.265 zero motion configuration, which also does not utilize motion estimation, our network was able to perform better as assessed by the perceptually relevant MS-SSIM, including when s = 0. Figure 11 shows the experimental results on the HEVC Class B dataset when using UNet and LSTM-UNet to reconstruct frames, respectively. As shown in the example, LSTM-UNet extends the advantage of UNet for extracting and representing spatial descriptors to include spatio-temporal descriptors using C-LSTM blocks, yielding better reconstruction performance. In addition, LSTM-UNet converges faster than the UNet counterparts, shortening the training time of the network. To verify that MOVI-Codec can capture large motions with the chosen set of displacement combination, we calculated the optical flow of adjacent frames in the testing datasets using a pre-trained network called SPynet [46] . To emphasize large motions, we calculated all motion vectors against adjacent frames, and only picked the minimum and maximum motion vectors in the x and y directions. As a result, we ended up with four values of motion vectors for each adjacent frames. Figure 12 shows the distribution of the picked motion vectors on all videos in the HEVC Class B dataset, which is the dataset having the highest resolution videos among our testing datasets. From the figure, we can conclude that our model produced a similar distribution as the original frame pairs, hence our model was able to capture large motions using a set of small displacements. Figure 13 and 14 illustrate the accuracy of our motion reconstruction. The test video in Figure 13 shows the x axis optical flow between two adjacent frames from the Kimono video, which is a video with a moving background and slow motion, whereas Figure 14 shows the optical flow images of two adjacent frames in Basketball Drive video, which has a static background and large motions. In both videos, our model was able to reconstruct motion accurately. To compare the computational complexity of the different codecs, we tested two deep learning models: the one proposed by Wu et al. [20] , the light version of DVC called DVC Lite [57] , and the commercial software x265 for H.265 compression, using a server with an Intel Core i9-9940X CPU and GTX 1080Ti on video sequences of resolution 1920 × 1080. The experimental results are provided in Figure 15 . The overall encoding speed of our framework is mostly invariant of bitrate, whereas since Wu's framework adopts a progressive coding scheme, its encoding speed varies with the target bitrate. In our framework, although we adopted an RNNbased compression method on displaced frame differences, we utilized the RNN unit to store temporal dependencies and did not use a progressive coding scheme for compression. DVC Lite is a lightweight version of DVC with a more efficient motion estimation module and a lightweight motion compression network, which can be twice as fast as the original DVC model in terms of encoding speed [57] . Our framework is faster than the lightweight model, further justifying the use of learned interpolation of displaced frame differences. Since the arithmetic coding at lower bitrates is faster than at larger ones, there is a slight slope to our encoding speed curve. But overall, the complexity of our model is invariant to bitrate, which means that our model maintains a stable encoding speed regardless of video content or bitrate for a given resolution. As shown in Figure 15 , compared with the traditional hybrid codec, our model is faster than the latest codec HEVC with slower setting. However, using the very fast setting on x264 and x265, the encoding speed can run at 110 fps and 30 fps, respectively. Of course, by applying model acceleration techniques such as model distillation, model quantization, or by decreasing the model size, it should be possible to similarly accelerate the encoding speed of our framework. From Figures 6, 7, 8, 9 , and 15, we can conclude that our model delivers better compression performance than LDP veryfast setting of traditional hybrid codecs like H.264 and HEVC in terms of MS-SSIM, at a low computational complexity. This justifies our use of displaced frame differences as motion information for video compression. Although our model was able to acheive competitive performances as lower settings of traditional codec having low computational complexity, and without the complicated motion estimation and compensation modules other deep learning-based models use, our model did not outperform all of the state-of-the-art models. Nonetheless, the performance achieved by our model provides a new way of motion computation that may prove quite useful for video compression. In our model, we designed the set of spatial displacements used by our network to cover a reasonable range of natural motions. A promising future direction is to automatically assign displacement combinations as a function of resolution. The encoding speed of our model is state-ofthe-art among deep learning models, but has not yet been optimized to match compute-optimized traditional codecs like HEVC or VVC, e.g. by model acceleration methods. In this paper, we proposed an end-to-end deep learning video compression framework that renovated motion prediction. To be specific, we proposed the use of displaced frame differences as indicators of motion information, and fed them into a deep space-time compression network, which learns optimal between-frame interpolated representations to achieve efficiency. Additionally, we proposed a new version of UNet, Fig. 8 . PSNR of HEVC test sequences for different compression codecs. The resolution of Class B is 1920 × 1080, of Class C is 832 × 480, of Class D is 416 × 240 and of Class E is 1280 × 720. Overall, our method is competitive with H.265, and is particularity good at lower bit rates on lower resolution datasets. called LSTM-UNet, that utilizes both spatial and temporal information to conduct frame reconstruction. Our experimental results show that our approach outperforms the LDP veryfast setting of the standard codecs H.264 and H.265 in terms of MS-SSIM. In addition, our network was able to outperform the latest H.266 codec at higher bitrates as assessed by the perceptual MS-SSIM algorithm, on high resolution videos. The reduced complexity of the framework and the avoidance of motion search could make it easier to implement on resroucelimited devices, such as smartphones, VR headsets, and AR glasses. Cisco visual networking index: Forecast and methodology Cisco annual internet report Mean squared error: Love it or leave it? a new look at signal fidelity measures Variable rate image compression with recurrent neural networks Full resolution image compression with recurrent neural networks Variational image compression with a scale hyperprior Soft-to-hard vector quantization for endto-end learning compressible representations End-to-end optimized image compression Improved lossy image compression with priming and spatially adaptive bit rates for recurrent networks Lossy image compression with compressive autoencoders Real-time adaptive image compression Generative adversarial networks for extreme learned image compression Conditional probability models for deep image compression Joint autoregressive and hierarchical priors for learned image compression Deep perceptual compression Context-adaptive entropy model for end-to-end optimized image compression Rethinking lossy compression: The ratedistortion-perception tradeoff Practical full resolution learned lossless image compression IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Deepcoder: A deep neural network based video compression Video compression through image interpolation Dvc: An end-to-end deep video compression framework Learning for video compression with hierarchical quality and recurrent enhancement Video quality assessment by reduced reference spatio-temporal entropic differencing On the space-time statistics of motion pictures Towards a theory of early visual processing Some informational aspects of visual perception Temporal decorrelation: a theory of lagged and nonlagged responses in the lateral geniculate nucleus The unsteady eye: an information-processing stage, not a bug Functional asymmetries in on and off ganglion cells of primate retina Microsaccades: A microcosm for research on oculomotor control, attention, and visual perception A compact field guide to the study of microsaccades: Challenges and functions Emergence of simple-cell receptive field properties by learning a sparse code for natural images The JPEG still picture compression standard The JPEG 2000 still image compression standard The latest open-source video codec vp9-an overview and preliminary results Convolutional networks for images, speech, and time series Long short-term memory On the properties of neural machine translation: Encoder-decoder approaches Video compression with rate-distortion autoencoders Learned video compression Learning for video compression Learning image and video compression through spatial-temporal energy compaction A new motion estimation method for motion-compensated frame interpolation using a convolutional neural network Deep frame prediction for video coding Optical flow estimation using a spatial pyramid network Flownet: Learning optical flow with convolutional networks Flownet 2.0: Evolution of optical flow estimation with deep networks Liteflownet: A lightweight convolutional neural network for optical flow estimation Convolutional lstm network: A machine learning approach for precipitation nowcasting The kinetics human action video dataset A short note about kinetics-600 Video enhancement with task-oriented flow VTL test sequences Common test conditions and software reference configurations UVG test sequences An endto-end learning framework for video compression Multiscale structural similarity for image quality assessment Overview of the h. 264/avc video coding standard Overview of the high efficiency video coding (HEVC) standard Overview of the versatile video coding (vvc) standard and its applications Fraunhofer Versatile Video Encoder (VVenC) Distributed video coding: Assessing the hevc upgrade