key: cord-0185589-nip4271g authors: Candila, Vincenzo; Gallo, Giampiero M.; Petrella, Lea title: Mixed-frequency quantile regression with realized volatility to forecast Value-at-Risk date: 2020-11-01 journal: nan DOI: nan sha: f6cc7c3b1ef2d85f779061c468311209b8359d0e doc_id: 185589 cord_uid: nip4271g The use of quantile regression to calculate risk measures has been widely recognized in the financial econometrics literature. When data are observed at mixed-frequency, the standard quantile regression models are no longer adequate. In this paper, we develop a model built on a mixed-frequency quantile regression to directly estimate the Value-at-Risk. In particular, the low-frequency component incorporates information coming from variables observed at, typically, monthly or lower frequencies, while the high-frequency component can include a variety of daily variables, like realized volatility measures or market indices. We derive the conditions for the weak stationarity of the daily return process suggested while the finite sample properties are investigated in an extensive Monte Carlo exercise. The validity of the proposed model is then explored through a real data application using the most important financial indexes. We show that our model outperforms other competing specifications, using backtesting and Model Confidence Set procedures. During the last decades, an extensive financial econometrics literature was devoted to developing model-based approaches to calculate regulatory capital in line with the various requirements imposed by the Basel-II and Basel-III agreements for risk management (Kinateder, 2016) . Among these methods, a special focus was given to forecasting methods for the Value-at-Risk (VaR) measure at a given confidence level τ, VaR(τ), defined as the worst portfolio value movement (return) to be expected at 1 − τ probability over a specific horizon (Jorion, 1997) , able to capture tail market risk in a perspective fashion. As the τ-quantile of a portfolio return distribution, the VaR(τ) can be predicted as the product of the portfolio volatility forecast times the quantile of the hypothesized distribution. For the first component, volatility clustering, modeled by conditionally autoregressive models (such as the ARCH/GARCH - Engle, 1982; Bollerslev, 1986) , produces good forecasts capable of reproducing well known stylized facts of financial time series, including skewed behavior and fat tails (Cont, 2001; Engle and Patton, 2001 , among others). Further improvements were made possible by the direct predictability of realized measures of financial volatility (Andersen et al., 2006b) . While a choice of a specific parametric distribution for the innovation term mey be uninfluential for model parameter estimation (Bollerslev and Wooldridge, 1992) , unless a few extreme events (e.g. the Flash Crash of May 2005) occur (Carnero et al., 2012) , a wrong choice of distribution for the innovation term delivers an inadequate VaR(τ) forecasting: see for example Manganelli and Engle (2001) and El Ghourabi et al. (2016) . As an alternative, the VaR(τ) can be directly derived through quantile regression methods (Koenker and Bassett, 1978; Engle and Manganelli, 2004) where no distributional hypothesis is required. A first suggestion in this direction comes from Koenker and Zhao (1996) who use quantile regression for a particular class of ARCH models, i.e., the Linear ARCH models (Taylor, 1986) , chosen for its ease of tractability in deriving theoretical properties. Subsequent refinements are, for instance, Xiao and Koenker (2009) , Lee and Noh (2013) , Zheng et al. (2018) for GARCH models, Noh and Lee (2016) who consider asymmetry, Chen et al. (2012) who consider nonlinear regression quantile approach with intra-day price, Bayer (2018) who combines VaR forecasts via penalized quantile regressions, Taylor (2019) who considers a quantile regression approach to joint estimates VaR and Expected Shortfall and the multivariate generalization of Merlo et al. (2021) . A relatively recent stream of literature investigates the value of information provided by data available at both high-and low-frequency incorporated into the same model in assessing the dynamics of financial market activity: this is the case of the GARCH-MIDAS model proposed by Engle et al. (2013) (building on the MI(xed)-DA(ta) Sampling approach by Ghysels et al., 2007) , the regime switching GARCH-MIDAS of Pan et al. (2017) and the recent paper by Xu et al. (2021) who consider a MIDAS component in the Conditional Autoregressive Value-at-Risk (CAViaR) of Engle and Manganelli (2004) . In what follows, we propose a Mixed-Frequency Quantile ARCH model (MF-Q-ARCH, extending Koenker and Zhao, 1996) : we show how the constant term in the quantile regression can be written as a function of data sampled at lower frequencies (and hence become a low-frequency component), while the high frequency component is regulated by the daily data. As a result, with the aim of capturing dependence on the business cycle, we benefit from the information contained in macroeconomic variables (cf. Mo et al., 2018; Conrad and Loch, 2015, among others) , and we achieve a quite flexible representation of volatility dynamics. Since both components enter additively, our model can be seen as a quantile model version of the Component GARCH by Engle and Lee (1999) . In the proposed model, we also include a predetermined variable observed daily, typically a realized measure: this adds the "-X" component in the resulting MF-Q-ARCH-X model. This variable can capture extra information useful in modeling and forecasting future volatility and may improve the accuracy of VaR forecasts. Such a use in the quantile regression framework is not new in itself: the paper by Gerlach and Wang (2020) jointly forecast VaR and Expected Shortfall and Zhu et al. (2020) forecast VaR by adopting a GARCH-X model for the volatility term. Also the work ofŽikeš and Baruník (2016) uses the realized measures in the context of quantile regressions to investigate the features of conditional quantiles of realized volatility and asset returns. From a theoretical point of view, we provide the conditions for the weak stationarity of the daily return process suggested and we propose a sequential test aimed at finding the optimal number of daily return lags to be included in MF-Q-ARCH-X model. The finite sample properties are investigated through an extensive Monte Carlo exercise. The empirical application is carried out on the VaR predictive capability for the S&P 500, NASDAQ, and Dow Jones indices, showing that the proposed MF-Q-ARCH-X outperforms all the other competing models considered across all market indices when the VaR forecasts are evaluated by backtesting and constructing the Model Confidence Set (MCS, Hansen et al., 2011) . The rest of the paper is organized as follows. In Section 2 we introduce the notation and the basis for a dynamic model for the VaR(τ) and we provide details of the conditional quantile regression approach. Section 3 presents our MF-Q-ARCH-X model. Section 4 is devoted to the Monte Carlo experiment. Section 5 details the backtesting and MCS procedures. Section 6 illustrates the empirical application. Conclusions follow. For the purposes of this paper we will adopt a double time index, i,t, where t = 1, . . . , T scans a low frequency time scale (i.e., monthly) and i = 1, . . . , N t identifies the day of the month, with a varying number of days N t in the month t, and an overall number N of daily observations N = ∑ T t=1 N t . Let the daily returns r i,t be, as customarily defined, the log-first differences of prices of an asset or a market index, and let the information available at time i,t be F i,t . In what follows, we are interested in the conditional distribution of returns, with the assumption: where z i,t iid ∼ (0, 1) having a cumulative distribution function denoted by F(·). The zero conditional mean assumption in Eq. (1) is not restrictive, in fact, when explicitly modeled, such a conditional mean is very close to zero, consistently with the market efficiency hypothesis. Based on this setup, the conditional (one-step-ahead) VaR for day i,t at τ level (VaR i,t (τ)) for r i,t is defined as i.e., the τ-th conditional quantile of the series r i,t , given F i−1,t ; consequently, we can write where For a given τ, the traditional volatility-quantile approach to estimate the VaR i,t (τ) is thus based on modeling σ i,t from a dynamic model of either the conditional variance of returns (following Engle, 1982; Bollerslev, 1986) or as a conditional expectation of a realized measure (Andersen et al., 2006a) and retrieving the constant F −1 (τ) either parametrically or nonparametrically. In either case, from an empirical point of view, it turns out that distribution tests mostly reject specific parametric choices, and that using the empirical distributions is prone to bias/variance problems and lack of stability through time. Alternatively, we can estimate directly Q r i,t (τ|F i−1,t ) using a quantile regression approach (Koenker and Bassett, 1978; Engle and Manganelli, 2004) which has become a widely used technique in many theoretical problems and empirical applications. While classical regression aims at estimating the mean of a variable of interest conditioned to regressors, quantile regression provides a way to model the conditional quantiles of a response variable with respect to a set of covariates in order to have a more robust and complete picture of the entire conditional distribution. This approach is quite suitable to be used in all the situations where specific features, like skewness, fat-tails, outliers, truncation, censoring and heteroskedasticity are present. The basic idea behind the quantile regression approach, as shown by Koenker and Bassett (1978) , is that the τ-th quantile of a variable of interest (in our case r i,t ), conditional on the information set F i−1,t , can be directly expressed as a linear combination of a q + 1 vector of variables x i−1,t (including a constant term), with parameters Θ τ , that is: An estimator for the (q+1) vector of coefficients Θ τ is obtained minimizing a suitable loss function (also known as check function):Θ with ρ τ (u) = u (τ − 1(u < 0)), where 1(·) denotes an indicator function. In our context, the advantage of such an approach is to avoid the need to specify the distribution of z i,t in Eq. (1), either parametrically or nonparametrically. Following the approach by Koenker and Zhao (1996) , we assume a Linear ARCH specification for r i,t , namely, a dependence of σ i,t on past absolute values of returns: where we assume that 0 < β 0 < ∞, β 1 , . . . , β q ≥ 0. Thus, substituting the generic term x i−1,t in Eq. (3) with the specific vector relative to the Linear ARCH model, we have The Linear ARCH choice turns out to be convenient, since it allows for a direct comparability of the two setups to estimate the VaR(τ) in Eq. (2): which establishes the equivalence Θ F −1 (τ) = Θ τ which will prove useful later in our Monte Carlo simulations. There is a second justification to adopt the Linear ARCH: the term σ i,t defining the volatility of returns, can also be seen as the conditional expectation of absolute returns in the Multiplicative Error Model representation used by Engle and Gallo (2006) : The term η i,t is an i.i.d. innovation with non-negative support and unit expectation, and the Eq. (7) can be used to derive an estimate of the VaR. The Linear ARCH is therefore a simple and convenient nonlinear autoregressive model for |r i,t | with multiplicative errors, which we hold as the maintained base specification to explore the merits of our proposal. Moreover, this lays the grounds for extending the approach, using other specifications for σ i,t in Eq. (5) as functions of past volatility-related observable variables. Alternatives can be considered, for example: with rv i,t the daily realized volatility. We will return on this point below. In what follows we will introduce a MIDAS extension to the Linear ARCH model in a quantile regression framework, taking advantage of the well-known predictive power of low-frequency variables for the volatility observed at a daily frequency (e.g. Conrad and Kleen, 2020) . We also augment the Linear ARCH to include a lagged realized measure of volatility (see also Gerlach and Wang, 2020, within a CAViAR context), in order to add the informational content of a more accurate measure to the volatility dynamics. 3 The MF-Q-ARCH-X model In order to take advantage of the information coming from variable(s) observed at different frequency we introduce a low-frequency component in model (5). This low-frequency term represents a one-sided filter of K lagged realizations of a given variable MV t (usually a macroeconomic variable), through a weighting function δ (ω), where ω = (ω 1 , ω 2 ). Our resulting Mixed-Frequency Quantile ARCH (MF-Q-ARCH) model becomes: where the parameter θ represents the impact of the weighted summation of the K past realizations of MV t , observed at each period t, that is, The importance of each lagged realization of MV t depends on δ (ω), which can be assumed as a Beta or Exponential Almon lag function (see, for instance, Ghysels and Qian, 2019) . Here we use the former function, that is: Eq. (10) is a rather flexible function able to accommodate various weighting schemes. Here we follow the literature and give a larger weight to the most recent observations, that is, we set ω 1 = 1 and ω 2 ≥ 1. The resulting weights δ k (ω) are at least zero and at most one, and their sum equals one, so that ∑ K k=1 δ k (ω)MV t−k is an affine combination of (MV t−1 , · · · , MV t−K ). In order to refine the VaR dynamics in our model, we include a predetermined variable X i,t , so that we can explore the empirical merits of such an extended specification, already present in the GARCH and MEM literature (Han and Kristensen, 2015; Engle and Gallo, 2006) . Such a variable may be the realized volatility of the asset or a market volatility index (see the use of the VIX in Amendola et al., 2021, among others) . The resulting eXtended Mixed-Frequency Quantile ARCH model, labelled MF-Q-ARCH-X, becomes: In either Eq. (9) or (11), the first component (including the constant) depends only on the low-frequency term (changing at every t, according to the term W S t−1 ), while the second comprises variables changing daily (i.e., every i,t) and include lagged returns and the high-frequency volatility term. In such a representation, the two components enter additively, in the spirit of the component model of Engle and Lee (1999) : which, for the MF-Q-ARCH-X, becomes In the following theorem we show that, under mild conditions, the process in (13) is weakly stationary: Theorem 1. Let MV t and X i,t be weakly stationary processes. Assume that has all roots λ inside the unit circle. Then the process r i,t in (13) is weakly stationary. Proof: see Appendix. In order to make inference on the MF-Q-ARCH-X model we need to solve Eq. (4) where The estimation of the vector Θ τ is encumbered by the fact that the mixed-frequency term W S t−1 is not observable, as it depends on the unknown ω 2 parameter of the weighting function δ k (ω), also to be estimated. To make estimation feasible, we resort to the expedient of profiling out 1 the parameter ω 2 , through a two-step procedure: we first fix ω 2 at an initial arbitrary value, say ω (b) 2 , which turns the vector x i−1,t into a completely observable counterpart, in short x i−1,t . This gives a solution to the minimization of the loss function, which is dependent on ω This procedure is repeated over a grid of B values for ω 2 , so that we have Θ , and the chosen overall estimator is ω * 2 , Θ ( * ) τ , corresponding to the smallest overall value of the loss function. Accordingly, the MF-Q-ARCH-X estimator of the VaR is To obtain reliable VaR estimates in our model (18), an important issue is the choice of the optimal number of lags q for the daily absolute returns in Eq. (5). To that end, we select the lag order suggested by a sequential likelihood ratio (LR) test on individual lagged coefficients (see also Koenker and Machado, 1999) . In particular, for a given τ, at each step j of the testing sequence over a range of J values, we compare the unrestricted model where the number of lags is set equal to j (labelled U, with an associated loss function V ( j) U,τ ), against a restricted model where the number of lags is j − 1 (labelled R, with an associated loss function V ( j−1) R,τ ). In this setup, the null hypothesis of interest is i.e., the coefficient on the most remote lag is zero. The procedure starts contrasting a lag-1 model against a model with just a constant, then a lag-2 against a lag-1, and so on. For a given τ, at each step j, we calculate the test statistic where s(τ) is the so-called sparsity function estimated accordingly to Siddiqui (1960) and Koenker and Zhao (1996) . Under the adopted configuration, LR τ is asymptotically distributed as a χ 2 1 , so that we select q to be the last value of j in the sequence, for which we reject the null hypothesis. The finite sample properties of the sequential test and of the estimator of the MF-Q-ARCH model can be investigated by means of a Monte Carlo experiment. In what follows we consider R = 5000 replications of the data generating process (DGP): where we assume a N (0, 1) distribution for z i,t and we set to zero the relevant initial values for r i,t . Moreover, the stationary variable MV t entering the weighted sum W S t−1 is assumed to be drawn from an autoregressive AR(1) process MV t = ϕMV t−1 + e t , with ϕ = 0.7 and the error term e t following a Skewed t-distribution (Hansen, 1994) , with degrees of freedom d f = 7 and skewing parameter sp = −6. The frequency of MV t is monthly and K = 24. The values of the parameters (collected in a vector Θ) are detailed in the first column of the Tables 2-4. For the simulation exercise we consider N = 1250, N = 2500 and N = 5000 observations, to mimic realistic daily samples. Finally, three different levels of the VaR coverage level τ are chosen: 0.01, 0.05, and 0.10. In the Monte Carlo experiment, we start by evaluating the features of the LR test for the lag selection in Eq. (20). To that end, we test sequentially H 0 : β j = 0 over J steps at a significance level α. Since the DGP is a fourth-order process, we expect to have a high rejection rate when the null involves a zero restriction on coefficients β j , j = 1, . . . , 4. In order to confirm the expected low rate of rejections, we extend the sequence of testing of further β j 's, up to J = 6. Looking at the Table 1 , where we report the percentages of rejections for different VaR coverage levels τ = 0.01, 0.05, 0.1 at the nominal significance level of α = 5% across replications, we validate the good behavior of the test. Overall, the sequential test procedure satisfactorily identifies the number of lags to be included in the MF-Q-ARCH model, with the performance improving with the number of observations, especially for H 0 : β 4 = 0; for the latter case, the percentage of rejections of the null increases considerably across coverage levels when N = 5000. Turning to the small sample properties of our estimator, the evaluation is done in terms of the original coefficients in the DGP, collected in the vector Θ = β 0 , θ , β 1 , . . . , β q , using the relation-ship with the quantile regression parameters Θ τ , i.e., Θ = Θ τ /F −1 (τ). 2 In Tables 2, 3 and 4 we report the Monte Carlo averages of the parameters (Θ) across replications for three levels of τ, and the estimated Mean Squared Errors relative to the true values. Overall, the proposed model presents good finite sample properties: independently of the τ level chosen, for small sample sizes, the estimates appear, in general, slightly biased, although, reassuringly, the MSE of the estimates relative to the true values always decreases as the sample period increases. The bulk of the evaluation of our model lies in its capability to capture the coverage of VaR at a given τ level, especially in comparison to other models. In the next sections we illustrate two of the main procedures used in literature to evaluate the quality of the Building models able to predict VaRs efficiently is our main objective in this paper. For this reason, it is quite relevant to evaluate the quality of the VaR estimates by performing a set of targeted tests. Nowadays, in the context of risk management, backtesting is one of most recognized model's evaluation procedure (see the reviews of Campbell, 2006; Nieto and Ruiz, 2016, among others) . For this reason, in this study, we evaluate the VaR predictive ability of our model by means of four backtesting procedures: the Actual over Expected (AE) exceedance ratio, the Unconditional Coverage (UC, Kupiec, 1995) , the Conditional Coverage (CC, Christoffersen, 1998) , and the Dynamic Quantile (DQ, Engle and Manganelli, 2004) (Berkowitz et al., 2011) to have more power over it. In particular, the DQ test consists of running a linear regression where the dependent variable is the sequence of VaR violations and the covariates are the past violations and possibly any other explanatory variables. Under the null hypothesis, the test statistic follows a χ 2 q , with q denoting the rank of the explanatory variable matrix. The results of the backtesting procedures may be ambiguous as to the superiority of one model over the others: we therefore propose to adopt a MCS procedure in order to find the subset of superior models according to a suitable loss function. By the MCS methodology we construct a "superior set of models" (SSM) among several competing ones. Initially proposed by Hansen et al. (2011) , it permits to discriminate within an initial set of models M 0 , labelled as l = 1, . . . , M, the SSM with dimension m * ≤ M, denoted by M * 1−δ . In other words, the SSM M * 1−δ is a subset of the full class of models containing the best specifications with a confidence level 1 − δ . In our VaR framework, such a procedure relies on loss functions based on quantiles mapping the distance between the observed returns, r i,t , and the estimated VaR. Let VaR l,i,t (τ) be the one-step-ahead VaR forecasts for day i and month t, obtained from model l at level τ. Formally, the loss function associated to the model l is: l,i,t = (r i,t , VaR l,i,t (τ)). In line with González-Rivera et al. (2004) and Laporta et al. (2018) , we use an asymmetric loss function 3 penalizing more heavily the negative returns falling below the VaR, that is: (r i,t , VaR l,i,t (τ)) = τ − 1 (ri,t< VaR l,i,t (τ)) r i,t − VaR l,i,t (τ) , where 1 (·) is an indicator function equals to one if the argument is true. The relative performance of two models, say l and k, is evaluated by means of the loss differential d lk,i,t , that is: Let µ lk ≡ E(d lk,i,t ) be the expectation of the loss differential. Clearly, when µ lk < 0, model l is preferred to model k. The MCS procedure consists of a sequence of equal predictive ability tests, with the null hypotesis given by: with M ⊂ M 0 . As in Cipollini et al. (2021) , the test used in the MCS procedure is based on the semi-quadratic (SQ) test statistic (Hansen et al. (2011) and Clements et al. (2009) ) given by: where d lk is the sample loss differential between model l and model k and var(d lk ) is the estimate variance of d lk . Intuitively, large values for T SQ provide evidence against model l with respect the others in M. Consequently, model l should be eliminated from M according to a defined elimination rule. For details on the proper elimination rule, we refer to Hansen et al. (2011) . Therefore, the steps of the MCS procedure can be synthesized as follows: 1. Configure the set of models to evaluate with M = M 0 . (24), with a significance level δ , using the loss function in (22) to calculate the loss differential. 3. If the null hypothesis is not rejected at the δ significance level, then M * 1−δ = M. Alternatively, eliminate a model from M according to the defined elimination rule and repeat Step 2. In this section we apply the MF-Q-ARCH-X model to estimate VaRs at different τ levels for three main daily financial indices: the S&P 500, the NASDAQ and the Dow Jones. We compare the estimate VaRs with several well-known competitive models: the Linear ARCH estimated in the quantile context (Q-ARCH), the Symmetric Absolute Value (SAV), Asymmetric Slope (AS) and Indirect GARCH (IG) specifications of the CAViaR (Engle and Manganelli, 2004 ) model, the GARCH, the GARCH with Student's t error term (GARCH-t), the GJR-GARCH (Glosten et al., 1993) , the GJR-GARCH with Student's t error term (GJR-t), the RiskMetrics (Riskmetrics, 1996) model and the GARCH-MIDAS, with the same low-frequency variables used in the proposed MF-Q-ARCH-X model. All the functional forms of these models are reported in Table 5 . The period of investigation covers almost 20 years, from January 2001 to February 2020 on a daily basis. The data have been collected from the realised library of the Oxford-Man Institute (Heber et al., 2009) , while the monthly variables come from the Federal Reserve Economic Data (FRED) database: the U.S. Industrial Production (IP t ), the Housing Starts (HS t ), taken as monthto-month percentage change, and the National Activity Index (NAI t ) (as done by Conrad and Loch, 2015 , for instance). Summary statistics for the daily log-returns, for the 5-minute realized volatilities (which is included in the MF-Q-ARCH-X model as "-X" variable) of the three indices and Notes: The table reports the functional forms for the MF-Q-ARCH-X, MF-Q-ARCH, Q-ARCH, Symmetric Absolute Value (SAV), Asymmetric Slope (AS), Indirect GARCH (IG), GARCH, GARCH with Student's t distribution for the errors (GARCH-t), GJR, GJR with Student's t distribution for the errors, and GARCH-MIDAS specifications. for the MIDAS variables are reported in Table 6 . In the same table we also report the estimated coefficients of the predictive regression proposed by Conrad and Schienle (2020) , aiming at assessing if the low-frequency component has some explanatory power for the volatility of the variable of interest. Since all the coefficients are statistically different from zero, the low frequency variables can be considered as contributing to the volatility predictions. ) and Housing Starts (HS t ), and the National Activity Index (NAI t ). The symbol ∆ refers to 100 · ((MV t /MV t−1 ) − 1), for a given Macro-Variable MV . π 1 (·) represents the estimated coefficient for the variable in parentheses of the predictive regression as proposed by Conrad and Schienle (2020) : log( RV t ) = c + π 1 x t−1 + π 2 log( RV t−1 ) + ζ t , where RV t is the volatility-adjusted realized variance, obtained summing monthly the squared deGARCHed returns, that is the returns divided by the estimated conditional volatility under the null of π 1 = 0, x t is the additional variable (the monthly variables) and ζ t is the iid innovation term. * , * * and * * * represent the significance at levels 10%, 5%, 1%, respectively, associated to HAC robust standard errors. In Figure 1 we present the patters of the daily log-returns (in percentage scale) of the three indices considered. Some commonalities can be found: first, the increased variability during the Dot-Com bubble (March 2001 to November 2001 and the Great Recession period (December 2007 to June 2009), identified by shared areas according to the U.S. recession periods (NBER dating); second, a tendency to a generalized downturn in the indices, due to the beginning of the Covid-19 pandemic. The plots of the monthly variables are reported in Figure 2 . The top and central panels show the first differences of IP t and HS t , while the bottom panel shows the NAI t pattern. It can be noted that, during the Great Recession, IP t (which is a proxy of the aggregate demand) and NAI t present some negative peaks, which in turn may correspond to an increased volatility in the stock markets. In what follows, we illustrate the results of the implementation of the MF-Q-ARCH-X model. In particular, we perform an in-sample analysis where running the LR test implemented in Section 3.2, we define the optimal number of daily lags in absolute log-returns to be included. Finally, in the out-of-sample analysis, we illustrate model performances in a comparative fashion using both backtesting and MCS procedures. Tables 7 reports the p-values of LR test using τ = 0.05, with an in-sample dataset from April 2016 to January 2001, i.e., approximately the 80% of the full sample period, for all the indices under investigation. The application of the LR test suggests the inclusion up to eight lagged daily logreturns in the models for all the three indices. As regards the number of lagged realizations entering the low-frequency component, we choose K = 12. The patterns of the resulting weighting schemes selected by the profiling procedure illustrated in Section 3.2 for the MF-Q-ARCH and MF-Q-ARCH-X models, and for all the monthly variables used in this work, are depicted in Figure 3 . Interestingly, all the patterns of the resulting weighting schemes are very reasonable: further observations have weights close to zero. Looking at the weighting schemes for MF-Q-ARCH-X and MF-Q-ARCH using ∆HS t in the low-frequency component (Panels 3c and 3d, respectively), the profiling procedure signals the same importance (approximately) to the macro-variable lagged realizations for the Dow Jones and S&P 500. For the other panels, there is no evidence of huge differences among the indices under investigation, in terms of the weights associated to the past macro-variable observations. The empirical analysis is completed by the out-of-sample analysis, where we compare the perfomances of the MF-Q-ARCH and the MF-Q-ARCH-X models in VaR prediction with those mentioned at the beginning of this section. In line with Lazar and Xue (2020) , the one-step-ahead VaR forecasts are obtained with parameters estimated every ten days, using a rolling window of size 1500 observations. The out-of-sample period represents approximatively 20% of the full sample, and covers the period from May 2016 to February 2020. As an example, the plots of the VaR estimates with the daily log-returns for the proposed MF-Q-ARCH-X, with ∆IP t as MIDAS variable Notes: Plots of weighting schemes, for the S&P 500 (green lines), NASDAQ (red lines), and Dow Jones (blue lines), according to the profiling procedure described in the text, for the in-sample period. and the realized volatility as "-X" term, are reported in Figure 4 . The results of the backtesting procedures are synthesized, for each index under investigation, in Tables from 8 to 10. In terms of the AE ratio, it can be noted that the MF-Q-ARCH models, with and without the "-X" component, generally perform better than the competing models, and more importantly, better than the main competitor (the Q-ARCH model). Sometimes, like for the S&P 500 index, the MF-Q-ARCH model with NAI t as low-frequency term has an AE ratio of one, which means that the expected and actual number of VaR violations coincide. Considering all the indices used in this work, the proposed models have the AE ratios ranging from 0.793 to 1.189. Conversely, the AR ratios of the GARCH, GARCH-t, GJR, GJR-t and GARCH-MIDAS models swing between 0.688 and 1.210, which represents a wider interval with respect to that of the MF-Q-ARCH-X models. Even the performance of the CAViaR specifications appears inferior than that of the Q-ARCH-based models with low-frequency and "-X" components. In fact, the AR ratios of the SAV, AS and IG range between 0.708 and 1.522 (which is indeed a severe underestimation of the VaR). These results witness the benefits of including the macro-variables within the Q-ARCH framework. With reference to the other three backtesting procedures, it is worth noting that the proposed MF-Q-ARCH-X and MF-Q-ARCH models always pass the tests used, at the 5% significance level (signalled by shades of gray in the tables), independently of the index considered. On the other hand, the standard GARCH-based models seem to produce inferior VaR forecasts, given that these models do not always pass the backtesting procedures. Interestingly, the specifications belonging to the CAViaR have an inconclusive performance. The SAV model overcomes the backtesting procedure only for the Dow Jones, the AS for S&P 500 and Dow Jones, and the IG only for the NASDAQ. To conclude, if the proposed models appear to brilliantly pass the backtesting procedures, irrespectively of the index considered, the other models do not have very good performances. The inclusion in the SSM is detailed in Table 11 , according to the MCS procedure and the loss function in (22), whose averages are reported in column . The table includes the SSM for all the indices and with two significance levels: 0.25 and 0.10. In line with what found earlier, also the MCS signals the superiority of the proposed MF-Q-ARCH-X model, independently of the index and significance level employed. Looking at the NASDAQ index, the only models entering the SSM are the three MF-Q-ARCH-X, based on the macro-variables variables considered. Interestingly, neither the GARCH-based models or the GARCH-MIDAS specifications enter the SSM. Instead, some models taking into account the leverage effect through some asymmetric terms (like the GJR-t and the AS) are sometimes included in the SSM. This paper explored the possibility of including mixed-frequency components in a quantile regression approach to VaR estimation, specifically introducing a low-frequency and a high-frequency component within a dynamic model of volatility. As an illustration, here we used a Linear ARCH-X which includes lagged realized volatility as an extended contribution to the short-run volatility dynamics. For the MF-Q-ARCH-X model, the accuracy of Value-at-Risk is increased by assuming that the underlying slow evolution of the volatility of returns is influenced by some macroeconomic variables (indeed observed at lower frequencies) and some additional predetermined daily variables. In estimating the MF-Q-ARCH-X by quantile regression, no distribution for the returns is necessary and robustness to outliers in the data is guaranteed. From a statistical point of view, we suggested an estimation procedure, having first established the conditions for the weak stationarity of our MF-Q-ARCH-X process. Moreover, we proposed a sequential test with the aim of finding the optimal number of daily lagged absolute returns to include in the high-frequency component of the MF-Q-ARCH-X model. The finite sample properties of the proposed model were investigated through an extensive Monte Carlo exercise. Overall, the estimates appear satisfactory and the resulting VaR forecasts are robust to some misspecification in the weighting parameter which may affect the low-frequency component. In the empirical application, the proposed MF-Q-ARCH-X model was extensively evaluated, both in-and out-of-sample, against several popular alternatives, with applications to the S&P 500, NASDAQ, and Dow Jones indices. All models based on the proposed MF-Q-ARCH-X specification brilliantly pass the backtesting procedures; the same does not hold for some popular competing models belonging to the parametric framework, like the GARCH and the GARCH-MIDAS. In addition, only our MF-Q-ARCH-X models consistently enter the set of superior models identified in the Model Confidence Sets. Further research may focus on the multivariate extension of the proposed specification, as done by Bernardi et al. (2017) , Petrella and Raponi (2019) , and Du and Pei (2020), among others. An- From the model in (13), we can write: r s p = σ s z s p = σ s p · z s p , (A.1) given the independence between σ s and z s . For p = 1, the right hand side (RHS) of (A.1) is zero, because z s i.i.d. ∼ (0, 1). Let us now focus on p = 2; let us replace the second term of the RHS of (A.1), having assumed that z s 2 = z * < ∞: r s r = z * E(β 0 + β 1 |r s−1 | + · · · + β q |r s−q | + θ |W S s−1 | + β X |X s−1 |) 2 1/2 (A.2) ≤ z * (β 0 + β 1 r s−1 2 + · · · + β q r s−q 2 + θ W S s−1 2 + β X X s−1 2 ). (A.3) Let us now translate this expression in matrix notation. Therefore, let us collect terms in a vector indexed by s, that is, ξ s = r s 2 , · · · , r s−q+1 2 , W S s 2 , X s 2 , and let the (q + 2) × (q + 2) dimensional companion matrix A, the vectors b and c z * β 1 z * β 2 · · · z * β q−1 z * β q z * θ z * β x 1 0 · · · 0 0 0 0 0 1 · · · 0 0 0 0 . . . . . . . . . . . . . . . . . . . . . 0 0 · · · 1 0 0 0 0 0 · · · 0 0 0 0 0 0 · · · 0 0 0 0 0 0 · · · 0 0 0 0 where we have made us of the fact that, because of the stationarity of W S s and X s , the vector c does not depend on time. Thus, we have: ξ s ≤ Aξ s−1 + b + c. (A.4) Substituting recursively ξ s−1 backwards, and letting I q+2 be the identity matrix of size (q + 2), ξ s ≤ A (Aξ s−2 + b + c) + b + c (A.5) ≤ A 2 ξ s−2 + Ab + b + Ac + c ≤ A 2 ξ s−2 + I q+2 + A b + I q+2 + A c (A.6) ≤ A 3 ξ s−3 + I q+2 + A + A 2 b + I q+2 + A + A 2 c (A.7) . . . ≤ A m ξ s−m + I q+2 + A + A 2 + · · · + A m−1 b + I q+2 + A + A 2 + · · · + A m−1 c. (A.8) Recall the characteristic polynomial of A is φ (λ ), defined by Eq. (14), namely, φ (λ ) = z * β 1 λ q+1 + β 2 λ q + · · · + β q λ q−2 − λ q+2 , (A.9) which has all eigenvalues λ lie inside the unit circle. When m → ∞, for the eigen-decomposition theorem, this implies that lim m→∞ A m = 0, (A.10) and that lim m→∞ (I q+2 + A + A 2 + · · · + A m−1 ) = (I q+2 − A) −1 . (A.11) Putting terms together, therefore, as m → ∞ we can say that that is the RHS converges to a finite expression not depending on time, establishing the result. Evaluation of volatility predictions in a VaR framework On the asymmetric impact of macro-variables on volatility Choosing the frequency of volatility components within the Double Asymmetric GARCH-MIDAS-X model Practical volatility and correlation modeling for financial market risk management Volatility and correlation forecasting Combining Value-at-Risk forecasts using penalized quantile regressions Evaluating value-at-risk models with desk-level data Multiple risk measures for multivariate dynamic heavy-tailed models Generalized autoregressive conditional heteroskedasticity Quasi-maximum likelihood estimation and inference in dynamic models with time-varying covariances A review of backtesting and backtesting procedures Estimating GARCH volatility in the presence of outliers Forecasting value-at-risk using nonlinear regression quantiles and the intra-day range Evaluating interval forecasts Realized volatility forecasting: Robustness to measurement errors On the efficacy of techniques for evaluating multivariate volatility forecasts Two are better than one: Volatility forecasting using multiplicative component GARCH-MIDAS models Anticipating long-term stock market volatility Testing for an omitted multiplicative long-term component in GARCH models Empirical properties of asset returns: stylized facts and statistical issues Backtesting portfolio value-at-risk with estimated portfolio weights Consistent Estimation of the Value at Risk when the Error Distribution of the Volatility Model is Misspecified Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflation A multiple indicators model for volatility using intra-daily data Stock market volatility and macroeconomic fundamentals Cointegration, Causality, and Forecasting: A Festschrift in Honor of CAViaR: Conditional autoregressive value at risk by regression quantiles What good is a volatility model? Higher order elicitability and Osband's principle Semi-parametric dynamic asymmetric Laplace models for tail risk forecasting, incorporating realized measures Estimating MIDAS regressions via OLS with polynomial parameter profiling MIDAS regressions: Further results and new directions On the relation between the expected value and the volatility of the nominal excess return on stocks Forecasting volatility: A reality check based on option pricing, utility function, value-at-risk, and predictive likelihood Semiparametric Multiplicative GARCH-X Model: Adopting Economic Variables to Explain Volatility Autoregressive conditional density estimation The Model Confidence Set Omi's realised library, version 0.1 Value at Risk. Chicago: Irwin Basel II versus III: a comparative assessment of minimum capital requirements for internal model approaches Regression quantiles Goodness of fit and related inference processes for quantile regression Conditional quantile estimation and inference for ARCH models Techniques for verifying the accuracy of risk measurement models Selection of value at risk models for energy commodities Forecasting risk measures using intraday data in a generalized autoregressive score framework Quantile regression estimator for GARCH models Value at risk models in finance Forecasting VaR and ES using a joint quantile regression and its implications in portfolio allocation The macroeconomic determinants of commodity futures volatility: Evidence from Chinese and Indian markets Frontiers in VaR forecasting and backtesting Quantile regression for location-scale time series models with conditional heteroscedasticity Macroeconomic uncertainty and expected shortfall (and value at risk): a new dynamic semiparametric model Oil price volatility and macroeconomic fundamentals: A regime switching GARCH-MIDAS model Joint estimation of conditional quantiles in multivariate linear regression models with an application to financial distress JP Morgan Technical Document Distribution of quantiles in samples from a bivariate population Forecasting value at risk and expected shortfall using a semiparametric approach based on the asymmetric Laplace distribution Modeling Financial Time Series Conditional quantile estimation for generalized autoregressive conditional heteroscedasticity models Quantile-based GARCH-MIDAS: Estimating value-at-risk using mixed-frequency information Hybrid quantile regression estimation for time series models with conditional heteroscedasticity Quantile estimation of regression models with GARCH-X errors. Forthcoming in Statistica Sinica Semi-parametric conditional quantile models for financial returns and realized volatility The table reports the inclusion in the SSM according to the MCS procedure. Column "LF var." reports the low-frequency variables. Column represents the averages of the loss function in (21), used in the MCS procedure. The out-of-sample period, which starts on May 2, 2016 and ends on February 28, 2020 (960, 959 e 958 days, respectively, for S&P 500, NASDAQ and Dow Jones). The models with the "-X" label include the realized volatility (see Eq. (13)). Symbols 25 and 10 denote inclusion in the SSM at significance level α = 0.25 and α = 0.10, respectively. other interesting point would be the inclusion of an asymmetric term in the proposed model, both for what concerns the daily returns and the low-frequency component, as done by Amendola et al. (2019) , for instance. Finally, the current approach involving the low-frequency and "-X" terms could be enriched by jointly estimating both the VaR and the Expected Shortfall measures, as recently developed by Fissler et al. (2016) , Taylor (2019) , and Pan et al. (2021) . Proof. Let x p = (E|x| p ) 1/p , and recall that MV t and X i,t are assumed to be weakly stationary processes. Let s be the compact time notation in lieu of i,t, that is,Moreover, let σ s = (β 0 + β 1 |r s−1 | + · · · + β q |r s−q | + θ |W S s−1 | + β X |X s−1 |). Note that W S t , obtained as an affine combination of (MV t−1 , · · · , MV t−K ), is weakly stationary.