key: cord-0635531-ncz6m63x authors: Lin, Jilei; Eck, Daniel J. title: Minimizing post-shock forecasting error through aggregation of outside information date: 2020-08-26 journal: nan DOI: nan sha: a6dad3c0cb982655337edf6cb0a9021ade2a618f doc_id: 635531 cord_uid: ncz6m63x We develop a forecasting methodology for providing credible forecasts for time series that have recently undergone a shock. We achieve this by borrowing knowledge from other time series that have undergone similar shocks for which post-shock outcomes are observed. Three shock effect estimators are motivated with the aim of minimizing average forecast risk. We propose risk-reduction propositions that provide conditions that establish when our methodology works. Bootstrap and leave-one-out cross validation procedures are provided to prospectively assess the performance of our methodology. Several simulated data examples, and a real data example of forecasting Conoco Phillips stock price are provided for verification and illustration. We provide forecasting adjustment techniques with the goal of lowering overall forecast error when the time series under study has undergone a structural shock. We focus on the specific setting in which a structural shock has occurred and one desires a prediction for the post-shock response at the next time point. Standard forecasting methods may not yield accurate predictions in the presence of such structural shocks [Baumeister and Kilian, 2014b] . This is a general problem that has many real life applications. For example, one may be interested in forecasting the stock price of a company tomorrow after hearing terrible or great news about the company after hours trading. Companies may be interested in forecasting the demand of their products to adjust production after they were involved in a brand crisis, but they only have recent sales data for which the company is operating well. All is not lost in this setting, one may be able to supplement the present forecast with past data borrowed from other time series which contain similar structural shocks. The core idea of our methodology is to sensibly aggregate similar past realized shock effects which arose from other time series, and then incorporate the aggregated shock effect estimator into the present forecast. Our method of combining shock effects embraces ideas from conditional forecasting Kilian, 2014b, Kilian and Lütkepohl, 2017] , time series pooling using cross-sectional panel data [Ramaswamy et al., 1993 , Pesaran et al., 1999 , Hoogstrate et al., 2000 , Baltagi, 2008 , Koop and Korobilis, 2012 , Liu et al., 2020 , forecasting with judgement and models [Svensson, 2005 , Monti, 2008 , synthetic control methodology [Abadie et al., 2010 , Agarwal et al., 2020 , and expectation shocks [Croushore and Evans, 2006 , Baumeister and Kilian, 2014a , Clements et al., 2019 . We study the post-shock forecasting problem in the context of additive shock effects in linear autoregressive models. In this post-shock forecasting setting, the researcher has a time series of interest which is known to have recently undergone a structural shock, and the post-shock response is not observed. In this setting, the additive shock effect is a random effect that is parameterized in the autoregressive model. The shock effect is then estimated using ordinary least squares (OLS). The researcher must move beyond the modeling paradigm that they were previously working under to accommodate this new shock effect [Monti, 2008 , Svensson, 2005 . One method for estimating the shock effect is to produce a conditional point forecast where a sequence of non-zero future structural shocks are conditioned upon and estimated [Baumeister and Kilian, 2014b] . Such conditional point forecasts are appropriate when the shock sequence considered is within the range of historical experience [Kilian and Lütkepohl, 2017] . On the other hand, our methodology allows for the inclusion of outside data sources and covariates into this conditional forecasting context provided that the shock effects from outside data sources are all thought to arise from a data generating process similar to that of the shock under study. For further differences of assumptions on shocks, our methodology allows for unprecedented shocks and no observation of past shocks. In our methodological framework, the researcher creates a synthetic panel of time series which have undergone similar structural shocks in the past. Construction of the donor pool that forms this synthetic panel is similar to that in synthetic control methodology (SCM) [Abadie et al., 2010] . As in SCM, care is needed when forming the donor pool of time series. However, there are key differences between our framework and SCM. We assume that the time series forming the donor pool are independent from the time series under study before the timing of the shock. Moreover, the shocks from candidate time series in the donor pool, together with the shock in the time series of interest, are assumed to be from a common family of distributions with existing first and second moments. We estimate the shock effects that are present in the time series forming the donor pool for which postshock responses are observed. We then aggregate these estimated shock effects and use this aggregated estimate as an estimator for the shock effect in the time series of interest. This estimator is then added to a forecast for the yet to be realized post-shock response corresponding to the time series of interest. Shock effects in our post-shock forecasting framework is similar to "expectation shocks" which are studied in Clements et al. [2019] . The context in Clements et al. [2019] allows for consistent estimation of expectation shocks under a vector autoregressive model, possibly involving an instrumental variable approach as in Croushore and Evans [2006] . In our context, the yet to be observed shock effect of interest is a random effect, and we can only partially estimate features of the random effect distribution using the time series forming the donor pool. In this article, we will assume a simple autoregressive data generating process similar to that in Blundell and Bond [1998] with a general random-effect structure. Therefore, our methodology is similar to the "K latent pooling" framework of Ramaswamy et al. [1993] . However, our model formulation is more general than Ramaswamy et al. [1993] . In our model, the donor pool can consist of dependent time series but time series within the donor pool should be independent of the time series of interest. However, mutual independence among time series in the donor pool can aid prospective evaluation of the reliability of our method. We consider three aggregation techniques: simple averaging, inverse-variance weighted averaging, and similarity weighting. The latter technique is similar to the weighting in synthetic control methodology [Abadie et al., 2010 ]. Our autoregressive model will consider present day covariates to better motivate similarity weighting. The considered adjustment strategies all target the mean of the shock effect distribution. Such an estimation strategy can reduce mean squared error (MSE) when variation in the shock effect distribution is small relative to the mean. We provide risk-reduction propositions that detail the conditions when the adjusted forecasts will work better than the original forecast. The involved parameters in the risk-reduction propositions can be estimated by a residual bootstrap procedure that we develop. We also motivate a simple leave-one-out cross validation procedure which can prospectively assess the performance of our shock effect adjustment estimators. This prospective assessment does not require the observation of the post-shock response. Our Monte Carlo simulation results show that the risk-reduction propositions are nearly perfectly correct when the model for the shock effects is identified well with appropriate covariates under a fixed design. We demonstrate the utility of our methodology in a real data analysis in which we forecast the stock price of Conoco Phillips shares that experienced a large structural shock on March 9th, 2020. We will show that our proposed adjustment estimators yield much better results than no adjustment in this setting. We also use this example to demonstrate settings in which the shock effect may be decomposed into separate estimable parts. We now motivate our framework for post-shock forecasting. t without shock effectŝ y 2 t with shock effectsα The time series experience a shock at T * 1 + 1 = 126 with true shock effect α = 9.21. (a) presents the comparison of the prediction without adjustment and one that uses simple averaging of estimated shock effects given a donor pool of n = 40 time series. (b) shows the histogram for the least square estimatesα i s used in estimating shock α 1 for i = 2, . . . , 41. The magenta dots representα i from the time series forming the donor pool. The prediction ofŷ 2 T * 1 +1 andŷ 1 T * 1 +1 differs only by an adjustmentα = 10.22. It is clear that y 2 T * 1 +1 performs better thanŷ 1 T * 1 +1 . We will suppose that a researcher has time series data (y i,t ,x i,t ), for t = 1, . . . , T i and i = 1, . . . , n + 1, where y i,t is a scalar response and x i,t is a vector of covariates that are revealed to the analyst prior to the observation of y 1,t . Suppose that the analyst is interested in forecasting y 1,t , the first time series in the collection. We will suppose that specific interest is in forecasting the response after the occurrence of a structural shock. To gauge the performance of forecasts, we consider forecast risk in the form of MSE, and root mean squared error (RMSE), given by √ R T , in our analyses. In this article, we focus on post-shock prediction where forecasts methods only differ at the next future time point. Thus the MSE reduces to the magnitude E(ŷ 1,t − y 1,t ) 2 . Our post-shock forecasting methodology will consist of selecting covariates x i,t , constructing a suitable donor pool of candidate time series that have undergone similar structural shocks to the time series under study, and specifying a model for the time series (y i,t ,x i,t ), for t = 1, . . . , T i and i = 1, . . . , n + 1. In this article, we consider a dynamic panel data model with autoregressive structure similar to that in Blundell and Bond [1998] . Our dynamic panel model includes an additional shock effect whose presence or absence is given by the binary variable D i,t , and we will assume that the donor pool time series are independent of the time series under study. The details of this model are in the next section. Figure 1 provides a simple intuition of the practical usefulness of our proposed methodology. This figure depicts a time series that experienced a shock at time point T * 1 + 1 = 126. It is supposed that the researcher does not have any information beyond T * 1 + 1, but does have observations from a donor pool of forty time series that have previously undergone a similar shock for which post-shock responses are recorded. Similarity in this context means that the shock effects are random variables that from a common distribution. In this example, the mean of the estimated shock effects is taken as a shock effect estimator for the time series under study. Forecasts are then made by adding this shock effect estimator to the estimated response values obtained from the estimation procedure that ignores the shock. It is apparent from Figure 1 that adjusting forecasts in this manner 1) leads to a reduction in forecasting risk; 2) does not fully recover the true shock effect. We evaluate the performance of this post-shock forecasting methodology throughout this article; we outline situations for when it is expected to work and when it is not. In this section, we will describe the assumed dynamic panel models for which post-shock aggregated estimators are provided. The basic structures of these models are the same for all time-series in the analysis, the differences between them lie in the setup of the shock effect distribution. Let I(·) be an indicator function, T i be the time length of the time series i for i = 1, . . . , n + 1, and T * i be the time point just before the one when the shock is known to occur, with T * i < T i . For t = 1, . . . , T i and i = 1, . . . , n + 1, the model M 1 is defined as where D i,t = I(t = T * i + 1) and x i,t ∈ R p with p ≥ 1. We assume that the x i,t 's are fixed. Let |x| denote the absolute value of x for x ∈ R. For i = 1, . . . , n + 1 and t = 1, . . . , T i , the random effects structure for M 1 is: Notice that M 1 assumes that α i are iid with E(α i ) = µ α for i = 1, . . . , n + 1. We also consider a model where the shock effects are linear functions of covariates with an additional additive mean-zero error. For i = 1, . . . , n + 1, the random effects structure for this model (model M 2 ) is: where the added random effects arẽ We further defineα i = µ α + δ i x i,T * i +1 . We will investigate the post-shock aggregated estimators in M 2 in settings where δ i is either fixed or random. We let M 21 denote model M 2 with δ i = δ for i = 1, . . . , n + 1, where δ is a fixed unknown parameter. We let M 22 denote model M 2 with the following random effects structure for δ i : We further define the parameter sets where Θ and Θ 1 can adapt to M 1 by dropping δ i . We assume this for notational simplicity. In this section we show how post-shock aggregate estimators improve upon standard forecasts that do not account for the shock effect. More formally, we will consider the following candidate forecasts: Forecast 1 :ŷ 1 1,T * 1 +1 =η 1 +φ 1 y 1,T * 1 +θ 1 x 1,T * 1 +1 , Forecast 2 :ŷ 2 1,T * 1 +1 =η 1 +φ 1 y 1,T * 1 +θ 1 x 1,T * 1 +1 +α, whereη 1 ,φ 1 , andθ 1 are all OLS estimators of η 1 , φ 1 , and θ 1 , respectively, andα is some form of estimator for the shock effect of time series of interest, i.e., α 1 . The first forecast ignores the presence of α 1 while the second forecast incorporates an estimate of α 1 that is obtained from the other independent forecasts under study. Note that the two forecasts do not differ in their predictions for y 1,t , t = 1, . . . T * 1 . Instead, they only differ in predicting y 1,T * 1 +1 . Throughout the rest of this article we show that the donor pool of time series {y i,t : t = 1, . . . , T i , i = 2, . . . , n + 1} has the potential to improve the forecasts for y 1,T * 1 +1 under different circumstances for the dynamic panel model M 1 , M 21 , and M 22 . Improvement will be measured by assessing the reduction in risk that Forecast 2 offers over Forecast 1. We will return to the theoretical details of risk-reduction in Section 3. We specifically focus on predictions for y 1,T * 1 +1 , the first post-shock response. It is important to note that in generalα does not converge to α 1 in any sense. Despite this shortcoming, adjustment of the forecast for y 1,T * 1 +1 through the addition ofα has the potential to lower forecast risk under several conditions corresponding to different estimators of α 1 . We now construct the aggregated estimators of the shock effects that appear in Forecast 2 (see Section 2.2). We use these to forecast response values y 1,T * 1 +1 assuming that T * 1 is known. First, we introduce the procedures of parameter estimation for M 1 , M 21 , and M 22 (see Section 2.1). For i = 2, . . . , n + 1, all parameters in this model will be estimated with ordinary least squares (OLS) using historical data of t = 1, . . . , T i . For i = 1, we estimate all the parameters but α 1 using OLS procedures for t = 1, . . . , T * 1 . In particular, letα i , i = 2, . . . , n + 1 be the OLS estimate of α i . Note that parameter estimation for M 1 is identically the same as that for M 21 or M 22 . We emphasize that α i s are random variables, but the OLS estimation is conditioned on the realizations from some distribution. Second, we introduce the candidate estimators for α 1 . Define the adjustment estimator for time series i = 1 byα where theα i s in (4) are OLS estimators of the α i s for i = 2, . . . , n+1. We can useα adj as an estimator for the unknown α 1 term for which no meaningful estimation information otherwise exists. It is intuitive that α adj should perform well under M 1 where we assume that α i 's share the same mean for i = 1, . . . , n + 1. However, it can also be shown thatα adj may be less favorable in M 21 and M 22 , which will be discussed in detail in Section 3. We also consider the inverse-variance weighted estimator in practical settings where the T i 's and T * i 's vary greatly across i = 2, . . . , n + 1. The inverse-variance weighted estimator is defined aŝ whereα i is the OLS estimator of α i ,σ i is the residual standard error from OLS estimation, and U i is the design matrix for OLS with respect to time series for i = 2, . . . , n + 1. Note that since σ is unknown, estimation is required and the numerator and denominator terms are dependent in general. However, α IVW can be a reasonable estimator in practical settings. We do not provide closed form expressions for E(α IVW ) and Var(α IVW ) but empirical performance ofα IVW is assessed via Monte Carlo simulation (see Section 4). We now motivate a weighted-adjustment estimator for model M 21 and M 22 . Our weighted-adjustment estimator is inspired by the weighting techniques in synthetic control methodology (SCM) developed in Abadie et al. [2010] . However, our weighted-adjustment estimator is not a causal estimator and our estimation premise is a reversal of that in SCM. Our objective is in predicting a post-shock response y 1,T * 1 +1 that is not yet observed using other time series whose post-shock responses are observed. We use similar notation as that in Abadie et al. [2010] to motivate our weighted-adjustment estimator. Consider a W ∈ R n weight vector W = (w 2 , . . . , w n+1 ) , where w i ∈ [0, 1] for all i = 2, . . . , n + 1. Construct where X 1 ,X 1 (W) ∈ R 1×p . Define W = {W ∈ [0, 1] n : 1 n W = 1}. Suppose there exists W * ∈ W with W * = (w * 2 , . . . , w * n+1 ) such that Note that (5) tries to find W * such that x 1,T * 1 +1 is a convex combination of x i,T * i +1 for i = 2, . . . , n + 1 with weights W * . Therefore, W * should exist as long as X 1 falls in the convex hull of Our weighted-adjustment estimator will therefore perform well when the pool of time series posses similar covariates to the time series for which no post-shock responses are observed. We compute W * as Abadie et al. [2010] commented that we can select W * so that (5) holds approximately and that weightedadjustment estimation techniques of this form are not appropriate when the fit is poor. Note that W * is not random since the covariates are assumed to be fixed. Since W is a closed and bounded subset of R n , W is compact. Because the objective function is continuous in W, W * will always exist. Our weighted-adjustment estimator for the shock effect α 1 iŝ We further define . Proposition 1. If V has full rank and it exists some W satisfies (5), the solution to (6) is unique. Proposition 1 details some conditions when W * is unique. Note that V is p × n. Therefore, if the covariates are of full rank and the true solution lies in the convex and compact W, a sufficient condition for W * to be unique is p ≥ n. However, when p < n, W * may not be unique. If it exists some W * satisfies (5) and p < n, there are infinitely many solutions to (5). The issue of non-uniqueness is further discussed in Section 3.2. Remark 1. In Section 2.1 we specify that x i,t , θ i ∈ R p . However, it is not necessary that the all p covariates are important for every time series under study. The regression coefficients θ i are nuisance parameters that are not of primary importance. It will be understood that structural 0s in θ i correspond to variables that are unimportant. Remark 2. Our forecasting premise and estimation construction shares similarities with Bayesian viewpoints. From a Bayesian perspective, if we assign a prior π to α 1 ,α adj ,α wadj , andα IVW can be interpreted as the Bayes rules with respect to π under different loss functions. If the sampling distribution of the data and π are known, it is possible to compute the Bayes risks ofα adj ,α wadj , andα IVW with respect to π, thus enabling comparisons among them. Additionally, from Theorem 2.4 in Chapter 5 of Lehmann and Casella [2006] ,α adj ,α wadj , andα IVW are admissible if they are unique with probability one. In this section, we discuss the properties that are related to forecast-risk reduction. In discussion of risk, it is useful to derive expressions for expectation and variance of the adjustment estimatorα adj and weighted-adjustment estimator. The expressions for the expectations are as follow, Formal justification for these results can be found in Appendix. Note thatα adj ,α wadj , andα IVW are not unbiased estimators for α 1 . However, under M 1 ,α adj andα adj are unbiased estimators for E(α 1 ) = µ α . Nevertheless,α adj is a biased estimator for E(α 1 ) butα wadj is an unbiased estimator for E(α 1 ) under both M 21 and M 22 . We collect these results in the following proposition. Proposition 2. (i) Under M 1 ,α adj is an unbiased estimator of E(α 1 ). Under M 21 and M 22 ,α adj is a biased estimator of E(α 1 ) in general. (ii) Suppose that W * satisfies (5). Under M 1 , M 21 and M 22 ,α wadj is an unbiased estimator of E(α 1 ). Unbiasedness properties for E(α 1 ) ofα adj andα wadj allow for simple conditions for risk-reduction to hold, and more importantly motivates a bootstrap estimation for evaluation of these conditions. These conditions and bootstrap will be discussed in Section 3.1 and 3.2, respectively. Next, we present the variance expressions forα adj andα wadj as below. (i) Under M 1 and M 21 , Formal justification for these results can be found in Appendix. Note that the variances are not comparable in closed-form because of the term E (U i U i ) −1 22 . This term exists because of the inclusion of the random lagged response in our autoregressive model formulation. Section 3.1 details conditions needed for risk-reduction and comparisons of adjustment estimators. These conditions involve variances and expectations which may be difficult to compute in practice. To make use of those conditions in practice, estimation is required. Sections 3.2 introduce a residual bootstrap procedure which estimates the involved parameters in those conditions and thus motivates prospective decision-making about whetherα i reduces the risk. Section 3.3 describes our leave-one-out cross validation procedures, which prospectively estimate the correctness of such decision without observation of the postshock response for the time series under study. Our simulations verify these procedures. In this section we will discuss the conditions for risk reduction for individual shock effects estimators under M 1 , M 21 , and M 22 . For an adjustment estimatorα, we will write the risk-reduction as ∆(α) = R T * 1 +1,1 − R T * 1 +1,2 where R T * 1 +1,2 is the risk of Forecast 2 calculated using the adjustment estimatorα. Recall that Proposition 2 implies that the adjustment estimatorα adj and weighted-adjustment estimator α wadj are unbiased for E(α 1 ) under M 1 . With this result, we will have the following propositions that specify the conditions that are necessary for risk reduction. Proposition 3 says that under M 1 if the variance of the estimator is smaller than the squared mean of α 1 , those estimators will enjoy the risk reduction properties. In this setting, under M 1 , ∆(α adj ) = µ 2 α −Var(α adj ) and ∆(α wadj ) = µ 2 α −Var(α wadj ). From Proposition 3, we obtain a risk-reduction condition Condition (7) implies two facts: (1) adjustment (Forecast 2) is preferable to no adjustment (Forecast 1) asymptotically in n whenever µ α = 0 (see Forecast in Section 2.2); (2) In finite donor pool settings, adjustment is preferable to no adjustment when µ α is large relative to its variability and overall regression variability. If W * does not satisfy (5), its unbiased properties for E(α 1 ) should hold approximately when the fit in (6) is appropriate as commented in Section 2.3. From Proposition 3 and the variance expression for α wadj , the risk-reduction condition forα wadj is In this case, adjustment is preferable to no adjustment when µ α is large relative to the weighted sum of variances for shock effects for other time series and overall regression variability. However, the above criteria are generally difficult to evaluate in practice. Sections 3.2 and 3.3 provide detailed treatments on how to estimate the sign of ∆(α) in practice. The shock effects α i s have different means under M 21 and M 22 unlike under M 1 . However, Proposition 2 implies thatα wadj is an unbiased estimator of E(α 1 ). We now state conditions for risk-reduction. Under Proposition 4, we can obtain a risk-reduction inequality that is similar to (8), The conclusions and intuitions will be identically the same as what we have in Section 3.1.1. Proposition 2 shows thatα adj is a biased estimator of E(α 1 ) under M 21 and M 22 generally. Hence, Proposition 3 no longer holds forα adj under M 21 and M 22 . As an alternative, we can derive similar risk-reduction conditions that are appropriate for this setting. By Lemma 1 (see Section 7) and risk decomposition, we will achieve risk-reduction as long as The above inequality simplifies to As mentioned in Section 2.3, it is difficult to evaluate the expectation and variance ofα IVW . We note thatα IVW is generally biased for E(α 1 ). That is to say we can adapt the above proof to derive the risk-reduction conditions forα IVW : under M 1 , M 21 , and In fact, more generally, using similar proof of Lemma 1, it can be shown that under M 2 , the risk-reduction quantities are where we estimate ∆(α) for estimatorα using bootstrap and leave-one-out cross validation procedures developed in Sections 3.2 and 3.3. In this section, we present bootstrap procedures that approximate the distribution of our shock effect estimators, checks the underlying conditions of our risk reduction propositions, and estimate risk-reduction quantity using plug-in approach in practice. Our procedure involves the resampling of residuals in the separate OLS fits. This procedure has its origins in Section 6 of Efron and Tibshirani [1986] and Chapter 12 of Kilian and Lütkepohl [2017] . Our procedure involves the resampling of the residuals which are assumed to be the realizations of an iid process. Our first bootstrap procedure is as follows: Let B be the bootstrap sample size. At iteration b, first resample the indices I = {2, . . . , n + 1} of the donor pool with replacement to form I (b) with cardinality n, where we note that the elements of I (b) may not be unique in terms of their indices in the donor pool. Initialize y , . . . , T i } using the model estimated by original data, and obtain shock effect estimators for each of the time series in the donor pool for all i ∈ I (b) . These shock effect estimators are then used to construct any of the adjustment estimatorsα We can then estimate distributional quantities of our shock effect estimators under our considered models with the bootstrap samplesα IVW , for b = 1, . . . , B. We denote this procedure by B u . We motivate a second bootstrap procedure B f which treats the the donor pool as fixed, and not a realization from an infinite super-population. Therefore, there is no resampling of the donor pool in B f , it is otherwise similar to B u . An algorithmic formulation of B u and B f are outlined in Section 2 in the Supplementary Materials. We will explicitly use these bootstrapped samples of shock effect estimators to check the risk-reduction conditions in Propositions 3 and 4. Recall thatα adj ,α wadj andα IVW are unbiased estimators of their expectations, andα wadj is an unbiased estimator of E(α 1 ) under M 1 and M 2 from Proposition 2. Our bootstrap procedure estimates the variance of our adjustment estimators. We can then estimate the risk-reduction propositions and inequalities. For example, we can estimate ∆(α adj ) under model M 21 or where S 2 α adj is the bootstrap sample variance estimator for Var(α adj ). We reiterate the philosophical distinction between B u and B f . B u treats the donor pool as realizations from some infinite super-population of potential donors. In contrast, B f treats the donor pool as being fixed and known before the analysis is conducted, where the randomness arises from parameters and idiosyncratic error. A double bootstrap procedure with similar steps can estimate the distribution of ∆(α) forα ∈ A. The double bootstrap, instead of checking whether ∆(α) > 0, can check whether a bootstrap percentile interval of resampled estimates of ∆(α) contain 0 at a desired error threshold. We investigated such a double bootstrap procedure and found that it produced inferences that were similar to those produced using the bootstrap techniques developed in the main text. We stress that our bootstrap approximations cannot alleviate the inherent bias of using our adjustment estimators as surrogates for α 1 . We caution that the bootstrapping residuals in OLS estimation may not provide valid inference in moderate or high dimension where p < T i but p/T i is not close to zero for i ∈ {2, . . . , n + 1} [El Karoui and Purdom, 2018] ; see alternatives for residual bootstrapping in linear models in El Karoui and Purdom [2018] . Recall that W * may not be unique if the conditions in Proposition 1 are not satisfied. Non-uniqueness might be a concern theoretically. This is due to the fact that infinitely many different weights can lead to infinitely many non-uniqueα wadj 's all targeting on the same α 1 . For example, consider the case where the size of donor pool to be 2, Var(α 2 ) = 1, Var(α 3 ) = 2, and there are two solutions to (5), say, W * 1 = (1, 0) and W * 2 = (0, 1). In this scenario, the weighted adjustment estimator induced by W * 1 has variance 1 whereas the one by W * 2 has variance 2. Nevertheless, even ifα i has the same variance across i = 2, . . . , n + 1, the same issue would occur if there were infinitely many W * with different norms. It is possible to resolve this issue by selecting a unique weight W * that optimizes a desirable objective function, prior to which one should find the bases spanning the subspace of W * satisfying (5). Simulations in the Section 4 of the Supplementary Materials provide some evidence that non-uniqueness of W * is not problematic for inferences. In this section, we adapt leave-one-out cross validation (LOOCV) to our estimation context in order to provide prospective evaluations of our adjustment techniques. Our proposed LOOCV procedure has its roots in Section 7.10 of Hastie et al. [2009] . Recall in Section 2.1 that we are given the data {(x i,t , y i,t ) : i = 1, . . . , n + 1, t = 1, . . . , T i }, where {(x 1,t , y 1,t ) : t = 1, . . . , T 1 } is the data of the time series of interest and the remaining observations form the donor pool. For iteration m ∈ {1, . . . , n} of our LOOCV procedure, we set aside {(x m+1,t , y m+1,t ) : t = 1, . . . , T m+1 } as the time series of interest, and construct a new donor pool {(x i,t , y i,t ) : i ∈ I m , t = 1, . . . , T i }, where I m = {2, . . . , n+1}\{m+1}. Since the post-shock response y m+1,T * m+1 +1 is observed, we can evaluate the performance of our adjustment estimators and the original forecast made without adjustment (i.e., Forecast 1 in Section 2.2). LOOCV can be very computationally intensive when n is large, especially when combined with bootstrapping. To alleviate these concerns we can perform LOOCV with a random subset of k ≤ n iterations selected without replacement. In this setting, we let J be the randomly sampled indices. For m ∈ J , we set aside {(x m+1,t , y m+1,t ) : t = 1, . . . , T m+1 } as the time series of interest, and construct a new donor pool {(x i,t , y i,t ) : i ∈ I, t = 1, . . . , T i }, where I = {2, . . . , n + 1} \ {m + 1}. Based on the new donor pool, we estimate relevant parameters using bootstrap procedures outlined in Section 3.2. In other words, k times of bootstrapping are nested in a LOOCV procedure. We find that k = 5 or k = 10 iterations of LOOCV performs well. We now outline how LOOCV can be used to prospectively assess the performance of adjustment estimators. Let A be the set of adjustment estimators. For eachα ∈ A, let δα = I(∆(α) > 0) be a decision rule where I(·) is the indicator function and a 1 corresponds to the decision to use estimator α. If ∆(α) > 0 (∆(α) < 0, respectively) but δα incorrectly reported 1 (0, respectively) so that it makes the decision not to useα (to useα, respectively), δα is said to be incorrect. If ∆(α) < 0 (∆(α) > 0, respectively) and δα correctly reported 0 (1, respectively) so that it makes the decision to useα, δα is said to be correct. These situations are depicted in the following table: We will use C(δα) = I(δα is correct) as a metric that evaluates the performance of forecasts made with the adjustment estimatorα. If E(C(δα)) > 0.5, we claim that δα is better than random guessing. Note that C(δα) can generally be computed only when the post-shock response is observed. However, it is possible to estimate E(C(δα)) using LOOCV. The LOOCV estimates for E(C(δα)) arē where C (−m) (δα) is computed with respect to donor pool with index set I m and the m + 1 time series is treated as the time series of interest. The LOOCV with k random draws estimates E(C(δα)) as where J is the set of the k randomly sampled indicies. Remark 3. Note that we allow the time series within the donor pool to be dependent but donor pool should be independent of the time series of interest. However, if we assume the mutual independence structure,C(δα) will be an almost unbiased estimator of E(C(δα)) [Marden, 2015, Page 222]. In this section we provide justification for our methods based on Monte Carlo simulation. We implemented our simulation based on M 22 with negligibly small Σ δ approximating the design of M 21 . We consider p = 25 and µ α = 2, where p = 25 is set to satisfy conditions in Proposition 1. Parameter setup of our simulations is detailed as follows: the φ i 's are sampled independently from Uniform(0, 1). We sampled T i 's independently from Gamma(15, 10) that are further rounded to integers, where the minimum allowable value of T i is fixed to be 90. We will randomly draw T * i from {p + 4, . . . , T i − 1}. The choices of T i and T * i are set up to satisfy a necessary condition for the design matrix of OLS estimation to have full rank. Moreover, it is designed to illustrate the performance ofα IVW that may perform well in time series with varying lengths. Additionally, we generated the covariates from Gamma(1, 2) to set up a setting when theα wadj may perform well. Last, we set δ i iid ∼ N (1, 0.5) and θ i ∼ N (0, 1). We will consider parameter setup by varying σ in the model of y i,t , n, the donor pool size, and σ α in the model of α i . We choose a Monte Carlo sample size of 30 replications and a bootstrap sample size of B = 200 for computation. Means and standard errors for estimated quantities will be recorded. Our LOOCV procedure will consider k = 5 random draws. Recall in Section 3.3 that B times of bootstrap are nested in a LOOCV with k random draws. It implies that B(k + 1) times of bootstrap replications are required for each Monte Carlo simulation. Our adjustment estimators will be evaluated by multiple criteria. We interpret δα = I(∆(α) > 0) for α ∈ A as the guess, with 1 indicating thatα provides risk-reduction over the simple no-adjustment forecast, and 0 indicates the converse. We will consider the LOOCV estimators (10) and (11) to assess correct decision making. We will also consider the Euclidean distance between the post-shock forecastŝ y 1,T * 1 +1 ,ŷ 1,T * 1 +1 +α adj ,ŷ 1,T * 1 +1 +α wadj , andŷ 1,T * 1 +1 +α IVW and the realized post-shock response y 1,T * 1 +1 . The first two metrics can combine to assess our forecasting methodology prospectively while the latter requires the realization of the post-shock response y 1,T * 1 +1 . In this section, we discuss simulation results for the bootstrap procedures used in estimating parameters for risk-reduction propositions and inequalities. We mainly discuss simulations under M 2 (see Section 2.1) for B u and B f (see Section 3.2) with comparisons to those under M 1 whose results are listed in Section 3 in the Supplementary Materials. Two simulation setups are investigated. In the first simulation setup, we consider the parameter combination of n ∈ {5, 10, 15, 25} and σ α ∈ {5, 10, 25, 50, 100} where we fix σ = 10. Note that E(E(α 1 )) = 52, where the last expectation is operated under the density of the covariates. In other words, data with σ α ∈ {5, 10, 25, 50, 100} should well represent the situations when the signal of the covariates is strong and when it is nearly lost. Results are displayed in Table 1 in the Appendix, Section 7.4. In the second simulation setup, we consider the parameter combination of σ, σ α ∈ {5, 10, 25, 50, 100} where we fix n = 10. Likewise, σ, σ α ∈ {5, 10, 25, 50, 100} will produce situations when the signal of the covariates is strong and when it is nearly lost in the model of both y i,t and α i . Results are in Table 2 in the Appendix, Section 7.4. First, assuming thatC (k) (δα) well estimates E(C(δα)) and fixing n, we observe from Table 1 that the decision making of δα is nearly correct forα ∈ A when σ α is small from Table 1. The reasons can be explained as follows. When σ α is small, the signal of the covariates is strong so thatα wadj will be expected to capture the signal according to construction ofα wadj in Section 2.3. Moreover, when σ α is small, M 22 approximates M 21 such that estimation of E(α 1 ) should be nearly unbiased according to Proposition 2. However, when the signal of the covariates is poor (σ α is large), the decision rule δα becomes unreliable forα ∈ A. It is to be expected since the bootstrap estimates become more biased. However, users can be warned byC (k) (δα) to have an idea of the effectiveness of δα. Second, fixing σ α , we can observe that the correctness of δα increases when n increases. It is due to the robustness gain in estimation when n increases. Additionally, we observe that in most cases δα wadj reportsα wadj reduces the risk even whenC (k) (δα wadj ) starts to break down, though they follow similar patterns. Recall from Section 3.2 that∆(α) contains the squared bias for estimating E(α 1 ). But it is not present for∆(α wadj ) since we applied the factα wadj is unbiased for E(α 1 ) from Proposition 2 in plugging it in with replacing E(α 1 ). Therefore, when the signal from covariates is poorer, δα wadj becomes less conservative. Besides, the averaged I ∆ (α) > 0 timesC (k) (δα) can provide an approximation for the probability thatα actually reduces the risk assuming an symmetry of correctness between the cases when∆(α) > 0 and when∆(α) < 0. For example, when n = 5 and σ α = 50, the probability thatα adj reduces the risk is approximately 0.83 × 0.59 = 0.490 from Table 1. In other words, the probability thatα reduces the risk has the same pattern asC (k) (δα) has with n and σ α forα ∈ A. From columns related to distance to y 1,T * 1 +1 in Table 1 , as σ α increases, the prediction appears to be poorer. When σ α = 5, 10, 25, forecasts usingα adj ,α wadj , andα IVW are always better than the original forecast significantly. But it does not hold generally for the case when σ α = 50, 100. It is reasonable in that when the σ α is large, it is difficult to find a reliable estimate of α 1 . Nevertheless, no statistical evidence has been found to support the claim that n matters in prediction. In other words, the size of the donor pool matters for producing reliable decision-making of δα rather than reliable prediction. From Table 2 , we observe that as σ α increases fixing σ,C(δα) decreases, which is a pattern similar to the one shown in the first experiment. Furthermore, as σ increases fixing σ α ,C(δα) decreases as well. Note that the correctness hinges on the estimation of the parameters. Sinceα wadj is a linear combination of OLS estimates, as σ increases, Var(α wadj ) increases as well. Therefore,α wadj become more volatile and its estimation of E(α 1 ) can be less reliable. Those reasons can explain why an increase of σ α contributes to a decrease ofC(δα). We observe similar patterns for distance to y 1,T * 1 +1 as well. When σ increases with fixing σ α , it is likely that the degree of variation of y 1,t exceeds the extent of adjustment improvementα can contribute to forα ∈ A. With respect to averaged I(∆(α) > 0) (i.e., the guess), it starts to decrease as σ increases. This is reasonable if we believe the bootstrap estimate S 2 α provides a good approximation for Var(α) forα ∈ A. The reasons are outlined as follows: Recall in Section 3.1.2, the conditions of risk-reduction propositions involve (E(α 1 )) 2 > Var(α) + (E(α) − E(α 1 )) 2 forα ∈ A. Notice that Var(α) is an increasing function of σ sinceα is estimated by OLS. Therefore, it explains the reason why the increase of σ would result in a decrease of averaged I(∆(α) > 0) since the inequality is not likely to hold when Var(α) increases. Simulation for B f with the same parameter setup as that of B u are implemented. See Table 3 and Table 4 for results in the Appendix, Section 7.4. Comparing Table 1 and Table 2 yields that when n is moderately small (n = 10) and σ α is small (σ α = 5), B u is better than B f with statistical evidence. For other situations, B u and B f are rather similar. It is likely that the extra randomness from sampling with replacement from donor pool compensates for the possible noises from a small donor pool. Concerning Table 2 and Table 4 , it appears that when n = 10 and σ α = 5, B u is better than B f when σ increases. It might be the case that additional layer of bootstrap in the donor pool buffers the negative effects on C(δα) introduced from increasing variation of y i,t . However, when σ α increases over 5 and n = 10, B f and B u are quite similar under situations of different σ and σ α . In conclusion, B u is better than B f when the signal of the covariates is strong and n is moderately small; otherwise, they are similar. Simulation results corresponding to model M 1 are listed in Section 3 in Supplementary Materials. Results under model M 1 are very similar to those of M 2 , except for the difference among estimators. The results show that (1) the performance ofα adj andα IVW are nearly the same and (2) in many situations, α adj andα IVW are better thanα wadj ; in other situations, they are mostly the same. Recall that in M 1 , the models for α 1 do not involve the covariates. Therefore, similarity weighting may not be informative when the model for α i is identified wrongly. Under M 1 , simple averaging, aimed for a reduction of variance, or inverse-variance weighting, targeting on reducing negative effects from varying time lengths, may work better. We have implicitly assumed that W * is non-degenerate in the population in these simulations. Recall that in Section 2.3 we noted that if there exists some W * which satisfies (5) and p < n, then there will be infinitely many solutions to W * . In this scenario, W * will take values on the boundary of W, in which case bootstrapping may fail to estimate the distribution ofα wadj [Andrews, 2000] . When p < n and there exists some W * ∈ W satisfies (5), B u fails since the non-uniqueness due to p < n will guarantee degeneracy of W * . However, this issue will not occur under B f since it takes W * as being fixed and the parameter space is Θ that does not involve the constrained W. Moreover, simulations in Section 4 of the Supplementary Materials show that non-uniqueness does not seriously compromise the inference. We demonstrate our post-shock forecasting methodology on a time series of Conoco Phillips share prices after the occurrence of a structural shock. Conoco Phillips is a large oil and gas resources company [ConocoPhillips, 2020] . The particular post-shock response that we predict happened after trading ended on Friday March 6th, 2020 and before trading began on Monday March 9th, 2020. It is reasonable that the timing of this shock is known, several events occurred over the trading weekend which had an impact on stock markets and the oil markets. For example, Russia and OPEC began a battle for global oil price control on Sunday, March 8th [Sukhankin, 2020] , and several US states began declaring state of emergencies in response to the evolving coronavirus pandemic [New York State Government, 2020 , Alonso, 2020 . In this analysis we make the following design considerations: (1) Selection of model. We will use an AR(1) model to forecast Conoco Phillips stock price. This model has been shown to beat no-change forecasts when predicting oil prices over time horizons of one and three months [Alquist et al., 2013] . For these reasons, we will consider 30 pre-shock trading days and we will forecast the immediate shock effect. All estimates will be adjusted for inflation. The model setup for AR(1) is exactly the same as what is stated in Section 2.1 with addition of shock effects. All the parameters are estimated using OLS. (2) Selection of covariates. We consider different covariates for the model of α i and y i,t . The model of α i incorporates daily S&P 500 index prices, West Texas Intermediate (WTI) crude oil prices, dollar index, 13 Week treasury bill rates, and Chicago Board Options Exchange volatility index (VIX). The model for y i,t disregards VIX. This is because VIX is a metric for capturing market risk and sentiment, which is highly influential for the shock effect model. [Shorter, 2008 , Ewing and Malik, 2013 , Dwyer and Tkac, 2009 , Longstaff, 2010 . The last shock effect corresponds to an OPEC induced supply side shock effect [Huppmann and Holz, 2015] . We assume that the five shocks are independent of the shock that Conoco Phillips experienced on March 9, 2020. The covariates and response of time series in the donor pool are adjusted for inflation. Note that there are three shock effects nested in the time series 2008 September, we assume that these three shocks are independent, where the assumption checks using likelihood ratio test are provided in the Section 1 in the Supplementary Materials. The estimated shock-effects for α i are −0.922, −7.063, −5.777, −6.395, −4.207 for i = 2, . . . , 6, respectively. Under M 2 , we computedα adj , weighted adjustmentα wadj , andα IVW . Note that non-uniqueness problems will not occur in this analysis since the conditions of Proposition 1 are satisfied. To avoid the effect of unit differences on weighting, we center and scaled the covariates in weights computation but not in the model of y i,t . Forα wadj , we observe that W * = (0.000, 0.000, 0.000, 0.273, 0.727) and X 1 −X 1 (W * ) 2 = 3.440. Note that the norm is computed using the k-dimensional Euclidean metric. The solution W * suggests that the shock effect of interest is very similar to the September 25, 2008 shock effect and the November 27, 2014 shock effect. The resulting shock effect estimates areα adj = −4.872,α wadj = −4.805, andα IVW = −4.384. Using the bootstrap procedure B f , we estimated parameters for risk-reduction propositions and risk-reduction quantities proposed in Section 3. The estimated bootstrap variances forα adj ,α wadj , andα IVW are 0.419, 0.559, and 0.667, respectively. We verify the consistency of the result yielded by risk-reduction propositions with the reality as below. We can see from Figure 2 thatα adj ,α wadj andα IVW perform decently well. Our estimation framework misses the yet to be observed pot-shock response, however they perform much better than unadjusted forecasts that do not account for shock effects. The unadjusted forecast misses the post-shock response by 9.870 dollars whereas the use ofα adj ,α wadj , andα IVW misses by 5.324, 5.392, and 5.813 dollars, respectively. The true shock effect is not fully recovered byα adj ,α wadj , andα IVW . This may be a result of a poorly constructed donor pool. The shock on March 9th, 2020 is in the midst of the COVID-19 pandemic and oil production volatility. It is difficult to find available stock market time series data that were generated under a similar setting. In any event, the shock on March 9th, 2020 was the largest price shock to Conoco Phillips shares by a wide margin, even after adjusting for inflation. From another perspective, it is possible that the stock of Conoco Phillips actually experienced multiple shocks on 2020 March 9th. For example, Kilian [2009] studied the effect that different supply and demand shocks have on oil prices through a vector autoregressive model. Their model postulates an additive nature of shock effects, although the additivity parameters requires estimation in their context. Motivated by Kilian [2009] , we also studied additive shock effect estimators where the shock effects corresponding to separate supply and demand shocks are added to estimate the unknown shock effect. The supply shock donor pool consists of the November 27, 2014 shock effect; and the demand shock donor pool consists of the remaining shock effects. The additive adjustment estimator computed by adding theα adj ,α wadj , and α IVW estimators for the demand and supply shock effects only, respectively, miss the post-shock Conoco Philips share price by 0.951, 0.405, and 1.460 dollars. These additive adjustment estimators do extremely well in this additive shock setting. We developed a methodology for forecasting post-shock response values after the occurrence of a structural shock. Our methodology is as follows: construct a synthetic panel of time series which have undergone similar shocks, estimate the shock effects in those series, aggregate them, and then adjust the original forecast by adding the aggregated shock effect estimator to the original forecast. There have been several other recent similar methods developed for forecasting COVID-19 cases. For example, Lee et al. [2020] constructed a Bayesian hierarchical model embracing data integration to improve predictive precision of COVID-19 infection trajectories for different countries. A similar setup may be appropriate for postshock forecasting but may be too dependent upon model specification for the shock distribution. Plessen [2020] employed a data-mining approach to combine COVID-19 data from different countries as input to predict global net daily infections and deaths of COVID-19 using a clustering approach. However, there is a tremendous amount of volatility in this form of COVID-19 data, and the fit of this prediction method may be improved with modeling structure or preprocessing of the donor pool. Agarwal et al. [2020] proposed a model-free synthetic intervention method to predict unobserved potential outcomes after different interventions given a donor pool of observed outcomes with given interventions. They also provided useful guidelines for how to estimate the effects of potential interventions by giving recommendations for choosing the metric of interest, the intervention of interest, time horizons, and the donor pool. Although the methodology in Agarwal et al. [2020] is quite general, there is no guarantee for theoretical properties in prediction without assuming any distributional structure. We provided risk-reduction propositions and empirical tools that can prospectively assess the effectiveness of our adjustment strategies in additive shock effect settings. The model, under which we verify these claims, is a simple AR(1) model. Similar results can be obtained for more general models such as AR(p), vector autoregression, and generalized autoregressive conditional heteroskedasticity models. Generally, multiple shock effects can be nested within a time series; and time series in the donor pool can be dependent. As an example, we considered a dependency structure for the September 2008 shock effects in our analysis of Conoco Phillips stock. But we note that consistency estimates from LOOCV may not work well if donor pool candidates are not mutually independent since the almost unbiased property hinges on the mutual independence among candidates in the donor pool. Although it is reflected in M 2 , we stress that our proposed methods allow α i to follow arbitrary distributions provided that its first and second moments exist. The covariates in the model for α i under M 2 can be different from the covariates in the model of y i,t . Additionally, α i can be heteroskedastic; i.e., they can have different variances. In this scenario, all the theoretical properties still hold though variance expressions in Section 3.1 may not apply. We also note that our post-shock framework can be extended to settings where the shock effect can be decomposed into separable estimable parts. An example of this is the additive shock effect estimators that we studied in our Conoco Phillips analysis. Although our work is developed for time-series or AR(p) models, in fact, it can be generalized to any similar setting with a model of the response, whose parameters can be estimated unbiasedly, an additive shock-effect structure, and the structure that the time series in the donor pool are independent of the one of interest. Our bootstrap procedures can be extended to approximate the distribution of shock effect estimators from more general time series. If the data are subject to heteroskedasticity of unknown form, bootstrapping tuples of regressands and regressors proposed by Freedman [1981] is robust in this situation with asymptotic validity in autoregressive models established by Gonçalves and Kilian [2004] . If serial correlation exists in the data, various block bootstrapping procedures [Künsch, 1989 , Liu et al., 1992 can be possible reasonable alternatives. Note that the pseudo time series generated by our proposed residual bootstrap are not stationary. If stationarity is of concern, one can be referred to the stationary bootstrap in Politis and Romano [1994] for stationary and weakly dependent time series. Nevertheless, it was shown that approximation accuracy might be a cost for the stationary bootstrap in autoregressive models in finite sample . More work related to bootstrapping time series can be referred to Chapters 3 and 4 in Politis et al. [1999] , , and Chapter 12 in Kilian and Lütkepohl [2017] . It is up to users in terms of selecting which procedure to choose but under different assumptions on the time series. Construction of donor pool is a critical step that directly matters in prediction. In our model, shocks from candidate time series in the donor pool, together with the shock in the time series of interest, should come from a family of distributions that allow for varying variances and varying means that are linear functions of common covariates. Remark 1 shows that common covariates are not restrictive with structural zero in coefficients. However, it is difficult to verify such assumption or structure in practice. In principle, one should select identities that come from a common population, which is also the one the identity under study come from, as candidates in donor pool. This selection criterion is often adopted in the literature of comparative case studies. For example, Card [1990] used a set of cities in southern U.S. to estimate the effect of 1980 Mariel Boatlift on the Miami labor market. Abadie and Gardeazabal [2003] utilized a donor pool of two Spanish regions to approximate the effect of terrorism on the economic growth of Baseque Country. To estimate the effect of Proposition 99, a large tabacco control program in California in 1988, on the annual per-capita cigarette sales of 2000, Abadie et al. [2010] constructed a donor pool consisting of US states, which did not implement any large-scale tobacco control program during their sampling period. Common features of those studies are that they all carefully constructed the donor pool and discarded any suspicious outliers. where we used the fact that n+1 i=2 w * i = 1. Case II: under M 21 and M 22 : Since E(ε i,T i ) = 0, E(α i ) = E(α i ) = E(α i ), it follows that (5)) Similarly, Notice that under the setting of OLS, the design matrix for M 2 is the same as the one for M 1 . Therefore, it follows that Under M 21 where δ i = δ are fixed unknown parameters, we will have Similarly, under M 22 where we assume δ i ⊥ ⊥ ε i,t , we have For the adjustment estimator, we simply replace W * with 1/n1 n . Thus, under M 21 we have Under M 22 , we shall have Notice that M 1 differs from M 21 only by its mean parameterization of α (see Section 2.1). In other words, the variances ofα adj andα wadj under M 1 are the same for those under M 21 . Proof of Proposition 1 The proof of Li [2019] in Appendix A.2 and A.3 adapts easily to Proposition 1. The proof for unbiasedness follows immediately from discussions related to expectation in Section 3. For the biasedness ofα adj under M 21 and M 22 , we write the bias term forα adj as below. But it may be unbiased in some special circumstances when the above bias turns out to be 0. Lemma 1. The forecast risk reduction is R T * 1 +1,1 − R T * 1 +1,2 = E(α 2 1 ) − E(α − α 1 ) 2 for all estimators of α 1 that are independent of Θ 1 (see Section 2.1). where Θ 1 is as defined in (3). Notice that It follows that Assuming S = (1 n , y 1,t−1 , x 1 ) has full rank, under OLS setting,η 1 ,φ 1 , andθ 1 are unbiased estimators of η 1 , φ 1 , and θ 1 , respectively under conditioning of Θ 1 . Since we assumeα is independent of Θ 1 , through the method of iterated expectation, It follows that which finishes the proof. Proof of Proposition 3 The proofs are arranged into two separate parts as below. Proof for statement (i): Under M 1 ,α adj is an unbiased estimator of E(α 1 ) because where we used the fact that OLS estimator is unbiased when the design matrix U i is of full rank for all i = 2, . . . , n + 1. Because α 1 ⊥ ⊥ ε i,t , E(α adj α 1 ) = E(α adj )E(α 1 ) = (E(α adj )) 2 . By Lemma 1, Therefore, as long as we have Var(α adj ) < µ 2 α , we will achieve the risk reduction. Proof for statement (ii): By Proposition 2, the property thatα wadj is an unbiased estimator of µ α holds for M 1 . The remainder of the proof follows a similar argument to the proof of statement (i). Proof of Proposition 4 By Proposition 2, the property thatα wadj is an unbiased estimator of E(α 1 ) holds for M 21 and M 22 . The remainder of the proof follows a similar argument to the proof of Proposition 3. The economic costs of conflict: A case study of the basque country Synthetic control methods for comparative case studies: Estimating the effect of california's tobacco control program Two burning questions on covid-19: Did shutting down the economy help? can we (partially) reopen the economy without risking the second wave? At least 8 us states have declared a state of emergency Forecasting the price of oil Forecasting with panel data Economic forecasts with the yield curve. Federal Reserve Bank of San Francisco Economic Letter A general approach to recovering market expectations from futures prices with an application to crude oil Real-time analysis of oil price risks using forecast scenarios Recent developments in bootstrapping time series On the finite sample accuracy of nonparametric resampling algorithms for economic time series Initial conditions and moment restrictions in dynamic panel data models The impact of the mariel boatlift on the miami labor market Measuring the effects of expectations shocks What we do Data revisions and the identification of monetary policy shocks The financial crisis of 2008 in fixed-income markets Bootstrap methods for standard errors, confidence intervals, and other measures of statistical accuracy Can we trust the bootstrap in high-dimensions? the case of linear models Volatility transmission between gold and oil futures under structural breaks Bootstrapping regression models Bootstrapping autoregressions with conditional heteroskedasticity of unknown form The elements of statistical learning: data mining, inference, and prediction Pooling in dynamic panel-data models: An application to forecasting gdp growth rates What about the opec cartel? Not all oil price shocks are alike: Disentangling demand and supply shocks in the crude oil market Structural vector autoregressive analysis Forecasting inflation using dynamic model averaging The jackknife and the bootstrap for general stationary observations. The Annals of Statistics Estimation of covid-19 spread curves integrating global data and borrowing information Theory of point estimation Statistical inference for average treatment effects estimated by synthetic control methods Forecasting with dynamic panel data models Moving blocks jackknife and bootstrap capture weak dependence. Exploring the limits of bootstrap The subprime credit crisis and contagion in financial markets Multivariate statistics: Old school Forecast with judgment and models. National Bank of Belgium Working Paper At novel coronavirus briefing, governor cuomo declares state of emergency to contain spread of virus Pooled mean group estimation of dynamic heterogeneous panels Integrated time series summarization and prediction algorithm and its application to covid-19 data mining The stationary bootstrap An empirical pooling approach for estimating marketing mix elasticities with pims data Bear Stearns: Crisis and" rescue" for a major provider of mortgage-related products Russian geopolitical objectives in the current oil price crisis, and implications for canada Monetary policy with judgment: Forecast targeting We are grateful to Forrest W. Crawford, Karl Oskar Ekvall, Soheil Eshghi, Lutz Kilian, Ziyu Liu, Dootika Vats, and Dave Zhao for helpful comments. The building block for the following proof is the fact that least squares is conditionally unbiased conditioned on Θ. Case I: under M 1 : It follows that under M 1 (see Section 2.1),