Home Combining Large Numbers of Density Predictions with Bayesian Predictive Synthesis
Article
Licensed
Unlicensed Requires Authentication

Combining Large Numbers of Density Predictions with Bayesian Predictive Synthesis

  • Tony Chernis EMAIL logo
Published/Copyright: December 18, 2023

Abstract

Bayesian Predictive Synthesis is a flexible method of combining density predictions. The flexibility comes from the ability to choose an arbitrary synthesis function to combine predictions. I study choice of synthesis function when combining large numbers of predictions – a common occurrence in macroeconomics. Estimating combination weights with many predictions is difficult, so I consider shrinkage priors and factor modelling techniques to address this problem. These techniques provide an interesting contrast between the sparse weights implied by shrinkage priors and dense weights of factor modelling techniques. I find that the sparse weights of shrinkage priors perform well across exercises.


Corresponding author: Tony Chernis, Bank of Canada, K1A 0G9, Ottawa, Canada, E-mail:

Appendix A: Technical Appendix

A.1 MCMC Algorithm

This section describes the Markov Chain Monte Carlo (MCMC) algorithm used to estimate the forecast combinations. It largely follows McAlinn and West (2019) for the BPS steps, Cadonna, Frühwirth-Schnatter, and Knaus (2020) for the global-local shrinkage priors combinations, and Lopes and West (2004) for the factor model combinations. The MCMC follows a two-component block Gibbs sampler: one component samples the synthesis function parameters, and the second samples from the expert forecast distributions or the agent states. As such, I discuss the estimation of each synthesis function separately, followed by details on sampling the agent states.

A.2 Global-local Shrinkage Combinations

This section describes the estimation of the global-local shrinkage synthesis functions. Knaus et al. (2021) provide an R package and the vignette is an excellent overview of the estimation and priors of these models. More details are available in Cadonna, Frühwirth-Schnatter, and Knaus (2020) and Bitto and Frühwirth-Schnatter (2019). I first describe the model, followed by the priors, and then describe the MCMC algorithm.

Starting with the centered parameterization of the synthesis function, for t = 1, …, T, we have that

(7) y t = x t β t + ϵ t β t = β t 1 + u t ϵ t N 0 , σ t 2 u t N ( 0 , Q )

where y t is a univariate response variable and x t = (x t0, x t1, …, x td ) is a d-dimensional row vector containing the regressors at time t, with x t1 corresponding to the intercept.

For simplicity, I assume here that Q = Diag(θ 1, …, θ d ) is a diagonal matrix, implying that the state innovations are conditionally independent. Moreover, I assume the initial value follows a normal distribution (i.e. β 0 N d ( β , Q ) ), with initial mean β = (β 1, …, β d ). Model (7) can be rewritten equivalently in the non-centered parametrization as

(8) y t = x t β + x t D i a g ( θ 1 , , θ d ) β t ̃ + ϵ t , ϵ t N 0 , σ t 2 β t ̃ = β ̃ t 1 + u t ̃ , u t ̃ N d ( 0 , I d )

with β ̃ 0 N d ( 0 , I d ) , where I d is the d-dimensional identity matrix. Furthermore, the model can accommodate stochastic volatility or constant volatility. In the former case, the log-volatility h t = log σ t 2 follows a random-walk. More specifically,

(9) h t | h t 1 , σ η 2 N h t 1 , σ η 2 ,

with initial state h 0 N a 0 , b 0 .

A.2.1 Shrinkage Priors on Variances and Model Parameters

This section describes the priors used in the previously discussed synthesis function. The triple gamma prior can be represented as a conditionally normal distribution, where the component specific variance is itself a compound probability distribution resulting from two gamma distributions. This results in independent normal-gamma-gamma (NGG) priors (Cadonna, Frühwirth-Schnatter, and Knaus 2020), both on the standard deviations of the innovations, that is the θ j ’s, and on the means of the initial value β j , for j = 1, …, d. Note that, in the case of the standard deviations, this can equivalently be seen as a triple gamma prior on the innovation variances θ j , for j = 1, …, d. In the constant parameterizations, I place an NGG prior on the β j using the centered parameterization:

(10) θ j | ξ j 2 N 0 , ξ j 2 , ξ j 2 | a ξ , κ j 2 G a ξ , a ξ κ j 2 2 , κ j 2 | c ξ , κ B 2 G c ξ , c ξ κ B 2

(11) β j | τ j 2 N 0 , τ j 2 , τ j 2 | a τ , λ j 2 N a τ , a τ λ j 2 2 λ j 2 | c τ , λ B 2 N c τ , c τ λ B 2 .

Letting c ξ and c τ go to infinity results in a normal-gamma (NG) prior (Brown and Griffin 2010) on the θ j ’s and β j ’s. It also has a representation as a conditionally normal distribution, with the component specific variance following a gamma distribution; that is

(12) θ j | ξ j 2 N 0 , ξ j 2 , ξ j 2 | a ξ , κ B 2 G a ξ , a ξ κ B 2 2 ,

(13) β j | τ j 2 N 0 , τ j 2 , τ j 2 | a τ , λ B 2 G a τ , a τ λ B 2 2 .

The parameters a ξ , a τ , c ξ , c τ , κ B 2 , and λ B 2 can be learned from the data through appropriate prior distributions. Results from Cadonna, Frühwirth-Schnatter, and Knaus (2020) motivate the use of different distributions for these parameters under the NGG and NG prior. In the NGG case, the scaled global shrinkage parameters conditionally follow F distributions, depending on their respective pole and tail parameters:

(14) κ B 2 2 | a ξ , c ξ F ( 2 a ξ , 2 c ξ ) , λ B 2 2 | a τ , c τ F ( 2 a τ , 2 c τ ) .

The scaled tail and pole parameters, in turn, follow beta distributions:

(15) 2 a ξ B α a ξ , β a ξ , 2 c ξ B α c ξ , β c ξ ,

(16) 2 a τ B α a τ , β a τ , 2 c τ B α c τ , β c τ .

These priors are chosen as they imply a uniform prior on a suitably defined model size; see Cadonna, Frühwirth-Schnatter, and Knaus (2020) for details. In the NG case, the global shrinkage parameters follow independent gamma distributions:

(17) κ B 2 G ( d 1 , d 2 ) , λ B 2 G ( e 1 , e 2 ) .

In order to learn the pole parameters in the NG case, I generalize the approach taken in Bitto and Frühwirth-Schnatter (2019) and place the following gamma distributions as priors:

(18) a ξ G α a ξ , α a ξ β a ξ , a τ G α a τ , α a τ β a τ ,

which correspond to the exponential priors used in Bitto and Frühwirth-Schnatter (2019) when α a ξ = 1 and α a τ = 1 . The parameters α a ξ and α a τ act as degrees of freedom and allow the prior to be bounded away from zero.

In the constant parameter case, I employ a hierarchical prior, where the scale of an inverse gamma prior for σ 2 follows a gamma distribution; that is,

(19) σ 2 | C 0 G 1 ( c 0 , C 0 ) , C 0 G ( c 0 + g 0 , G 0 + σ 2 1 ) ,

with hyperparameters c 0, g 0, and G 0.

In the case of stochastic volatility, the priors on the parameters σ η 2 in Equation (9) are,

(20) σ η 2 G 1 ( ν , S h ) , h 0 N ( a 0 , b 0 )

with hyperparameters ν, S h , a 0 and b 0.

A.2.2 MCMC Sampling Algorithm

This next section describes the MCMC Gibbs sampling algorithm with Metropolis–Hastings steps to obtain draws from the posterior distribution of the global-local shrinkage prior synthesis function parameters. This is meant to be an overview of the algorithm; for more details, please refer to Cadonna, Frühwirth-Schnatter, and Knaus (2020) and Bitto and Frühwirth-Schnatter (2019).

Algorithm 1

Gibbs Sampling Algorithm

  1. If in TVP specification, sample the latent states β ̃ = ( β ̃ 0 , , β ̃ T ) in the non-centered parametrization from a multivariate normal distribution using precision sampling (Chan and Jeliazkov 2009). Otherwise skip.

  2. If in TVP specification, sample jointly β 1, …, β d , and θ 1 , , θ d in the non-centered parametrization from a multivariate normal distribution. Otherwise, sample β 1, …, β d , in the centered parameterization from a multivariate normal distribution.

  3. If in TVP specification, perform an ancillarity-sufficiency interweaving step and redraw each β 1, …, β d from a normal distribution and each θ 1, …, θ d from a generalized inverse Gaussian distribution using the MATLAB implementation (Hartkopf 2022) of Hörmann and Leydold (2014). Otherwise skip.

  4. Sample (where required) the prior variances ξ 1 2 , ξ d 2 and τ 1 2 , τ d 2 and the component specific hyper-parameters. Sample the pole, tail, and global shrinkage parameters. In the NGG case, this is done by employing steps (b)–(f) from Algorithm 1 in Cadonna, Frühwirth-Schnatter, and Knaus (2020). In the NG case, use steps (d) and (e) from Algorithm 1 in Bitto and Frühwirth-Schnatter (2019).

  5. Sample the error variance σ 2 from an inverse gamma distribution in the homoscedastic case or, in the SV case, sample the volatility of the volatility σ η 2 and the log-volatilities h = (h 0, …, h T ).

Step 4 presents a fork in the algorithm, as different parameterizations are used in the NGG and NG case, to improve mixing. For details on the exact parameterization used in the NGG case, see Cadonna, Frühwirth-Schnatter, and Knaus (2020). One key feature of the algorithm is the joint sampling of the time-varying parameters β ̃ t , for t = 0, …, T in step 1 of Algorithm 1. I employ the procedure described in Chan and Jeliazkov (2009) and McCausland, Miller, and Pelletier (2011) from Rue and Held (2005), which exploits the sparse, block tri-diagonal structure of the precision matrix of the full conditional distribution of β ̃ = ( β 0 ̃ , , β T ̃ ) , to speed up computations.

Step 3, as described in Bitto and Frühwirth-Schnatter (2019), makes use of the ancillarity-sufficiency interweaving strategy (ASIS) introduced by Yu and Meng (2011). ASIS is well known to improve mixing by sampling certain parameters both in the centered and non-centered parameterization.

A.3 Factor Model Combinations

The second synthesis function considered in this paper is a Bayesian Factor Model similar to that of Lopes and West (2004), and Lopes (2014) provides an overview of Modern Bayesian Factor Analysis. Please refer to those references for detailed discussion on the methods. Here I provide a brief overview of the model and estimation technique.[11]

(21) y t = F t γ t + ϵ t γ t = γ t 1 + u t x t = Λ f t + ν t

(22) ϵ t N 0 , σ t 2 u t N ( 0 , θ ) ν t N ( 0 , R )

where f t is a k × 1 vector of factors, F t = (1, f t ), γ t is k + 1 vector of coefficients, Λ is a J × k vector of loadings, and R is a diagonal covariance matrix with elements σ ν J 2 . In order to derive combination weights, I need to identify the factors. This is done by the following restriction f t f t = I J and by restricting the first k elements of the loadings matrix to be positive block lower diagonal. This is a common identification scheme used to fix indeterminacy in the estimation of the factors.

To complete model specifications, I need priors for Λ, R, σ t 2 , and θ. The factor loadings have independent priors Λ i j N ( 0 , C 0 ) when ij and Λ i j N ( 0 , C 0 ) 1 ( Λ i i > 0 ) for the upper-diagonal elements of positive loadings i = 1, …, k. Each of the prior variances are independent and modeled as σ ν J 2 I G ( ν / 2 , ν s 2 / 2 ) , similarly θ I G ν θ / 2 , ν θ s θ 2 / 2 . Initial conditions for the γ t are γ 0 N ( 0 , P 0 ) , where P 0 I G ( ν P , ( ν P 1 ) × c P ) .

With the model specified, the next section provides a sketch of the MCMC routine. Interested readers can refer to Lopes and West (2004).

Algorithm 2

Gibbs Sampling Algorithm

  1. Sample f t from independent normal distributions for every t, namely, f t N ( I k + Λ R 1 Λ 1 Λ R 1 x t , I k + Λ R 1 Λ 1 ) .

  2. Sample Λ for i = 1 , , k Λ i N ( m i , C i ) 1 ( Λ i i > 0 ) where m i = C i C 0 1 μ 0 1 i + σ ν i 2 F i x i and C i 1 = C 0 1 I i + σ ν i 2 F i F i .

  3. Sample Λ for i = k + 1 , , J Λ i N ( m i , C i ) 1 ( Λ i i > 0 ) where m i = C i C 0 1 μ 0 1 k + σ ν i 2 F x i and C i 1 = C 0 1 I k + σ ν i 2 F F .

  4. Sample σ ν i 2 I G ( ( ν + T ) / 2 , ( ν s 2 + d i ) / 2 ) where d i = x i F Λ x i F Λ .

  5. If in TVP specification, sample the latent states γ 1, …, γ d , jointly from a multivariate normal distribution using the precision sampler of Chan and Jeliazkov (2009). Otherwise, sample γ = (γ 0, …, γ T ) from a multivariate normal distribution.

  6. Sample the error variance σ 2 from an inverse gamma distribution in the homoscedastic case or, in the SV case, sample the volatility of the volatility σ 2 and the log-volatilities h = (h 0, …, h T ).

A.4 Sampling the Agent States

After estimating the synthesis function parameters, the next step in BPS is to draw x 1:t from p ( x 1 : t | Φ 1 : t , y 1 : t , H 1 : t ) where Φ is the model parameters, y t is the target variable, and H 1 : t is the set of agent densities. As shown in McAlinn and West (2019), the x t , draws from agent densities, are conditionally independent over t with time t conditionals:

(23) p ( x t | Φ t , y t , H t ) N y t | X t β t , ϵ t j = 1 : J h t j ( x t j ) with X t = ( 1 , x t 1 , , x t J )

If the agents provide normal forecast densities, then 23 yields a multivariate normal distribution for x t . The posterior distribution for each x t is:

(24) p ( x t | Φ t , y t , H t ) = N h t + b t c t , H t b t b t g t

where c t = y t β t 0 h t β t , 1 : J , g t = σ t 2 + β t , 1 : J H t β t , 1 : J , and b t = H t β t,1:J /g t . Unfortunately, the applications in this paper do not have analytical forms; instead, histograms represent the agent densities. With no analytical form, I use a Block Metropolis–Hastings step with 24 as a proposal distribution. Since the number of agent densities can be large, I break the MH step into blocks of five experts that are sampled at a time.

There are a few details for Bayesian Factor Model combinations that warrant explanation. First, the model has to be re-parameterized in terms of the x t so that I can use the proposal distribution from 24 in the MH step. The model is straightforward to re-parameterize with the following steps:

(25) y t = x t γ t + ϵ t x t = Λ f t + ν t

(26) f t = ( Λ Λ ) 1 Λ x t ( Λ Λ ) 1 Λ ν t where, Ω = ( Λ Λ ) 1 Λ

(27) y t = x t Ω γ t ν t Ω γ t + ϵ t y t = x t γ t * + ϵ t *

(28) where, ϵ t * = ν t Ω γ t + ϵ t and γ t * = Ω γ t

Now that the model has been re-parameterized, I can use the equation (24) in the MH step by substituting in β t = γ t * , and error variance ϵ t * N 0 , γ t Ω R Ω γ t + σ t 2 .

The second issue that the data (x t ) used to estimate Bayesian Factor Models is standardized to be mean 0 and variance 1. Since the agents provide forecast distributions, I calculate mean and variance used to standardized draws from the agent densities using the marginal density of each expert over all T (h(x)1:T ). Each x t draw is standardized during each MCMC iteration.

Appendix B: Calibration Appendix

This section assesses the calibration of the BPS predictions. Calibration (also referred to as absolute accuracy) is achieved when a predictive density properly characterizes the probability of the events that it is predicting. For example, events predicted to occur with a 20 percent probability should be observed in the data roughly 20 percent of the time. More formally, calibration refers to the statistical consistency between the predictive distributions and the observations of the data they are predicting (Gneiting and Raftery 2007). I assess calibration with a test based off of the probability integral transforms (PITs) (Diebold, Gunther, and Tay 1998) as proposed in Knüppel (2015). In general, I find little evidence to suggest that the predictions from any of the synthesis functions are not calibrated.

Figures 13 and 14 show results from the nowcasting application and the SPF forecasting application. For the most part in the nowcasting applications, the factor model combinations show little evidence of being uncalibrated. However, the shrinkage approaches have slightly different results. The LASSO synthesis function does not appear to produce calibrated predictions, and the test rejects calibration for the time-varying double gamma specification at most horizons. In contrast, the constant parameter specifications produce calibrated predictions at most horizons, the exception being the shortest horizons where calibration is rejected at the 10 percent level. The SPF application has more straightforward results – there is little evidence to suggest the BPS predictions are uncalibrated from any synthesis function. In only two cases is the null hypothesis rejected at the 10 percent level.

Figure 13: 
Knüpple test for probabilistic calibration: nowcasting application. Results from the Knüpple test for probabilistic calibration. Null hypothesis is for calibration and values in the table correspond to p-values. Red shading corresponds to rejection of calibration at 5 percent level and yellow at 10 percent level.
Figure 13:

Knüpple test for probabilistic calibration: nowcasting application. Results from the Knüpple test for probabilistic calibration. Null hypothesis is for calibration and values in the table correspond to p-values. Red shading corresponds to rejection of calibration at 5 percent level and yellow at 10 percent level.

Figure 14: 
Knüpple test for probabilistic calibration: survey forecast application. Results from the Knüpple test for probabilistic calibration. Null hypothesis is for calibration and values in the table correspond to p-values. Red shading corresponds to rejection of calibration at 5 percent level and yellow at 10 percent level.
Figure 14:

Knüpple test for probabilistic calibration: survey forecast application. Results from the Knüpple test for probabilistic calibration. Null hypothesis is for calibration and values in the table correspond to p-values. Red shading corresponds to rejection of calibration at 5 percent level and yellow at 10 percent level.

Appendix C: MCMC Convergence Appendix

In this section, I assess the convergence of the MCMC algorithms. This is done using the Gelman-Rubin diagnostic (Rubin et al. 2015) and implemented through the MATLAB MCMC Diagnostics Toolbox (Vehtari and Särkkä 2014). The Gelman-Rubin diagnostic compares within-chain variance to across-chain variance to estimate a potential scale reduction factor (R), which can be used to assess convergence of the MCMC chain. As a rule of thumb, values below 1.1 suggest convergence. The diagnostic is performed with five chains and on four specifications of BPS using the SPF “tall” dataset. I focus on the constant and time-varying versions of the triple gamma and one factor synthesis functions. This is because the other shrinkage priors are special cases of the triple gamma prior and reduce to simpler versions of the sampler. Results for other shrinkage are available upon request. Since the number of parameters, state variables, and hyper-parameters sampled can number in the thousands, I report box plots of the potential scale reduction factor in Figure 15. These results show reasonable convergence of the MCMC algorithms.

Figure 15: 
Box plots of potential scale reduction factors. The above shows the potential scale reduction factors (R) from the Gelman-Rubin diagnostic with 5 chains on the SPF ‘Tall’ dataset. Values less than 1.1 provide evidence that the MCMC has converged.
Figure 15:

Box plots of potential scale reduction factors. The above shows the potential scale reduction factors (R) from the Gelman-Rubin diagnostic with 5 chains on the SPF ‘Tall’ dataset. Values less than 1.1 provide evidence that the MCMC has converged.

References

Aastveit, K. A., K. Gerdrup, and A. S. Jore. 2011. Short-term Forecasting of GDP and Inflation in Real-Time: Norges Bank’s System for Averaging Models, Vol.9. Oslo: Norges Bank Staff Memo.Search in Google Scholar

Aastveit, K. A., F. Ravazzolo, and H. K. Van Dijk. 2016. “Combined Density Nowcasting in an Uncertain Economic Environment.” Journal of Business & Economic Statistics 36 (1): 131–45.10.1080/07350015.2015.1137760Search in Google Scholar

Aastveit, K. A., J. Mitchell, F. Ravazzolo, and H. K. Van Dijk. 2018. The Evolution of Forecast Density Combinations in Economics. Tinbergen Institute. Technical Report 18-069/III.10.2139/ssrn.3243003Search in Google Scholar

Aastveit, K. A., J. L. Cross, and H. K. V. Dijk. 2023. “Quantifying Time-Varying Forecast Uncertainty and Risk for the Real Price of Oil.” Journal of Business & Economic Statistics 41 (2): 523–37. https://doi.org/10.1080/07350015.2022.2039159.Search in Google Scholar

Bache, I. W., J. Mitchell, F. Ravazzolo, and S. P. Vahey. 2009. Macro Modelling with Many Models. Norges Bank. Technical Report 2009/15.10.1017/CBO9780511779770.016Search in Google Scholar

Belmonte, M. A. G., G. Koop, and D. Korobilis. 2014. “Hierarchical Shrinkage in Time-Varying Parameter Models.” Journal of Forecasting 33 (1): 80–94. https://doi.org/10.1002/for.2276.Search in Google Scholar

Billio, M., R. Casarin, F. Ravazzolo, and H. K. van Dijk. 2013. “Time-varying Combinations of Predictive Densities Using Nonlinear Filtering.” Journal of Econometrics 177 (2): 213–32. https://doi.org/10.1016/j.jeconom.2013.04.009.Search in Google Scholar

Bitto, A., and S. Frühwirth-Schnatter. 2019. “Achieving Shrinkage in a Time-Varying Parameter Model Framework.” Journal of Econometrics 210 (1): 75–97. https://doi.org/10.1016/j.jeconom.2018.11.006.Search in Google Scholar

Bjèrnland, H. C., K. Gerdrup, A. S. Jore, C. Smith, and L. A. Thorsrud. 2012. “Does Forecast Combination Improve Norges Bank Inflation Forecasts?” Oxford Bulletin of Economics & Statistics 74 (2): 163–79. https://doi.org/10.1111/j.1468-0084.2011.00639.x.Search in Google Scholar

Brown, P. J., and J. E. Griffin. 2010. “Inference with Normal-Gamma Prior Distributions in Regression Problems.” Bayesian Analysis 5 (1): 171–88. https://doi.org/10.1214/10-ba507.Search in Google Scholar

Cadonna, A., S. Frühwirth-Schnatter, and P. Knaus. 2020. “Triple the Gamma—A Unifying Shrinkage Prior for Variance and Variable Selection in Sparse State Space and TVP Models.” Econometrics 8 (2), https://doi.org/10.3390/econometrics8020020.Search in Google Scholar

Carvalho, C. M., N. G. Polson, and J. G. Scott. 2010. “The Horseshoe Estimator for Sparse Signals.” Biometrika 97 (2): 465–80. https://doi.org/10.1093/biomet/asq017.Search in Google Scholar

Casarin, R., S. Grassi, F. Ravazzollo, and H. K. Van Dijk. 2019. Forecast Density Combinations with Dynamic Learning for Large Data Sets in Economics and Finance. SSRN Scholarly Paper 3363556. Rochester: Social Science Research Network.10.2139/ssrn.3363556Search in Google Scholar

Chan, J. C., and I. Jeliazkov. 2009. “Efficient Simulation and Integrated Likelihood Estimation in State Space Models.” International Journal of Mathematical Modelling and Numerical Optimisation 1 (1–2): 101–20. https://doi.org/10.1504/ijmmno.2009.030090.Search in Google Scholar

Chernis, T., and R. Sekkel. 2017. “A Dynamic Factor Model for Nowcasting Canadian GDP Growth.” Empirical Economics 53 (1): 217–34. https://doi.org/10.1007/s00181-017-1254-1.Search in Google Scholar

Chernis, T., and R. Sekkel. 2018. Nowcasting Canadian Economic Activity in an Uncertain Environment. Bank of Canada: Discussion Paper.Search in Google Scholar

Chernis, T., and T. Webley. 2022. Nowcasting Canadian GDP with Density Combinations. Technical Report 2022-12, Bank of Canada, publication Title: Discussion Papers.Search in Google Scholar

Coletti, D., and S. Murchison. 2002. “Models in Policy-Making.” In Bank of Canada Review, Vol. 2002, 19–26. Spring.Search in Google Scholar

Conflitti, C., C. De Mol, and D. Giannone. 2015. “Optimal Combination of Survey Forecasts.” International Journal of Forecasting 31 (4): 1096–103. https://doi.org/10.1016/j.ijforecast.2015.03.009.Search in Google Scholar

Cross, J. L., C. Hou, and A. Poon. 2020. “Macroeconomic Forecasting with Large Bayesian VARs: Global-Local Priors and the Illusion of Sparsity.” International Journal of Forecasting 36 (3): 899–915. https://doi.org/10.1016/j.ijforecast.2019.10.002.Search in Google Scholar

Del Negro, M., R. B. Hasegawa, and F. Schorfheide. 2016. “Dynamic Prediction Pools: An Investigation of Financial Frictions and Forecasting Performance.” Journal of Econometrics 192 (2): 391–405. https://doi.org/10.1016/j.jeconom.2016.02.006.Search in Google Scholar

Diebold, F. X., T. A. Gunther, and A. S. Tay. 1998. “Evaluating Density Forecasts with Applications to Financial Risk Management.” International Economic Review 39 (4): 863–83. https://doi.org/10.2307/2527342.Search in Google Scholar

Diebold, F. X., M. Shin, and B. Zhang. 2022. “On the Aggregation of Probability Assessments: Regularized Mixtures of Predictive Densities for Eurozone Inflation and Real Interest Rates.” Journal of Econometrics 237 (2): 105321, https://doi.org/10.2139/ssrn.4004848.Search in Google Scholar

ECB. 2019. Results Of the Third Special Questionnaire for Participants in the ECB Survey of Professional Forecasters. Frankfurt: European Central Bank. https://www.ecb.europa.eu/stats/ecb_surveys/survey_of_professional_forecasters/html/ecb.spf201902_specialsurvey~7275f9e7e6.en.html.Search in Google Scholar

Fava, B., and H. F. Lopes. 2021. “The Illusion of the Illusion of Sparsity: An Exercise in Prior Sensitivity.” Brazilian Journal of Probability and Statistics 35 (4): 699–720. https://doi.org/10.1214/21-bjps503.Search in Google Scholar

Frühwirth-Schnatter, S., and H. Wagner. 2010. “Stochastic Model Specification Search for Gaussian and Partial Non-gaussian State Space Models.” Journal of Econometrics 154 (1): 85–100. https://doi.org/10.1016/j.jeconom.2009.07.003.Search in Google Scholar

García, J. A. 2003. An Introduction to the ECB’s Survey of Professional Forecasters. Occasional Paper Series 8. European Central Bank.Search in Google Scholar

Geweke, J., and G. Amisano. 2011. “Optimal Prediction Pools.” Journal of Econometrics 164 (1): 130–41. https://doi.org/10.1016/j.jeconom.2011.02.017.Search in Google Scholar

Giannone, D., M. Lenza, and G. E. Primiceri. 2021. “Economic Predictions with Big Data: The Illusion of Sparsity.” Econometrica 89 (5): 2409–37. https://doi.org/10.3982/ecta17842.Search in Google Scholar

Gneiting, T., and A. E. Raftery. 2007. “Strictly Proper Scoring Rules, Prediction, and Estimation.” Journal of the American Statistical Association 102 (477): 359–78. https://doi.org/10.1198/016214506000001437.Search in Google Scholar

Hall, S. G., and J. Mitchell. 2007. “Combining Density Forecasts.” International Journal of Forecasting 23 (1): 1–13. https://doi.org/10.1016/j.ijforecast.2006.08.001.Search in Google Scholar

Hartkopf, J. 2022. “Gigrnd.” In MATLAB Central File Exchange. Also available at: Https://ww2.mathworks.cn/matlabcentral/fileexchange/78805-gigrnd.Search in Google Scholar

Hörmann, W., and J. Leydold. 2014. “Generating Generalized Inverse Gaussian Random Variates.” Statistics and Computing 24 (4): 547–57. https://doi.org/10.1007/s11222-013-9387-3.Search in Google Scholar

Jore, A. S., J. Mitchell, and S. P. Vahey. 2010. “Combining Forecast Densities from VARs with Uncertain Instabilities.” Journal of Applied Econometrics 25 (4): 621–34. https://doi.org/10.1002/jae.1162.Search in Google Scholar

Kastner, G., and S. Frühwirth-Schnatter. 2017. “Ancillarity-sufficiency Interweaving Strategy (ASIS) for Boosting MCMC Estimation of Stochastic Volatility Models.” Paper, arXiv.org.Search in Google Scholar

Kastner, G., S. Frühwirth-Schnatter, and H. F. Lopes. 2017. “Efficient Bayesian Inference for Multivariate Factor Stochastic Volatility Models.” Journal of Computational & Graphical Statistics 26 (4): 905–17. https://doi.org/10.1080/10618600.2017.1322091.Search in Google Scholar

Knaus, P., A. Bitto-Nemling, A. Cadonna, and S. Frühwirth-Schnatter. 2021. “Shrinkage in the Time-Varying Parameter Model Framework Using the R Package shrinkTVP.” Journal of Statistical Software 100 (13): 1–32. https://doi.org/10.18637/jss.v100.i13.Search in Google Scholar

Knotek, E. S., and S. Zaman. 2023. “Real-time Density Nowcasts of US Inflation: A Model Combination Approach.” International Journal of Forecasting 39 (4): 1736–60. https://doi.org/10.1016/j.ijforecast.2022.04.007.Search in Google Scholar

Knüppel, M. 2015. “Evaluating the Calibration of Multi-Step-Ahead Density Forecasts Using Raw Moments.” Journal of Business & Economic Statistics 33 (2): 270–81. https://doi.org/10.1080/07350015.2014.948175.Search in Google Scholar

Lopes, H. F. 2014. “Modern Bayesian Factor Analysis.” In Bayesian Inference in the Social Sciences, edited by I. Jeliazkov and X. S. Yang, 115–53. Hoboken: John Wiley & Sons, Inc.10.1002/9781118771051.ch5Search in Google Scholar

Lopes, H. F., and M. West. 2004. “Bayesian Model Assessment in Factor Analysis.” Statistica Sinica 14 (1): 41–67.Search in Google Scholar

McAlinn, K. 2021. “Mixed-frequency Bayesian Predictive Synthesis for Economic Nowcasting.” Journal of the Royal Statistical Society: Series C (Applied Statistics) 70 (5): 1143–63. https://doi.org/10.1111/rssc.12500.Search in Google Scholar

McAlinn, K., and M. West. 2019. “Dynamic Bayesian Predictive Synthesis in Time Series Forecasting.” Journal of Econometrics 210 (1): 155–69. https://doi.org/10.1016/j.jeconom.2018.11.010.Search in Google Scholar

McAlinn, K., K. A. Aastveit, J. Nakajima, and M. West. 2020. “Multivariate Bayesian Predictive Synthesis in Macroeconomic Forecasting.” Journal of the American Statistical Association 115 (531): 1092–110. https://doi.org/10.1080/01621459.2019.1660171.Search in Google Scholar

McCausland, W. J., S. Miller, and D. Pelletier. 2011. “Simulation Smoothing for State–Space Models: A Computational Efficiency Analysis.” Computational Statistics & Data Analysis 55 (1): 199–212. https://doi.org/10.1016/j.csda.2010.07.009.Search in Google Scholar

Mitchell, J., and S. Hall. 2005. “Evaluating, Comparing and Combining Density Forecasts Using the KLIC with an Application to the Bank of England and NIESR ‘Fan’ Charts of Inflation.” Oxford Bulletin of Economics & Statistics 67 (s1): 995–1033. https://doi.org/10.1111/j.1468-0084.2005.00149.x.Search in Google Scholar

Onorante, L., and A. E. Raftery. 2016. “Dynamic Model Averaging in Large Model Spaces Using Dynamic Occam’s Window.” European Economic Review 81: 2–14. https://doi.org/10.1016/j.euroecorev.2015.07.013.Search in Google Scholar PubMed PubMed Central

Prado, R., and M. West. 2010. Time Series: Modeling, Computation, and Inference, 1st ed. Boca Raton: Chapman and Hall/CRC.Search in Google Scholar

Ray, P., and A. Bhattacharya. 2018. “Signal Adaptive Variable Selector for the Horseshoe Prior.” Technical Report arXiv:1810.09004, arXiv.Search in Google Scholar

Rubin, D. B., A. Gelman, J. B. Carlin, H. S. Stern, D. B. Dunson, and A. Vehtari. 2015. Bayesian Data Analysis, 3rd ed. New York: Chapman and Hall/CRC.Search in Google Scholar

Rue, H., and L. Held. 2005. Gaussian Markov Random Fields: Theory And Applications. New York: Chapman and Hall/CRC.10.1201/9780203492024Search in Google Scholar

Takanashi, K., and K. McAlinn. 2021. “Predictions with Dynamic Bayesian Predictive Synthesis Are Exact Minimax.” arXiv:1911.08662 [econ, math, stat] ArXiv: 1911.08662.Search in Google Scholar

Vehtari, A., and S. Särkkä. 2014. MCMC Diagnostics Toolbox for Matlab 6.x. Also available at: Https://users.aalto.fi/ave/code/mcmcdiag/.Search in Google Scholar

Wallis, K. F. 2005. “Combining Density and Interval Forecasts: A Modest Proposal.” Oxford Bulletin of Economics & Statistics 67 (s1): 983–94. https://doi.org/10.1111/j.1468-0084.2005.00148.x.Search in Google Scholar

West, M. 1992. “Modelling Agent Forecast Distributions.” Journal of the Royal Statistical Society: Series B 54 (2): 553–67. https://doi.org/10.1111/j.2517-6161.1992.tb01896.x.Search in Google Scholar

West, M., and J. Crosse. 1992. “Modelling Probabilistic Agent Opinion.” Journal of the Royal Statistical Society: Series B 54 (1): 285–99. https://doi.org/10.1111/j.2517-6161.1992.tb01882.x.Search in Google Scholar

Yu, Y., and X. L. Meng. 2011. “To Center or Not to Center: That Is Not the Question—An Ancillarity—Sufficiency Interweaving Strategy (ASIS) for Boosting MCMC Efficiency.” Journal of Computational & Graphical Statistics 20 (3): 531–70. https://doi.org/10.1198/jcgs.2011.203main.Search in Google Scholar


Supplementary Material

This article contains supplementary material (https://doi.org/10.1515/snde-2022-0108).


Received: 2022-11-28
Accepted: 2023-11-01
Published Online: 2023-12-18

© 2023 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 15.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/snde-2022-0108/html
Scroll to top button