Page 360 - Special Topic Session (STS) - Volume 3
P. 360

STS550 Matteo Mogliani
                  aggregating  (weighting)  functions.  The  inclusion  of  many  high-frequency
                  variables into MIDAS regressions may nevertheless lead to overparameterized
                  models, with poor predictive performance. This happens because the MIDAS
                  regression approach can efficiently address the dimensionality issue arising
                  from the number of high-frequency lags in the model, but not that arising
                  from the number of high-variables. Hence, recent literature has focused on
                  MIDAS penalized regressions, based mainly on the so-called Lasso and Elastic-
                  Net penalizations (Marsilli, 2014; Siliverstovs, 2017; Uematsu and Tanaka, in
                  press).
                      In the present paper, we propose a similar approach, but we depart from
                  the existing literature on several points. First, we consider MIDAS regressions
                  resorting to Almon lag polynomial weighting schemes, which depend only on
                  a  bunch  of  functional  parameters  governing  the  shape  of  the  weighting
                  function and keep linearity in the regression model. Second, we consider a
                  Group Lasso penalty, which operates on distinct groups of regressors, and we
                  set as many groups as the number of high-frequency predictors, allowing each
                  group  to  include  the  entire  Almon  lag  polynomial  of  each  predictor.  This
                  grouping  structure  is  motivated  by  the  fact  that  if  one  high-frequency
                  predictor is irrelevant, it should be expected that zero-coefficients occur in all
                  the parameters of its lag polynomial. Third, we implement Bayesian techniques
                  for the estimation of our penalized MIDAS regressions. The Bayesian approach
                  offers two attractive features in our framework. The first one is the inclusion of
                  spike-and-slab priors that, combined with the penalized likelihood approach,
                  aim at improving the selection ability of the model by adding a probabilistic
                  recovery layer to the hierarchy. The second one is the estimation of the penalty
                  hyper-parameters through an automatic and data-driven approach that does
                  not resort to extremely time-consuming pilot runs. In this paper we consider
                  an  algorithm  based  on  stochastic  approximations,  which  consists  in
                  approximating the steps necessary to estimate the hyper-parameters in such
                  a  way that simple analytic solutions can  be used. It turns out that penalty
                  hyper-parameters  can  be  automatically  tuned  with  a  small  computational
                  effort compared to existing and very popular alternative algorithms.

                  2.  Methodology
                      Consider the variable t, which is observed at discrete times (i.e. only once
                  between t − 1 and t), and suppose that we want to use information stemming
                                                    ()
                                                            ()
                  from a set of  predictors  ()  =  1,  ,... ,  ,  )’, which are observed  times
                                             
                                                                                     ()
                  between t − 1 and t, for forecasting purposes. The variables t and    , for
                                                                                     ,
                   = 1, … ,   are  said  to  be  sampled  at  different  frequencies.  For  instance,
                  quarterly and monthly frequencies, respectively, in which case  = 3. Let us
                                                                              
                  define  the  high-frequency  lag  operator  1/m ,  such  that  L 1/m ()  =  ()  .
                                                                               ,  ,−1/
                                                                     349 | I S I   W S C   2 0 1 9
   355   356   357   358   359   360   361   362   363   364   365