Hi,

  If you run a regression like Y= XB + GH + error where GH is unobserved so the actual regression you will run is Y = XB + (GH + error) and then use the output to forecast various things. If G has variables in it that trend in one direction or another over time ( i.e population in a given area) then will the forecast get less accurate as we get away from the time period used to estimate the coefficients and if so, is there some test that can be used to give an idea of how often the model should be re-estimated? This would also mean that the R^2 should decrease as we get away from the time period used for estimation right?

Thanks,

Chris