Dear all,
I have been having some troubles computing a Monte Carlo simulation.
I am a student, so please don't be too horrified by my short experience :)
The problem is the following:
I am requested to perform a MC simulation on a lagged-dependent variable
model with autocorrelated residuals. The estimator for the lagged-dependent
variable, the only regressor, should be inconsistent and biased.
My concern is that decreasing the sample size the bias decreases, and I
guess that that's just wrong.
(I also tried using robust std errors, and nothing changes!)
here's the code I used:
nulldata 200
setobs --time-series
scalar n=$nobs
scalar beta1=0.8
scalar rho=0.4
smpl 1 1
series u=0
series y=0
smpl 2 200
series e=randgen(N,0,1)
series u=rho*u(-1)+e
series y=beta1*y(-1)+u
ols y y(-1) --simple
loop 5000 --progressive --quiet
series e=randgen(N,0,1)
series u=rho*u(-1)+e
series y=beta1*y(-1)+u
ols y y(-1)
scalar beta1hat= $coeff(y_1)
scalar s1= $stderr(y_1)
scalar bias=$coeff(y_1)-beta1
genr LB = beta1hat - critical(t,$df,0.025)*s1
genr UB = beta1hat + critical(t,$df,0.025)*s1
print beta1 beta1hat bias LB UB
endloop
smpl 50 --random --replace
loop 5000 --progressive --quiet
series e=randgen(N,0,1)
series u=rho*u(-1)+e
series y=beta1*y(-1)+u
ols y y(-1)
scalar beta1hat= $coeff(y_1)
scalar s1= $stderr(y_1)
scalar bias=$coeff(y_1)-beta1
genr LB = beta1hat - critical(t,$df,0.025)*s1
genr UB = beta1hat + critical(t,$df,0.025)*s1
print beta1 beta1hat bias LB UB
endloop
Thank you so much