Previous Student Question
Wawro takes the T=3 case and defines the true model to be: y_{i,t} = \gamma y_{i,t-1} + \alpha_{i} + u_{i,t}
What does this allow us to write?
\Delta_{3,2} y_{i} = \underbrace{\gamma [ \overbrace{\gamma y_{i,1} + \alpha_{i} + u_{i,2}}^{y_{i,2}} ] + \alpha_{i} + u_{i,3}}_{y_{i,3}} - \underbrace{\gamma y_{i,1} + \alpha_{i} + u_{i,2}}_{y_{i,2}} (\gamma^{2} - \gamma)y_{i,1} + \underbrace{\gamma \alpha_{i}}_{\delta_{i}} + \underbrace{u_{i,3} + (\gamma - 1) u_{i,2}}_{u_{\Delta(3,2)}^{*}}
In other words, this is an estimating equation without regressors.
So how does he get \Delta y_{i,2} = (\gamma - 1) y_{i,1} + \alpha_{i} + u_{i,2}? It relies on how we treat the initial condition.
\Delta_{2,1} y_{i} = \overbrace{\gamma y_{i,1} + \alpha_{i} + u_{i,2}}^{y_{i,2}} - y_{i,1} where the last term is \gamma y_{i,0} + \alpha_{i} + u_{i,1}. If, as is commonly the case, we treat the initial condition as fixed, we get what he gets. If we treat y_{i,1} as the fixed effect plus random noise, then we would get, after some manipulation, \Delta_{2,1} y_{i} = (\gamma - 1)y_{i,1} + (u_{i,2} - u_{i,1}) Because of the assumption of no serial correlation, the latter, under normality, is also normal (as a difference in independent normals) because the unit effects cancel under differencing.