Академический Документы
Профессиональный Документы
Культура Документы
2, 46-58
Available online at http://pubs.sciepub.com/ajams/4/2/4
Science and Education Publishing
DOI:10.12691/ajams-4-2-4
Abstract This paper provides a generalized model for the random-coefficients panel data model where the errors
are cross-sectional heteroskedastic and contemporaneously correlated as well as with the first-order autocorrelation
of the time series errors. Of course, the conventional estimators, which used in standard random-coefficients panel
data model, are not suitable for the generalized model. Therefore, the suitable estimator for this model and other
alternative estimators have been provided and examined in this paper. Moreover, the efficiency comparisons for
these estimators have been carried out in small samples and also we examine the asymptotic distributions of them.
The Monte Carlo simulation study indicates that the new estimators are more reliable (more efficient) than the
conventional estimators in small samples.
Keywords: classical pooling estimation, contemporaneous covariance, first-order autocorrelation,
heteroscedasticity, mean group estimation; monte carlo simulation, random coefficient regression
Cite This Article: Mohamed Reda Abonazel, Generalized Random Coefficient Estimators of Panel Data
Models: Asymptotic and Small Sample Properties. American Journal of Applied Mathematics and Statistics, vol.
4, no. 2 (2016): 46-58. doi: 10.12691/ajams-4-2-4.
1. Introduction
Statistical methods can be characterized according to
the type of data to which they are applied. The field of
survey statistics usually deals with cross-sectional data
describing each of many different individuals or units at a
single point in time. Econometrics commonly uses time
series data describing a single entity, usually an economy
or market. The econometrics literature reveals another
type of data called panel data, which refers to the
pooling of observations on a cross-section of households,
countries, and firms over several time periods. Pooling
this data achieves a deep analysis of the data and gives a
richer source of variation which allows for more efficient
estimation of the parameters. With additional, more
informative data, we can get more reliable estimates and
test more sophisticated behavioral models with less
restrictive assumptions. Another advantage of panel data
sets is their ability to control for individual heterogeneity. 1
Panel data sets are also more effective in identifying
and estimating effects that are simply not detectable in
pure cross-sectional or pure time series data. In particular,
panel data sets are more effective in studying complex
issues of dynamic behavior. For example, in a crosssectional data set, we can estimate the rate of
unemployment at a particular point in time. Repeated
cross sections can show how this proportion changes over
time. Only panel data sets can estimate what proportion of
those who are unemployed in one period remain
0F
47
k =1 ki xkit + uit =
K
x it i + uit ,
i =1, 2, , N ; t =1, 2, , T ,
(1)
3
Cooley and Prescott [14] suggested a model where coefficients vary
from one time period to another on the basis of a non-stationary process.
Similar models have been considered by Sant [39] and Rausser et al. [38]
4
The RCR model has been applied also in different sciences fields, see,
e.g., [9].
=
yi X i i + ui ,
)
where = (1 , , ) ,
= (x1
, , x
(1 , , ) , and = (1 , , ) .
i = X i' X i
X i' yi .
(3)
ii IT
'
E ui u=
j
0
( )
if i = j
i=
, j 1, 2, , N .
if i j
Assumption 3: The exogenous variables are nonstochastic and the ( ) = for every =
1, 2, , , where < .
Assumption 4: The exogenous variables and the errors are
independent, i.e., ( ) = 0 , .
These conditions are sufficient but not necessary for the
optimality of the OLS estimator. 5 When OLS is not
optimal, estimation can still proceed equation by equation
in many cases. For example, if variance of is not
constant, the errors are either serially correlated and/or
heteroskedastic, and the GLS method will provide
relatively more efficient estimates than OLS, even if GLS
was applied to each equation separately as in OLS.
If the covariances between and (for every
i=
, j 1, 2, , N ) do not equal to zero, then contemporaneous
correlation is present, and we have what Zellner [51]
termed as seemingly unrelated regression (SUR) equations,
where the equations are related through cross-equation
correlation of errors. If the ( = 1, 2, , ) matrices do
not span the same column space 6 and contemporaneous
correlation exists, a relatively more efficient estimator of
than equation by equation OLS is the GLS estimator
applied to the entire equation system as shown in Zellner
[51].
With either separate equation estimation or the SUR
methodology, we obtain parameter estimates for each
individual unit in the database. Now suppose it is
necessary to summarize individual relationships and to
draw inferences about certain population parameters.
Alternatively, the process may be viewed as building a
single model to describe the entire group of individuals
4F
5F
5
For more information about the optimality of the OLS estimators, see,
e.g., [36,40]
6
In case of involves exactly the same elements and/or no crossequation correlation of the errors, then no gain in efficiency is achieved
by using Zellner's SUR estimator and OLS can be applied equation by
equation. Dwivedi and Srivastava [19] showed further that whenever
spans the same column space, OLS can be applied equation by equation
without a loss in efficiency.
=
Y X + u,
where = (1 , , ) , = (1 , , ) ,
=
)
(1 , ,
, and = (1 , , ) is a vector of fixed
coefficients which to be estimated. Here we will
differentiate between three cases based on the variancecovariance structure of . In the first case, the errors have
the same variance for each individual as given in the
following assumption:
Assumption 6:
=
2
0
=
, = 1,2, , .
CP1
= X X
'
'
X Y.
(5)
X ' ( I )1 Y , (6)
H
T
=
, = 1,2, , .
u=
yi X i i ,
i
where is defined in (3). 7
6F
3. Random-Coefficients Models
In this section, we review the standard randomcoefficients model, proposed by Swamy [41]. Moreover,
we present the random-coefficients model in the general
case; when the errors are cross-sectional heteroskedastic
and contemporaneously correlated as well as with the
first-order autocorrelation of the time series errors.
X ' ( I )1 Y , (7)
HC
T
12
22
ui' u j
T K
; i, j = 1, 2, , N ,
j
if i =
'
E ( =
i=
, j 1, 2, , N ,
i ) 0, and E i =
j
0 if i j
RCR = X ' 1 X
(8)
(11)
X ' 1Y ,
1
2
ij =
(10)
i= + i ,
where
11
21
=
1
(9)
7F
48
( H IT ) + D ( I N ) D ' .
(12)
(13)
7
The in (8) is unbiased estimator, because we assume, in the first,
that the number of exogenous variables of each equation is equal, i.e.,
= for = 1,2, , . However, in the general case, , the
unbiased estimator is + ( ) , where =
( )1 . See [6,14].
8 This means that the individuals in our database are drowning from a
population with a common regression parameter , which is fixed
component, and a random component , which will allow the
coefficients to differ from unit to unit.
49
1
RCR X i' X i X i' + ii IT
Xi
=
i =1
X i' ( X i X i' + ii IT )
N
+
=
1 N
1 N N
i i' i i' ,
N 1=
N=i 1 =i 1
i 1
(18)
(15)
i =1
N
(14)
yi
= Wi i ,
i =1
N
=
Wi + ii X i' X i
i =1
N
'
+ ii X i X i
i =1
) (
N
= + ii X i' X i
i =1
)
)
1 1
1 1
1 1
(16)
1 N ^ ^ ' 1 N ^ ? N ^ '
i i i i
=
i 1
N 1=
N
=i 1 =i 1
N ^
ii X i' X i
i =1
(17)
ii IT
'
E i =
j
ij IT
if i = j
( )
i=
, j 1, 2, , N ,
if i j
GRCR = X '*1 X
X '*1Y ,
(19)
Where
*
X1X1' + 11
11
21 21
N1 N1
12
12
'
X 2 X 2 + 22
22
N2
N 2
(20)
2 N
2N
'
X N X N + NN
NN
1N
1N
with
1
j
1
ij =
1 i j
T 1
j
i2
iT 1
iT 2
Tj 2
Tj 3
. (21)
11
This suggestion was been used by Stata program, specifically in xtrchh
and xtrchh2 Statas commands. See [34].
uit ui,t 1 ,
i = t =T2
t =2ui2,t 1
T
(22)
ij =
i' j
T K
where = (1 , 2 , , ) ; 1 = 1 1 2 , and
= ,1 for = 2, , .
By replacing by in (21), we get consistent
. And then we will use and
estimators of , say
1 N * *' 1 N * N *'
i i i i
N
N 1=
i 1
=i 1 =i 1
ii ( X iii
N
1
'
Xi
i =1
ij ( X iii
N
'
i j
Xi
(24)
N ( N 1)
1
N2
i , j =1
where
i* = X i'ii1 X i
X i'ii1 yi .
i* ,
(29)
i =1
^
1 *
var MG =
'
1
1
'
1
X i ii ij jj X j X j jj X j
1
N
GMG =
(23)
50
(25)
ii ( X i' X i )
N
i =1
N
ij ( X i' X i )
i j
i , j =1
var GMG
X i'ii X i X i' X i
X i'ij X j X 'j X j
(30)
1
N
N '
'
** 1 * *
ii
) in
By using the consistent estimators ( ,
, and
i i
N
=i 1 =i 1
i 1
. Then
(20), we have a consistent estimator of , say
N
(31)
1
1
to get the generalized RCR (GRCR) estimator=
we use
+ ij X i'ii1 X i
.
N ( N 1) i j
of :
i , j =1
1 ' *1
' *1
(26)
GRCR = X X
X Y.
1
' 1
1
'
1
X iii ij jj X j X j jj X j
(27)
var GRCR = X X
.
the general RCR assumptions are satisfied. In other words,
the GMG estimator is more efficient than the MG
estimator. But under RCR assumptions, we have:
) (
MG =
1
N
i .
(28)
i =1
12
The estimator of in (22) is consistent, but it is not unbiased. See
Srivastava and Giles ([40], p. 211) for other suitable consistent
estimators of that are often used in practice.
13
For more information about the estimation methods for DPD models,
see, e.g., [2,6,48,49,50].
(32)
N
1
1 N N 1 +
= i i' i? i' =
.
N ( N 1)=
N=i 1 =i 1 N
i 1
5. Efficiency Comparisons
In this section, we examine the efficiency gains from
the use of GRCR estimator. Moreover, the asymptotic
variances (as with fixed) of GRCR, RCR, GMG,
and MG estimators have been derived.
Under the general RCR assumptions, It is easy to verify
that the classical pooling estimators (1 , 2 , and 3 )
and Swamys estimator ( ) are unbiased for and with
variance-covariance matrices:
51
( )
=
var ( CP3 )
( )
* '
G=
3 G3 ,?var ( RCR )
* '
=
var CP1 G=
G2 *G2' ,
1 G1 ,?var CP 2
(33)
G4 *G4' ,
(34)
where 1 = ( )1 , 2 = [ (1 )]1 (1 ) ,
1
1
)]1 (
) ,
and
4 =
3 = [ (
1
1
1
( ) . The efficiency gains, from the use of
GRCR estimator, it can be summarized in the following
equation:
( )
=
EGa var a var GRCR
(35)
=
1, , 4,
( Gh G0 ) ( Gh G0 ) , h =
'
finite and positive definite for all and for 0 < | | < 1.
Lemma 2.
If the general RCR assumptions and assumption 10 are
satisfied, then the estimated asymptotic variancecovariance matrices of GRCR, RCR, GMG, and MG
estimators are equal:
^
)
)
1 +
.
N
ki xkit + uit =
k =1
x it + x it i + uit ,
(36)
i =1, 2, , N ; t =1, 2, , T .
ij
ij
ij
ii
ij
ij
.
ij
ii
N N
1000 l =1
(0, 0)
52
(0.95, 0.85)
(5, 8)
(5, 12)
(10, 15)
(10, 20)
(5, 8)
(5, 12)
(10, 15)
(10, 20)
(5, 8)
(5, 12)
(10, 15)
(10, 20)
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
2.579
2.739
2.875
2.793
2.055
14.467
2.394
1.812
1.819
1.795
1.912
1.479
3.074
1.728
0.965
0.950
0.904
1.068
0.904
2.333
0.839
0.765
0.746
0.657
0.813
0.701
2.127
0.672
2.970
3.087
3.235
2.925
2.207
13.457
2.527
1.764
1.773
1.723
1.917
1.218
5.275
1.623
1.071
1.052
0.955
1.165
0.846
4.653
0.812
0.893
0.882
0.785
0.960
0.684
4.487
0.714
5.016
5.483
5.796
5.337
3.441
12.508
4.165
2.881
2.875
2.756
2.935
1.531
21.747
2.255
1.473
1.493
1.344
1.594
0.785
9.985
0.992
1.337
1.324
1.144
1.267
0.613
7.719
0.810
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
4.849
5.204
5.607
4.222
4.187
16.589
4.007
4.387
4.633
4.835
3.892
3.886
4.543
3.869
2.598
2.767
3.133
2.332
2.330
2.306
2.227
3.415
3.602
3.872
3.127
3.127
3.126
3.095
5.235
5.671
6.216
4.508
4.524
9.822
4.287
4.275
4.534
4.648
3.697
3.629
5.695
3.546
3.613
3.978
4.530
3.231
3.203
3.227
3.126
2.638
2.801
2.960
2.417
2.388
2.489
2.330
5.904
6.504
6.900
6.058
5.432
15.662
5.042
4.929
5.376
5.467
4.697
4.518
12.161
4.323
3.217
3.730
3.951
2.947
2.836
4.955
2.675
3.528
4.017
4.063
3.147
3.074
4.513
3.009
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
11.791
13.194
14.553
9.483
9.469
9.797
9.329
10.687
11.391
12.095
9.145
9.143
9.863
9.107
8.097
8.719
10.108
6.812
6.812
6.810
6.781
6.234
6.583
7.155
5.736
5.736
5.735
5.718
9.382
10.605
11.417
7.836
7.850
70.360
7.726
8.687
9.250
9.591
7.185
7.143
10.059
7.107
9.483
10.443
11.928
7.993
7.980
8.042
7.946
6.166
6.621
7.098
5.568
5.556
5.568
5.533
10.457
11.714
12.595
9.170
8.935
11.511
8.353
7.060
7.942
8.199
6.431
6.278
20.520
6.155
7.520
9.039
9.714
6.711
6.665
6.725
6.612
6.983
8.115
8.220
6.208
6.172
6.235
6.142
2.
53
(0, 0)
(0.95, 0.85)
(5, 5)
(10, 10)
(15, 15)
(20, 20)
(5, 5)
(10, 10)
(15, 15)
(20, 20)
(5, 5)
(10, 10)
(15, 15)
(20, 20)
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
4.015
5.107
6.626
4.078
2.848
5.362
3.376
1.398
1.451
2.038
1.573
1.302
2.368
1.152
0.704
0.682
0.858
0.791
0.701
1.203
0.541
0.496
0.478
0.548
0.551
0.501
1.554
0.330
10.555
13.245
14.811
9.155
6.401
9.809
8.166
1.385
1.434
1.888
1.605
1.120
7.191
1.045
0.810
0.802
0.989
0.907
0.681
3.232
0.549
0.580
0.569
0.608
0.632
0.453
2.256
0.335
10.411
14.354
16.655
10.010
6.880
14.884
8.778
2.371
2.549
3.202
2.612
1.402
14.094
1.600
1.314
1.325
1.501
1.318
0.747
10.858
0.735
0.907
0.892
0.830
0.896
0.455
18.453
0.402
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
5.789
7.578
10.048
5.203
4.948
7.719
4.762
3.435
3.879
6.187
3.054
3.051
3.101
2.823
2.077
2.248
3.930
1.915
1.914
1.897
1.809
2.039
2.165
3.971
1.869
1.869
1.865
1.812
9.953
12.696
14.156
8.545
7.302
10.074
7.761
3.464
3.972
6.277
3.118
3.070
3.710
2.876
2.370
2.641
4.454
2.148
2.129
2.137
2.023
2.252
2.452
4.423
2.073
2.052
2.067
1.999
10.443
14.440
16.836
10.005
7.742
15.432
11.464
3.261
3.722
5.301
3.216
3.080
7.726
2.551
2.842
3.362
5.285
2.558
2.510
3.317
2.332
2.419
2.829
4.622
2.176
2.117
2.217
2.027
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
12.123
16.067
21.362
9.441
9.357
11.912
9.041
7.455
8.605
14.099
6.325
6.323
6.297
6.218
5.439
5.958
10.719
4.876
4.876
4.875
4.837
5.141
5.477
10.258
4.639
4.639
4.639
4.617
11.900
15.172
16.722
9.652
9.348
10.657
8.910
7.637
8.912
14.102
6.465
6.441
6.450
6.359
6.373
7.183
12.534
5.599
5.591
5.599
5.553
4.987
5.448
9.985
4.530
4.521
4.528
4.497
13.839
19.262
22.554
11.947
11.803
26.889
11.524
6.262
7.604
11.238
5.229
5.141
6.663
4.800
5.750
6.980
11.359
4.994
4.962
5.019
4.867
4.680
5.596
9.333
4.197
4.166
4.214
4.123
( , )
(, )
(0, 0)
(0.95, 0.85)
(8, 5)
(12, 5)
(15, 10)
(20, 10)
(8, 5)
(12, 5)
(15, 10)
(20, 10)
(8, 5)
(12, 5)
(15, 10)
(20, 10)
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
8.059
12.611
12.098
7.346
5.085
7.583
6.269
5.011
9.223
8.479
4.968
3.780
6.827
3.781
0.915
0.914
1.037
1.048
0.912
1.963
0.594
1.286
1.474
1.790
1.497
1.161
3.424
0.984
5.775
8.959
8.614
5.346
4.694
21.049
4.661
8.819
15.237
14.618
7.303
4.072
7.390
5.896
1.215
1.272
1.472
1.386
1.019
3.765
0.780
1.020
1.106
1.472
1.228
0.944
7.005
0.673
10.427
15.700
18.234
10.191
5.637
16.782
7.861
9.936
18.193
17.588
9.075
8.109
42.044
7.448
2.104
2.455
2.734
2.266
1.636
12.592
1.469
1.597
1.789
2.279
1.875
1.100
10.106
0.937
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
7.211
11.436
10.724
6.429
6.011
7.966
5.929
4.939
9.220
8.292
4.963
4.623
7.216
3.838
2.659
3.138
3.822
2.360
2.359
2.363
2.173
2.498
2.956
3.592
2.346
2.343
2.801
1.938
6.885
10.504
10.083
6.065
6.043
9.943
5.356
6.820
12.145
11.084
5.477
5.124
10.356
4.909
2.132
2.475
3.014
2.001
1.969
3.427
1.602
2.285
2.735
3.324
2.107
2.082
69.747
1.797
9.652
14.789
17.059
9.610
6.398
19.301
7.570
9.851
18.384
17.539
9.036
8.538
35.246
7.515
2.663
3.233
3.642
2.658
2.712
6.077
1.997
2.811
3.642
4.099
2.698
2.614
5.216
2.122
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
8.409
13.196
12.464
7.703
7.762
11.761
6.661
7.200
13.419
12.334
6.546
6.554
7.170
5.629
5.316
6.278
7.654
4.555
4.554
4.547
4.462
5.907
7.128
8.452
4.956
4.954
4.882
4.782
10.697
16.445
16.636
8.304
8.312
28.804
7.712
9.053
16.848
14.188
7.363
7.512
8.898
7.055
4.732
5.724
6.895
4.022
4.007
4.002
3.846
5.113
6.255
7.413
4.418
4.407
4.399
4.286
10.190
15.927
17.419
10.221
9.875
14.425
8.354
11.609
21.264
20.728
10.246
10.139
14.960
8.680
5.723
7.436
8.426
4.907
4.946
4.997
4.554
5.620
7.688
8.309
4.849
4.804
4.805
4.557
3.
( , )
(, )
(0, 0)
4.
54
(0.95, 0.85)
(5, 8)
(5, 12)
(10, 15)
(10, 20)
(5, 8)
(5, 12)
(10, 15)
(10, 20)
(5, 8)
(5, 12)
(10, 15)
(10, 20)
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
4.700
4.854
5.109
4.823
3.652
7.652
4.324
2.869
2.876
2.822
3.074
2.410
10.706
2.725
1.578
1.564
1.505
1.747
1.480
2.723
1.389
1.344
1.316
1.178
1.466
1.258
8.070
1.191
6.294
6.823
7.166
6.259
4.985
16.169
5.674
2.990
3.020
2.941
3.127
2.204
5.969
2.717
1.827
1.805
1.667
1.979
1.474
8.925
1.502
1.522
1.502
1.339
1.663
1.118
5.743
1.202
9.733
10.431
10.790
9.745
4.269
11.531
7.352
4.994
5.022
4.959
5.422
2.336
15.708
3.872
2.793
2.758
2.460
2.946
1.436
13.279
1.801
2.177
2.167
1.880
2.049
1.041
38.349
1.320
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
6.069
6.311
6.704
5.598
5.461
11.318
5.476
4.812
4.969
5.101
4.489
4.462
6.401
4.308
3.119
3.279
3.651
2.874
2.871
3.760
2.659
3.565
3.720
3.948
3.274
3.273
3.452
3.143
6.382
6.996
7.415
6.331
5.948
10.609
5.996
3.283
3.349
3.290
3.337
3.027
13.571
3.116
4.274
4.619
5.165
3.836
3.787
4.511
3.581
4.306
4.615
4.883
3.998
3.919
4.017
3.829
8.993
9.682
9.905
9.174
5.693
16.977
7.382
4.950
5.095
5.076
5.334
4.178
31.590
4.398
3.200
3.271
3.151
3.286
2.852
19.676
2.430
3.396
3.745
3.664
3.147
2.948
10.222
2.770
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
11.783
12.614
13.791
9.398
9.395
12.364
9.239
10.693
11.288
11.705
9.171
9.156
10.120
9.030
8.316
8.920
10.160
7.055
7.054
7.048
6.973
7.119
7.496
8.070
6.387
6.386
6.382
6.331
13.570
14.942
15.989
11.139
11.228
474.873
10.788
8.748
9.425
9.956
7.758
7.717
12.815
7.600
7.442
8.219
9.417
6.555
6.520
6.559
6.411
7.734
8.342
9.007
6.899
6.852
6.889
6.802
8.176
9.083
9.310
8.718
6.889
88.890
7.734
14.887
16.391
16.943
12.302
11.999
18.314
12.024
7.895
9.390
10.113
7.244
7.085
7.466
6.904
6.279
7.273
7.413
5.824
5.711
8.117
5.628
( , )
(, )
(0, 0)
(0.95, 0.85)
(5, 5)
(10, 10)
(15, 15)
(20, 20)
(5, 5)
(10, 10)
(15, 15)
(20, 20)
(5, 5)
(10, 10)
(15, 15)
(20, 20)
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
25.198
31.269
41.189
20.301
12.441
21.118
18.106
2.054
2.081
2.802
2.302
1.946
4.029
1.687
1.214
1.172
1.463
1.336
1.184
2.303
0.876
0.882
0.852
0.992
0.966
0.872
1.519
0.592
12.304
15.469
16.359
10.396
8.180
35.396
9.674
2.575
2.659
3.599
2.818
2.118
7.438
1.950
1.408
1.385
1.701
1.526
1.198
35.939
0.949
1.033
1.014
1.150
1.129
0.849
4.282
0.618
22.924
29.981
55.404
21.736
13.756
23.866
18.606
3.645
3.913
4.976
4.045
2.422
14.154
2.711
2.181
2.216
2.490
2.296
1.149
12.892
1.203
1.554
1.551
1.454
1.584
0.785
8.994
0.702
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
24.857
30.642
40.026
19.204
13.204
24.814
17.694
3.789
4.151
6.472
3.492
3.487
5.061
3.031
2.731
2.931
5.114
2.541
2.540
2.509
2.305
2.660
2.814
5.173
2.458
2.457
2.445
2.323
12.342
15.877
16.719
10.361
9.071
18.642
9.887
3.594
3.930
5.935
3.527
3.469
8.365
2.903
2.648
2.878
4.771
2.486
2.451
2.945
2.136
2.424
2.601
4.447
2.228
2.185
2.243
2.012
21.516
28.305
49.204
20.526
14.664
24.831
17.352
3.445
3.605
4.579
3.896
3.427
19.997
2.669
2.948
3.288
4.572
2.880
2.638
18.780
2.198
2.504
2.821
4.208
2.351
2.166
4.708
1.895
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
22.111
28.169
37.528
16.156
15.764
24.433
16.830
8.161
9.273
14.875
6.873
6.872
6.837
6.674
6.346
6.914
12.451
5.690
5.690
5.687
5.596
4.752
5.056
9.510
4.300
4.299
4.297
4.225
15.841
20.204
21.343
12.892
13.272
27.430
12.785
8.101
9.567
15.129
7.112
7.084
7.613
6.805
7.383
8.273
14.478
6.385
6.372
6.392
6.269
5.726
6.237
11.181
5.011
4.992
5.016
4.919
20.627
27.543
51.439
19.940
18.283
29.796
18.204
7.499
9.081
13.459
6.696
6.546
31.041
5.921
6.586
7.973
12.643
5.842
5.727
5.860
5.536
4.702
5.573
9.041
4.253
4.150
4.287
4.020
55
( , )
(, )
(0, 0)
(0.95, 0.85)
(8, 5)
(12, 5)
(15, 10)
(20, 10)
(8, 5)
(12, 5)
(15, 10)
(20, 10)
(8, 5)
(12, 5)
(15, 10)
(20, 10)
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
8.099
12.381
12.232
7.742
5.447
8.382
6.386
17.393
32.968
29.385
14.751
9.402
17.489
12.973
1.731
1.781
2.033
2.034
1.768
3.876
1.153
1.392
1.406
1.963
1.648
1.463
10.630
0.834
10.036
16.362
18.922
10.003
6.250
15.198
8.263
9.281
16.928
15.942
9.046
7.273
48.547
7.059
2.099
2.229
2.556
2.453
2.045
6.812
1.423
1.675
1.727
2.392
1.892
1.736
46.391
1.010
12.098
18.230
19.356
10.226
10.228
20.562
9.115
58.422
95.939
93.663
44.144
38.853
48.053
37.422
3.198
3.496
3.873
3.628
2.075
19.644
1.908
2.578
2.705
3.693
2.836
1.775
21.881
1.403
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
7.977
12.251
12.069
7.550
6.193
9.369
6.490
15.698
29.797
26.622
12.435
9.803
15.712
11.975
3.145
3.544
4.361
2.977
2.975
3.497
2.384
2.695
3.100
3.805
2.522
2.520
2.553
1.995
9.307
15.449
17.208
9.329
7.059
12.705
8.071
9.106
16.513
15.601
8.838
7.670
21.261
6.935
2.874
3.210
3.799
2.927
2.915
3.835
2.060
2.892
3.379
4.140
2.704
2.731
2.992
2.101
12.425
18.659
20.114
10.485
10.795
18.461
9.445
55.988
92.529
89.044
42.576
37.501
47.773
35.999
3.053
3.340
3.635
3.558
3.151
26.250
2.038
2.948
3.272
4.271
3.085
2.916
22.414
1.799
=
CP1
CP2
CP3
MG
GMG
RCR
GRCR
10.148
15.623
15.672
9.006
8.838
11.771
8.098
14.075
26.924
23.191
11.305
11.598
13.046
11.092
6.294
7.411
9.144
5.418
5.417
5.377
5.132
5.831
6.918
8.111
4.844
4.843
4.813
4.607
9.455
15.729
17.441
9.752
8.971
14.957
8.488
9.717
17.896
17.000
9.346
9.206
11.915
7.649
6.780
5.270
13.786
57.674
6.578
5.433
8.220
6.437
20.662
91.990
8.384
7.082
9.768
7.650
22.626
91.419
9.488
7.981
5.856
4.467
11.409
43.289
6.030
4.925
5.853
4.489
11.916
38.975
5.877
4.826
5.896
4.437
21.958
42.733
8.370
4.872
5.477
4.130
10.302
37.793
5.172
4.239
for firm heterogeneity, predicting firm-specific coefficients, and
estimating Strategy Trade-Offs. Working Paper, No. 14-022.
Harvard Business School Strategy Unit.
Anh, V. V., Chelliah, T. (1999). Estimated generalized least
squares for random coefficient regression models. Scandinavian
journal of statistics 26(1):31-46.
Baltagi, B. H. (2011). Econometrics. 5th ed. Berlin: SpringerVerlag Berlin Heidelberg.
Baltagi, B. H. (2013). Econometric Analysis of Panel Data. 5th ed.
Chichester: John Wiley and Sons.
Beck, N., Katz, J. N. (2007). Random coefficient models for timeseriescross-section data: Monte Carlo experiments. Political
Analysis 15(2):182-195.
Beran, R., Millar, P. W. (1994). Minimum distance estimation in
random coefficient regression models. The Annals of Statistics
22(4):1976-1992.
Bodhlyera, O., Zewotir, T., Ramroop, S. (2014). Random
coefficient model for changes in viscosity in dissolving pulp.
Wood Research 59(4):571-582.
Boness, A. J., Frankfurter, G. M. (1977). Evidence of NonHomogeneity of capital costs within risk-classes. The Journal of
Finance 32(3):775-787.
Boot, J. C., Frankfurter, G. M. (1972). The dynamics of corporate
debt management, decision rules, and some empirical evidence.
Journal of Financial and Quantitative Analysis 7(04):1957-1965.
Chelliah, N. (1998). A new covariance estimator in random
coefficient regression model. The Indian Journal of Statistics,
Series B 60(3):433-436.
Cheng, J., Yue, R. X., Liu, X. (2013). Optimal Designs for
Random Coefficient Regression Models with Heteroscedastic
Errors. Communications in Statistics-Theory and Methods
42(15):2798-2809.
Cooley, T. F., Prescott, E. C. (1973). Systematic (non-random)
variation models: varying parameter regression: a theory and some
applications. Annals of Economic and Social Measurement 2(4):
463-473.
Dielman, T. E. (1983). Pooled cross-sectional and time series data:
a survey of current statistical methodology. The American
Statistician 37(2):111-122.
Dielman, T. E. (1989). Pooled Cross-Sectional and Time Series
Data Analysis. New York: Marcel Dekker.
Dielman, T. E. (1992a). Misspecification in random coefficient
regression models: a Monte Carlo simulation. Statistical Papers
33(1):241-260.
7. Conclusion
In this paper, the classical pooling (CP1, CP2, and CP3),
random-coefficients (RCR and GRCR), and alternative
(MG and GMG) estimators of stationary RCPD models
were examined in different sample sizes in case the errors
are cross-sectionally and serially correlated. Efficiency
comparisons for these estimators indicate that the mean
group and random-coefficients estimators are equivalent
when sufficiently large. Moreover, we carried out
Monte Carlo simulations to investigate the small samples
performance for all estimators given above.
The Monte Carlo results show that the classical pooling
estimators are not suitable for random-coefficients models
absolutely. Also, the MG and GMG estimators are more
efficient than RCR estimator in random- and fixedcoefficients models especially when is small ( 12).
Moreover, the GMG and GRCR estimators perform well
in small samples if the coefficients are random or fixed.
The MG, GMG, and GRCR estimators are approximately
equivalent when 20. However, the GRCR estimator
performs well than the GMG estimator in most situations
especially in moderate samples. Therefore, we conclude
that the GRCR estimator is suitable to stationary RCPD
models whether the coefficients are random or fixed.
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
[13]
[14]
References
[1]
[2]
[3]
[15]
[16]
[17]
56
[19]
[20]
[21]
[22]
[23]
[24]
[25]
[26]
[27]
[28]
[29]
[30]
[31]
[32]
[33]
[34]
[35]
[36]
[37]
[38]
[39]
[40]
[41]
[42]
[44]
[45]
[46]
[47]
[48]
[49]
[50]
[51]
Appendix
A.1 Proof of Lemma 1
) = (
) =
:
a. Show that (
By substituting (25) into (29), we can get
GMG =
( X i'ii1 X i )
N
1
N
i =1
X i'ii1 yi ,
(A.1)
1
N
i + ( X i'ii1 X i )
N
i =1
X i'ii1ui . (A.2)
MG
=
1 '
1 N
i + X i' X i
X i ui .
N i =1
(A.3)
) (
E =
E=
MG
GMG
1 N
=
i .
N i =1
:
b. Derive the variance-covariance matrix of
Beginning, note that under assumption 8, we have
= + . Let us add to the both sides:
i + i* = + i + i* ,
i* = + i + i* i ,
(A.4)
i* = + i + i ,
(A.5)
57
1 N *
1 N
1 N
i =
+ ii + i ,
N
N
N
=i 1
=i 1 =i 1
ij ( X i'ii1 X i )
(A.8)
1
N2
ii (
i =1
ij
i j
i , j =1
1
X i' X i
X i'ij X j
X i'ij X j X 'j X j
(A.12)
X i'ii1ij jj1
=
0.
N
N
N
* = 1 ? ' 1 ? ' = + . (A.13)
plim
i
i
i
i
N 1=
N=i 1 =i 1
T
i 1
By substitute (A.11)-(A.13) into (30), (31), and (27),
we get
^
plim
T
plim T ( X i X i )
2
T ii
N
^*
MG
plim var =
1
'
i =1 T
N
plim
N2
i j T T
T
ij
X i ii X i X i X i
'
1
'
Xi Xi
'
'
X i ij X j
1
'
X jX j
(A.14)
i , j =1
,
^
N ( N 1) T =
i 1
1
=
(A.10)
X i' X i
'
plim i*i*
'
*
i i
N
*
=i 1 =i 1
1
1
ij
1
T
+
T
N ( N 1) i j ' 1
1
'
1
X X X j jj X j
i ii ij jj j
i , j =1
1
.
X 'j X j
X i'ii X i X i' X i
(A.9)
var =
MG var ( ) + var ( )
X j X 'j jj1 X j
1
ij T X i'ii1 X i
T T
= plim
where =
, and , =
. From (A.9) and
=1
=1
using the general RCR assumptions, we get
X i'ii X i
1
ij T X i' X i
T
T
= plim
1
X i' X i
1
ii T X i' X i
T T
1
1
1
i =
+ ii + i ,
N
N
N
=i 1
=i 1 =i 1
= plim
MG = + + ,
(A.11)
1
ii T X i'ii1 X i
T
T
i = + i + i ,
1
1
=
+
N
N2
plim
:
c. Derive the variance-covariance matrix of
As above, we can rewrite the equation (3) as follows:
=
plim ij =
ij and plim ij ij ,
(A.7)
=
i plim
=
i* i ,?plim
=
ij? ij ,?
plim
N
N
N '
'
i*i* 1 i* i*
=
N
i 1
=i 1 =i 1
N
^
1
1
=
var
(GMG ) N ( N 1) + ij X i'ii1 X i
i j
i , j =1
X ' 1 1 X X ' 1 X
j jj
j
i ii ij jj j
X i'ij X j X 'j X j
i j
i , j =1
and then,
i j
i , j =1
X i'ii X i X i' X i
1
1
1 N
+
X i'ii1 X i
ii
N
N 2 i =1
N
ij ( X i' X i )
N
(A.6)
where =
and =
. From (A.6) and
=1
=1
using the general RCR assumptions, we get
N i =1
GMG = + + ,
var =
var ( ) + var (
GMG
ii ( X i' X i )
1 N
^
1 *
var MG =
1
N
+ .
(A.15)
(
T
*1 X
plim var GRCR = plim X '
N +1
=
i =1
(A.16)
' 1
X
= plim X
1
= + .
N
1
N
58
' 1 * 1
' 1
X
X
X X
(A.17)
+
.
)
)
1 +
.
N