UNIVERSITY OF

ILLINOIS LIBRARY

AT URBANA-CHAMPAIGN

BOOKSTACKS

UNIVERSITY LIBRARY

UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

The person charging this material is responsible for its

renewal or return to the library on or before the due date.

The minimum fee for a lost item is $1 25.00, $300.00 for

bound journals.

Theft, mutilation, and underlining of books are reasons

for disciplinary action and may result in dismissal from

the University. Please note: self-stick notes may result in

torn pages and lift some inks.

Renew via the Telephone Center at 217-333-8400,

846-262-1510 (toll-free) orcirclib@uiuc.edu.

Renew online by choosing the My Account option at:

http://www.library.uiuc.edu/catalog/

330 B385

No. 1668 COPY 2

STX

BEBR

FACULTY WORKING PAPER NO. 90-1668

The Risk Properties of a Pre-Test Estimator for Zellner's Seemingly Unrelated Regression Model

Ahmet Ozcam George Judge Anil Bera Thomas Yancey

The Library of the AUG 2 1 J99Q

College of Commerce and Business Administration Bureau of Economic and Business Research University of Illinois Urbana-Champaign

BEBR

FACULTY WORKING PAPER NO. 90-1668

College of Commerce and Business Administration

University of Illinois at Urbana-Champaign

July 1990

The Risk Properties of a Pre-Test Estimator for Zellner's Seemingly Unrelated Regression Model

Ahmet Ozcam

Research Division of Turkish Parliment

Ankara, Turkey

George Judge University of California, Berkeley

Anil Bera & Thomas Yancey

Department of Economics

University of Illinois at Urbana-Champaign

Digitized by the Internet Archive

in 2011 with funding from

University of Illinois Urbana-Champaign

http://www.archive.org/details/riskpropertiesof1668z

Abstract

In the case of Zellner's seemingly unrelated statistical model it is well known that the efficiency of the generalized least squares estimator (GLSE) relative to that of the least squares estimator (LSE) is conditional on the magnitude of the correlation between the equa- tion errors. Using a relevant test statistic, we analytically eva- luate the risk, characteristics of a seemingly unrelated regressions pre-test estimator (SURPE) that is the GLSE if a preliminary test, based on the data at hand, indicates that the correlation between equation errors is significantly different from zero, and the LSE if we accept the null hypothesis of no correlation. The small sample distribution of the test statistic, used in defining SURPE is also derived.

Key Words: Risk, Pre-Test estimator, Least squares estimator,

Generalized least squares estimator, Seemingly unrelated regression model, Test statistic.

THE RISK PROPERTIES OF A PRE-TEST ESTIMATOR FOR ZELLNER'S SEEMINGLY UNRELATED REGRESSION MODEL

1. Introduction

Since Zellner (1962) proposed the use of Aitker's generalized least squares estimator (GLSE) for a set of disturbance related regression equations, the efficiency of this estimator relative to that of the least squares estimator (LSE) has received much attention. For the uncorrelated regressors case, Zellner (1963) derived the small sample properties of the seemingly unrelated regression estimator (SURE) and noted that the distribution of the estimator converges rapidly toward a normal density. Mehta and Swamy (1976) derived the exact second moment matrix of Zellner 's estimator conditional on an estimate of the variance-covariance matrix of the error terms and found that (i) the LSE is more efficient than Zellner 's estimator if the correlation in the errors of the two equations is zero, or small and (ii) Zellner's estimator is better if the contemporaneous correla- tion is high (also see Kunitomo (1977)). They also indicate that the gain in efficiency in using Zellner's estimator is especially high when the equation error correlation coefficient is close to one, and the loss is small when the errors are mildly correlated and the degrees of freedom is greater than 12.

In this paper, we examine under a squared error loss measure the risk of the seemingly unrelated regression pre-test estimator, (SURPE),

The authors acknowledge the helpful comments of David Giles and Helga Hessenius. This work was partially supported by National Science Foundation grant, SES-86-96152.

-2-

which is the GLSE if a preliminary test indicates that the correlation coefficient is significantly different from zero, and the LSE if we accept the null hypothesis of no correlation. The motivation for this research comes from Zellner's suggestion that it is possible to develop a decision procedure for deciding whether to use the LSE, or the GLSE.

In section 2, we present the statistical model and the various estimators. Our main interest is to derive the risk function of the SURPE with respect to the joint distribution of the test statistic r =

s „//s s and v = s ./s. , where the s.. (i,j = 1,2), which are de- fined later, are consistent estimators of the variances and the covariances of the errors. The small sample distribution of r as a function of the population correlation coefficient <j> is given in section 3. The marginal distribution of r is obtained from the joint distribution of r and v. In section 4, we derive the risk, function of the SURPE and compare it with those of LSE and GLSE. Section 5 sum- marizes and discusses the implications of the paper.

2. Statistical Model and Estimators

Consider the following two sample regression model

Zi

xl °

0 X,

' f

r

2i

+

Si

?2

?2

11

1 -1

, or y = Xa + e

(2.1)

where y. is a (nxl) vector of observations, X. is a (nxp) matrix of fixed regressors of rank p, a. is a (pxl) unknown location vector, and

e. is an (nxl) random error vector for i = 1,2. For expository

-3-

purposes we assume that X 'X„ = X0'Xn = 0 . We further assume that v * 1 2 2 1 p

the equation errors are distributed as multivariate normal random variables with zero means and covariance matrix

E = E

?1

e-2

[e^ e2'] = E[ee'

a I a I 11 n 12 n

o I a I 21 n 22 n

(2.2)

where I is an identity matrix of dimension n. The LSE for this model

n

is

a*(l) =

-1,

(X^) X1,y1 (X2'X2)"1X2'y2

(2.3)

The Zellner SUR estimator

o*(2) = (X'E LX) 1X'5: *y

(2.4)

is obtained by applying Aitken's GLSE to the whole system (2.1). The estimator in (2.4) is not feasible since it depends on unknown param- eters of the £ matrix. Replacing £ by a consistent estimator S produces Zellner's feasible GLSE, a*(4). One choice for the elements

of S =

Sll S12

-S21 S22

is s. . = - (y. - X.a*(l))'(y. - X.a*(l)),

i,j = 1,2.

-4-

Now the feasible GLSE is given by

a*(4) =

X ' 0

o x21

11T 12_ s I s I

n n

21T 22T s I s I

n n

Xl ° 0 X„

X ' 0

o x2'

11T 12T s I s I

n n

21 22T s I s I

n n

(\'\) l x1'y1 - (s12/s1I)(x1,xl) \'y2

(X2'X2)_1 X2'y2 - (s12/s22)(X2,X2)~1X2'y1

(2.5)

where we have used the assumption X 'X = X 'X = 0 and the s are

12 2 1 p

the elements of S

-1

11 12

s s

21 22

s s

The estimates of the variances and

the covariances are obtained from the restricted residuals, that are obtained from regressing y. on X. (i=l,2), i.e., implicitly assuming $ = 0.

The SUR pre-test estimator (SURPE) is based on the test statistic

r = s.,//s,.s„ that is used to test the null hypothesis H : <$> = 0 12 11 22 0

that the population correlation coefficient <f> is zero, versus a one- sided alternative H : <J> > 0. We reject the null hypothesis if r > c,

a

where c is the critical value chosen for the test. If we suspect a negative correlation then we reject the H , if r < -c. A two-sided alternative can also be set up and this would of course have impli- cations for the properties of the implied pretest estimator. This test statistic is similar to the locally best invariant test statistic given by Kariya (1981) and the Lagrange multiplier statistic of Breusch and Pagan (1980) and Shiba and Tsurumi (1988). The pretest estimator (Judge and Bock (1978)) is defined as follows: if we accept

-5-

H , the SURPE is the LSE, and otherwise it is the GLSE. This means o

the SURPE is

o*(3) = I (r)o*(l) + I (r)a*(4) (2.6)

[-l,c] (c,+l]

where I, N(*) is a zero-one indicator function. (•)

3. The Small Sample Distribution of r

The distribution of SURPE a*(3) and hence its risk, depends on the distribution of r. Therefore, in this section we derive the small sample distribution of r. First, we find the joint distribution of the test statistic r and v. It is well known that ns =x, ns =y and ns =z are distributed according to the Wishart distribution with covariance matrix £, and degrees of freedom t = n-2p. The joint density of x, y and z is given by

W(E,t) = k(xy-z2)(t"3)/2exp[-(x/an - l^z/Ja^o^ + y / a n) /2(l-<fr2)

(3.1)

where k = 1/ [2C | I | t/2/^ T(t/2) T( (t-1 )/2) ]. In the evaluation we make a transformation from the variables x, y and z to r = z//xy , v = z/y

and w = z. The density, in these new variables with Jacobian =

2 3 ' 1S 2w /vr

e, v ,/, 2. 3W 2. 2 2,(t-3)/2

f(r,v,w) = k(2w /vr )(w /r -w )

exp{-w(v/anr2 - 2$//a^0^ + l/a22v)/2( 1-<T ) }

(3.2)

when w, v e R, and 1 < r < +1.

-6-

Due to the nature of the transformation, the density in (3.2) is defined only when r, v, w are either all positive or all negative. As we see later, for our purpose, it is sufficient to consider only positive values of r. Therefore, from now on, we consider f(r,v,w) only when r, v, w are all positive and this means we assume a positive critical value.

To obtain the joint density of r and w, we integrate out w by using the gamma function

f(r,v) = 2k(l-r2)(t_3)/2r(t)/((v/r2a11-24>/alia22+l/va22)/2(l-*2))tvrt

(3.3) If we define

g = l/2(l-<t>2)an > 0,

h = -*/(l-4> )/^11°22 G R' and q = l/2(l-<j>2)a 22 > 0, (3.4)

the density in (3.3) may be written compactly as

f(r,v) = 2k(l-r2)(t"3)/2r(t)vC~1/rt((gv2/r2)+hv+q)t

= 2k(l-r2)(c"3)/2r(t)vt_1/rtgt((v2/r2)+(hv/g)+q/g)t

= 2k(l-r2)(t~3)/2r(t)vt_1/rtgt[((v/r)+hr/2g)2+(q/g)-h2r2/4g2]t

(3.5)

-7-

This completes Che derivation of the joint density of r and v.

To obtain the marginal distribution of r we define m(r) =

2 2 2 1/2 2

((q/g)-h r /4g ) and make a change of variable in v, x = v + hr /2g

and r = r. This gives

f(r,x) = 2k(l-r2)(C"3)/2r(t)(x-hr2/2g)t"1rt/gtU2+r2m(r)2)t

ni n/ Wl 2.(t-3)/2 t t~1/t-lw . 2.„ Nt-l-j j. t, 2^ 2 , ,2.t = 2kT(t)(l-r y r £ ( . )(-hr /2g) JxJ/g (x +r m(r) )

j=0 J

2 where x > hr /2g (3.6)

Next we substitute x = rm(r) tan 9 and obtain

ft Q) - 2kr(t)(l-r2)(t"3)/2 t;1(t-l)(-hr2/2g)t"1"jsinjecosW~:ie

t _ j , Nw-j+l w/2-j

g j=0 J m(r) r

where w = 2t-2 and arctg(hr/2gra(r ) ) < 9 < tt/2. (3.7)

To integrate out Q, we use successive integration by parts. This method depends heavily on j being even since the reduction from the integration by parts is by two at each step. Hence we distinguish two cases i) j is even, and ii) j is odd. The value of the integrals for even j is given by it/2

I = / (sin 0)J(cos 0)w Jd0

e*

j/2

= I (j-l)!!(-l)/(j-2i+l)!!

i = l

x ((w-j-l)!!/(w-j-l+2i)!!)sin(9*)j+1 2lcos(G*)W j 1+21

tt/2 + (j-1)! !(w-j-l)! !/(w-l)! !/ (cos G)V'dG (3.8)

9*

-8-

where Q* = arctg hr/2gm(r) and ! ! means double factorial. The

integral in (3.8) can be evaluated by using the value given in

Gradshteyn and Ryznik (1980).

tt/2 _ t-2

/ (cos 0)Wd6 = ((1/2W)( W )Q* + 1/2" L) Z (*)sin{(w-2k)Q*/(w-2k)} Q* Z k=0

(3.9)

When j is odd, the odd terras of the summation indexed by j in (3.7)

can be integrated using

tt/2 I = / (sin G)J(cos G)w Jd0

° Q*

1+1

2

= Z ((-l)(j-l)!!/(j-2i+l)!!) i=l

((w-j-1)! !/(w-j-l+2i)!!)sin(Q*)j+1"2icos(Q*)W"j_1+2i

(3.10)

Finally using I or I depending on whether j is even or odd, Q in (3.7) can be integrated out to compute the marginal distribution of the test statistic. This is given by

2(l-r2)(t-3)/2r(t)(l-*2)t/2 ti1(t-i)(,rt-1^(I ,1 ,j)/(H2r2)t-1/2-3/2 j=0 J l_2

/7 r(t/2)r((t-i)/2)

(3.11)

f(r)

where (I ,1 ,j) means that we pick either I or I depending on whether e o e o

j is even or odd.

-9-

In Figures 1 and 2, this distribution is plotted as a function of t=n-2p and <(>. In Figure 1 where (j> = 0, the distribution is symmetric for t = 10, 15. The distribution for the larger t has more probabil- ity mass around zero, but goes to zero faster on either side as r differs from zero. In Figure 2, we show for t = 15, the same dis- tribution with <)> = . 2 and 4> = .4. Under this scenario, as <(> gets larger there is more probability to the right. For example, P(r>0|4>=.2)=.72, whereas (P(r>0| <j>=.4)=. 88.

4. The Risk of the Pre-test Estimator (SURPE)

Since the derivation is symmetric and the calculations for the second sample are exactly similar, we can reduce the dimensionality of the coefficient vectors by two without affecting the results. Therefore, henceforth ct*(l), a*(3) and a*(4) are (pxl) vectors of estimators of the coefficients of the first sample only. Under squared error loss the risk of the SURPE is given by

p(o*(3),a ) = trE||l (r)o*(l)+I (r)a*(4)-a | |2

[-l,c] Cc,+1] " -1

= trE| | [I (r)(X 'X )_1X 'y - I (r)o. ] [-l,c] l l l ~l [-l,c] -1

+ [I (r)f(X 'X )_1X 'v - v(X 'X )_1X 'y }

Cc+i] L l l _1 l 1 1 ~l

- I (Oal||2 (4.1)

(c,+l] "X

Using (X1'X1)"1X1'y1 = ^HX^X^X^e^ and (X{ ' X{ )~lX{ 'y2 = (X1'X1)~1X1'e2 we have

■10-

-0.8 -0.4 0 0.4 0.8

Sample Correlation Coefficient

QP-267

FIG. 1. THE SMALL SAMPLE DISTRIBUTION OF R (t=10, 15: ~p=0)

■11-

-0.8 -0.4 0 0.4

Sample Correlation Coefficient

QP-268

FIG. 2. THE SMALL SAMPLE DISTRIBUTION OF R (t=15: $=0.2, 0.4)

-12-

p(a*(3),a1) =trE||[I (r)(X1 '^T^l '^

[-l,c]

+ I (r)(X '^)~\ 'e (c,+l] X i l ~l

- I (r)v(X 'X ) 1X1 'e ]||2 (c,+l] l l l ~L

= trE||(X'X. ) V'e. - I (r)v(X 'X ) LX »e | |2 (4.2) 11 L ~L (c,+l]

where we use the fact that I, , ,(r) + I, . , •. (r) = 1, since r e [-1,1].

l~l,cj (,c,+lj

Also, because the domains of the indicator functions are disjoint, this

means that Ir , ,(r)I, ,nl(r) = 0 and we obtain [-l,cj (c,+lj

p(a*(3),a1) = a11tr(X1*X1)"1

»X1) 1X1'e1e2'X1(X1'Xi) l

? -l -1

- 2trE{l (r)v(X 'X ) XX 'e. e„ 'X (X 'X ) l]

(c,+l] l L ~l~l l l L

+trE{l (r)v (X.'X.) LX 'e_e 'X. (X 'X ) X} (4.3)

(c,+l] X X X ~Z"Z X X l

Using the independence of the following vectors, (a*(l), (X 'X ) X 'y ,

(X 'X ) X 'y ) and the scale parameter estimates (s..., s , s ), yields

p(a*(3),aL) = ontr(X1'X1)"1

- 2E{l(C)+1](r)v}trE{(X1'X1)"1X11e1e2'X1(X1*X1)"1}

+ E{l (r)v2}trE{(X 'X. )_1X 'e9e ,'X (X 'X )_1 } (c,+l] X

= a tr(X. »X. )_1 - 2a E{l (r)v}tr(X 'X )_1 1 l (c,+l]

+ a tr(X 'X )_1E{l (r)v2} (4.4)

1 L (c,+l]

■13-

In order to compare the risks of SURPE , Zellner's GLSE and LSE, all risk evaluations are made with respect to the LSE risk, a tr(X 'X ) Therefore, the relative risk is

p(S*(3),a ) .

i ±TT\ \ = l~2 E1I (r)v}(a /a ) + E{l (r)v}(a Jo ) (4.5) PC^Cl).^) (Cf+1] l2 11 (Cj+1] 22 11

Here we should note that the r in the argument of the indicator func- tion in (4.5) is positive unless we choose a negative value of c. That is why, in section 2 the joint distribution f(r,v,w) is considered only for the positive values of r, v and w [see equation (3.2)].

The relative risk values of the SURPE with respect to that of LSE are given as a function of the population correlation coefficient $ and the critical value of the test c, in Table 1, for t = 10, 15, and 20 respectively, when a = a„„ = 1. These values are obtained by calculating the expectations in (4.5) with respect to the joint distribution of r and v that is derived in Section 3.

From the tabled values of the relative risk of SURPE, that is a function of <j> and the critical value c used in the preliminary testing, we notice that over the range of the (4>,c) parameter space, the relative risks of the pretest estimators cross. As larger and larger critical values are used, the LSE is used more frequently and this causes the relative risk of the SURPE to decrease for $ close to zero, and to increase for <j> close to one. The effect of degrees of freedom on these results is minimal.

The critical values of the SURPE for significance levels .05 and .10 are respectively .60 and .45. The relative risks of LSE and Zellner's GLSE for t = 10 are presented in Figure 3. The risk values

-14-

TABLE 1

Relative risk values of SURPE as a function of the population correlation coefficient <f» and the critical value c

10

t = 15

t = 20

c_

.1

.3

.5

.7

.9

9

1.0004

1.0009

1.0002

0.9775

0.5551

8

1.0040

1.0072

0.9967

0.8753

0.3030

7

1.0133

1.0180

0.9803

0.7652

0.2413

6

1.0273

1.0273

0.9517

0.6837

0.2247

5

1.0425

1.0303

0.9187

0.6332

0.2196

4

1.0552

1.0263

0.8887

0.6050

0.2179

3

1.0630

1.0178

0.8660

0.5907

0.2174

0

1.0648

0.9997

0.8426

0.5815

0.2172

9

1.0000

1.0000

1.0000

0.9924

0.5623

8

1.0001

1.0005

0.9870

0.8163

0.2563

7

1.0017

1.0041

0.9807

0.7554

0.2129

6

1.0064

1.0085

0.9436

0.6459

0.2128

5

1.0146

1.0085

0.8967

0.5880

0.2048

4

1.0240

1.0011

0.8553

0.5626

0.2047

3

1.0310

0.9885

0.8271

0.5530

0.2046

0

1.0307

0.9651

0.8049

0.5491

0.2046

9

1.0000

1.0000

1.0000

0.9972

0.5665

8

1.0000

1.0002

0.9987

0.9192

0.2348

7

1.0004

1.0015

0.9848

0.7528

0.2200

6

1.0022

1.0040

0.9450

0.6266

0.2195

5

1.0070

1.0031

0.8979

0.5675

0.2135

4

1.0143

0.9942

0.8413

0.5465

0.2090

3

1.0207

0.9790

0.8107

0.5402

0.2088

0

1.0212

0.9524

0.7907

0.5376

0.2086

1.2

1.0

CO

i_

O

"50.8

E

"to

LU

So.6

_*:

to

ir

>

0.4

CD

0.2

0

Risk of GLSE Risk of SURPE(c-0.45) Risk of SURPE (c-0.60) Risk of LSE

0.2 0.4 0.6 0.8

Population Correlation Coefficient QP-27o

FIG. 3. RISK VALUES OF SURPE ESTIMATORS (t=10]

-16-

of Zellner's estimator are taken from Zellner's paper (1963, p. 983). We observe that the relative risk of the SURPE with c = .60, starts below that of c = .45, crosses the latter around ij> = .3, and remains above for all <+> > .3. This means that throughout the (c,v) parameter space, no one SURPE is risk superior to the other. The SURPE with c = .6 is risk superior to SURPE with c = .45, for 4> close to zero. In turn it is risk inferior once 4> exceeds .3. This relationship between the SURPE 's with different critical values holds true throughout. In general, as can be observed from Table 1, the SURPE with a larger critical value has a small sampling variability when 4> is small, but then performs worse after its risk crosses that of the SURPE with a smaller critical value.

The relative risk function of Zellner's GLSE is also presented in Figure 3. Its risk is highest for small 4>, and then crosses the risks of LSE, SURPE (c=. 6) and finally SURPE (c=. 45) as <f> gets larger. Therefore, under squared error loss, none of the estimators in Figure 3 dominates. However, it is interesting to note that there is a range of <}> where SURPE is better than both LSE and GLSE. This is not the case in the regression coefficient pretesting. A possible reason for this might be the fact that 0 <_ <t> £ 1 prevents the pretest from making any disastrous type I and type II errors. The SURPE with 0 < c < 1 at cf> = 0 starts with a risk in between that of the LSE and the GLSE. It ends with a risk in between these two estimators when <$ = 1. One can also see that the SURPE has a substantial risk gain over the LSE for large ■+>, and the risk loss is modest when <J> is close to zero. When the critical value c takes on extreme values, the risk of SURPE

-17-

approaches the risk, of the LSE or the risk, of the GLSE depending whether c tends to 1 or to -1. Similar comparisons can be made for the same estimators in Figure 4 with t = 20 where the critical values .5 and .35 correspond to significance levels .05 and .1 respectively. As t increases, Zellner's GLSE becomes more efficient, and in fact approaches asymptotic efficiency levels.

5. Summary and Conclusions

We have made risk comparisons between the SURPE, LSE and Zellner's GLSE in the two sample seemingly unrelated regression model and found that no one estimator is uniformly superior. However, we can now determine the risk gains that accrue when the pre-test estimator is used to take advantage of the risk superiority of LSE, when <j> is close to zero, and the GLSE is used when $ is close to 1. Alternatively, we can determine the risk consequences of always using the pre-test rule. Finally, we examined the distribution of the test statistic r, evalu- ated some probabilities by numerical methods and found that the distribution when <$> = 0, is symmetric around zero, but skewed to the left for 0 > 0.

The applied statistician can gain insight into the nature of the correlation of disturbances of the underlying model, by conducting a preliminary test. Consequently in many situations, the risk advan- tages of the SUR pre-test estimator over the LSE and the GLSE can be exploited. For example in a somewhat different context, Stanek (1988) considers an experimental design which permits a variety of hypotheses

-18-

\J.<L

1 > 1 < I > 1

1.0

^m w^ ™™*^^^^II^ ^T^^^»

*T^.^«w.'"**'*«»

^K x-

%v N-

^v \

1_

£0.8

x * x %

0

*V \

E

"^Z

"\\

in

•X\

LU

u—

X*

S0.6

\>t

_*:

CO

X_

ir

X X

•x

«X

CD

•X •X •X

>

•X •X

«x

"-»—

•X

•x

-2 0.4

(D

rr

p;-i. _r pi op

jy^

r\ISK 0T oLotl

/"\ Ol

Risk of SURPE (c = 0.35)

0.2

Risk of SURPE (c = 0.50)

Risk of LSE

0

1 < 1 1 1 1 1

0.2 0.4 0.6 0.8

Population Correlation Coefficient

QP-269

FIG. 4. RISK VALUES OF SURPE ESTIMATORS (t=20]

-19-

to be tested some of which use SUR estimation to reduce variances. The SURPE procedure could be used to determine if SUR estimation is justified.

-20-

References

Breusch, T. S. and A. P. Pagan (1980). "The Lagrange Multiplier Test and Its Applications to Model Specification in Econometrics," The Review of Economic Studies, 47, 239-254.

Grandshteyn, I. S. and I. M. Ryznik (1980). "Table of Integrals, Series and Products," New York: Academic Press.

Judge, G. G. and M. E. Bock (1978). "The Statistical Implications of Pretest and Stein^rule Estimators in Econometrics," Amsterdam: North-Holland.

Kariya, T. (1981). "Tests for the Independence Between Two Seemingly Unrelated Regression Equations," Annals of Statistics, 9, 381-390.

Kunitomo, N. (1977). "A Note on the Efficiency of Zellner's Estima- tor for the Case of Two Seemingly Unrelated Regression Equations," Economic Studies Quarterly, 28, 73-77.

Mehta, J. S. and P. A. V. B. Swamy (1976). "Further Evidence on the Relative Efficiencies, of Zellner's Seemingly Unrelated Regressions Estimator," Journal of American Statistical Association, 71, 634- 639.

Shiba, T. and H. Tsurumi (1988). "Bayesian and Non-Bayesian Tests of Independence in Seemingly Unrelated Regressions," International Economic Review, 20, 377-395.

Stanek, Edward J. (1988). "Choosing a Pretest-Posttest Analysis," The American Statistician, 42, 178-183.

Zellner, A. (1962). "An Efficient Method of Estimating Seemingly Un- related Regressions and Tests of Aggregation Bias," Journal of American Statistical Association, 57, 348-368.

(1963). "Estimators of Seemingly Unrelated Regression

Equations: Some Exact Finite Sample Results," Journal of American Statistical Association, 58, 977-992.

D/508

NOTICE: Return or renew all Library Materials! The Minimum Fee for each Lost Book Is $50.00.

The person charging this material is responsible for its return to the library from which it was withdrawn on or before the Latest Date stamped below.

Theft, mutilation, and underlining of books are reasons for discipli- nary action and may result in dismissal from the University. To renew call Telephone Center, 333-8400

UNIVERSITY OF ILLINOIS LIBRARY AT URBANA-CHAMPAIGN

. . 1

w§r

L16I— O-1096

HECKMAN

BINDERY INC.

JUN95

Bound. T„. pi^ N.MANCHESTER